Posted in

Magic 公司与谷歌云合作训练前沿级别的大型语言模型_AI阅读总结 — 包阅AI

包阅导读总结

1. `Magic`、`Google Cloud`、`LLMs`、`AI 平台`、`创新生态`

2. Magic 与 Google Cloud 合作,利用其平台和技术训练前沿规模的 LLMs,构建超级计算机。众多 AI 初创公司基于 Google Cloud 构建应用和模型,共同推动创新。

3.

– Magic 与 Google Cloud 合作

– 目标:训练具有超长上下文窗口和先进能力的前沿规模 LLMs,打造“自动化 AI 软件工程师和研究员”。

– 行动:构建基于云的超级计算机,包括 G4 采用 NVIDIA H100 Tensor Core GPUs 的 A3 Mega VMs,G5 将迁移至 NVIDIA Grace Blackwell 平台。

– Google Cloud 的支持

– 提供信任的 AI 平台、可靠的计算能力和 AI 产品扩展经验。

– 创新生态

– 众多初创公司在 Google Cloud 上构建模型、工具和应用,如 Arize AI、Character AI 等。

思维导图:

文章地址:https://cloud.google.com/blog/products/ai-machine-learning/magic-ai-100m-tokens-cloud-supercomputer/

文章来源:cloud.google.com

作者:Amin Vahdat

发布时间:2024/8/30 0:00

语言:英文

总字数:932字

预计阅读时间:4分钟

评分:88分

标签:AI开发,谷歌云,大型语言模型,英伟达GPU,AI超级计算机


以下为原文内容

本内容来源于用户推荐转载,旨在分享知识与观点,如有侵权请联系删除 联系邮箱 media@ilingban.com

More than half of the world’s generative AI startups, including more than 90% of generative AI unicorns, are building on Google Cloud — utilizing our trusted infrastructure, a variety of hardware systems, the Vertex AI platform, and much more. These startups are building the next generation of AI applications, models, and tooling that will be used by millions of businesses, entrepreneurs, developers, students, and more in the coming months and years.

Supporting Magic with Google Cloud’s AI Platform

Today, Magic announced it is partnering with Google Cloud to build two new cloud-based supercomputers that will support Magic’s mission of developing code assistants with a context window reaching 100 million tokens (enough information to equal 10 years of human speech). Magic has selected Google Cloud as its preferred cloud provider, and the San Francisco-based startup will utilize Google Cloud’s AI Hypercomputer architecture and tooling, which will help build frontier-scale AI models that can automate aspects of software engineering.

With Google Cloud, Magic will build its G4 supercomputer utilizing A3 Mega VMs powered by NVIDIA H100 Tensor Core GPUs. For their next-generation G5 supercomputer, they will be one of the first users to migrate to the NVIDIA Grace Blackwell platform on Google Cloud when they become available early next year, scaling up to tens of thousands of GPUs. These computers will be able to achieve 160 exaflops, a measure of computing performance so large, it’s roughly equal to 160 billion people each holding one billion calculators and running a computation at the same exact moment.

Magic’s goal is to build an “automated AI software engineer and researcher” that can see and understand an organization’s entire code repository and complete large tasks over long time horizons. To do this, they are training frontier-scale LLMs with ultra-long context windows and other advanced capabilities. The compute required to train these models, and perform AI inference, is significant; Google Cloud is providing the trusted AI platform, reliable access to compute power, and first-hand experience scaling AI products to help the Magic team bring its products to market.

Magic CEO and founder Eric Steinberger said: “Magic’s goal is to build AGI, and that will take a lot of compute. Google Cloud will be a valuable partner to Magic as we train and serve our next-gen models. Google’s experience operating the largest infrastructure systems in the world will help our team be maximally effective, and their team has been incredibly supportive as we started ramping up.”

Driving Innovation at the World’s Most Exciting AI Startups

Magic joins a thriving ecosystem of exciting startup businesses who are building foundational models, AI tooling, and applications on Google Cloud. For example:

  • Arize AI, which offers a platform for model observability and evaluation, continues to utilize Google Cloud services including GKE, Vertex AI, and Google Cloud Marketplace to build and scale its platform and bring it to market.
  • Character AI named Google Cloud as its preferred cloud infrastructure in 2023, and utilizes TPUs and GPUs for faster training and inference of its models, as well as our AlloyDB database to exponentially scale its database load.
  • Cohere has partnered with Google Cloud since 2021 to utilize TPUs for training and inference with its enterprise-grade frontier AI models, and to bring its security and privacy focused platform to market on Google Cloud Marketplace.
  • Labelbox powers Google Cloud’s LLM evaluation service, and Google Cloud recently partnered with Labelbox to allow Vertex AI customers to seamlessly leverage human raters to evaluate LLM responses while handling the entire workforce and labeling orchestration.
  • Mistral began working with Google Cloud in 2023, using our AI-optimized infrastructure, including TPUs, to scale up its LLMs and offering its foundational model, Mistral-7B, on Vertex AI.
  • Glean uses a mix of Google Cloud services for its AI assistant and enterprise search platform, including BigQuery for data analytics, TPUs for model training, and App Engine and GKE to scale its platform reliability.
  • Higgsfield is using Gemini and our AI-optimized infrastructure, including GPUs, to power its AI video creation platform and for training and inference of its proprietary model. Google Cloud is also helping Higgsfield implement AI safety standards, including watermarking, to help prevent the production of malicious content.
  • Jasper is working with Google Cloud to power its marketing content creation tools, including utilizing Gemini models to help users automatically generate content like blog posts or product descriptions for their customers.
  • Repl.it is a popular AI-powered software development and deployment platform. The company utilizes Google Cloud services, infrastructure, and Gemini models to help its 20 million-plus users create high-quality code more quickly.
  • ThoughtSpot’s search and AI-powered analytics platform makes it simple to ask and answer questions with data. ThoughtSpot integrates Gemini models to power its AI features, helping customers tap into new levels of productivity.
  • Typeface, the generative AI platform for enterprise content creation, empowers Fortune 500 brands and enterprise marketers to create multimodal branded content. An early partner of Google’s GenAI foundational models, Typeface delivers end-to-end content workflows across Google platforms, including Google Cloud, Google Ads, and Google Workspace.
  • Weights & Biases’ Weave is a user-friendly, lightweight toolkit designed to help developers track and evaluate their Gemini family of multi-modal LLMs in a more organized and efficient manner. The Gemini ecosystem includes some of the most powerful models, featuring extremely long context and multimodal capabilities that allow reasoning across text, images, audio and video.
  • Writer uses a variety of Google Cloud services to power its generative AI enterprise content platform, including AlloyDB, GKE, BigQuery, and GPUs on Google Cloud.

You can learn more about Magic’s work to train large scale AI models on Google Cloud here.