share_log

GROQ SUPERCHARGES FAST AI INFERENCE FOR META LLAMA 3.1

GROQ SUPERCHARGES FAST AI INFERENCE FOR META LLAMA 3.1

GROQ为Meta Llama 3.1的FASt AI推理提供强力支持
PR Newswire ·  07/23 11:15

New Largest and Most Capable Openly Available Foundation Model 405B Running on GroqCloud

在 GroqCloud 上运行的最大、功能最强大的全新公开基础模型 4050

MOUNTAIN VIEW, Calif., July 23, 2024 /PRNewswire/ -- Groq, a leader in fast AI inference, launched Llama 3.1 models powered by its LPU AI inference technology. Groq is proud to partner with Meta on this key industry launch, and run the latest Llama 3.1 models, including 405B Instruct, 70B Instruct, and 8B Instruct, at Groq speed. The three models are available on GroqCloud Dev Console, a community of over 300K developers already building on Groq systems, and on GroqChat for the general public.

加利福尼亚州山景城,2024年7月23日 /PRNewswire/ — 快速人工智能推理领域的领导者Groq推出了由其LPU AI推理技术提供支持的Llama 3.1模型。Groq 很荣幸能与 Meta 合作推出这一重要的行业产品,并以 Groq 的速度运行最新的 Llama 3.1 车型,包括 4050Instruct、700Instruct 和 80Instruct。这三种模式可在GroqCloud开发者控制台上使用,GroqCloud开发控制台是一个已经在Groq系统上构建的由超过3万名开发人员组成的社区,也可以在GroqChat上面向公众开放。

"I'm really excited to see Groq's ultra-low-latency inference for cloud deployments of the Llama 3.1 models. This is an awesome example of how our commitment to open source is driving innovation and progress in AI. By making our models and tools available to the community, companies like Groq can build on our work and help push the whole ecosystem forward." - Mark Zuckerberg, Founder & CEO, Meta
“看到Groq对Llama 3.1模型云部署的超低延迟推断,我感到非常兴奋。这是一个很好的例子,说明我们对开源的承诺如何推动人工智能的创新和进步。通过向社区提供我们的模型和工具,像 Groq 这样的公司可以在我们的工作基础上再接再厉,帮助推动整个生态系统向前发展。”-Meta 创始人兼首席执行官马克·扎克伯格

Mark Zuckerberg, Founder & CEO, Meta, shared:
"I'm really excited to see Groq's ultra-low-latency inference for cloud deployments of the Llama 3.1 models. This is an awesome example of how our commitment to open source is driving innovation and progress in AI. By making our models and tools available to the community, companies like Groq can build on our work and help push the whole ecosystem forward."

Meta 创始人兼首席执行官马克·扎克伯格分享说:
“看到Groq对Llama 3.1模型云部署的超低延迟推断,我感到非常兴奋。这是一个很好的例子,说明我们对开源的承诺如何推动人工智能的创新和进步。通过向社区提供我们的模型和工具,像Groq这样的公司可以在我们的工作基础上再接再厉,帮助推动整个生态系统向前发展。”

"Meta is creating the equivalent of Linux, an open operating system, for AI – not only for the Groq LPU which provides fast AI inference, but for the entire ecosystem. In technology, open always wins, and with this release of Llama 3.1, Meta has caught up to the best proprietary models. At this rate, it's only a matter of time before they'll pull ahead of the closed models," said Jonathan Ross, CEO and Founder of Groq. "With every new release from Meta, we see a significant surge of developers joining our platform. In the last five months we've grown from just a handful of developers to over 300,000, attracted by the quality and openness of Llama, as well as its incredible speed on the Groq LPU."

“Meta 正在为人工智能创建相当于 Linux 的开放操作系统,不仅适用于提供快速人工智能推理的 Groq LPU,还适用于整个生态系统。在技术领域,开放永远是赢家,随着Llama 3.1的发布,Meta已经赶上了最好的专有模型。按照这种速度,他们领先于封闭模式只是时间问题。” Groq首席执行官兼创始人乔纳森·罗斯说。“随着 Meta 发布的每一个新版本,我们都会看到大量开发人员加入我们的平台。在过去的五个月中,受Llama的质量和开放性以及其在Groq LPU上令人难以置信的速度所吸引,我们已经从仅有的几名开发人员发展到超过30万人。”

The Llama 3.1 models are a significant step forward in terms of capabilities and functionality. As the largest and most capable openly available Large Language Model to date, Llama 3.1 405B rivals industry-leading closed-source models. For the first time, enterprises, startups, researchers, and developers can access a model of this scale and capability without proprietary restrictions, enabling unprecedented collaboration and innovation. With Groq, AI innovators can now tap into the immense potential of Llama 3.1 405B running at unprecedented speeds on GroqCloud to build more sophisticated and powerful applications.

就能力和功能而言,Llama 3.1 模型是向前迈出的重要一步。作为迄今为止最大和最强大的公开语言模型,Llama 3.1 4050可与业界领先的闭源模型相媲美。企业、初创企业、研究人员和开发人员首次能够不受专有限制地访问这种规模和能力的模型,从而实现前所未有的协作和创新。有了Groq,人工智能创新者现在可以利用Llama 3.1 4050在GroqCloud上以前所未有的速度运行的巨大潜力,来构建更复杂、更强大的应用程序。

With Llama 3.1, including 405B, 70B, and 8B Instruct models, the AI community gains access to increased context length up to 128K and support across eight languages. Llama 3.1 405B is in a class of its own, with unmatched flexibility, control, and state-of-the-art capabilities in general knowledge, steerability, math, tool use, and multilingual translation. Llama 3.1 405B will unlock new capabilities, such as synthetic data generation and model distillation, and deliver new security and safety tools to further the shared Meta and Groq commitment to build an open and responsible AI ecosystem.

借助Llama 3.1,包括4050、700和80Instruct模型,人工智能社区可以获得更长的上下文长度,最高可达12.8万,并支持八种语言。Llama 3.1 4050独树一帜,在常识、操控性、数学、工具使用和多语言翻译方面具有无与伦比的灵活性、控制力和最先进的能力。Llama 3.1 4050将解锁新功能,例如合成数据生成和模型提炼,并提供新的安全和安全工具,以进一步履行Meta和Groq对建立开放和负责任的人工智能生态系统的共同承诺。

With unprecedented inference speeds for large openly available models like Llama 3.1 405B, developers are able to unlock new use cases that rely on agentic workflows to provide a seamless, yet personalized, human-like response for use cases such as: patient coordination and care; dynamic pricing by analyzing market demand and adjusting prices in real-time; predictive maintenance using real-time sensor data; and customer service by responding to customer inquiries and resolving issues in seconds.

借助诸如Llama 3.1 4050之类的大型公开模型前所未有的推理速度,开发人员能够解锁新的用例,这些用例依赖于代理工作流程为用例提供无缝但个性化、类似人类的响应,例如:患者协调和护理;通过分析市场需求和实时调整价格进行动态定价;使用实时传感器数据进行预测性维护;通过响应客户查询并在几秒钟内解决问题来提供客户服务。

GroqCloud has grown to over 300,000 developers in five months, underscoring the importance of speed when it comes to building the next generation of AI-powered applications at a fraction of the cost of GPUs.

GroqCloud 在五个月内已发展到超过 300,000 名开发人员,这突显了速度对于以低于 GPU 的成本构建下一代人工智能应用程序的重要性。

To experience Llama 3.1 models running at Groq speed, visit groq.com, and learn more about this launch from Groq and Meta.

要体验以 Groq 速度运行的 Llama 3.1 模型,请访问 groq.com,详细了解 Groq 和 Meta 本次发布的相关信息。

About Groq
Groq builds fast AI inference technology. Groq LPU AI inference technology is a hardware and software platform that delivers exceptional AI compute speed, quality, and energy efficiency. Groq, headquartered in Silicon Valley, provides cloud and on-prem solutions at scale for AI applications. The LPU and related systems are designed, fabricated, and assembled in North America. Try Groq speed at .

关于 Groq
Groq 构建了快速的人工智能推理技术。Groq LPU 人工智能推理技术是一种硬件和软件平台,可提供卓越的人工智能计算速度、质量和能效。Groq 总部位于硅谷,为 AI 应用提供大规模的云和本地解决方案。LPU 和相关系统在北美设计、制造和组装。试试Groq的速度.

Media Contact
Megan Hartwick
[email protected]

媒体联系人
梅根·哈特威克
[电子邮件保护]

SOURCE Groq

来源 Groq

声明:本内容仅用作提供资讯及教育之目的,不构成对任何特定投资或投资策略的推荐或认可。 更多信息
    抢沙发