share_log

GROQ SUPERCHARGES FAST AI INFERENCE FOR META LLAMA 3.1

GROQ SUPERCHARGES FAST AI INFERENCE FOR META LLAMA 3.1

GROQ爲Meta Llama 3.1的FASt AI推理提供強力支持
PR Newswire ·  07/23 11:15

New Largest and Most Capable Openly Available Foundation Model 405B Running on GroqCloud

新的最大、最能幹的開源基礎模型4050億在GroqCloud上運行

MOUNTAIN VIEW, Calif., July 23, 2024 /PRNewswire/ -- Groq, a leader in fast AI inference, launched Llama 3.1 models powered by its LPU AI inference technology. Groq is proud to partner with Meta on this key industry launch, and run the latest Llama 3.1 models, including 405B Instruct, 70B Instruct, and 8B Instruct, at Groq speed. The three models are available on GroqCloud Dev Console, a community of over 300K developers already building on Groq systems, and on GroqChat for the general public.

加利福尼亞州山景城,2024年7月23日 / PRNewswire / - Groq,快速AI推斷領域的領導者,推出由LPU AI推斷技術提供動力的Llama 3.1模型。Groq很自豪能與Meta合作推出此關鍵行業發佈,並以Groq速度運行最新的Llama 3.1模型,包括4050億指令、700億指令和80億指令。這三種模型可以在GroqCloud Dev Console上獲得,這是一個由超過30萬名開發人員構建在Groq系統上的社區,也可以在GroqChat上爲大衆所用。

"I'm really excited to see Groq's ultra-low-latency inference for cloud deployments of the Llama 3.1 models. This is an awesome example of how our commitment to open source is driving innovation and progress in AI. By making our models and tools available to the community, companies like Groq can build on our work and help push the whole ecosystem forward." - Mark Zuckerberg, Founder & CEO, Meta
“我真的很興奮地看到Groq對Llama 3.1模型雲部署的超低延遲推理。這是我們致力於開源推動AI創新和進步的絕佳示例。通過向社區提供我們的模型和工具,像Groq這樣的公司可以在我們的基礎上構建,並幫助推動整個生態系統向前發展。”-馬克·扎克伯格,Meta的創始人和CEO

Mark Zuckerberg, Founder & CEO, Meta, shared:
"I'm really excited to see Groq's ultra-low-latency inference for cloud deployments of the Llama 3.1 models. This is an awesome example of how our commitment to open source is driving innovation and progress in AI. By making our models and tools available to the community, companies like Groq can build on our work and help push the whole ecosystem forward."

Meta的創始人兼CEO馬克·扎克伯格(Mark Zuckerberg)表示:
“我真的很興奮地看到Groq對Llama 3.1模型雲部署的超低延遲推理。這是我們致力於開源推動AI創新和進步的絕佳示例。通過向社區提供我們的模型和工具,像Groq這樣的公司可以在我們的基礎上構建,並幫助推動整個生態系統向前發展。”

"Meta is creating the equivalent of Linux, an open operating system, for AI – not only for the Groq LPU which provides fast AI inference, but for the entire ecosystem. In technology, open always wins, and with this release of Llama 3.1, Meta has caught up to the best proprietary models. At this rate, it's only a matter of time before they'll pull ahead of the closed models," said Jonathan Ross, CEO and Founder of Groq. "With every new release from Meta, we see a significant surge of developers joining our platform. In the last five months we've grown from just a handful of developers to over 300,000, attracted by the quality and openness of Llama, as well as its incredible speed on the Groq LPU."

“Meta正在爲AI打造相當於Linux的開源操作系統,不僅針對提供快速AI推斷的Groq LPU,而且針對整個生態系統。在技術上,開源始終勝出,在發佈Llama 3.1後,Meta已經追上了最好的專有模型。以這樣的速度,它只是時間問題,它將超過封閉模型,”Groq的首席執行官和創始人喬納森·羅斯(Jonathan Ross)說。“隨着Meta的每一次新發布,我們看到有大量開發人員加入我們的平台。在過去的五個月中,我們已經從只有少數開發人員增加到300,000多名,他們被Llama的質量和開放性以及在Groq LPU上的難以置信的速度所吸引。”

The Llama 3.1 models are a significant step forward in terms of capabilities and functionality. As the largest and most capable openly available Large Language Model to date, Llama 3.1 405B rivals industry-leading closed-source models. For the first time, enterprises, startups, researchers, and developers can access a model of this scale and capability without proprietary restrictions, enabling unprecedented collaboration and innovation. With Groq, AI innovators can now tap into the immense potential of Llama 3.1 405B running at unprecedented speeds on GroqCloud to build more sophisticated and powerful applications.

Llama 3.1模型在功能和功能方面邁出了重要的一步。作爲迄今爲止最大、最能幹的公開可用的大語言模型,Llama 3.1 4050億可以與業界領先的專有模型相媲美。企業、初創公司、研究人員和開發人員首次可以在沒有專有限制的情況下訪問此規模和功能的模型,從而實現前所未有的協作和創新。利用Groq,AI創新者現在可以利用在GroqCloud上以前所未有的速度運行的Llama 3.1 4050億的巨大潛力,構建更復雜和強大的應用。

With Llama 3.1, including 405B, 70B, and 8B Instruct models, the AI community gains access to increased context length up to 128K and support across eight languages. Llama 3.1 405B is in a class of its own, with unmatched flexibility, control, and state-of-the-art capabilities in general knowledge, steerability, math, tool use, and multilingual translation. Llama 3.1 405B will unlock new capabilities, such as synthetic data generation and model distillation, and deliver new security and safety tools to further the shared Meta and Groq commitment to build an open and responsible AI ecosystem.

使用Llama 3.1,包括4050億、700億和80億指令模型,AI社區可以獲得增加到12.8萬的上下文長度,並在八種語言上獲得支持。Llama 3.1 4050億是獨一無二的,具有無與倫比的靈活性、控制和先進的能力,涵蓋了一般知識、可操縱性、數學、工具使用和多語種翻譯方面的最新技能。Llama 3.1 4050億將拓展新的功能,例如合成數據生成和模型蒸餾,並提供新的安全和安全工具,進一步推動Meta和Groq共同承諾建立開放和負責任的AI生態系統。

With unprecedented inference speeds for large openly available models like Llama 3.1 405B, developers are able to unlock new use cases that rely on agentic workflows to provide a seamless, yet personalized, human-like response for use cases such as: patient coordination and care; dynamic pricing by analyzing market demand and adjusting prices in real-time; predictive maintenance using real-time sensor data; and customer service by responding to customer inquiries and resolving issues in seconds.

藉助像Llama 3.1 4050億這樣大型公開可用的模型的超前推斷速度,開發人員能夠解鎖依賴代理工作流程爲用例提供無縫、個性化、類人響應的新用例,例如:患者協調和護理;通過分析市場需求並實時調整價格來實現動態定價;使用實時傳感器數據進行預測性維護;以及通過響應客戶查詢並在幾秒鐘內解決問題來提供客戶服務。

GroqCloud has grown to over 300,000 developers in five months, underscoring the importance of speed when it comes to building the next generation of AI-powered applications at a fraction of the cost of GPUs.

GroqCloud在五個月內已經發展成爲擁有300,000名開發人員的平台,這證明了在以GPU的一小部分成本構建下一代AI動力應用程序時速度的重要性。

To experience Llama 3.1 models running at Groq speed, visit groq.com, and learn more about this launch from Groq and Meta.

要體驗以Groq速度運行的Llama 3.1模型,請訪問groq.com,並從Groq和Meta了解更多有關此發佈的信息。

About Groq
Groq builds fast AI inference technology. Groq LPU AI inference technology is a hardware and software platform that delivers exceptional AI compute speed, quality, and energy efficiency. Groq, headquartered in Silicon Valley, provides cloud and on-prem solutions at scale for AI applications. The LPU and related systems are designed, fabricated, and assembled in North America. Try Groq speed at .

關於Groq
Groq構建了快速的AI推斷技術。Groq LPU AI推斷技術是一種硬件和軟件平台,可提供出色的AI計算速度、質量和能源效率。總部位於硅谷的Groq爲AI應用程序提供大規模的雲和本地解決方案。 LPU和相關係統在北美設計、製造和組裝。嘗試Groq速度。

Media Contact
Megan Hartwick
[email protected]

媒體聯繫人
梅根·哈特威克
[email protected]

SOURCE Groq

來源:Groq

声明:本內容僅用作提供資訊及教育之目的,不構成對任何特定投資或投資策略的推薦或認可。 更多信息
    搶先評論