NVIDIA Corp. (NVDA) Q2 2024 Earnings Call Transcript
NVIDIA Corp. (NVDA) Q2 2024 Earnings Call Transcript
Company Management
公司管理
Jensen Huang - Co-Founder, CEO & President |
Colette Kress - EVP & CFO |
Simona Jankowski - VP, IR |
黄仁森-联合创始人、首席执行官兼总裁 |
Colette Kress-执行副总裁兼首席财务官 |
西蒙娜·扬科夫斯基——投资者关系副总裁 |
Analysts
分析师
Joseph Moore - Morgan Stanley |
Vivek Arya - Bank of America |
Toshiya Hari - Goldman Sachs |
Matt Ramsay - Cowen |
Mark Lipacis - Jefferies |
Atif Malik - Citi |
Timothy Arcuri - UBS |
Benjamin Reitzes - Melius |
Stacy Rasgon - Bernstein Research |
约瑟夫摩尔-摩根士丹利 |
Vivek Arya-美国银行 |
Toshiya Hari-高盛 |
Matt Ramsay-Cowen |
马克·利帕西斯-杰富瑞 |
Atif Malik-花旗 |
Timothy Arcuri-瑞银 |
Benjamin Reitzes-Melius |
Stacy Rasgon-伯恩斯坦研究 |
Operator
操作员
Good afternoon. My name is David, and I'll be your conference operator today. At this time, I'd like to welcome everyone to NVIDIA's Second Quarter Earnings Call. Today's conference is being recorded. All lines have been placed on mute to prevent any background noise. After the speakers’ remarks, there will be a question-and-answer session. [Operator Instructions]
下午好。我叫大卫,今天我将成为你的会议接线员。此时,我想欢迎大家参加 NVIDIA 的第二季度财报电话会议。今天的会议正在录制中。为了防止任何背景噪音,所有线路都已设置为静音。发言者发言后,将进行问答环节。[操作员说明]
Thank you. Simona Jankowski, you may begin your conference.
谢谢。西蒙娜·扬科夫斯基,你可以开始你的会议了。
Simona Jankowski
西蒙娜·扬科夫斯基
Thank you. Good afternoon, everyone and welcome to NVIDIA's conference call for the second quarter of fiscal 2024. With me today from NVIDIA are Jensen Huang, President and Chief Executive Officer; and Colette Kress, Executive Vice President and Chief Financial Officer. I'd like to remind you that our call is being webcast live on NVIDIA's Investor Relations website. The webcast will be available for replay until the conference call to discuss our financial results for the third quarter of fiscal 2024. The content of today's call is NVIDIA's property. It can't be reproduced or transcribed without our prior written consent.
谢谢。大家下午好,欢迎参加英伟达2024财年第二季度的电话会议。今天和我一起来自 NVIDIA 的有总裁兼首席执行官黄仁森和执行副总裁兼首席财务官科莱特·克雷斯。我想提醒大家,我们的电话会议将在NVIDIA的投资者关系网站上进行网络直播。在讨论我们2024财年第三季度财务业绩的电话会议之前,网络直播将可供重播。今天的电话会议内容是 NVIDIA 的财产。未经我们事先书面同意,不得对其进行复制或转录。
During this call, we may make forward-looking statements based on current expectations. These are subject to a number of significant risks and uncertainties, and our actual results may differ materially. For a discussion of factors that could affect our future financial results and business, please refer to the disclosure in today's earnings release, our most recent Forms 10-K and 10-Q and the reports that we may file on Form 8-K with the Securities and Exchange Commission. All our statements are made as of today, August 23, 2023, based on information currently available to us. Except as required by law, we assume no obligation to update any such statements.
在本次电话会议上,我们可能会根据当前的预期发表前瞻性陈述。这些风险和不确定性会受到许多重大风险和不确定性的影响,我们的实际结果可能存在重大差异。有关可能影响我们未来财务业绩和业务的因素的讨论,请参阅今天财报中的披露、我们最新的10-K和10-Q表格以及我们可能在8-K表格上向美国证券交易委员会提交的报告。截至今天,即2023年8月23日,我们所有的声明均基于我们目前可用的信息作出。除非法律要求,否则我们不承担更新任何此类声明的义务。
During this call, we will discuss non-GAAP financial measures. You can find a reconciliation of these non-GAAP financial measures to GAAP financial measures in our CFO commentary, which is posted on our website.
在本次电话会议上,我们将讨论非公认会计准则财务指标。您可以在我们网站上发布的首席财务官评论中找到这些非公认会计准则财务指标与GAAP财务指标的对账情况。
And with that, let me turn the call over to Colette.
然后,让我把电话交给 Colette。
Colette Kress
Colette Kress
Thanks, Simona. We had an exceptional quarter. Record Q2 revenue of $13.51 billion was up 88% sequentially and up 101% year-on-year, and above our outlook of $11 billion.
谢谢,西蒙娜。我们度过了一个非同寻常的季度。创纪录的第二季度收入为135.10亿美元,环比增长88%,同比增长101%,高于我们预期的1100万美元。
Let me first start with Data Center. Record revenue of $10.32 billion was up 141% sequentially and up 171% year-on-year. Data Center compute revenue nearly tripled year-on-year, driven primarily by accelerating demand from cloud service providers and large consumer Internet companies for HGX platform, the engine of generative AI and large language models.
让我先从数据中心开始。创纪录的103.2亿美元收入环比增长141%,同比增长171%。数据中心计算收入同比增长近三倍,这主要是由于云服务提供商和大型消费互联网公司对HGX平台(生成式人工智能和大型语言模型的引擎)的需求不断增长。
Major companies, including AWS, Google Cloud, Meta, Microsoft Azure and Oracle Cloud as well as a growing number of GPU cloud providers are deploying, in volume, HGX systems based on our Hopper and Ampere architecture Tensor Core GPUs. Networking revenue almost doubled year-on-year, driven by our end-to-end InfiniBand networking platform, the gold standard for AI.
包括AWS、谷歌云、Meta、微软 Azure 和甲骨文云在内的各大公司以及越来越多的 GPU 云提供商正在批量部署基于我们的 Hopper 和 Ampere 架构 Tensor Core GPU 的 HGX 系统。在我们的端到端 InfiniBand 网络平台(人工智能的黄金标准)的推动下,网络收入同比几乎翻了一番。
There is tremendous demand for NVIDIA accelerated computing and AI platforms. Our supply partners have been exceptional in ramping capacity to support our needs. Our data center supply chain, including HGX with 35,000 parts and highly complex networking has been built up over the past decade. We have also developed and qualified additional capacity and suppliers for key steps in the manufacturing process such as [indiscernible] packaging.
对NVIDIA加速计算和人工智能平台的需求巨大。我们的供应合作伙伴在提高能力以满足我们的需求方面表现出色。我们的数据中心供应链,包括拥有 35,000 个零件和高度复杂网络的 HGX,在过去十年中已经建立起来。我们还为制造过程中的关键步骤(例如 [难以辨别的] 包装)开发并认证了额外的产能和供应商。
We expect supply to increase each quarter through next year. By geography, data center growth was strongest in the U.S. as customers direct their capital investments to AI and accelerated computing. China demand was within the historical range of 20% to 25% of our Data Center revenue, including compute and networking solutions.
我们预计,到明年,每个季度的供应量都将增加。按地域划分,美国的数据中心增长最为强劲,因为客户将其资本投资用于人工智能和加速计算。中国需求占我们数据中心收入(包括计算和网络解决方案)的20%至25%的历史区间内。
At this time, let me take a moment to address recent reports on the potential for increased regulations on our exports to China. We believe the current regulation is achieving the intended results. Given the strength of demand for our products worldwide, we do not anticipate that additional export restrictions on our Data Center GPUs, if adopted, would have an immediate material impact to our financial results.
此时,让我花点时间谈谈最近关于可能加强对中国出口监管的报道。我们认为,目前的法规正在实现预期的结果。鉴于全球对我们产品的需求强劲,我们预计如果采用对数据中心GPU的额外出口限制,不会对我们的财务业绩产生直接的重大影响。
However, over the long term, restrictions prohibiting the sale of our Data Center GPUs to China, if implemented, will result in a permanent loss and opportunity for the U.S. industry to compete and lead in one of the world's largest markets.
但是,从长远来看,禁止向中国销售我们的数据中心GPU的限制措施如果得到实施,将给美国行业带来永久损失,并有机会在全球最大的市场之一中竞争并处于领先地位。
Our cloud service providers drove exceptional strong demand for HGX systems in the quarter, as they undertake a generational transition to upgrade their data center infrastructure for the new era of accelerated computing and AI. The NVIDIA HGX platform is culminating of nearly two decades of full stack innovation across silicon, systems, interconnects, networking, software and algorithms.
我们的云服务提供商在本季度推动了对HGX系统的异常强劲需求,因为他们进行了代际过渡,以升级其数据中心基础设施,以迎接加速计算和人工智能的新时代。NVIDIA HGX 平台是将近二十年来在芯片、系统、互连、网络、软件和算法方面的全栈创新的结晶。
Instances powered by the NVIDIA H100 Tensor Core GPUs are now generally available at AWS, Microsoft Azure and several GPU cloud providers, with others on the way shortly. Consumer Internet companies also drove the very strong demand. Their investments in data center infrastructure purpose-built for AI are already generating significant returns. For example, Meta, recently highlighted that since launching Reels, AI recommendations have driven a more than 24% increase in time spent on Instagram.
由NVIDIA H100 Tensor Core GPU提供支持的实例现已在AWS、Microsoft Azure和多家GPU云提供商上市,其他提供商也即将推出。消费互联网公司也推动了非常强劲的需求。他们对专为人工智能构建的数据中心基础设施的投资已经产生了可观的回报。例如,Meta最近强调,自推出Reels以来,人工智能推荐已使在Instagram上花费的时间增加了24%以上。
Enterprises are also racing to deploy generative AI, driving strong consumption of NVIDIA powered instances in the cloud as well as demand for on-premise infrastructure. Whether we serve customers in the cloud or on-prem through partners or direct, their applications can run seamlessly on NVIDIA AI enterprise software with access to our acceleration libraries, pre-trained models and APIs.
企业也在竞相部署生成式人工智能,这推动了云端NVIDIA支持的实例的强劲消费以及对本地基础设施的需求。无论我们通过合作伙伴还是直接在云端或本地为客户提供服务,他们的应用程序都可以在NVIDIA AI企业软件上无缝运行,并可以访问我们的加速库、预训练模型和API。
We announced a partnership with Snowflake to provide enterprises with accelerated path to create customized generative AI applications using their own proprietary data, all securely within the Snowflake Data Cloud. With the NVIDIA NeMo platform for developing large language models, enterprises will be able to make custom LLMs for advanced AI services, including chatbot, search and summarization, right from the Snowflake Data Cloud.
我们宣布与Snowflake建立合作伙伴关系,为企业提供更快的途径,让他们使用自己的专有数据创建定制的生成式人工智能应用程序,所有这些应用程序都安全地在Snowflake数据云中完成。借助用于开发大型语言模型的 NVIDIA NeMo 平台,企业将能够直接从 Snowflake 数据云为高级人工智能服务(包括聊天机器人、搜索和摘要)制作自定义 LLM。
Virtually, every industry can benefit from generative AI. For example, AI Copilot such as those just announced by Microsoft can boost the productivity of over 1 billion office workers and tens of millions of software engineers. Billions of professionals in legal services, sales, customer support and education will be available to leverage AI systems trained in their field. AI Copilot and assistants are set to create new multi-hundred billion dollar market opportunities for our customers.
实际上,每个行业都可以从生成式人工智能中受益。例如,像微软刚刚宣布的那样的人工智能副驾驶可以提高超过100亿办公室工作人员和数千万软件工程师的工作效率。数十亿法律服务、销售、客户支持和教育领域的专业人员将可以利用在其领域接受过培训的人工智能系统。AI Copilot和助手将为我们的客户创造数千亿美元的新市场机会。
We are seeing some of the earliest applications of generative AI in marketing, media and entertainment. WPP, the world's largest marketing and communication services organization, is developing a content engine using NVIDIA Omniverse to enable artists and designers to integrate generative AI into 3D content creation. WPP designers can create images from text prompts while responsibly trained generative AI tools and content from NVIDIA partners such as Adobe and Getty Images using NVIDIA Picasso, a foundry for custom generative AI models for visual design.
我们看到了生成式人工智能在营销、媒体和娱乐领域的一些最早应用。全球最大的营销和传播服务组织WPP正在使用NVIDIA Omniverse开发内容引擎,使艺术家和设计师能够将生成式人工智能集成到3D内容创作中。WPP设计师可以根据文本提示创建图像,同时使用NVIDIA Picasso(一家为视觉设计定制生成人工智能模型的代工厂)提供来自Adobe和Getty Images等NVIDIA合作伙伴的负责任的生成人工智能工具和内容。
Visual content provider Shutterstock is also using NVIDIA Picasso to build tools and services that enables users to create 3D scene background with the help of generative AI. We've partnered with ServiceNow and Accenture to launch the AI Lighthouse program, fast tracking the development of enterprise AI capabilities. AI Lighthouse unites the ServiceNow enterprise automation platform and engine with NVIDIA accelerated computing and with Accenture consulting and deployment services.
视觉内容提供商Shutterstock还使用NVIDIA Picasso开发工具和服务,使用户能够在生成式人工智能的帮助下创建3D场景背景。我们已经与ServiceNow和埃森哲合作启动了人工智能灯塔计划,以快速跟踪企业人工智能能力的发展。AI Lighthouse 将 ServiceNow 企业自动化平台和引擎与 NVIDIA 加速计算以及埃森哲咨询和部署服务相结合。
We are collaborating also with Hugging Face to simplify the creation of new and custom AI models for enterprises. Hugging Face will offer a new service for enterprises to train and tune advanced AI models powered by NVIDIA HGX cloud. And just yesterday, VMware and NVIDIA announced a major new enterprise offering called VMware Private AI Foundation with NVIDIA, a fully integrated platform featuring AI software and accelerated computing from NVIDIA with multi-cloud software for enterprises running VMware.
我们还与Hugging Face合作,以简化为企业创建新的和自定义的人工智能模型。Hugging Face将为企业提供一项新服务,用于训练和调整由NVIDIA HGX云支持的高级人工智能模型。就在昨天,VMware和NVIDIA宣布与NVIDIA推出一项名为VMware Private AI Foundation的大型新企业产品。NVIDIA是一个完全集成的平台,具有人工智能软件和来自NVIDIA的加速计算,适用于运行VMware
VMware's hundreds of thousands of enterprise customers will have access to the infrastructure, AI and cloud management software needed to customize models and run generative AI applications such as intelligent chatbot, assistants, search and summarization. We also announced new NVIDIA AI enterprise-ready servers featuring the new NVIDIA L40S GPU built for the industry standard data center server ecosystem and BlueField-3 DPU data center infrastructure processor.
VMware成千上万的企业客户将可以使用定制模型和运行生成式人工智能应用程序(例如智能聊天机器人、助手、搜索和摘要)所需的基础架构、人工智能和云管理软件。我们还宣布推出全新 NVIDIA AI 企业级服务器,该服务器采用专为行业标准数据中心服务器生态系统构建的全新 NVIDIA L40S GPU 和 BlueField-3 DPU 数据中心基础设施处理器。
L40S is not limited by [indiscernible] supply and is shipping to the world's leading server system makers (ph). L40S is a universal data center processor designed for high volume data center standing out to accelerate the most compute-intensive applications, including AI training and inventing through the designing, visualization, video processing and NVIDIA Omniverse industrial digitalization.
L40S不受 [难以辨别的] 供应的限制,而是向世界领先的服务器系统制造商(ph)出货。L40S 是一款通用数据中心处理器,专为大容量数据中心而设计,可通过设计、可视化、视频处理和 NVIDIA Omniverse 工业数字化加速计算密集型应用程序,包括人工智能训练和发明。
NVIDIA AI enterprise ready servers are fully optimized for VMware, Cloud Foundation and Private AI Foundation. Nearly 100 configurations of NVIDIA AI enterprise ready servers will soon be available from the world's leading enterprise IT computing companies, including Dell, HP and Lenovo. The GH200 Grace Hopper Superchip which combines our ARM-based Grace CPU with Hopper GPU entered full production and will be available this quarter in OEM servers. It is also shipping to multiple supercomputing customers, including Atmos (ph), National Labs and the Swiss National Computing Center.
NVIDIA AI 企业就绪型服务器已针对 VMware、云基础和私有 AI 基金会进行了全面优化。包括戴尔、惠普和联想在内的全球领先的企业IT计算公司将很快推出近100种NVIDIA AI企业级就绪型服务器配置。将我们基于 ARM 的 Grace CPU 与 Hopper GPU 相结合的 GH200 Grace Hopper Superchip 已进入全面生产,并将于本季度在 OEM 服务器上市。它还向多个超级计算客户发货,包括Atmos(ph)、国家实验室和瑞士国家计算中心。
And NVIDIA and SoftBank are collaborating on a platform based on GH200 for generative AI and 5G/6G applications. The second generation version of our Grace Hopper Superchip with the latest HBM3e memory will be available in Q2 of calendar 2024. We announced the DGX GH200, a new class of large memory AI supercomputer for giant AI language model, recommendator systems and data analytics. This is the first use of the new NVIDIA [indiscernible] switch system, enabling all of its 256 Grace Hopper Superchips to work together as one, a huge jump compared to our prior generation connecting just eight GPUs over [indiscernible]. DGX GH200 systems are expected to be available by the end of the year, Google Cloud, Meta and Microsoft among the first to gain access.
NVIDIA 和软银正在合作开发一个基于 GH200 的平台,用于生成式人工智能和 5G/6G 应用程序。配备最新HBM3e内存的Grace Hopper Superchip的第二代版本将于2024年第二季度上市。我们发布了 DGX GH200,这是一款用于巨型 AI 语言模型、推荐器系统和数据分析的新型大内存 AI 超级计算机。这是首次使用新的NVIDIA [难以辨认] 交换机系统,它使所有256个Grace Hopper Superchips能够作为一个整体协同工作,与我们的上一代仅连接八个GPU相比,这是一个巨大的飞跃 [难以辨认]。DGX GH200 系统预计将于今年年底上市,谷歌云、Meta和微软是最早获得访问权限的系统之一。
Strong networking growth was driven primarily by InfiniBand infrastructure to connect HGX GPU systems. Thanks to its end-to-end optimization and in-network computing capabilities, InfiniBand delivers more than double the performance of traditional Ethernet for AI. For billions of dollar AI infrastructures, the value from the increased throughput of InfiniBand is worth hundreds of [indiscernible] and pays for the network. In addition, only InfiniBand can scale to hundreds of thousands of GPUs. It is the network of choice for leading AI practitioners.
网络的强劲增长主要是由连接HGX GPU系统的InfiniBand基础设施推动的。得益于其端到端优化和网络内计算能力,InfiniBand 为人工智能提供的性能是传统以太网的两倍多。对于价值数十亿美元的人工智能基础设施来说,InfiniBand增加吞吐量所带来的价值相当于数百个 [难以辨认],并且为网络付出了代价。此外,只有 InfiniBand 可以扩展到数十万个 GPU。它是领先的人工智能从业者的首选网络。
For Ethernet-based cloud data centers that seek to optimize their AI performance, we announced NVIDIA Spectrum-X, an accelerated networking platform designed to optimize Ethernet for AI workloads. Spectrum-X couples the Spectrum or Ethernet switch with the BlueField-3 DPU, achieving 1.5x better overall AI performance and power efficiency versus traditional Ethernet. BlueField-3 DPU is a major success. It is in qualification with major OEMs and ramping across multiple CSPs and consumer Internet companies.
对于寻求优化 AI 性能的基于以太网的云数据中心,我们宣布推出了 NVIDIA Spectrum-X,这是一款旨在针对人工智能工作负载优化以太网的加速网络平台。Spectrum-X 将 Spectrum 或以太网交换机与 BlueField-3 DPU 相结合,与传统以太网相比,整体 AI 性能和能效提高了 1.5 倍。Bluefield-3 DPU 取得了重大成功。它已获得主要原始设备制造商的资格,并已扩展到多家通信服务提供商和消费互联网公司。
Now moving to gaming. Gaming revenue of $2.49 billion was up 11% sequentially and 22% year-on-year. Growth was fueled by GeForce RTX 40 Series GPUs for laptops and desktop. End customer demand was solid and consistent with seasonality. We believe global end demand has returned to growth after last year's slowdown. We have a large upgrade opportunity ahead of us. Just 47% of our installed base have upgraded to RTX and about 20% of the GPU with an RTX 3060 or higher performance.
现在转向游戏。博彩收入为24.9亿美元,环比增长11%,同比增长22%。适用于笔记本电脑和台式机的 GeForce RTX 40 系列 GPU 推动了增长。终端客户需求稳健,且与季节性一致。我们认为,在去年的经济放缓之后,全球终端需求已恢复增长。我们面前还有很大的升级机会。我们只有 47% 的安装量升级到了 RTX,大约 20% 的 GPU 升级到了 RTX 3060 或更高的性能。
Laptop GPUs posted strong growth in the key back-to-school season, led by RTX 4060 GPUs. NVIDIA's GPU-powered laptops have gained in popularity, and their shipments are now outpacing desktop GPUs from several regions around the world. This is likely to shift the reality of our overall gaming revenue a bit, with Q2 and Q3 as the stronger quarters of the year, reflecting the back-to-school and holiday build schedules for laptops.
在RTX 4060 GPU的带动下,笔记本电脑GPU在关键的返校季节实现了强劲的增长。NVIDIA基于GPU的笔记本电脑越来越受欢迎,其出货量现在已超过全球多个地区的台式机GPU。这可能会稍微改变我们总体游戏收入的现实,第二季度和第三季度是今年最强劲的季度,这反映了笔记本电脑的返校和假日建造时间表。
In desktop, we launched the GeForce RTX 4060 and the GeForce RTX 4060 TI GPUs, bringing the Ada Lovelace architecture down to price points as low as $299. The ecosystem of RTX and DLSS games continue to expand. 35 new games added to DLSS support, including blockbusters such as Diablo IV and Baldur’s Gate 3.
在台式机方面,我们推出了GeForce RTX 4060和GeForce RTX 4060 TI GPU,使Ada Lovelace架构的价格降至299美元。RTX和DLSS游戏的生态系统继续扩大。增加了35款新游戏支持DLSS,包括诸如《暗黑破坏神IV》和《博德之门3》等大片。
There's now over 330 RTX accelerated games and apps. We are bringing generative AI to gaming. At COMPUTEX, we announced NVIDIA Avatar Cloud Engine or ACE for games, a custom AI model foundry service. Developers can use this service to bring intelligence to non-player characters. And it harnesses a number of NVIDIA Omniverse and AI technologies, including NeMo, Riva and Audio2Face.
现在有 330 多款 RTX 加速游戏和应用程序。我们正在将生成式人工智能引入游戏中。在COMPUTEX上,我们宣布推出用于游戏的NVIDIA Avatar云引擎或ACE,这是一项定制的人工智能模型铸造服务。开发者可以使用此服务为非玩家角色提供情报。它还利用了许多 NVIDIA Omniverse 和人工智能技术,包括 NeMo、Riva 和 Audio2Face。
Now moving to Professional Visualization. Revenue of $375 million was up 28% sequentially and down 24% year-on-year. The Ada architecture ramp drove strong growth in Q2, rolling out initially in laptop workstations with a refresh of desktop workstations coming in Q3. These will include powerful new RTX systems with up to 4 NVIDIA RTX 6000 GPUs, providing more than 5,800 teraflops of AI performance and 192 gigabytes of GPU memory. They can be configured with NVIDIA AI enterprise or NVIDIA Omniverse inside.
现在转向专业可视化。收入为3.75亿美元,环比增长28%,同比下降24%。Ada架构的升级推动了第二季度的强劲增长,最初在笔记本电脑工作站中推出,第三季度将更新台式机工作站。这将包括功能强大的全新 RTX 系统,配备多达 4 个 NVIDIA RTX 6000 GPU,可提供超过 5,800 万亿次浮点运算的人工智能性能和 192 GB 的 GPU 内存。它们可以在内部配置 NVIDIA AI 企业版或 NVIDIA Omniverse。
We also announced three new desktop workstation GPUs based on the Ada generation. The NVIDIA RTX 5000, 4500 and 4000, offering up to 2x the RT core throughput and up to 2x faster AI training performance compared to the previous generation. In addition to traditional workloads such as 3D design and content creation, new workloads in generative AI, large language model development and data science are expanding the opportunity in pro visualization for our RTX technology.
我们还宣布推出三款基于 Ada 一代的新台式机工作站 GPU。与上一代产品相比,NVIDIA RTX 5000、4500 和 4000 可提供高达 2 倍的 RT 核心吞吐量和高达 2 倍的人工智能训练性能。除了 3D 设计和内容创作等传统工作负载外,生成式 AI、大型语言模型开发和数据科学领域的新工作负载也在扩大我们的 RTX 技术在专业可视化领域的机会。
One of the key themes in Jensen's keynote [indiscernible] earlier this month was the conversion of graphics and AI. This is where NVIDIA Omniverse is positioned. Omniverse is OpenUSD's native platform. OpenUSD is a universal interchange that is quickly becoming the standard for the 3D world, much like HTML is the universal language for the 2D [indiscernible]. Together, Adobe, Apple, Autodesk, Pixar and NVIDIA form the Alliance for OpenUSD. Our mission is to accelerate OpenUSD's development and adoption. We announced new and upcoming Omniverse cloud APIs, including RunUSD and ChatUSD to bring generative AI to OpenUSD workload.
詹森本月早些时候的主题演讲 [难以辨认] 的关键主题之一是图形和人工智能的转换。这就是 NVIDIA Omniverse 的定位。Omniverse 是 OpenUSD 的原生平台。OpenUSD是一种通用交换,正在迅速成为3D世界的标准,就像HTML是2D的通用语言一样 [难以辨认]。Adobe、苹果、Autodesk、Pixar和NVIDIA共同组成了OpenUSD联盟。我们的使命是加快 OpenUSD 的开发和采用。我们宣布了新的和即将推出的Omniverse云API,包括runUSD和ChatUSD,为OpenUSD工作负载带来生成人工智能。
Moving to automotive. Revenue was $253 million, down 15% sequentially and up 15% year-on-year. Solid year-on-year growth was driven by the ramp of self-driving platforms based on [indiscernible] or associated with a number of new energy vehicle makers. The sequential decline reflects lower overall automotive demand, particularly in China. We announced a partnership with MediaTek to bring drivers and passengers new experiences inside the car. MediaTek will develop automotive SoCs and integrate a new product line of NVIDIA's GPU chiplet. The partnership covers a wide range of vehicle segments from luxury to entry level.
转向汽车。收入为2.53亿美元,环比下降15%,同比增长15%。基于 [难以辨认] 或与多家新能源汽车制造商相关的自动驾驶平台的兴起推动了稳健的同比增长。连续下降反映了整体汽车需求的下降,尤其是在中国。我们宣布与联发科建立合作伙伴关系,为驾驶员和乘客带来全新的车内体验。联发科将开发汽车 SoC,并整合 NVIDIA GPU 芯片的新产品线。该合作伙伴关系涵盖了从豪华车到入门级的广泛汽车细分市场。
Moving to the rest of the P&L. GAAP gross margins expanded to 70.1% and non-GAAP gross margin to 71.2%, driven by higher data center sales. Our Data Center products include a significant amount of software and complexity, which is also helping drive our gross margin. Sequential GAAP operating expenses were up 6% and non-GAAP operating expenses were up 5%, primarily reflecting increased compensation and benefits. We returned approximately $3.4 billion to shareholders in the form of share repurchases and cash dividends. Our Board of Directors has just approved an additional $25 billion in stock repurchases to add to our remaining $4 billion of authorization as of the end of Q2.
转向损益的其余部分。在数据中心销售增加的推动下,GAAP毛利率扩大至70.1%,非公认会计准则毛利率扩大至71.2%。我们的数据中心产品包括大量的软件和复杂性,这也有助于提高我们的毛利率。连续的GAAP运营费用增长了6%,非公认会计准则运营费用增长了5%,这主要反映了薪酬和福利的增加。我们以股票回购和现金分红的形式向股东返还了约34亿美元。截至第二季度末,我们的董事会刚刚批准了另外2500亿美元的股票回购,以增加我们剩余的400亿美元授权。
Let me turn to the outlook for the third quarter of fiscal 2024. Demand for our Data Center platform where AI is tremendous and broad-based across industries on customers. Our demand visibility extends into next year. Our supply over the next several quarters will continue to ramp as we lower cycle times and work with our supply partners to add capacity. Additionally, the new L40S GPU will help address the growing demand for many types of workloads from cloud to enterprise.
让我来谈谈2024财年第三季度的展望。对我们的数据中心平台的需求,在这个平台上,人工智能非常庞大,并且在客户身上有广泛的基础。我们的需求可见度将延续到明年。随着我们缩短周期时间并与供应合作伙伴合作增加产能,我们在未来几个季度的供应将继续增加。此外,全新 L40S GPU 将有助于满足从云到企业的多种工作负载不断增长的需求。
For Q3, total revenue is expected to be $16 billion, plus or minus 2%. We expect sequential growth to be driven largely by Data Center with gaming and ProViz also contributing. GAAP and non-GAAP gross margins are expected to be 71.5% and 72.5%, respectively, plus or minus 50 basis points. GAAP and non-GAAP operating expenses are expected to be approximately $2.95 billion and $2 billion, respectively.
第三季度的总收入预计为1600亿美元,正负2%。我们预计,连续增长将在很大程度上由数据中心推动,游戏和ProviZ也将做出贡献。GAAP和非GAAP毛利率预计分别为71.5%和72.5%,正负50个基点。GAAP和非GAAP运营费用预计分别约为29.5亿美元和200亿美元。
GAAP and non-GAAP other income and expenses are expected to be an income of approximately $100 million, excluding gains and losses from non-affiliated investments. GAAP and non-GAAP tax rates are expected to be 14.5%, plus or minus 1%, excluding any discrete items. Further financial details are included in the CFO commentary and other information available on our IR website.
GAAP和非GAAP其他收入和支出预计约为1亿美元,其中不包括非关联投资的损益。GAAP和非GAAP税率预计为14.5%,正负1%,不包括任何离散项目。更多财务细节包含在我们的投资者关系网站上的首席财务官评论和其他信息中。
In closing, let me highlight some upcoming events for the financial community. We will attend the Jefferies Tech Summit on August 30 in Chicago, the Goldman Sachs Conference on September 5 in San Francisco, the Evercore Semiconductor Conference on September 6 as well as the Citi Tech Conference on September 7, both in New York. And the BofA Virtual AI conference on September 11. Our earnings call to discuss the results of our third quarter of fiscal 2024 is scheduled for Tuesday, November 21.
最后,让我重点介绍金融界即将开展的一些活动。我们将参加8月30日在芝加哥举行的杰富瑞科技峰会、9月5日在旧金山举行的高盛会议、9月6日的Evercore半导体会议以及9月7日在纽约举行的花旗科技会议。还有9月11日的美国银行虚拟人工智能会议。我们将讨论2024财年第三季度业绩的财报电话会议定于11月21日星期二举行。
Operator, we will now open the call for questions. Could you please poll for questions for us? Thank you.
接线员,我们现在将开放提问电话。你能为我们投票提问吗?谢谢。
Question-and-Answer Session
问答环节
Operator
操作员
Thank you. [Operator Instructions] We'll take our first question from Matt Ramsay with TD Cowen. Your line is now open.
谢谢。[操作员说明] 我们将从 TD Cowen 的 Matt Ramsay 那里回答第一个问题。您的线路现已开放。
Matt Ramsay
马特·拉姆齐
Yes. Thank you very much. Good afternoon. Obviously, remarkable results. Jensen, I wanted to ask a question of you regarding the really quickly emerging application of large model inference. So I think it's pretty well understood by the majority of investors that you guys have very much a lockdown share of the training market. A lot of the smaller market -- smaller model inference workloads have been done on ASICs or CPUs in the past.
是的。非常感谢。下午好。显然,效果显著。Jensen,我想问你一个关于大型模型推理迅速出现的应用的问题。因此,我认为大多数投资者都很清楚,你们在培训市场中占有很大的封锁份额。过去,许多较小的市场——较小的模型推理工作负载都是在ASIC或CPU上完成的。
And with many of these GPT and other really large models, there's this new workload that's accelerating super-duper quickly on large model inference. And I think your Grace Hopper Superchip products and others are pretty well aligned for that. But could you maybe talk to us about how you're seeing the inference market segment between small model inference and large model inference and how your product portfolio is positioned for that? Thanks.
而且,对于其中许多GPT和其他非常大的模型,这种新的工作负载正在快速加速大型模型推理。而且我认为你的 Grace Hopper Superchip 产品和其他产品在这方面非常一致。但是,你能否和我们谈谈你如何看待小模型推理和大型模型推理之间的推理细分市场,以及你的产品组合在这方面的定位如何?谢谢。
Jensen Huang
黄仁森
Yeah. Thanks a lot. So let's take a quick step back. These large language models are fairly -- are pretty phenomenal. It does several things, of course. It has the ability to understand unstructured language. But at its core, what it has learned is the structure of human language. And it has encoded or within it -- compressed within it a large amount of human knowledge that it has learned by the corpuses that it studied. What happens is, you create these large language models and you create as large as you can, and then you derive from it smaller versions of the model, essentially teacher-student models. It's a process called distillation.
是的。非常感谢。因此,让我们快速退后一步。这些大型语言模型相当——非常了不起。当然,它做了几件事。它能够理解非结构化语言。但从本质上讲,它所学到的是人类语言的结构。它已经编码了或者在里面—— 在里面压缩了它从研究的尸体中学到的大量人类知识。发生的事情是,你创建了这些大型语言模型,然后你创建了尽可能大的语言模型,然后你从中衍生出模型的较小版本,本质上是师生模型。这是一个叫做蒸馏的过程。
And so when you see these smaller models, it's very likely the case that they were derived from or distilled from or learned from larger models, just as you have professors and teachers and students and so on and so forth. And you're going to see this going forward. And so you start from a very large model and it has a large amount of generality and generalization and what's called zero-shot capability. And so for a lot of applications and questions or skills that you haven't trained it specifically on, these large language models miraculously has the capability to perform them. That's what makes it so magical.
因此,当你看到这些较小的模型时,它们很可能是从更大的模型中衍生出来的、提炼出来的,或者从中学到的,就像你有教授、老师和学生等等一样。而且你会看到这种情况向前发展。因此,你从一个非常大的模型开始,它有大量的概括性和概括性,以及所谓的零射能力。因此,对于许多你还没有专门训练过的应用程序、问题或技能,这些大型语言模型奇迹般地有能力执行它们。这就是它如此神奇的原因。
On the other hand, you would like to have these capabilities in all kinds of computing devices, and so what you do is you distill them down. These smaller models might have excellent capabilities on a particular skill, but they don't generalize as well. They don't have what is called as good zero-shot capabilities. And so they all have their own unique capabilities, but you start from very large models.
另一方面,你想在各种计算设备中拥有这些功能,所以你要做的就是把它们提炼出来。这些较小的模型在特定技能上可能具有出色的能力,但它们不能一概而论。他们没有所谓的良好零射能力。因此,它们都有自己独特的功能,但你要从非常大的模型开始。
Operator
操作员
Okay. Next, we'll go to Vivek Arya with BofA Securities. Your line is now open.
好吧。接下来,我们将和美国银行证券一起去Vivek Arya。您的线路现已开放。
Vivek Arya
Vivek Arya
Thank you. Just had a quick clarification and a question. Colette, if you could please clarify how much incremental supply do you expect to come online in the next year? You think it's up 20%, 30%, 40%, 50%? So just any sense of how much supply because you said it's growing every quarter.
谢谢。刚才有一个简短的澄清和一个问题。Colette,如果你能澄清一下你预计明年上线的供应量将增加多少?你认为它上涨了20%、30%、40%、50%?所以只要知道供应量有多大,因为你说供应量每个季度都在增长。
And then Jensen, the question for you is, when we look at the overall hyperscaler spending, that buy is not really growing that much. So what is giving you the confidence that they can continue to carve out more of that pie for generative AI? Just give us your sense of how sustainable is this demand as we look over the next one to two years? So if I take your implied Q3 outlook of Data Center, $12 billion, $13 billion, what does that say about how many servers are already AI accelerated? Where is that going? So just give some confidence that the growth that you are seeing is sustainable into the next one to two years.
然后詹森,你的问题是,当我们看超大规模的总体支出时,购买量并没有真正增长那么多。那么,是什么让你有信心他们可以继续为生成人工智能开辟更多资源呢?在我们展望未来一到两年时,请告诉我们你对这种需求的可持续性的看法?那么,如果我以你暗示的第三季度数据中心展望为例,即1200万美元、1300万美元,那么这说明有多少服务器已经在人工智能加速呢?那要去哪里?因此,只要有一定的信心,相信你所看到的增长将在未来一到两年内持续下去。
Colette Kress
Colette Kress
So thanks for that question regarding our supply. Yes, we do expect to continue increasing ramping our supply over the next quarters as well as into next fiscal year. In terms of percent, it's not something that we have here. It is a work across so many different suppliers, so many different parts of building an HGX and many of our other new products that are coming to market. But we are very pleased with both the support that we have with our suppliers and the long time that we have spent with them improving their supply.
因此,感谢您提出有关我们供应的问题。是的,我们确实预计在接下来的几个季度以及下一财年将继续增加供应。就百分比而言,这不是我们这里的数字。这项工作涉及许多不同的供应商、建造HGX的许多不同部分以及我们即将上市的许多其他新产品。但是,我们对供应商提供的支持以及我们花了很长时间与他们一起改善供应感到非常满意。
Jensen Huang
黄仁森
The world has something along the lines of about $1 trillion worth of data centers installed, in the cloud, in enterprise and otherwise. And that $1 trillion of data centers is in the process of transitioning into accelerated computing and generative AI. We're seeing two simultaneous platform shifts at the same time. One is accelerated computing. And the reason for that is because it's the most cost-effective, most energy effective and the most performant way of doing computing now.
世界上已经安装了价值约1万亿美元的云端、企业和其他数据中心。而这1万亿美元的数据中心正在向加速计算和生成式人工智能过渡。我们看到两个平台同时移动。一是加速计算。之所以如此,是因为它是目前最具成本效益、最节能、性能最高的计算方式。
So what you're seeing, and then all of a sudden, enabled by generative AI, enabled by accelerated compute and generative AI came along. And this incredible application now gives everyone two reasons to transition to do a platform shift from general purpose computing, the classical way of doing computing, to this new way of doing computing, accelerated computing. It's about $1 trillion worth of data centers, call it, $0.25 trillion of capital spend each year.
因此,你所看到的,然后突然之间,由生成式人工智能提供支持,由加速计算和生成人工智能提供支持。现在,这个令人难以置信的应用程序为每个人提供了两个理由,让他们从通用计算(传统的计算方式)过渡到这种新的计算方式,即加速计算。这相当于每年价值约1万亿美元的数据中心,即0.25万亿美元的资本支出。
You're seeing the data centers around the world are taking that capital spend and focusing it on the two most important trends of computing today, accelerated computing and generative AI. And so I think this is not a near-term thing. This is a long-term industry transition and we're seeing these two platform shifts happening at the same time.
你会看到世界各地的数据中心正在利用这笔资本支出,并将其重点放在当今计算的两个最重要的趋势上,即加速计算和生成式人工智能。所以我认为这不是短期的事情。这是一个长期的行业转型,我们看到这两个平台的转变同时发生。
Operator
操作员
Next, we go to Stacy Rasgon with Bernstein Research. Your line is open.
接下来,我们去找伯恩斯坦研究中心的 Stacy Rasgon。您的线路已打开。
Stacy Rasgon
斯泰西·拉斯贡
Hi, guys. Thanks for taking my question. I was wondering, Colette, if you could tell me like how much of Data Center in the quarter, maybe even the guide is like systems versus GPU, like DGX versus just the H100? What I'm really trying to get at is, how much is like pricing or content or however you want to define that [indiscernible] versus units actually driving the growth going forward. Can you give us any color around that?
嗨,伙计们。感谢你回答我的问题。我想知道,Colette,你能否告诉我本季度有多少数据中心,也许指南就像系统对比 GPU,比如 DGX 对比 H100?我真正想说的是,在多大程度上像定价或内容,或者你想如何定义 [难以辨认] 与实际推动未来增长的单位相比。你能给我们点什么颜色吗?
Colette Kress
Colette Kress
Sure, Stacy. Let me help. Within the quarter, our HGX systems were a very significant part of our Data Center as well as our Data Center growth that we had seen. Those systems include our HGX of our Hopper architecture, but also our Ampere architecture. Yes, we are still selling both of these architectures in the market. Now when you think about that, what does that mean from both the systems as a unit, of course, is growing quite substantially, and that is driving in terms of the revenue increases. So both of these things are the drivers of the revenue inside Data Center.
当然,Stacy。让我来帮忙。在本季度内,我们的HGX系统是我们数据中心的重要组成部分,也是我们所看到的数据中心增长的重要组成部分。这些系统包括我们的 Hopper 架构的 HGX,还有我们的 Ampere 架构。是的,我们仍在市场上销售这两种架构。现在,当你考虑这个问题时,这两个系统作为一个单元意味着什么,当然,它们正在大幅增长,这推动了收入的增长。因此,这两个因素都是数据中心内部收入的驱动力。
Our DGXs are always a portion of additional systems that we will sell. Those are great opportunities for enterprise customers and many other different types of customers that we're seeing even in our consumer Internet companies. The importance there is also coming together with software that we sell with our DGXs, but that's a portion of our sales that we're doing. The rest of the GPUs, we have new GPUs coming to market that we talk about the L40S, and they will add continued growth going forward. But again, the largest driver of our revenue within this last quarter was definitely the HGX system.
我们的 DGX 始终是我们将要出售的其他系统的一部分。对于企业客户和许多其他不同类型的客户来说,这些都是绝佳的机会,即使在我们的消费互联网公司中也是如此。将我们与 DGX 一起出售的软件整合在一起也很重要,但这是我们正在进行的销售的一部分。其余的GPU,我们有新的GPU上市,我们谈论的是L40S,它们将在未来增加持续增长。但同样,上个季度我们收入的最大驱动力肯定是HGX系统。
Jensen Huang
黄仁森
And Stacy, if I could just add something. You say it’s H100 and I know you know what your mental image in your mind. But the H100 is 35,000 parts, 70 pounds, nearly 1 trillion transistors in combination. Takes a robot to build – well, many robots to build because it’s 70 pounds to lift. And it takes a supercomputer to test a supercomputer. And so these things are technology marvels, and the manufacturing of them is really intensive. And so I think we call it H100 as if it’s a chip that comes off of a fab, but H100s go out really as HGXs sent to the world’s hyperscalers and they’re really, really quite large system components, if you will.
还有 Stacy,如果我能补充点东西的话。你说现在是 H100 我知道你知道自己在脑海中的心理形象。但是H100是35,000个零件,70磅,总共将近1万亿个晶体管。需要一个机器人来建造 —— 好吧,要建造很多机器人,因为它要举起 70 磅。而且,测试超级计算机需要一台超级计算机。因此,这些东西是技术奇迹,它们的制造非常密集。所以我想我们称之为 H100,就好像它是一款从晶圆厂生产的芯片一样,但是 H100 实际上是 HGX 发往世界超大规模公司的样子,如果你愿意的话,它们确实是相当大的系统组件。
Operator
操作员
Next, we go to Mark Lipacis with Jefferies. Your line is now open.
接下来,我们去找杰富瑞集团的 Mark Lipacis。您的线路现已开放。
Mark Lipacis
马克·利帕西斯
Hi. Thanks for taking my question and congrats on the success. Jensen, it seems like a key part of the success -- your success in the market is delivering the software ecosystem along with the chip and the hardware platform. And I had a two-part question on this. I was wondering if you could just help us understand the evolution of your software ecosystem, the critical elements. And is there a way to quantify your lead on this dimension like how many person years you've invested in building it? And then part two, I was wondering if you would care to share with us your view on the -- what percentage of the value of the NVIDIA platform is hardware differentiation versus software differentiation? Thank you.
你好。感谢你回答我的问题,恭喜你成功了。詹森,这似乎是成功的关键部分——你在市场上的成功在于提供软件生态系统以及芯片和硬件平台。我对此有一个由两部分组成的问题。我想知道你能否帮助我们了解你的软件生态系统的演变,即关键要素。有没有办法量化你在这个维度上的领先优势,比如你花了多少人年时间来建造这个维度?然后是第二部分,我想知道你是否愿意和我们分享你对 NVIDIA 平台价值的看法 —— 硬件差异化与软件差异化在NVIDIA平台的价值中所占的百分比是多少?谢谢。
A – Jensen Huang
A — Jensen Huang
Yeah, Mark, I really appreciate the question. Let me see if I could use some metrics, so we have a run time called AI Enterprise. This is one part of our software stack. And this is, if you will, the run time that just about every company uses for the end-to-end of machine learning from data processing, the training of any model that you like to do on any framework you'd like to do, the inference and the deployment, the scaling it out into a data center. It could be a scale-out for a hyperscale data center. It could be a scale-out for enterprise data center, for example, on VMware.
是的,马克,我真的很感激这个问题。让我看看能否使用一些指标,所以我们有一个名为 AI Enterprise 的运行时间。这是我们软件堆栈的一部分。如果你愿意的话,这就是几乎每个公司用于端到端机器学习的运行时间,包括数据处理、在你想做的任何框架上训练你想做的任何模型、推理和部署、将其扩展到数据中心。它可能是超大规模数据中心的横向扩展。它可能是企业数据中心的横向扩展,例如,在 VMware 上。
You can do this on any of our GPUs. We have hundreds of millions of GPUs in the field and millions of GPUs in the cloud and just about every single cloud. And it runs in a single GPU configuration as well as multi-GPU per compute or multi-node. It also has multiple sessions or multiple computing instances per GPU. So from multiple instances per GPU to multiple GPUs, multiple nodes to entire data center scale. So this run time called NVIDIA AI enterprise has something like 4,500 software packages, software libraries and has something like 10,000 dependencies among each other.
你可以在我们的任何 GPU 上执行此操作。我们在现场有数亿个 GPU,云端有数百万个 GPU,几乎每个云都有。而且它可以在单个 GPU 配置中运行,也可以为每个计算或多节点运行多 GPU。每个 GPU 也有多个会话或多个计算实例。因此,从每个 GPU 的多个实例到多个 GPU,从多个节点到整个数据中心规模。因此,这个名为 NVIDIA AI Enterprise 的运行时有大约 4,500 个软件包、软件库,彼此之间有大约 10,000 个依赖关系。
And that run time is, as I mentioned, continuously updated and optimized for our installed base for our stack. And that's just one example of what it would take to get accelerated computing to work. The number of code combinations and type of application combinations is really quite insane. And it's taken us two decades to get here. But what I would characterize as probably our -- the elements of our company, if you will, are several. I would say number 1 is architecture.
正如我所提到的,该运行时间会不断更新和优化,以适应我们堆栈的安装量。而这只是让加速计算发挥作用需要做的事情的一个例子。代码组合的数量和应用程序组合的类型实在是太疯狂了。我们花了二十年的时间才到达这里。但是我想描述的可能是我们的 —— 如果你愿意的话,我们公司的要素有几个。我想说第一是建筑。
The flexibility, the versatility and the performance of our architecture makes it possible for us to do all the things that I just said, from data processing to training to inference, for preprocessing of the data before you do the inference to the post processing of the data, tokenizing of languages so that you could then train with it. The amount of -- the workflow is much more intense than just training or inference. But anyways, that's where we'll focus and it's fine. But when people actually use these computing systems, it's quite -- requires a lot of applications. And so the combination of our architecture makes it possible for us to deliver the lowest cost ownership. And the reason for that is because we accelerate so many different things.
我们架构的灵活性、多功能性和性能使我们能够完成我刚才所说的所有事情,从数据处理到训练再到推理,在推理之前对数据进行预处理,再到数据的后期处理,对语言进行标记化以便你可以用它进行训练。量 —— 工作流程要紧张得多,而不仅仅是训练或推理。但不管怎样,这就是我们要关注的地方,没关系。但是,当人们真正使用这些计算系统时,确实如此 —— 需要大量的应用程序。因此,我们的架构组合使我们能够实现最低的拥有成本。其原因是因为我们加速了很多不同的事情。
The second characteristic of our company is the installed base. You have to ask yourself, why is it that all the software developers come to our platform? And the reason for that is because software developers seek a large installed base so that they can reach the largest number of end users, so that they could build a business or get a return on the investments that they make.
我们公司的第二个特点是安装量。你得问问自己,为什么所有的软件开发人员都来我们的平台?其原因是因为软件开发人员寻求庞大的安装基础,以便他们能够接触到最大数量的最终用户,这样他们就可以建立业务或从所做的投资中获得回报。
And then the third characteristic is reach. We're in the cloud today, both for public cloud, public-facing cloud because we have so many customers that use -- so many developers and customers that use our platform. CSPs are delighted to put it up in the cloud. They use it for internal consumption to develop and train and to operate recommender systems or search or data processing engines and whatnot all the way to training and inference. And so we're in the cloud, we're in enterprise.
然后第三个特征是触及率。如今,我们处于云端,无论是公共云,还是面向公众的云,因为我们有太多的客户在使用,很多开发人员和客户都在使用我们的平台。CSP 很高兴将其放在云端。他们将其用于内部消费,用于开发和训练以及操作推荐系统、搜索或数据处理引擎等等,一直到训练和推理。因此,我们在云中,我们在企业中。
Yesterday, we had a very big announcement. It's really worthwhile to take a look at that. VMware is the operating system of the world's enterprise. And we've been working together for several years now, and we're going to bring together -- together, we're going to bring generative AI to the world's enterprises all the way out to the edge. And so reach is another reason. And because of reach, all of the world's system makers are anxious to put NVIDIA's platform in their systems. And so we have a very broad distribution from all of the world's OEMs and ODMs and so on and so forth because of our reach.
昨天,我们有一个非常重要的公告。看一看真的很值得。VMware 是全球企业的操作系统。我们已经合作了好几年,我们将团结起来—— 团结起来,我们将把生成式人工智能带给世界各地的企业,一直到边缘。因此,覆盖面是另一个原因。而且,由于覆盖范围的原因,世界上所有的系统制造商都急于将NVIDIA的平台应用到他们的系统中。因此,由于我们的影响力,我们拥有来自全球所有原始设备制造商和原始设计制造商等的非常广泛的分布。
And then lastly, because of our scale and velocity, we were able to sustain this really complex stack of software and hardware, networking and compute and across all of these different usage models and different computing environments. And we're able to do all this while accelerating the velocity of our engineering. It seems like we're introducing a new architecture every two years. Now we're introducing a new architecture, a new product just about every six months. And so these properties make it possible for the ecosystem to build their company and their business on top of us. And so those in combination makes us special.
最后,由于我们的规模和速度,我们能够在所有这些不同的使用模式和不同的计算环境中维持这个非常复杂的软件和硬件、网络和计算堆栈。而且,我们能够在加快工程速度的同时完成所有这些工作。看来我们每两年推出一次新的架构。现在,我们要推出一种新架构,大约每六个月推出一款新产品。因此,这些属性使生态系统有可能将他们的公司和业务建立在我们之上。因此,这些结合起来使我们与众不同。
Operator
操作员
Next, we'll go to Atif Malik with Citi. Your line is open.
接下来,我们将和花旗一起去阿蒂夫·马利克。您的线路已打开。
Atif Malik
阿蒂夫·马利克
Hi. Thank you for taking my question. Great job on results and outlook. Colette, I have a question on the core L40S that you guys talked about. Any idea how much of the supply tightness can L40S help with? And if you can talk about the incremental profitability or gross margin contribution from this product? Thank you.
你好。感谢你回答我的问题。在业绩和前景方面做得很好。Colette,我有一个关于核心 L40S 的问题,你们谈过。知道L40S能在多大程度上缓解供应紧张吗?你能否谈谈这个产品的增量盈利能力或毛利率贡献?谢谢。
Jensen Huang
黄仁森
Yeah, Atif. Let me take that for you. The L40S is really designed for a different type of application. H100 is designed for large-scale language models and processing just very large models and a great deal of data. And so that's not L40S' focus. L40S' focus is to be able to fine-tune models, fine-tune pretrained models, and it'll do that incredibly well. It has a transform engine. It's got a lot of performance. You can get multiple GPUs in a server. It's designed for hyperscale scale-out, meaning it's easy to install L40S servers into the world's hyperscale data centers. It comes in a standard rack, standard server, and everything about it is standard and so it's easy to install.
是的,Atif。让我来帮你拿吧。L40S 实际上是为不同类型的应用而设计的。H100 专为大型语言模型而设计,仅处理非常大的模型和大量数据。因此,这不是 L40 的重点。L40S的重点是能够微调模型,微调预训练的模型,它会做得非常好。它有一个变换引擎。它的性能很高。你可以在一台服务器中安装多个 GPU。它专为超大规模横向扩展而设计,这意味着可以轻松地将 L40S 服务器安装到全球超大规模数据中心中。它采用标准机架和标准服务器,其所有内容均为标准配置,因此易于安装。
L40S also is with the software stack around it and along with BlueField-3 and all the work that we did with VMware and the work that we did with Snowflakes and ServiceNow and so many other enterprise partners. L40S is designed for the world's enterprise IT systems. And that's the reason why HPE, Dell, and Lenovo and some 20 other system makers building about 100 different configurations of enterprise servers are going to work with us to take generative AI to the world's enterprise. And so L40S is really designed for a different type of scale-out, if you will. It's, of course, large language models. It's, of course, generative AI, but it's a different use case. And so the L40S is going to -- is off to a great start and the world's enterprise and hyperscalers are really clamoring to get L40S deployed.
L40S 还包括围绕它的软件堆栈、BlueField-3、我们在 VMware 上所做的所有工作以及我们与 Snowflakes 和 ServiceNow 以及许多其他企业合作伙伴所做的工作。L40S 专为全球企业 IT 系统而设计。这就是为什么慧与、戴尔和联想以及其他约20家构建约100种不同企业服务器配置的系统制造商将与我们合作,将生成式人工智能带入全球企业的原因。因此,如果你愿意的话,L40S 实际上是为另一种类型的横向扩展而设计的。当然,这是大型语言模型。当然,这是生成式人工智能,但它是一个不同的用例。因此,L40S将如此 —— 有了一个良好的开端,全世界的企业和超大规模企业都在大声疾呼要部署 L40S。
Operator
操作员
Next, we'll go to Joe Moore with Morgan Stanley. Your line is open.
接下来,我们将和摩根士丹利一起去找乔·摩尔。您的线路已打开。
Joseph Moore
约瑟夫摩尔
Great. Thank you. I guess the thing about these numbers that's so remarkable to me is the amount of demand that remains unfulfilled, talking to some of your customers. As good as these numbers are, you sort of more than tripled your revenue in a couple of quarters. There's a demand, in some cases, for multiples of what people are getting. So can you talk about that? How much unfulfilled demand do you think there is? And you talked about visibility extending into next year. Do you have line of sight into when you get to see supply-demand equilibrium here?
太棒了。谢谢。我想这些数字对我来说非常引人注目的是,在与你的一些客户交谈时,需求量仍未得到满足。尽管这些数字不错,但你的收入在几个季度内增长了三倍多。在某些情况下,人们的需求是人们所得到的东西的倍数。那你能谈谈吗?你认为有多少需求未得到满足?你还谈到了能见度将延续到明年。你知道什么时候能看到这里的供需平衡吗?
Jensen Huang
黄仁森
Yeah. We have excellent visibility through the year and into next year. And we're already planning the next-generation infrastructure with the leading CSPs and data center builders. The demand – easiest way to think about the demand, the world is transitioning from general-purpose computing to accelerated computing. That's the easiest way to think about the demand. The best way for companies to increase their throughput, improve their energy efficiency, improve their cost efficiency is to divert their capital budget to accelerated computing and generative AI. Because by doing that, you're going to offload so much workload off of the CPUs, but the available CPUs is -- in your data center will get boosted.
是的。我们全年和明年的知名度都非常好。而且,我们已经在与领先的通信服务提供商和数据中心建设商一起规划下一代基础架构。需求——考虑需求的最简单方法,世界正在从通用计算向加速计算过渡。这是考虑需求的最简单方法。企业提高吞吐量、提高能源效率、提高成本效益的最佳方法是将资本预算转用于加速计算和生成式人工智能。因为这样做,你将从 CPU 上卸下这么多工作负载,但可用的 CPU 确实如此 —— 数据中心中的可用CPU将得到提升。
And so what you're seeing companies do now is recognizing this -- the tipping point here, recognizing the beginning of this transition and diverting their capital investment to accelerated computing and generative AI. And so that's probably the easiest way to think about the opportunity ahead of us. This isn't a singular application that is driving the demand, but this is a new computing platform, if you will, a new computing transition that's happening. And data centers all over the world are responding to this and shifting in a broad-based way.
因此,你看到的公司现在所做的就是意识到这一点 —— 这里的转折点,认识到这种过渡的开始,并将资本投资转移到加速计算和生成人工智能上。因此,这可能是思考摆在我们面前的机会的最简单方法。这不是推动需求的单一应用程序,但如果你愿意的话,这是一个新的计算平台,正在发生新的计算过渡。世界各地的数据中心都在对此做出回应,并以广泛的方式进行转变。
Operator
操作员
Next, we go to Toshiya Hari with Goldman Sachs. Your line is now open.
接下来,我们和高盛一起去看哈里俊也。您的线路现已开放。
Toshiya Hari
Hari Toshiya
Hi. Thank you for taking the question. I had one quick clarification question for Colette and then another one for Jensen. Colette, I think last quarter, you had said CSPs were about 40% of your Data Center revenue, consumer Internet at 30%, enterprise 30%. Based on your remarks, it sounded like CSPs and consumer Internet may have been a larger percentage of your business. If you can kind of clarify that or confirm that, that would be super helpful.
你好。感谢你回答这个问题。我有一个简短的澄清问题要给 Colette 然后又有一个问题要给 Jensen。Colette,我想上个季度,你曾说过,通信服务提供商约占你的数据中心收入的40%,消费互联网占30%,企业占30%。根据你的评论,听起来通信服务提供商和消费者互联网在你的业务中所占的比例可能更大。如果你能澄清这一点或确认这一点,那将非常有帮助。
And then Jensen, a question for you. Given your position as the key enabler of AI, the breadth of engagements and the visibility you have into customer projects, I'm curious how confident you are that there will be enough applications or use cases for your customers to generate a reasonable return on their investments. I guess I ask the question because there is a concern out there that there could be a bit of a pause in your demand profile in the out years. Curious if there's enough breadth and depth there to support a sustained increase in your Data Center business going forward. Thank you.
然后是 Jensen,一个问题要问你。考虑到你作为人工智能关键推动者的地位、参与的广度以及你对客户项目的知名度,我很好奇你对有足够的应用程序或用例让你的客户产生合理的投资回报有多有信心。我想我之所以问这个问题,是因为有人担心在过去的几年中,你的需求状况可能会有一些停顿。想知道那里是否有足够的广度和深度来支持您的数据中心业务未来持续增长。谢谢。
Colette Kress
Colette Kress
Okay. So thank you, Toshiya, on the question regarding our types of customers that we have in our Data Center business. And we look at it in terms of combining our compute as well as our networking together. Our CSPs, our large CSPs are contributing a little bit more than 50% of our revenue within Q2. And the next largest category will be our consumer Internet companies. And then the last piece of that will be our enterprise and high performance computing.
好吧。因此,Toshiya,感谢你提出有关我们在数据中心业务中拥有的客户类型的问题。我们从将计算和网络结合在一起的角度来看待它。在第二季度,我们的通信服务提供商、大型通信服务提供商对我们收入的贡献略高于我们收入的50%。第二大类别将是我们的消费互联网公司。然后最后一部分将是我们的企业和高性能计算。
Jensen Huang
黄仁森
Toshi, I'm reluctant to guess about the future and so I'll answer the question from the first principle of computer science perspective. It is recognized for some time now that general purpose computing is just not and brute forcing general purpose computing. Using general purpose computing at scale is no longer the best way to go forward. It's too energy costly, it's too expensive, and the performance of the applications are too slow.
Toshi,我不愿猜测未来,所以我会从计算机科学的第一原理的角度回答这个问题。一段时间以来,人们已经认识到,通用计算不是暴力强迫的通用计算。大规模使用通用计算不再是向前迈进的最佳方式。它的能源成本太高,太昂贵了,而且应用程序的性能也太慢了。
And finally, the world has a new way of doing it. It's called accelerated computing and what kicked it into turbocharge is generative AI. But accelerated computing could be used for all kinds of different applications that's already in the data center. And by using it, you offload the CPUs. You save a ton of money in order of magnitude, in cost and order of magnitude and energy and the throughput is higher and that's what the industry is really responding to.
最后,世界有了一种新的方法来做到这一点。它被称为加速计算,推动它进入涡轮增压状态的是生成式人工智能。但是加速计算可以用于数据中心中已经存在的各种不同应用程序。通过使用它,你可以卸载 CPU 的负载。在数量级、成本、数量级和能源方面,您可以节省大量资金,吞吐量更高,这才是该行业真正做出的回应。
Going forward, the best way to invest in the data center is to divert the capital investment from general purpose computing and focus it on generative AI and accelerated computing. Generative AI provides a new way of generating productivity, a new way of generating new services to offer to your customers, and accelerated computing helps you save money and save power. And the number of applications is, well, tons. Lots of developers, lots of applications, lots of libraries. It's ready to be deployed.
展望未来,投资数据中心的最佳方法是将资本投资从通用计算中转移出来,将其重点放在生成式人工智能和加速计算上。生成式人工智能提供了一种提高生产力的新方式,一种生成新服务以提供给客户的新方式,而加速计算可以帮助您节省资金和节省电力。而且,应用程序的数量很多。很多开发人员,很多应用程序,很多库。它已准备就绪,可以部署了。
And so I think the data centers around the world recognize this, that this is the best way to deploy resources, deploy capital going forward for data centers. This is true for the world's clouds and you're seeing a whole crop of new GPU specialty -- GPU specialized cloud service providers. One of the famous ones is CoreWeave and they're doing incredibly well. But you're seeing the regional GPU specialist service providers all over the world now. And it's because they all recognize the same thing, that the best way to invest their capital going forward is to put it into accelerated computing and generative AI.
因此,我认为世界各地的数据中心都意识到这一点,这是为数据中心部署资源和部署资金的最佳方式。世界上的云就是这样,你会看到一大堆新的GPU专业——GPU专业的云服务提供商。其中一个著名的是CoreWeave,他们的表现非常出色。但是你现在看到的区域性 GPU 专业服务提供商遍布世界各地。正是因为他们都认识同样的事情,所以未来投资资金的最佳方法是将其投入到加速计算和生成人工智能中。
We're also seeing that enterprises want to do that. But in order for enterprises to do it, you have to support the management system, the operating system, the security and software-defined data center approach of enterprises, and that's all VMware. And we've been working several years with VMware to make it possible for VMware to support not just the virtualization of CPUs but a virtualization of GPUs as well as the distributed computing capabilities of GPUs, supporting NVIDIA's BlueField for high-performance networking.
我们还看到,企业希望这样做。但是,为了让企业做到这一点,你必须支持企业的管理系统、操作系统、安全和软件定义的数据中心方法,仅此而已 VMware。多年来,我们一直在与VMware合作,使VMware不仅能够支持CPU的虚拟化,还支持GPU的虚拟化以及GPU的分布式计算功能,从而支持NVIDIA的BlueField进行高性能联网。
And all of the generative AI libraries that we've been working on is now going to be offered as a special SKU by VMware's sales force, which is, as we all know, quite large because they reach some several hundred thousand VMware customers around the world. And this new SKU is going to be called VMware Private AI Foundation. And this will be a new SKU that makes it possible for enterprises. And in combination with HP, Dell, and Lenovo's new server offerings based on L40S, any enterprise could have a state-of-the-art AI data center and be able to engage generative AI.
我们一直在研究的所有生成式人工智能库现在都将由VMware的销售人员作为特殊的SKU提供,众所周知,销售人员规模相当大,因为它们覆盖了全球约数十万VMware客户。这个新的 SKU 将被命名为 VMware Private AI 基金会。这将是一个新的SKU,使企业成为可能。再加上惠普、戴尔和联想基于L40S的新服务器产品,任何企业都可以拥有最先进的人工智能数据中心,并能够使用生成式人工智能。
And so I think the answer to that question is hard to predict exactly what's going to happen quarter-to-quarter. But I think the trend is very, very clear now that we're seeing a platform shift.
因此,我认为这个问题的答案很难准确预测每个季度会发生什么。但我认为,既然我们看到了平台的转变,趋势非常非常明显。
Operator
操作员
Next, we'll go to Timothy Arcuri with UBS. Your line is now open.
接下来,我们将和瑞银一起去找蒂莫西·阿库里。您的线路现已开放。
Timothy Arcuri
蒂莫西·阿库里
Thanks a lot. Can you talk about the attach rate of your networking solutions to your -- to the compute that you're shipping? In other words, is like half of your compute shipping with your networking solutions more than half, less than half? And is this something that maybe you can use to prioritize allocation of the GPUs? Thank you.
非常感谢。你能否谈谈你的网络解决方案对你的 —— 你所运送的计算机 —— 的附着率?换句话说,您的网络解决方案大约有一半的计算出货量是否超过一半,不到一半?也许你可以用这个来确定GPU分配的优先顺序吗?谢谢。
Jensen Huang
黄仁森
Well, working backwards, we don't use that to prioritize the allocation of our GPUs. We let customers decide what networking they would like to use. And for the customers that are building very large infrastructure, InfiniBand is, I hate to say it, kind of a no-brainer. And the reason for that because the efficiency of InfiniBand is so significant, some 10%, 15%, 20% higher throughput for $1 billion infrastructure translates to enormous savings. Basically, the networking is free.
好吧,向后看,我们不会用它来确定我们的 GPU 分配的优先级。我们让客户决定他们想要使用哪种网络。对于正在建设非常大型基础设施的客户来说,我不想这么说,InfiniBand有点不费吹灰之力。其原因是,由于InfiniBand的效率非常高,100亿美元基础设施的吞吐量提高了约10%、15%、20%,这意味着可以节省大量资金。基本上,网络是免费的。
And so, if you have a single application, if you will, infrastructure or it’s largely dedicated to large language models or large AI systems, InfiniBand is really a terrific choice. However, if you’re hosting for a lot of different users and Ethernet is really core to the way you manage your data center, we have an excellent solution there that we had just recently announced and it’s called Spectrum-X. Well, we’re going to bring the capabilities, if you will, not all of it, but some of it, of the capabilities of InfiniBand to Ethernet so that we can also, within the environment of Ethernet, allow you to – enable you to get excellent generative AI capabilities.
因此,如果你有一个应用程序(如果你愿意的话)、基础架构,或者它主要用于大型语言模型或大型人工智能系统,那么InfiniBand确实是一个不错的选择。但是,如果你为许多不同的用户托管,而以太网确实是你管理数据中心的核心,那么我们有一个很好的解决方案,我们最近刚刚宣布了这个解决方案,它叫做 Spectrum-X。好吧,如果你愿意的话,我们会将InfiniBand的全部功能,而是部分功能,带到以太网,这样我们也可以在以太网环境中允许你——使你能够获得出色的生成人工智能功能。
So Spectrum-X is just ramping now. It requires BlueField-3 and it supports both our Spectrum-2 and Spectrum-3 Ethernet switches. And the additional performance is really spectacular. BlueField-3 makes it possible and a whole bunch of software that goes along with it. BlueField, as all of you know, is a project really dear to my heart, and it’s off to just a tremendous start. I think it’s a home run. This is the concept of in-network computing and putting a lot of software in the computing fabric is being realized with BlueField-3, and it is going to be a home run.
所以 Spectrum-X 现在才刚刚起步。它需要 BlueField-3,并且同时支持我们的 Spectrum-2 和 Spectrum-3 以太网交换机。而且额外的表现真的很壮观。BlueField-3 使之成为可能,还有一大堆随之而来的软件。众所周知,BlueField是一个我非常珍视的项目,而且开局不错。我认为这是本垒打。这是网络内计算的概念,BlueField-3正在实现将大量软件放入计算结构,这将是本垒打。
Operator
操作员
Our final question comes from the line of Ben Reitzes with Melius. Your line is now open.
我们的最后一个问题来自 Ben Reitzes 和 Melius 的台词。您的线路现已开放。
Benjamin Reitzes
Benjamin Reitzes
Hi. Good afternoon. Good evening. Thank you for the question, putting me in here. My question is with regard to DGX Cloud. Can you talk about the reception that you're seeing and how the momentum is going? And then Colette, can you also talk about your software business? What is the run rate right now and the materiality of that business? And it does seem like it's already helping margins a bit. Thank you very much.
你好。下午好。晚上好。谢谢你的提问,让我来这里。我的问题是关于 DGX Cloud 的。你能谈谈你所看到的接待情况以及势头如何吗?然后 Colette,你还能谈谈你的软件业务吗?现在的运行率是多少,该业务的重要性如何?而且看来它已经在一定程度上提高了利润率。非常感谢。
Jensen Huang
黄仁森
DGX Cloud's strategy, let me start there. DGX Cloud's strategy is to achieve several things: number one, to enable a really close partnership between us and the world's CSPs. We recognize that many of our -- we work with some 30,000 companies around the world. 15,000 of them are startups. Thousands of them are generative AI companies and the fastest-growing segment, of course, is generative AI. We're working with all of the world's AI start-ups. And ultimately, they would like to be able to land in one of the world's leading clouds. And so we built DGX Cloud as a footprint inside the world's leading clouds so that we could simultaneously work with all of our AI partners and help blend them easily in one of our cloud partners.
DGX Cloud 的策略,让我从这里开始。DGX Cloud的战略是实现几件事:第一,在我们和全球的CSP之间建立非常紧密的伙伴关系。我们认识到,我们的许多公司——我们与全球约3万家公司合作。其中15,000家是初创公司。其中成千上万是生成式人工智能公司,当然,增长最快的细分市场是生成式人工智能。我们正在与世界上所有的人工智能初创企业合作。最终,他们希望能够降落在世界领先的云层之一。因此,我们将 DGX Cloud 打造成了全球领先云端的足迹,这样我们就可以同时与所有的人工智能合作伙伴合作,帮助他们轻松融入我们的云合作伙伴之一。
The second benefit is that it allows our CSPs and ourselves to work really closely together to improve the performance of hyperscale clouds, which is historically designed for multi-tenancy and not designed for high-performance distributed computing like generative AI. And so to be able to work closely architecturally to have our engineers work hand in hand to improve the networking performance and the computing performance has been really powerful, really terrific.
第二个好处是,它使我们的通信服务提供商和我们自己能够非常紧密地合作,以提高超大规模云的性能。超大规模云历来是为多租户设计的,而不是为生成式人工智能等高性能分布式计算而设计的。因此,能够在架构上紧密合作,让我们的工程师携手合作,提高网络性能和计算性能,真的很强大,真的很棒。
And then thirdly, of course, NVIDIA uses very large infrastructures ourselves. And our self-driving car team, our NVIDIA research team, our generative AI team, our language model team, the amount of infrastructure that we need is quite significant. And none of our optimizing compilers are possible without our DGX systems. Even compilers these days require AI, and optimizing software and infrastructure software requires AI to even develop. It's been well publicized that our engineering uses AI to design our chips.
第三,当然,NVIDIA自己使用非常大的基础架构。而我们的自动驾驶汽车团队、我们的 NVIDIA 研究团队、我们的生成人工智能团队、我们的语言模型团队,我们需要的基础设施数量相当可观。如果没有我们的 DGX 系统,我们的任何优化编译器都不可能实现。如今,即使是编译器也需要人工智能,而优化软件和基础架构软件甚至需要人工智能才能开发。众所周知,我们的工程使用人工智能来设计我们的芯片。
And so the internal -- our own consumption of AI, our robotics team, so on and so forth, Omniverse teams, so on and so forth, all needs AI. And so our internal consumption is quite large as well, and we land that in DGX Cloud. And so DGX Cloud has multiple use cases, multiple drivers, and it's been off to just an enormous success. And our CSPs love it, the developers love it and our own internal engineers are clamoring to have more of it. And it's a great way for us to engage and work closely with all of the AI ecosystem around the world.
因此,内部 —— 我们自己对人工智能的消费、我们的机器人团队等等、Omniverse 团队等等,都需要人工智能。因此,我们的内部消耗也相当大,我们将其存放在DGX Cloud中。因此,DGX Cloud有多个用例,多个驱动因素,并且取得了巨大的成功。而且我们的 CSP 喜欢它,开发人员也喜欢它,我们自己的内部工程师也在大声疾呼要拥有更多。这是我们与全球所有人工智能生态系统互动和密切合作的绝佳方式。
Colette Kress
Colette Kress
And let's see if I can answer your question regarding our software revenue. In part of our opening remarks that we made as well, remember, software is a part of almost all of our products, whether they're our Data Center products, GPU systems or any of our products within gaming and our future automotive products. You're correct, we're also selling it in a standalone business. And that stand-alone software continues to grow where we are providing both the software services, upgrades across there as well.
让我们看看我能否回答你关于我们软件收入的问题。在我们也发表的开幕词中,请记住,软件是我们几乎所有产品的一部分,无论是我们的数据中心产品、GPU系统,还是我们在游戏领域的任何产品和未来的汽车产品。你说得对,我们也在独立企业中出售它。而且,这种独立软件继续增长,我们既提供软件服务,也提供升级。
Now we're seeing, at this point, probably hundreds of millions of dollars annually for our software business, and we are looking at NVIDIA AI enterprise to be included with many of the products that we're selling, such as our DGX, such as our PCIe versions of our H100. And I think we're going to see more availability even with our CSP marketplaces. So we're off to a great start, and I do believe we'll see this continue to grow going forward.
现在,我们看到我们的软件业务每年可能有数亿美元的收入,我们正在考虑将NVIDIA人工智能企业版纳入我们销售的许多产品中,例如我们的DGX,例如我们的PCIe版本的H100。而且我认为,即使在我们的 CSP 市场上,我们也会看到更多的可用性。因此,我们有了一个良好的开端,我确实相信未来我们会看到这种情况继续增长。
Operator
操作员
And that does conclude today's question-and-answer session. I'll turn the call back over to Jensen Huang for any additional or closing remarks.
今天的问答环节到此结束。我会把电话转回给黄仁森,听听任何补充或闭幕词。
Jensen Huang
黄仁森
A new computing era has begun. The industry is simultaneously going through 2 platform transitions, accelerated computing and generative AI. Data centers are making a platform shift from general purpose to accelerated computing. The $1 trillion of global data centers will transition to accelerated computing to achieve an order of magnitude better performance, energy efficiency and cost. Accelerated computing enabled generative AI, which is now driving a platform shift in software and enabling new, never-before possible applications. Together, accelerated computing and generative AI are driving a broad-based computer industry platform shift.
一个新的计算时代已经开始。该行业正在同时经历两个平台转型,即加速计算和生成式人工智能。数据中心正在实现平台从通用计算向加速计算的转变。1万亿美元的全球数据中心将过渡到加速计算,以实现更高的性能、能源效率和成本。加速计算使生成式人工智能成为可能,它现在正在推动软件的平台转变,并实现前所未有的新应用。加速计算和生成式人工智能共同推动了基础广泛的计算机行业平台转变。
Our demand is tremendous. We are significantly expanding our production capacity. Supply will substantially increase for the rest of this year and next year. NVIDIA has been preparing for this for over two decades and has created a new computing platform that the world’s industry -- world’s industries can build upon. What makes NVIDIA special are: one, architecture. NVIDIA accelerates everything from data processing, training, inference, every AI model, real-time speech to computer vision, and giant recommenders to vector databases. The performance and versatility of our architecture translates to the lowest data center TCO and best energy efficiency.
我们的需求是巨大的。我们正在显著扩大我们的生产能力。今年剩余时间和明年的供应量将大幅增加。二十多年来,NVIDIA一直在为此做准备,并创建了一个新的计算平台,世界各行各业都可以在此基础上再接再厉。NVIDIA的特别之处在于:第一,架构。NVIDIA 加速了从数据处理、训练、推理、每个 AI 模型、实时语音到计算机视觉、巨型推荐器到矢量数据库的所有内容。我们架构的性能和多功能性可转化为最低的数据中心总拥有成本和最佳能效。
Two, installed base. NVIDIA has hundreds of millions of CUDA-compatible GPUs worldwide. Developers need a large installed base to reach end users and grow their business. NVIDIA is the developer’s preferred platform. More developers create more applications that make NVIDIA more valuable for customers. Three, reach. NVIDIA is in clouds, enterprise data centers, industrial edge, PCs, workstations, instruments and robotics. Each has fundamentally unique computing models and ecosystems. System suppliers like OEMs, computer OEMs can confidently invest in NVIDIA because we offer significant market demand and reach. Scale and velocity. NVIDIA has achieved significant scale and is 100% invested in accelerated computing and generative AI. Our ecosystem partners can trust that we have the expertise, focus and scale to deliver a strong road map and reach to help them grow.
二、安装基地。英伟达在全球拥有数亿个兼容 CUDA 的 GPU。开发人员需要庞大的安装量来吸引最终用户并发展他们的业务。NVIDIA 是开发者的首选平台。更多的开发人员创建了更多的应用程序,从而让 NVIDIA 对客户来说更有价值。第三,伸手可及。NVIDIA 涉足云端、企业数据中心、工业边缘、PC、工作站、仪器和机器人技术。每个模型都有本质上独特的计算模型和生态系统。像原始设备制造商、计算机原始设备制造商这样的系统供应商可以放心地投资NVIDIA,因为我们提供了巨大的市场需求和覆盖范围。比例和速度。NVIDIA 已经实现了可观的规模,并将 100% 投资于加速计算和生成式人工智能。我们的生态系统合作伙伴可以相信,我们拥有专业知识、专注力和规模,可以提供强有力的路线图和影响力,帮助他们发展。
We are accelerating because of the additive results of these capabilities. We’re upgrading and adding new products about every six months versus every two years to address the expanding universe of generative AI. While we increased the output of H100 for training and inference of large language models, we’re ramping up our new L40S universal GPU for scale, for cloud scale-out and enterprise servers. Spectrum-X, which consists of our Ethernet switch, BlueField-3 Super NIC and software helps customers who want the best possible AI performance on Ethernet infrastructures. Customers are already working on next-generation accelerated computing and generative AI with our Grace Hopper.
我们之所以加速,是因为这些功能的累加成果。我们大约每六个月升级和添加一次新产品,而不是每两年更新一次,以应对不断扩大的生成人工智能领域。虽然我们增加了 H100 用于训练和推断大型语言模型的输出,但我们正在为扩展、云横向扩展和企业服务器增加全新 L40S 通用 GPU。Spectrum-X 由我们的以太网交换机、BlueField-3 Super NIC 和软件组成,可帮助希望在以太网基础设施上获得最佳人工智能性能的客户。客户已经在使用我们的 Grace Hopper 开发下一代加速计算和生成式人工智能。
We’re extending NVIDIA AI to the world’s enterprises that demand generative AI but with the model privacy, security and sovereignty. Together with the world’s leading enterprise IT companies, Accenture, Adobe, Getty, Hugging Face, Snowflake, ServiceNow, VMware and WPP and our enterprise system partners, Dell, HPE, and Lenovo, we are bringing generative AI to the world’s enterprise. We’re building NVIDIA Omniverse to digitalize and enable the world’s multi-trillion dollar heavy industries to use generative AI to automate how they build and operate physical assets and achieve greater productivity. Generative AI starts in the cloud, but the most significant opportunities are in the world’s largest industries, where companies can realize trillions of dollars of productivity gains. It is an exciting time for NVIDIA, our customers, partners and the entire ecosystem to drive this generational shift in computing. We look forward to updating you on our progress next quarter.
我们正在将 NVIDIA AI 扩展到全球需要生成人工智能但具有隐私、安全和主权模型的企业。我们与全球领先的企业IT公司埃森哲、Adobe、Getty、Hugging Face、Snowflake、ServiceNow、VMware和WPP以及我们的企业系统合作伙伴戴尔、慧与和联想一起,为全球企业带来生成式人工智能。我们正在构建 NVIDIA Omniverse,以实现数字化,使全球价值数万亿美元的重工业能够使用生成人工智能来自动构建和运营有形资产,从而提高生产力。生成式人工智能始于云端,但最重要的机会来自世界上最大的行业,在这些行业中,公司可以实现数万亿美元的生产率提高。对于 NVIDIA、我们的客户、合作伙伴和整个生态系统来说,这是一个激动人心的时刻,推动计算领域的这一代际转变。我们期待在下个季度向你通报最新进展情况。
Operator
操作员
This concludes today's conference call. You may now disconnect.
今天的电话会议到此结束。您现在可以断开连接。
(Tips:This transcript is converted by recording, we will do our best, but cannot fully guarantee the accuracy of the conversion, it is for reference only.)
(提示:此笔录是通过录音转换的,我们会尽力而为,但不能完全保证转换的准确性,仅供参考。)