Nvidia's new B200 GPU boasts up to 20 petaflops of FP4 performance with 208 billion transistors, while the GB200 "superchip," which pairs two B200 GPUs with a single Grace CPU, can deliver up to 30 times the performance for Large Language Model (LLM) inference workloads compared to previous models, and is significantly more efficient, reducing cost and energy by up to 25 times than an H100. For training an AI model with 1.8 trillion parameters, where 8,000 Hopper GPUs once needed 15 megawatts of power, only 2,000 Blackwell GPUs are now required, using just four megawatts. The GB200 offers seven times the performance and four times the training speed on a GPT-3 LLM benchmark with 175 billion parameters. Nvidia also highlights that its new technology significantly reduces the time spent on inter-GPU communication, allowing for more actual computing.
frank Crane_3546 : Love it. Nvidia is a hyper market offering platform chip solution software SaaS solutions!!!!!!