pwshub.com

Blackwell will land in Q4, Nvidia CEO assures AI faithful

Nvidia CEO Jensen Huang has attempted to quell concerns over the reported late arrival of the Blackwell GPU architecture, and the lack of ROI from AI investments.

"Demand is so great that delivery of our components and our technology and our infrastructure and software is really emotional for people because it directly affects their revenues, it directly affects their competitiveness," Huang explained, according to a transcript of remarks he made at the Goldman Sachs Tech Conference on Wednesday. "It's really tense. We've got a lot of responsibility on our shoulders and we're trying the best we can."

The comments follow reports that Nvidia's next-generation Blackwell accelerators won't ship in the second half of 2024, as Huang has previously promised. The GPU giant's admission of a manufacturing defect – which necessitated a mask change – during its Q2 earnings call last month hasn't helped this perception. However, speaking with Goldman Sachs's Toshiya Hari on Wednesday, Huang reiterated that Blackwell chips were already in full production and would begin shipping in calendar Q4.

Unveiled at Nvidia's GTC conference last northern spring, the GPU architecture promises between 2.5x and 5x higher performance and more than twice the memory capacity and bandwidth of the H100-class devices it replaces. At the time, Nvidia said the chips would ship sometime in the second half of the year.

Despite Huang's reassurance that Blackwell will ship this year, talk of delays has sent Nvidia's share price on a roller coaster ride – made more chaotic by disputed reports that the GPU giant had been subpoenaed by the DoJ and faces a patent suit brought by DPU vendor Xockets.

According to Huang, demand for Blackwell parts has exceeded that for the previous-generation Hopper products which debuted in 2022 – before ChatGPT's arrival made generative AI a must-have.

Huang told the conference that extra demand appears to be the source of many customers' frustrations.

"Everybody wants to be first and everybody wants to be most … the intensity is really, really quite extraordinary," he said.

Accelerating ROI

Huang also addressed concerns about the ROI associated with the pricey GPU systems powering the AI boom.

From a hardware standpoint, Huang's argument boils down to this: the performance gains of GPU acceleration far outweigh the higher infrastructure costs.

"Spark is probably the most used data processing engine in the world today. If you use Spark and you accelerate it, it's not unusual to see a 20:1 speed-up," he claimed, adding that even if that infrastructure costs twice as much, you're still looking at a 10x savings.

According to Huang, this also extends to generative AI. "The return on that is fantastic because the demand is so great that every dollar that they [service providers] spend with us translates to $5 worth of rentals."

However, as we've previously reported, the ROI on the applications and services built on this infrastructure remains far fuzzier – and the long-term practicality of dedicated AI accelerators, including GPUs, is up for debate.

Addressing AI use cases, Huang was keen to highlight his own firm's use of custom AI code assistants. "I think the days of every line of code being written by software engineers, those are completely over."

Huang also touted the application of generative AI on computer graphics. "We compute one pixel, we infer the other 32," he explained – an apparent reference to Nvidia’s DLSS tech, which uses frame generation to boost frame rates in video games.

Technologies like these, Huang argued, will also be critical for the success of autonomous vehicles, robotics, digital biology, and other emerging fields.

  • Oracle boasts zettascale 'AI supercomputer,' just don't ask about precision
  • Amazon to pour £8B into UK datacenters through to 2028
  • Mainframes aren't dead, they're just learning AI tricks
  • We're in the brute force phase of AI – once it ends, demand for GPUs will too

Densified, vertically integrated datacenters

While Huang remains confident the return on investment from generative AI technologies will justify the extreme cost of the hardware required to train and deploy it, he also suggested smarter datacenter design could help drive down costs.

"When you want to build this AI computer people say words like super-cluster, infrastructure, supercomputer for good reason – because it's not a chip, it's not a computer per se. We're building entire datacenters," Huang noted in apparent reference to Nvidia's modular cluster designs, which it calls SuperPODs.

Accelerated computing, Huang explained, allows for a massive amount of compute to be condensed into a single system – which is why he says Nvidia can get away with charging millions of dollars per rack. "It replaces thousands of nodes."

However, Huang made the case that putting these incredibly dense systems – as much as 120 kilowatts per rack – into conventional datacenters is less than ideal.

"These giant datacenters are super inefficient because they're filled with air, and air is a lousy conductor of [heat]," he explained. "What we want to do is take that few, call it 50, 100, or 200 megawatt datacenter which is sprawling, and you densify it into a really, really small datacenter."

Smaller datacenters can take advantage of liquid cooling – which, as we've previously discussed, is often a more efficient way to cool systems.

How successful Nvidia will be at driving this datacenter modernization remains to be seen. But it's worth noting that with Blackwell, its top-specced parts are designed to be cooled by liquids. ®

Source: theregister.com

Related stories
2 days ago - Data centres gobble up roughly 2% of global electricity, which translates to around 1% of energy-related greenhouse gas emissions. Streaming Netflix, storing stuff in the cloud, and meeting up on Zoom are just some of the online...
2 weeks ago - A prominent leaker claims Nvidia's upcoming RTX 5000 series graphics cards will consume significantly more power than their predecessors. The rumors might not represent final TDPs, so readers should take the information with a grain of...
3 weeks ago - Recent comments from the chairman of Chinese PC manufacturer Hasee suggest that a mainstream mobile GPU in Nvidia's upcoming Blackwell lineup will draw significantly less power than its direct predecessor – the notebook GeForce RTX 4060....
3 weeks ago - Nvidia is determined to redefine the AI landscape with its Blackwell platform, positioning it as a comprehensive ecosystem that goes beyond traditional GPU capabilities. Nvidia will showcase the setup and configuration of its Blackwell...
1 month ago - Nvidia has reportedly invested some $10 billion developing the Blackwell platform – an effort involving around 25,000 people. With all the performance packed into a single Blackwell GPU, it's no surprise these products command...
Other stories
8 minutes ago - Experts at the Netherlands Institute for Radio Astronomy (ASTRON) claim that second-generation, or "V2," Mini Starlink satellites emit interference that is a staggering 32 times stronger than that from previous models. Director Jessica...
8 minutes ago - The PKfail incident shocked the computer industry, exposing a deeply hidden flaw within the core of modern firmware infrastructure. The researchers who uncovered the issue have returned with new data, offering a more realistic assessment...
8 minutes ago - Nighttime anxiety can really mess up your ability to sleep at night. Here's what you can do about it right now.
8 minutes ago - With spectacular visuals and incredible combat, I cannot wait for Veilguard to launch on Oct. 31.
8 minutes ago - Finding the perfect pair of glasses is difficult, but here's how to do so while considering your face shape, skin tone, lifestyle and personality.