Nvidia prepares for exponential growth in AI inference | Computer Weekly
Briefly

Nvidia prepares for exponential growth in AI inference | Computer Weekly
"CEO Jensen Huang said the company continued to see growth in artificial intelligence (AI) workloads, which require the high performance graphics processor units (GPUs) that Nvidia specialises in. According to Huang, AI inference is scaling exponentially due to advancements in pre-training, post-training, and reasoning capabilities. He said that that inference is becoming increasingly complex as AI systems now "read, think and reason" before generating answers. Huang claimed this exponential growth in computation requirements is driving demand for Nvidia's platforms."
"Huang said: " Customer interest in NVLink Fusion continues to grow. We announced a strategic collaboration with Fujitsu in October where we will integrate Fujitsu's CPUs and Nvidia GPUs via NVL Fusion, connecting our large ecosystems. We also announced a collaboration with Intel to develop multiple generations of custom datacentre and PC products connecting Nvidia and Intel's ecosystems using NVLink.""
"Among the areas Nvidia sees as a big differentiator is the power to watt metric, which is directly linked to the running costs of high-performance compute in datacentres. Discussing the breakthroughs in GPUs, he said: "In each generation, from Ampere to Hopper, from Hopper to Blackwell, Blackwell to Rubin, our part of the datacentre increases." He said that each generation of GPU sees a major increase in performance, but that performance needs to be delivered within the power limits of the datacentre."
Nvidia reported $57bn revenue for Q3 2026, with datacentre revenue at $51bn, a 66% year-over-year increase. AI workloads are driving demand for high-performance GPUs as AI inference scales exponentially through improved pre-training, post-training and reasoning capabilities. NVLink AI networking revenue rose 162% to $8.2bn, supported by collaborations with Fujitsu and Intel to integrate CPUs and GPUs and connect large ecosystems. Nvidia highlights power-to-watt as a key differentiator tied to datacentre operating costs. GPU generations deliver major performance gains, but those gains must be realized within fixed datacentre power constraints.
Read at ComputerWeekly.com
Unable to calculate read time
[
|
]