Design & Reuse

AI in 2025: Chips, robots, and the race for scale

Dec. 23, 2025 – 

2025 saw a dramatic change in the fortunes of AI chip makers, with Nvidia in particular benefitting. The growth in AI by companies such as OpenAI catapulted Nvidia into the top slot as the world’s most valuable company, while others fell by the wayside.

 

At CES in January Nvidia announced its desktop AI machine, then called Digits. This uses the GB10 chip with the Grace processor and Blackwell graphics processing unit (GPU) and 128GB of unified memory to run models with up to 200bn parameters. This evolved through the year to become the DGX Spark, and the first version of this was delivered to Elon Musk, CEO of Tesla and SpaceX, in September. 

Other early recipients of DGX Spark included Anaconda, Cadence Design Systems, ComfyUI Docker, Google, Hugging Face, JetBrains, LM Studio, Meta, Microsoft, Ollama and Roboflow who are validating and optimizing their tools, software and models for the desktop unit. The 240W design is also being made by Acer, ASUS, Dell Technologies, GIGABYTE, HP, Lenovo and MSI.

The successor to Blackwell, called Rubin CPX, is also emerging to power the next generation of AI factories. Announced in September, the chip is scheduled to be available by the end of 2026. Combined with the Vera CPU and 100TBytes of high bandwidth memory (HBM4), Rubin CPX will provide 8 exaflops of AI performance in a single rack in the datacentre, 7.5x more  performance than the GB300 NVL72 systems that are shipping today.

 

“The Vera Rubin platform will mark another leap in the frontier of AI computing — introducing both the next-generation Rubin GPU and a new category of processors called CPX,” said Jensen Huang, founder and CEO of NVIDIA. “Just as RTX revolutionized graphics and physical AI, Rubin CPX is the first CUDA GPU purpose-built for massive-context AI, where models reason across millions of tokens of knowledge at once.”

Rubin CPX delivers up to 30 petaflops of 4bit compute with 128GB of cost-efficient GDDR7 memory. It also integrates video decoders and encoders, as well as long-context inference processing, in a single chip for higher performance in long-format applications such as video search and high-quality generative video.

Click here to read more