Nvidia CEO Jensen Huang arrives to attend the opening ceremony of Siliconware Precision Industries Co. (SPIL)’s Tan Ke Plant site in Taichung, Taiwan Jan. 16, 2025.
Nvidia announced new chips for building and deploying artificial intelligence models at its annual GTC conference on Tuesday.
CEO Jensen Huang revealed Blackwell Ultra, a family of chips shipping in the second half of this year, as well as Vera Rubin, the company’s next-generation graphics processing unit, or GPU, that is expected to ship in 2026.
Nvidia’s sales are up more than sixfold since its business was transformed by the release of OpenAI’s ChatGPT in late 2022. That’s because its “big GPUs” have most of the market for developing advanced AI, a process called training.
Software developers and investors are closely watching the company’s new chips to see if they offer enough additional performance and efficiency to convince the company’s biggest end customers — cloud companies including Microsoft, Google and Amazon — to continue spending billions of dollars to build data centers based around Nvidia chips.
“This last year is where almost the entire world got involved. The computational requirement, the scaling law of AI, is more resilient, and in fact, is hyper-accelerated,” Huang said.
Tuesday’s announcements are also a test of Nvidia’s new annual release cadence. The company is striving to announce new chip families on an every-year basis. Before the AI boom, Nvidia released new chip architectures every other year.
The GTC conference in San Jose, California, is also a show of strength for Nvidia.
The event, Nvidia’s second in-person conference since the pandemic, is expected to have 25,000 attendees and hundreds of companies discussing the ways they use the company’s hardware for AI. That includes Waymo, Microsoft and Ford, among others. General Motors also announced that it will use Nvidia’s service for its next-generation vehicles.
The chip architecture after Rubin will be named after physicist Richard Feynman, Nvidia said on Tuesday, continuing its tradition of naming chip families after scientists. Nvidia’s Feynman chips are expected to be available in 2028, according to a slide displayed by Huang.
Nvidia will also showcase its other products and services at the event.
For example, Nvidia announced new laptops and desktops using its chips, including two AI-focused PCs called DGX Spark and DGX Station that will be able to run large AI models such as Llama or DeepSeek. The company also announced updates to its networking parts for tying hundreds or thousands of GPUs together so they work as one, as well as a software package called Dynamo that helps users get the most out of their chips.
Jensen Huang, co-founder and chief executive officer of Nvidia Corp., speaks during the Nvidia GPU Technology Conference (GTC) in San Jose, California, US, on Tuesday, March 18, 2025.
Nvidia expects to start shipping systems on its next-generation GPU family in the second half of 2026.
The system has two main components: a CPU, called Vera, and a new GPU design, called Rubin. It’s named after astronomer Vera Rubin.
Vera is Nvidia’s first custom CPU design, the company said, and it’s based on a core design they’ve named Olympus.
Previously when it needed CPUs, Nvidia used an off-the-shelf design from Arm. Companies that have developed custom Arm core designs, such as Qualcomm and Apple, say that they can be more tailored and unlock better performance.
The custom Vera design will be twice as fast as the CPU used in last year’s Grace Blackwell chips, the company said.
When paired with Vera, Rubin can manage 50 petaflops while doing inference, more than double the 20 petaflops for the company’s current Blackwell chips. Rubin can also support as much as 288 gigabytes of fast memory, which is one of the core specs that AI developers watch.
Nvidia is also making a change to what it calls a GPU. Rubin is actually two GPUs, Nvidia said.
The Blackwell GPU, which is currently on the market, is actually two separate chips that were assembled together and made to work as one chip.
Starting with Rubin, Nvidia will say that when it combines two or more dies to make a single chip, it will refer to them as separate GPUs. In the second half of 2027, Nvidia plans to release a “Rubin Next” chip that combines four dies to make a single chip, doubling the speed of Rubin, and it will refer to that as four GPUs.
Nvidia said that will come in a rack called Vera Rubin NVL144. Previous versions of Nvidia’s rack were called NVL72.
Jensen Huang, co-founder and chief executive officer of Nvidia Corp., speaks during the Nvidia GPU Technology Conference (GTC) in San Jose, California, US, on Tuesday, March 18, 2025.
Nvidia also announced new versions of its Blackwell family of chips that it calls Blackwell Ultra.
That chip will be able to produce more tokens per second, which means that the chip can generate more content in the same amount of time as its predecessor, the company said in a briefing.
Nvidia says that means that cloud providers can use Blackwell Ultra to offer a premium AI service for time-sensitive applications, allowing them to make as much as 50 times the revenue from the new chips as the Hopper generation, which shipped in 2023.
Blackwell Ultra will come in a version with two paired to an Nvidia Arm CPU, called GB300, and a version with just the GPU, called B300. It will also come in versions with eight GPUs in a single server blade and a rack version with 72 Blackwell chips.
The top four cloud companies have deployed three times the number of Blackwell chips as Hopper chips, Nvidia said.
China’s DeepSeek R1 model may have scared Nvidia investors when it was released in January, but Nvidia has embraced the software. The chipmaker will use the model to benchmark several of its new products.
Many AI observers said that DeepSeek’s model, which reportedly required fewer chips than models made in the U.S., threatened Nvidia’s business.
But Huang said earlier this year that DeepSeek was actually a good sign for Nvidia. That’s because DeepSeek uses a process called “reasoning,” which requires more computing power to provide users better answers.
The new Blackwell Ultra chips are better for reasoning models, Nvidia said.
It’s developed its chips to more efficiently do inference, so when new reasoning models require more computing power at the time of deployment, Nvidia’s chips will be able to handle it.
“In the last 2 to 3 years, a major breakthrough happened, a fundamental advance in artificial intelligence happened. We call it agentic AI,” Huang said. “It can reason about how to answer or how to solve a problem.”
WATCH: Nvidia kicks off its GTC Conference: The Committee debate how to trade it
Page 2
Jensen Huang, co-founder and chief executive officer of Nvidia Corp., speaks during the Nvidia GPU Technology Conference (GTC) in San Jose, California, US, on Tuesday, March 18, 2025.
General Motors and Nvidia have agreed to a strategic collaboration that includes the automaker using several products and artificial intelligence services from the tech giant for its next-generation vehicles, advanced driver-assistance systems and factories.
The companies on Tuesday announced that the new initiatives include building custom artificial intelligence systems using Nvidia compute platforms, including “Omniverse with Cosmos,” for optimizing GM’s factory planning and robotics.
The Detroit automaker also said it will use “Nvidia Drive AGX” for “in-vehicle hardware for future advanced driver-assistance systems and in-cabin enhanced safety driving experiences.”
GM declined to disclose a cost for the new tools with Nvidia. The tech company has been attempting to diversify its automotive business, which has notably included substantial work in data centers and GPUs.
“The era of physical AI is here, and together with GM, we’re transforming transportation, from vehicles to the factories where they’re made,” Jensen Huang, Nvidia founder and CEO, said in a release. “We are thrilled to partner with GM to build AI systems tailored to their vision, craft and know-how.”
GM has been using Nvidia graphics processing units, or GPUs, for training AI models across various areas of its business, including simulation and validation. The new business expands to in-vehicle hardware, automotive plant design and operations, the companies said.
The automaker also had been testing Nvidia’s Omniverse since at least 2022. Some of GM’s testing was in designing a “digital twin,” or replica, of its new design center and processes to assist virtual vehicle development. It also acted as a single digital environment for employees to work and collaborate in, according to a video last year featuring GM for Nvidia’s GTC developer conference in 2023.
Jensen Huang, co-founder and chief executive officer of Nvidia Corp., speaks during the Nvidia GPU Technology Conference (GTC) in San Jose, California, US, on Tuesday, March 18, 2025.
Nvidia anticipated it would strike a deal with GM mid-last year for Omniverse, according to an internal company email viewed by CNBC. At that time, two sources with GM signaled the automaker wasn’t sure Nvidia’s software and GPUs were worth the high cost compared with other companies.
It wasn’t immediately clear what sealed the deal for GM. But since that time, both companies have experienced increased competition from China and uncertain regulatory changes such as tariffs. GM’s stock is off roughly 8% during in 2025, while Nvidia is off about 12% this year.
“AI not only optimizes manufacturing processes and accelerates virtual testing but also helps us build smarter vehicles while empowering our workforce to focus on craftsmanship,” GM CEO Mary Barra said in Nvidia’s release. “By merging technology with human ingenuity, we unlock new levels of innovation in vehicle manufacturing and beyond.”
Stock chart icon
GM and Nvidia stock prices
The companies announced the new initiatives in connection with Nvidia’s GTC AI conference this week in California.
Nvidia describes Omniverse as a platform for “developing and deploying physically based industrial digitalization applications.” It’s essentially connecting a physical environment with a digital, or software, world to optimize processes using a “digital twin” of a physical environment such as a GM design facility or plant.
Users of Nvidia’s Omniverse have included BMW, Amazon Robotics and Samsung, Rev Lebaredian, Nvidia vice president of Omniverse and simulation technology, said during a media briefing a year ago. He said the company was licensing Omniverse for $4,500 per GPU, per year.
It’s unclear how many GPUs GM will need. But given the amount of robotics, sensors and other systems needed to operate a modern assembly plant, it would likely be quite a bit.
More than 20 other automakers have used Nvidia’s “system on a chip” hardware in the central computing units of their smart vehicles, including Mercedes Benz, Volvo, Audi, Volkswagen and BYD, according to an industry equity research note from Jeffries in November 2023.
In recent years, Nvidia has seen soaring demand for its GPUs, which are used for everything from bitcoin mining to AI inference and training.