Refresh
Nvidia preshow kicks off CES festivities
While Jensen Huang hasn't taken to the stage yet, Nvidia has a 'pre-game' show, where its showing a manner of applications that use Nvidia tech, complete with an interview panel ahead of time. While it's a neat warm-up, don't expect anything too staggering while things are still kicking off.
Huang is set to take the stage at 1 p.m. PT, so there's not too long to wait until we get to the real meat of today's keynote.
— Sayem Ahmed
The preshow is over: T-minus 3 minutes until the keynote begins
The preshow is now over, and Huang is expected to take the stage in three minutes to kick things off in earnest. Being rolled onto the stage is a high-resolution screen, with some semi-dramatic music playing, and we'll expect to hear more shortly. While this is fun, what are the odds that the music is AI-generated?
There may be some expectation that Huang will emerge carrying some kind of object. We've seen him roll out with robots and even a shield made of chips before; surely the company has the cash for more than just a new leather jacket.
— Sayem Ahmed
Jensen Huang takes to the stage
Slightly later than expected, the show has now officially begun, with Nvidia CEO Jensen Huang taking the stage to lay out what 2026 and beyond might hold for Nvidia. Expect Nvidia to press on topics like AI, data centers, and robotics. While it might not be quite as exciting as the launch of a new consumer graphics card, whatever is announced over the course of the keynote will likely set the tone for the rest of the industry to follow for the rest of the year.
— Sayem Ahmed
Inventing a new world with AI
100 trillion dollars of industry investment is shifting over to AI, or so Jensen Huang claims. The CEO states that R&D budgets are now pouring in, perhaps to cool down claims of circular funding in the AI industry. The CEO is now shifting the conversation into a history lesson for AI models, reiterating that Agentic models like Cursor are a focus for 2026 and beyond.
— Sayem Ahmed
AI tops leaderboards across the industry
Nvidia's models are topping leaderboards, with the hopes that it can make developers utilize and create AI agents in order to research, models, and more easily. With reasoning capabilities, AI models no longer have to understand everything in a single day.
Jensen mentioned Perplecity as an example of how it would create an AI reasoning chain to create a multi-model, making just about anything fit a certain task, which has empowered AI startups — all thanks to Nvidia's technology, of course.
— Jeffrey Kampman
Nvidia Brev: Creating closed agentic sandboxes
Using a new tool called Nvidia Brev, the company has demoed an environement that can hook up to a DGX spark inside a closed environment for a variety of AI applications, or connecting it to frontier models via HuggingFace. A couple of examples were given, including integrating it with robotics.
"Isn't that incredible?", Jensen remarks, before mentioning that just a few years ago, it would have been impossible. Thanks to using a variety of applications, named an Agentic framework, Brev creates the architecture of applications across a closed sandbox.
— Sayem Ahmed
Full-stack Phsyical AI: Compute becomes data
Jensen is explaining how several computers, connected through Nvidia Omniverse, can create a full-stack AI platform through training (GB300), Simulation (RTX Pro), and inference (labelled THOR). By utilizing multiple models aligned with language, compute power has become data, Jensen says.
Synthetic data, grounded by physics, can be used in a Cosmos foundation model to become physically-based, digestible data to train a new AI platform. This could curtail the requirement for real-world training data for Physical AI platforms, aligning elements like trajectory predictions into real-world, usable data for training.
— Jeff Kampman
Nvidia Alpamayo: A thinking, reasoning AI model for vehicles
Powered by the Cosmos AI model, Nvidia has created Alpamayo, an AI model used for autonomous vehicles. It can reason every action: From the actions it's about to take, how it reasoned that action, and then developing a trajectory.
Huang claims that it's based on a large amount of human training and supplemented by Cosmos' synthetic data. The keynote has now shifted to a real-world example of the model working in action, with a car (in San Francisco) taking turns and making decisions on the fly. Huang notes that it's impossible to get every single possible scenario, but breaking these tasks down can be understood by the AI model if it reasons through it.
— Jeff Kampman
More Alpamayo - coming to Mercedes Benz CLA
"Not only does your car drive as you would expect it to drive, it reasons about every scenario and tells you what it's about to do."
The first AV car powered by Alpamayo is the 2025 Mercedes-Benz CLA, shipping to the U.S. in Q1. It has been rated as the world's safest car by NCAP.
"Use some Nvidia wherever you can"
Jensen says that in the next 10 years, a high percentage of the world's cars will be autonomous. Stating we can all agree that this market has very much arrived.
IT'S ROBOT TIME
The moment you've all been waiting for, it's time for Jensen's robot hour! Not just for entertainment, Nvidia says that the principles that underpin its Alpamayo stack for autonomous vehicles will also inform other AI industries, including robotics. This is great cinema.
"Nobody's as cute as you guys"
Caterpillar and Uber Eats, as well as Boston Dynamics, join Jensen on stage to showcase the different types of robots that Nvidia's AI stack is being use to train.
2026-01-05T22:08:46.159Z
Insane Demand for AI computing
Jensen says the demand for AI is "skyrocketing," with model size growing by 10x parameters per year, test time scaling needs increasing 5x per year, and token cost becoming 10x cheaper per year. The latter signals the race that everyone is trying to get to the next level, which means it's a computing problem.
"We have to advance the state of computation every single year."
"An order of magnitude every single year"
Vera CPU in all its glory, with 88 Olympus Custom cores, 227 Billion Transistors. Bottom right, two dejected robots leave the stage, no longer the center of attention.
Jensen says Nvidia has a rule: "No new generation should have more than one or two chips change." However, "the number of transistors we get year after year after year can't possibly keep up with the models, with 5x more tokens per year generated, it can't keep up with the cost decline of the tokens." Nvidia can no longer keep up with the demands of AI unless it innovates all of the chips, the entire stack, all at the same time.
"I think this weighs a couple of hundred pounds"
Vera Rubin vs Blackwell
Here's how Vera Rubin stacks up against Blackwell. Jensen tells another water-based joke: "It weighs two and a half tons today, because we forgot to drain the water out of it." A much better pop than his previous attempt.
So how well does Vera Rubin perform?
As you can see, quite well. According to Nvidia, Vera Rubin throughput is through the roof compared to Blackwell, using 1/4th fewer GPUs in a time to train test of 100T tokens in one month. Rubin factory throughput is around 10x that of Blackwell, and with just 1/10th token cost.

1 week ago
16







English (US) ·