Unknown

0 views11015 WordsCopy TextShare
Unknown
Video Transcript:
[Music] n n [Music] a [Music] oh [Applause] a [Music] [Music] [Music] [Music] [Music] [Music] [Music] [Music] [Music] [Music] [Music] [Music] [Music] [Music] [Music] [Music] [Music] [Music] [Music] [Music] [Music] [Music] n [Music] [Music] [Applause] [Music] p n [Music] n [Music] [Applause] n a [Music] [Music] C [Music] [Music] [Music] [Music] [Music] [Music] [Music] for e e e e [Music] [Music] [Music] [Music] [Music] [Music] [Music] [Music] [Music] w [Music] [Applause] [Music] [Applause] [Music] [Music] [Applause] [Music] [Applause] [Music] [Applause] [Music] [Applause] [Music] [Music] [Music] be [Music] [Music] e e and when technology and Humanity intersect the answer is anything
that's because Tech doesn't just solve challenges it transforms them into opportunities it helps us move smarter live healthier and experience the world in ways we never thought possible we are not just here for a tech event we're here to connect to solve to discover together Tech isn't just advancing it's uniting bringing us closer to an autonomous future connecting us to Better Care making life more connected more Dynamic and more human today's challenges demand bold Solutions and CES is where they start to take shape breakthroughs and sustainability advancements to help feed our growing world this week
isn't just a stage for breakthroughs it's the spark for Discovery every screen every pixel every part of the tech you will see here showcases the extraordinary potential of human Ingenuity meeting technological power now we begin this celebration of what connects us has the power to solve our greatest challenges and offers endless possibil ities we're yet to discover Right Here Right Now the world is watching so let's dive [Music] in ladies and gentlemen welcome to CES [Music] 2025 I'm Gary Shapiro CEO and vice chairman of the consumer technology Association the producer of CES and I am
so thrilled to kick off this show with a keynote by one of the most consequential companies in the world Nvidia exemplifies The Cutting Edge Innovation we celebrate at CES and founder in CEO Jensen Wong is a true Visionary demonstrating the power of ideas technology and conviction to drive Innovation and reshape our industry and our society I always like to say that if I listened a little bit more closely the last time Jensen spoke at a CTA event I could have retired already but over the past three decades he has established Nvidia as a force driving
change across the globe in Industries ranging from Health Care to automotive and entertainment today Nvidia is pioneering breakthroughs and AI and accelerated Computing that touched nearly every person and every business thanks to his leadership nvidia's Innovations enable Advanced chatbots robots for softwar defined Vehicles huge Virtual Worlds hypers synchronized Factory floors and so much more Wong has been named the world's best CEO by fortune and The Economist as well as one of Time magazine's 100 most influential people in the world but you know the fact is like for all of us in this room our success
and his success was not pre-ordained Jensen started out working at a Denny's as a dishwasher and a bus boy so be nice to them in the future and he said that the lessons he's learned there the value of hard work humility and Hospitality are what helped them keep the faith and persevere through some of nidia's early challenges in just a few minutes we'll hear from the video founder and CEO Jensen Wong on his unwavering vision of the future and where we're headed next stay tuned and have a great CES [Applause] [Music] [Applause] [Music] [Music] [Music]
e e e some go [Music] [Music] [Music] [Music] [Music] this is how intelligence is made a new kind of factory generator of tokens the building blocks of AI tokens have opened a new frontier the first step into an extraordinary world where endless possibilities are [Music] born tokens transform words into knowledge and breathe life into images they turn ideas into videos and help us safely navigate any environment tokens teach robots to move like the m [Music] Masters Inspire new ways to celebrate our victories a martini please calling R up thank you Adam and give us peace
of mind when we need it most hioka hi Emma it's good to see you again hi Emma we're going to take a blood sample today okay don't worry I'm going to be here the whole time they bring meaning to numbers to help us better understand the world around [Music] us predict the dangers that surround [Music] us and find cures for the threats within us [Music] tokens can bring our Visions to life and restore what we've lost Zachary I got my voice back buddy they help us move forward [Music] one small step at a [Music] time
and one giant leap [Music] together and here is where it all begins welcome to the stage Nvidia founder and CEO Jensen [Applause] [Music] [Applause] Wong welcome to CES are you excited to be in Las Vegas do you like my jacket I thought I'd go the other way from Gary Shapiro I'm in Las Vegas after all if does if this doesn't work out if all of you object well just get used to it I think I really think you have to let this sink in in another hour or so you're going to feel good about it
well uh welcome to Nvidia in fact you're inside nvidia's digital twin and we're going to take you to Nvidia ladies and gentlemen welcome to Nvidia your inside our digital twins everything here is generated by AI it has been an extraordinary Journey extraordinary year and uh it started in 1993 ready go with mv1 we wanted to build computers that can do things that normal computers couldn't and mv1 made it possible to have a game console in your PC our programming architecture was called UD missing the letter c until a little while later but UDA uni unified
device architecture and the first developer for UDA and the first application that ever worked on UDA was sega's Virtual Fighter six years later we invented in 1999 the programmable GPU and it started 20 years 20 plus years of incredible advance in this incredible processor called the GPU it made modern computer Graphics possible and now 30 years later sega's Virtual Fighter is completely cinematic this is the new Virtual Fighter project that's coming I just can't wait absolutely incredible six years after that six year six years after 1999 we invented Cuda so that we could explain or
Express the programmability of our gpus to a rich set of algorithms that could benefit from it Cuda initially was difficult to explain and it took years in fact it took approximately six years somehow six years later six years later or so 2012 Alex kvki ilas suser and Jeff Hinton discovered Cuda used it to process alexnet and the rest of it is history AI has been advancing at an incredible Pace sense started with perception AI we now can understand images and words and sounds to generative AI we can generate images and texts and sounds and now
a gentic ai ai that can perceive reason plan and act and then the next phase some of which we'll talk about tonight physical AI 2012 now magically 2018 something happened that was pretty incredible Google's Transformer was released as Bert and the world of AI really took off Transformers as you know completely changed the landscape for artificial intelligence in fact it completely changed the landscape for computing allog together we recognized properly that AI was not just a new application with a new business opportunity but AI more importantly machine learning enabled by Transformers was going to fundamentally
change how Computing works and today Computing is revolutionized in every single layer from hand coding instructions that run on CPUs to create software tools that humans use we now have machine learning that creates and optimizes newo networks that processes on gpus and creates artificial intelligence every single layer of the technology stack has been completely changed an incredible transformation in just 12 years well we can Now understand information of just about any modality surely you've seen text and images and sounds and things like that but not only can we understand those we can understand amino acids
we can understand physics we understand them we can translate them and generate them the applications are just completely endless in fact almost any AI application that you you see out there what modality is the input it learned from what modality of information did it translate to and what modality of information is it generating if you ask these three fundamental questions just about every single application could be inferred and so when you see application after applications that are Aid driven AI native at the core of it this fundamental concept is there machine learning has changed how
every application is going to be built how Computing will be done and the possibilities Beyond well gpus GeForce in a lot of ways all of this with AI is the house that GeForce built GeForce enabled AI to reach the masses and now ai is coming home to GeForce there are so many things that you can't do without AI let me show you some of it now [Music] that was realtime computer Graphics no computer Graphics researcher no computer scientists would have told you that it is possible for us to R trce every single pixel at this
point we Ray tracing is a simulation of light the amount of geometry that you saw was absolutely insane it would have been impossible without artificial intelligence there are two fundamental things that we did we used of course programmable shading and Ray traced acceleration to produce incredibly beautiful pixels but then we have artificial intelligence be conditioned be controlled by that pixel to generate a whole bunch of other pixels not only is it able to generate other pixels spatially because it's aware of what the color should be it has been trained on a supercomputer back in Nvidia
and so the neuron Network that's running on the GPU can infer and predict the pixels that we did not render not only can can we do that it's called dlss the latest generation of dlss also generates Beyond frame frames it can predict the future generating three additional frames for every frame that we calculate what you saw if we just said four frames of what you saw because we're going to render one frame and generate three if I said four frames at full HD 4K that's 33 million pixels or so out of that 33 million pixels
we computed only two it is an absolute miracle that we can computationally computationally using programmable shaders and our Ray traced engine race racing engine to compute 2 million pixels and have ai predict all of the other 33 and as a result we're able to render at incredibly high performance because AI does a lot less computation it takes of course an enormous amount of training to produce that but once you train it the generation is extremely efficient so this is one of the incredible capabilities of artificial intelligence and that's why there's so many amazing things that
are happening we used GeForce to enable artificial intelligence and now artificial intelligence is revolutionizing GeForce everyone today we're announcing our next Generation the RTX blackw family let's take a look [Music] here it is our brand new GeForce RTX 5050 series Blackwell architecture the GPU is just a beast 92 billion transistors 4,000 tops four pedop flops of AI three times higher than the last generation Ada and we need all of it to generate those pixels that I showed you 380 Ray tracing Tera flops so that we could for the pixels that we have to compute compute
the most beautiful image you possibly can and of course 125 Shader Tera flops there's actually a concurrent Shader Tera flops as well as an ingri unit of equal performance so two dual shaders one is for floating point one is for integer G7 memory from Micron 1.8 terabytes Per Second Twice the performance of our last generation and we now have the ability to intermix AI workloads with computer graphics workloads and one of the amazing things about this generation is the programmable Shader is also able to now process neuron networks the Shader is able to carry these
neuron networks and as a result we invented neurot texture compression and neurom material shading as a result of that you get these amazingly beautiful images that are only possible because we use AIS to learn the texture learned a compression algorithm and as a result get extraordinary results okay so this is this is uh the brand new RTX Blackwell 90 now even even the even the mechanical design is a miracle look at this it's got two fans this whole graphics card is just one giant fan you know so the question is where's the graphics card is
it literally this big the voltage regular design is state-of-the-art incredible design the engineering team did a great job so here it is thank you okay so those are the speeds and feeds so how does it compare well this is RTX 490 I know I know many of you have one I know it look it's $1,599 it is one of the best investments you could possibly make you for $15.99 you bring it home to your $110,000 PC entertainment Command Center isn't that right don't tell me that's not true don't be ashamed it's liquid cooled fancy lights
all over it you lock it when you leave it's it's the modern home theater it makes perfect sense and now for $1,500 and 99 $15.99 you get to upgrade that and turbocharge the living data lights out of it well now with the Blackwell family RTX 570 490 performance at 549 [Applause] impossible without artificial intelligence impossible without the Four Tops four ter Ops of AI tensor cores impossible without the G7 memories okay so 5070 4090 performance $549 and here's the whole family starting from 5070 all the way up to 590 5090 twice the performance of a
4090 starting of course we're producing a very large scale availability starting January well it is incredible but we managed to put these in in gigantic performance gpus into a laptop this is a 55070 laptop for $12.99 this 570 laptop has a 4090 performance I think there's one here somewhere let me show you this this is a look at this thing here let me here there's only so many pockets ladies and gentlemen Janine [Applause] Paul so can you imagine you get this incredible graphics card here Blackwell we're going to shrink it and put it in put
it in there there does that make any sense well you can't do that without artificial intelligence and the reason for that is because we're generating most of the pixels using pixels using our tensor cores so we retrace only the pixels we need and we generate using artificial intelligence all the other pixels we have as a result the amount of the Energy Efficiency is just off the charts the future of computer Graphics is neural rendering the fusion of artificial intelligence and computer graphics and what's really amazing is oh here we go thank you this is a
surprisingly kinetic keynote and and uh what's really amazing is the family of gpus we're going to put in here and so the 1590 the 1590 will fit into a laptop a thin laptop that last laptop was 14 14.9 mm you got a 5080 5070 TI and 5070 okay so ladies and gentlemen the RTX Blackwell [Applause] family well GeForce uh brought AI to to the world democratized AI now ai has come back and revolutionized gForce let's talk about artificial intelligence let's go to somewhere else at Nvidia this is literally our office this is literally nvidia's headquarters
okay so let's talk about let's talk about AI the industry is chasing and racing to scale artificial intelligence int artificial intelligence and the scaling law is a powerful model it's an empirical law that has been oberved and demonstrated by researchers and Industry over several generations and this the the scale the scaling laws says that the more data you have the training data that you have the larger model that you have and the more compute that you apply to it therefore the more effective or the more capable your model will become and so the scaling law
continues what's really amazing is that now we're moving towards of course and the internet is producing about twice twice the amount of data every single year as it did last year I think the in the next couple years we produce uh Humanity will produce more data than all of humanity has ever produced uh since the beginning and so we're still producing a gigantic amount of data and it's becoming more multimodal video and images and sound all of that data could be used to train the fundamental knowledge the foundational knowledge of an AI but there are
in fact two other scaling laws that has now emerged and it's somewhat intuitive the second scaling law is posttraining scaling law posttraining scaling law uses technology techniques like reinforcement learning human feedback basically the AI produces and generates answers the hum based on a human query the human then of course gives a feedack back um it's much more complicated than that but that reinforcement learning system uh with a fair number of very high quality prompts causes the AI to refine its skills it could fine-tune its skills for particular domains it could be better at solving math
problems better at reasoning so on so forth and so it's essentially like having a mentor or having a coach give you feedback um after you're done going to school and so you you get test you get feedback you improve yourself we also have reinforcement learning AI feedback and we have synthetic data generation uh these techniques are rather uh uh akin to if you will uh self-practice uh you know you know the answer to a particular problem and uh you continue to try it until you get it right and so an AI could be presented with
a very complicated and a difficult problem that has that is verifiable u functionally and has a has an answer that we understand maybe proving a theorem maybe solving a solving a geometry problem and so these problems uh would cause the AI to produce answers and using reinforcement learning uh it would learn how to improve itself that's called post training post training requires an enormous amount of computation but the end result produces incredible models we now have a third scaling law and this third scaling law has to do with uh what's called test time scaling test
time scaling is basically when you're being used when you're using the AI uh the AI has the ability to now apply a different resource allocation instead of improving its parameters now it's focused on deciding how much computation to use to produce the answers uh it wants to produce reasoning is a way of thinking about this uh long thinking is a way to think about this instead of a direct inference or one-hot answer you might reason about it you might break down the problem into multiple steps you might uh generate multiple ideas and uh evaluate you
know your AI system would evaluate which one of the ideas that you generated was the best one maybe it solves the problem step by step so on so forth and so now test time scaling has proven to be incredibly effective you're watching this sequence of technology and this all of these scale saing loss emerge as we see incredible achievements from chat GPT to 01 to 03 and now Gemini Pro all of these systems are going through this journey step by step by step of pre-training to post trining to test time scaling well the amount of
computation that we need of course is incredible and we would like in fact we would like in fact that Society has the ability to scale the amount of computation to produce more and more novel and better intelligence intelligence of course is the most valuable asset that we have and it can be applied to solve a lot of very challenging problems so scaling law it's driving enormous demand for NVIDIA Computing is driving an enormous demand for this incredible chip we call Blackwell let's take a look at Blackwell well Blackwell is in full production it is in
incredible what it looks like so first of all there's some uh every every single cloud service provider now have systems up and running uh we have systems here from about 15 uh 15 15 U uh excuse me 15 computer makers it's being made uh about 200 different SKS 200 different configurations they're liquid cooled air cooled x86 Nvidia gray CPU versions MV link 36 by two MV links 70 2 by one whole bunch of different types of systems so that we can accommodate just about every single data center in the world well this these systems are
being currently manufactured in some 45 factories it tells you how pervasive artificial intelligence is and how much the industry is jumping onto artificial intelligence in this new Computing model well the reason why we're driving it so hard is because we need a lot more computation and it's very clear it's very clear that that um Janine you know I it's hard to tell you don't ever want to reach your hands into a dark place hang second is this a good idea all right [Music] wait for it wait for it I thought I was Worthy apparently yor
didn't think I was worthy all right this is my show and tell this a show in tell so uh this MV link system this right here this MV link system this is gb200 MV link 72 it is 1 and2 tons 600,000 Parts approximately equal to 20 cars 12 12 120 KW it has um a spine behind it that connects all of these GPU together two miles of copper cable 5,000 cables this is being manufactured in 45 factories around the world we build them we liquid cool them we test them we disassemble them ship them parts
to the data centers because it's 1 and A2 tons we reassemble it outside the data centers and install them the manufacturing is insane but the goal of all of this is because the scaling laws are driving Computing so hard that this level of computation Blackwell over our last generation improves the performance per watt by a factor of four performance per wat by a factor of four perform performance per dollar by a factor of three that's basically says that in one generation we reduce the cost of training these models by a factor of three or if
you want to increase um the size of your model by a factor three it's about the same cost but the important thing is this these are generating tokens that are being used by all of us when we use Chad GPT or when we use Gemini use our phones in the future just about all of these applications are going to be consuming these AI tokens and these AI tokens are being generated by these systems and every single data center is limited by power and so if the per per watt of Blackwell is four times our last
generation then the revenue that could be generated the amount of business that can be generated in the data center is increased by a factor of four and so these AI Factory systems really are factories today now the goal of all of this is to so that we can create one giant chip the amount of computation we need is really quite incredible and this is basically one giant chip if we would have had to build a chip one here we go sorry guys you see that that's cool look at that disco lights in here right if
we had to build this as one chip obviously this would be the size of the wafer but this doesn't include the impact of yield it would have to be probably three or four times the size but what we basically have here is 72 Blackwell gpus or 144 dieses this one chip here is 1.4 xof flops the world's largest supercomputer fastest supercomputer only recently this entire room supercomputer only recently achieved an xof flop plus this is 1.4 xof flops of AI floating Point performance it has 14 terabytes of memory but here's the amazing thing the memory
bandwidth is 1.2 pedabytes per second that's basically basically the entire internet traffic that's happening right now the entire world's internet traffic is being processed across these chips okay and we have um 103 130 trillion transistors in total 2592 CPU cores whole bunch of networking and so these I wish I could do this I don't think I will so these are the black Wells these are our connectx networking chips these are the MV link and we're trying to pretend about the mvy the the MV link spine but that's not possible okay and these are all of
the HPM memories 12 ter 14 terabytes of hbm memory this is what we're trying to do and this is the miracle this is the miracle of the black wall system the blackwall dies right here it is the largest single chip the world's ever made but yet the miracle is really in addition to that this is uh the grace black wall system well the goal of all of this of course is so that we can thank you thanks boy is there a chair I could sit down for a second can I have a michelou Ultra how
is it possible that we're in the micho ultra Stadium it's like coming to Nvidia and we don't have a GPU for you so so we need an enormous amount of computation because we want to train larger and larger models and these inferences these inferences used to be one inference but in the future the AI is going to be talking to itself it's going to be thinking it's going to be internally reflecting processing so today when the tokens are being generated at you so long as is coming out at 20 or 30 tokens per second it's
basically as fast as anybody can read however in the future and right now with uh gp01 you know with with the new the pre uh Gemini Pro and and the new GP the the 0103 models they're talking to themselves we reflecting they're thinking and so as you can imagine the rate at which the tokens could be ingested is incredibly high and so we need the token rates the token generation rates to go way up and we also have to drive the cost way down simultaneously so that the C the quality of service can be extraordinary
the cost to customers can continue to be low and AI will continue to scale and so that's the fundamental purpose the reason why we created MV link well one of the most important things that's happening in the world of Enterprise is agentic ai agentic ai basically is a perfect example of test time scaling it's a AI is a system of models some of it is understanding interacting with the customer interacting with the user some of is maybe retrieving information retrieving information from Storage a semantic AI system like a rag uh maybe it's going on to
to the internet uh maybe it's uh studying a PDF file and so it might be using tools it might be using a calculator and it might be using a generative AI to uh generate uh charts and such and it's iter it's taking the the problem you gave it breaking it down step by step and it's iterating through all these different models well in order to respond to a customer in the future in order for AI to respond it used to be ask a question answer start spewing out in the future you ask a question a
whole bunch of models are going to be working in the background and so test time scaling the amount of computation used for inferencing is going to go through the roof it's going to go through the roof because we want better and better answers well to help the the industry build agentic AI our our go to market is not direct to Enterprise customers our go to market is is we work with software developers in the it ecosystem to integrate our technology to make possible new capabilities just like we did with Cuda libraries we now want to
do that with AI libraries and just as the Computing model of the past has apis that are uh doing computer Graphics or doing linear algebra or doing fluid dynamics in the future on top of those acceleration libraries C acceleration libraries will have ai libraries we've created three things for helping the ecosystem build agentic AI Nvidia Nims which are essentially AI microservices all packaged up it takes all of this really complicated Cuda software Cuda DNN cutless or tensor rtlm or Triton or all of these different really complicated software and the model itself we package it up
we optimize it we put it into a container and you could take it wherever you like and so we have models for vision for understanding languages for for speech for animation for digital biology and we have some new new exciting models coming for physical Ai and these AI models run in every single Cloud because nvidia's gpus are now available in every single Cloud it's available in every single OEM so you could literally take these models integrate it into your software packages create AI agents that run on Cadence or they might be S uh service now
agents or they might be sap agents and they could deploy to their customers and run it wherever the customers want to run the software the next layer is what we call Nvidia Nemo Nemo is essentially a digital employee onboarding and training evaluation system in the future these AI agents are essentially digital Workforce that are working alongside your employees um working Al doing things for you on your behalf and so the way that you would bring these specialized agents into your these special agents into your company is to on board them just like you on board
an employee and so we have different libraries that helps uh these AI agents be uh trained for the type of you know language in your company maybe the vocabulary is unique to your company the business process is different the way you work is different so you would give them examples of what the work product should look like and they would try to generate it and you would give a feedback and then you would evaluate them so on so forth and so that and you would guardrail them you say these are the things that you're not
allowed to do these are things you're not allowed to say this and and we even give them access to certain information okay so that entire pipeline a digital employee pipeline is called Nemo in a lot of ways the IT department of every company is going to be the HR department of AI agents in the future today they manage maintain a bunch of software from uh from the IT industry in the future they'll maintain you know nurture on board and uh improve a whole bunch of digital agents and provision them to the companies to use okay
and so your H your it department is going to become kind of like AI agent HR and on top of that we provide a whole bunch of blueprints that our ecosystem could could uh take advantage of all of this is completely open source and so you could take take it and uh modifi the blueprints we have blueprints for all kinds of different different types of Agents well today we're also announcing that we're doing something that's really cool and I think really clever we're announcing a whole family of models that are based off of llama the
Nvidia llama neotron language Foundation models llama 3.1 is a complete phenomenon the download of llama 3.1 from meta 350 650,000 times something like that it has been derived and turned into other models uh about 60,000 other different models it it is singularly the reason why just about every single Enterprise and every single industry has been activated to start working on AI well the thing that we did was we realized that the Llama models really could be better fine-tuned for Enterprise use and so we find tun them using our expertise and our capabilities and we turn
them into the Llama neotron Suite of open models there are small ones that interact and uh very very fast response time extremely small uh they're uh Su what we call Super llama neotron supers they're basically your mainstream versions of your models or your Ultra model the ultra model could be used uh to be a teacher model for a whole bunch of other models it could be a reward model evaluator uh a judge for other models to create answers and decide whether it's a good answer or not give basically give feedback to other models it could
be distilled in a lot of different ways basically a teacher model a knowledge distillation uh uh model very large very capable and so all of this is now available online well these models are incredible it's a uh number one in leaderboards for chat leaderboard for instruction uh lead leaderboard for retrieval um so the different types of functionalities necessary that are used in AI agents around the world uh these are going to be incredible models for you we're also working with uh the ecosystem these te all of our Nvidia AI Technologies are integrated into uh uh
the it in Industry uh we have great partners and really great work being done at service now at saap at Seaman uh for industrial AI uh Cadence is doing great works and is doing great work I'm really proud of the work that we do with perplexity as you know they revolutionize search yeah really fantastic stuff a codium uh every every software engineer in the world this is going to be the next giant AI application next giant AI service period is software coding 30 million software Engineers around the world everybody is going to have a software
assistant uh helping them code uh if if um if not obviously you're just you're going to be way less productive and create lesser good code and so this is 30 million there's a billion knowledge workers in the world it is very very clear AI agents is probably the next robotics industry and likely to be a multi-trillion dollar opportunity well let me show you some of the uh blueprints that we've created and some of the work that we've done with our partners uh with these AI agents [Music] e e e e that was the first pitch
at a baseball that was not generated I just felt that none of you were impressed okay so ai ai was was created in the cloud and for the cloud yeah I was created in the cloud for the cloud and for uh enjoying AI on on phones of course it's perfect um very very soon we're going to have a Contin AI that's going to be with you and when you use those metag glasses you could of course uh point at something look at something and and ask it you know whatever information you want and so AI
is is perfect in the CL it was creating the cloud is perfect in the cloud however we would love to be able to take that AI everywhere I've mentioned already that you could take Nvidia AI to any Cloud but you could also put it inside your company but the thing that we want to do more than anything is put it on our PC as well and so as you know Windows 95 revolutionized the computer industry it made possible this new suite of multimedia services and it changed the way that applications was created forever um Windows
95 this this model of computing of course is not perfect for AI and so the thing that we would like to do is we would like to have in the future your AI basically become your AI assistant and instead of instead of just the the 3D apis and the sound apis and the video apis you would have generative apis generative apis for 3D and generative apis for language and generative AI for sound and so on so forth and we need a system that makes that possible while leveraging the massive investment that's in the cloud there's
no way that we could the world can create yet another way of programming AI models it's just not going to happen and so if we could figure out a way to make Windows PC a world class AI PC um it would be completely awesome and it turns out the answer is Windows it's Windows wsl2 Windows wsl2 Windows wsl2 basically is two operating systems within one it works perfectly it's developed for developers and it's developed uh uh so that you can have access to Bare Metal it's been wsl2 has been optimized optimized for cloud native applications
it is optimized for and very importantly it's been optimized for Cuda and so wsl2 supports Cuda perfectly out of the box as a result everything that I showed you with Nvidia Nims Nvidia Nemo the blueprints that we develop that are going to be up in ai. nvidia.com so long as the computer fits it so long as you can fit that model and we're going to have many models that that fit whether it's Vision models or language models or speech models or these animation human digital human models all kinds of different different types of models are
going to be perfect for your PC and it would you download it and it should just run and so our focus is to turn Windows wsl2 Windows PC into a Target first class platform that we will support and maintain for as long as we shall live and so this is an incredible thing for engineers and developers everywhere let me show you something that we can do with that this is one of the examples of a blueprint we just made for [Music] you e Nvidia AI for your PCS hundreds of millions of PCS in the world
with Windows and so we could get them ready for AI uh oems all the PCS we work with just basically all of the world's leading PCMS are going to get their PCS ready for this step back and so aips are coming to a home near you Linux is good okay let's talk about physical AI speaking of Linux let's talk about physical AI So Physical AI imagine imagine whereas your large language model you give it your context your prompt on the left and it generates tokens one at a time to produce the output that's basically how
it works the amazing thing is this model in the middle is quite large has billions of parameters the context length is incredibly large because you might decide to load in a PDF and my case I might load in several PDFs before I ask it a question those PDFs are turned into tokens the attention the basic attention characteristic of a transformer has every single token find its relationship and relevance against every other token so you could have hundreds of thousands of tokens and the computational load increases quadratically and it does this that all of the parameters
all of the input sequence process it through every single single layer of the Transformer and it produces one token that's the reason why we need a Blackwell and then the next token is produced when the current token is done it puts the current token into the input sequence and takes that whole thing and generates the next token it does it one at a time this is the Transformer model it's the reason why it is so so incredibly effective computationally demanding What If instead of PDFs it's your surrounding and what If instead of the prompt a
question it's a request go over there and pick up that you know that box and bring it back and instead of what is produced in tokens its text it produces action tokens well that I just described is a very sensible thing for the future of Robotics and the technology is right around the corner but what we need to do is we need to create the effective effectively the world model of you know as opposed to GPT which is a language model and this world model has to understand the language of the world it has to
understand physical Dynamics things like gravity and friction and inertia it has to understand geometric and spatial relationships it has to understand cause and effect if you drop something if lost to the ground if you you know poke at it tips over it has still understand object permanence if you roll a ball over to Kitchen en counter when it goes off the other side the ball didn't leave into another quantum universe that's still there and so all of these types of understanding as intuitive understanding that we know that most models today have a very hard time
with and so we would like to create a world we need a world Foundation model today we're announcing a very big thing we're announcing Nvidia Cosmos a world Foundation model that is designed that was created to understand the physical world and the only way for you to really understand this is to see it let's play [Music] it the next Frontier of AI is physical AI model performance is directly related to data availability but physical world data is costly to C e e e bringing the power of foresight and Multiverse simulation to AI models generating every
possible future to help the model select the right path working with the world's developer ecosystem Nvidia is helping Advance the next wave of physical [Music] AI Nvidia Cosmos Nvidia Cosmos Nvidia Cosmos the world's first world Foundation model it is trained on 20 million hours of video the 20 million hours of video focuses on physical Dynamic things so n n Dynamic nature nature themes themes uh humans uh walking uh hands moving uh manipulating things uh you know things that are uh fast camera movements it's really about teaching the AI not about generating creative content but teaching
the AI to understand the physical world and from this with this physical AI there are many Downstream things that we could uh do as a result we could do synthetic data generation to train uh models we could distill it and turn it into effectively the seed the beginnings of a robotics model you could have it generate multiple physically based physically plausible uh scenarios of the future basically do a doctor strange um you could uh because because this model understands the physical world of course you saw a whole bunch of images generated this model understanding the
physical world it also uh could do of course captioning and so it could take videos caption it incredibly well and that captioning and the video could be used to train large language models multimodality large language models and uh so you could use this technology to uh use this Foundation model to train robotics robots as well as larger language models and so this is the Nvidia Cosmos the platform has an autor regressive model for real-time applications has diffusion model for a very high quality image generation it's incredible tokenizer basically learning the vocabulary of uh real world
and a data pipeline so that if you would like to take all of this and then train it on your own data this data pipeline because there's so much data involved we've accelerated everything end to end for you and so this is the world's first data processing pipeline that's Cuda accelerated as well as AI accelerated all of this is part of the cosmos platform and today we're announcing that Cosmos is open licensed it's open available on GitHub we hope we hope that this moment and there's a there's a small medium large for uh uh very
fast models um you know mainstream models and also teacher models basically not knowledge transfer models Cosmo cosmos's World Foundation model being open we really hope will do for the world of Robotics and Industrial AI what llama 3 has done for Enterprise AI the magic happens when you connect Cosmos to Omniverse and the reason fundamentally is this Omniverse is a physics grounded not physically grounded but physics grounded it's algorithmic physics principled physics simulation grounded system it's a simulator when you connect that to Cosmos it provides the grounding the ground truth that can control and to condition
the Osmos generation as a result what comes out of Osmos is grounded on Truth this is exactly the same idea as connecting a large language model to a rag to a retrieval augmented generation system you want to ground the AI generation on ground truth and so the combination of the two gives you a physically simulated a physically grounded Multiverse generator and the application the use cases are really quite exciting and of course uh for robotics for industrial applications uh it is very very clear this Cosmos plus Omniverse plus Cosmos represents the Third computer that's necessary
for building robotic systems every robotics company will ultimately have to build three computers a robotics the robotic system could be a factory the robotic system could be a car it could be a robot you need three fundamental computers one computer of course to train the AI we call it the djx computer to train the AI another of course when you're done to deploy the AI we call that agx that's inside the car in the robot or in an AMR or you know at the uh in a in a stadium or whatever it is these computers
are at the edge and they're autonomous but to connect the two you need a digital twin and this is all the simulations that you were seeing the digital twin is where the AI that has been trained goes to practice to be refined to do its synthetic data generation reinforcement learning AI feedback such and such and so it's the digital twin of the AI these three computers are going to be working interactively nvidia's strategy for uh the industrial world and we've been talking about this for sometime is this three computer system you know instead of a
three three body problem we have a three Computer Solution and so it's the Nvidia robotics [Applause] so let me give you three examples all right so the first example is uh uh uh how we apply apply all of this to Industrial digitalization there millions of factories hundreds of thousands of warehouses that's basically it's the backbone of A50 trillion doll manufacturing industry all of that has to become software defined all of it has has to have Automation and the future and all of it will be infused with robotics well we're partnering with Keon the world's leading
Warehouse automation Solutions provider and Accenture the world's largest professional services provider and they have a big focus in digital manufacturing and we're working together to create something that's really special and I'll show you that in a second but our go to market is essentially the same as all of the other software uh platforms and all technology platforms that we have through the uh developers and ecosystem Partners uh and we have just just a growing number of ecosystem Partners connecting to Omniverse and the reason for that is very clear everybody wants to digitalize the future of
Industries there's so much waste so much opportunity for Automation in that $50 trillion doar of the world's GDP so let's take a look at that this one one p one example that we're doing with Keon and Accenture Keon the supply chain solution company Accenture a global leader in Professional Services and Nvidia are bringing physical AI to the $1 trillion warehouse and Distribution Center Market managing High Performance Warehouse Logistics involves navigating a complex web of decisions influenced by constantly shifting variables these include daily and seasonal demand changes space constraints Workforce availability and the integration of diverse
robotic and automated systems and predicting operational kpis of a physical Warehouse is nearly impossible today to tackle these challenges Keon is adopting Mega an Nvidia Omniverse blueprint for building industrial digital twins to test and optimize robotic fleets first Keon's warehouse management solution assigns tab s to the industrial AI brains in the digital twin such as moving a load from a buffer location to a shuttle storage solution the robot's brains are in a simulation of a physical Warehouse digitalized into Omniverse using open USD connectors to aggregate CAD video and image to 3D Light Art to point
cloud and AI generated data the fleet of robots execute tasks by perceiving and reason in about their Omniverse digital twin environment planning their next motion and acting the robot brains can see the resulting State through sensor simulations and decide their next action the loop continues while Mega precisely tracks the state of everything in the digital twin now Keon can simulate infinite scenarios at scale while measuring operational kpis such as throughput efficiency and utilization all before deploying changes to the physical Warehouse together with Nvidia Keon and Accenture are Reinventing industrial autonomy in the future is that
that's incredible everything is in simulation in the future in the future every Factory will have a digital twin and that digital twin operates exactly like the real factory and in fact you could use Omniverse with Cosmos to to generate a whole bunch of future scenarios and you pick then an AI decides which one of the scenarios are the most optimal for whatever kpis and that becomes the programming constraints the program if you will the AIS that will be uh deployed into the real factories the next example autonomous vehicles the AV revolution has arrived after so
many years with wayo success and Tesla's success it is very very clear autonomous vehicles has finally arrived well our offering to this industry is the three computers the training systems to train the AIS the simulation systems and and the and the synthetic data generation systems Omniverse and now Cosmos and also the computer that's inside the car each car company might might work with us in a different way use one or two or three of the computers we're working with just about every major car company around the world weo and zuk and Tesla of course in
their data center byd the largest uh e company the world jlr has got a really cool car coming Mercedes because a fleet of cars coming with Nvidia starting with this starting this year going to production and I'm super super pleased to announce that today Toyota and Nvidia are going to partner together to create their next Generation AVS just so many so many cool companies uh lucid and rivan and xiaomi and of course Volvo just so many different companies wabby is uh building self-driving trucks Aurora we announced this week also that Aurora is going to use
Nvidia to build self-driving trucks autonomous 100 million cars build each year a billion cars vehicles on a road all over the world a trillion miles that are driven around the world each year that's all going to be either highly autonomous or you know fully autonomous coming up and so this is going to be a very large very large industry I predict that this will likely be the first multi-trillion dollar robotics industry this IND this business for us um notice in just just a few uh of these cars that are starting to ramp into the world
uh our business is already $4 billion and this year probably on a run rate of about5 billion so really significant business already this is going to be very large well today we're announcing that our next generation processor for the car our next generation computer computer for the car is called Thor I have one right here hang on a second okay this is Thor this is Thor this is this is a robotics computer this is a robotics computer takes sensors and just a Madness amount of sensor information process it you know een teen cameras high resolution
radar Lars they're all coming into this chip and this chip has to process all that sensor turn them into tokens put them into a Transformer and predict the next PATH and this AV computer is now in full production Thor is 20 times the processing capability of our last generation Orin which is really the standard of autonomous vehicles today and so this is just really quite quite incredible Thor is in full production this robotics processor by the way way also goes into a full robot and so it could be an AMR it could be a human
or robot it could be the brain it could be the manipulator uh this Ro this processor basically is a universal robotics computer the second part of our drive system that I'm incredibly proud of is the dedication to safety Drive OS I'm pleased to announce is now the first softwar defined programmable AI computer computer that has been certified up to asold D which is the highest standard of functional safety for automobiles the only and the highest and so I'm really really proud of this asold ISO 26262 it is um the work of some 15,000 engineering years
this is just extraordinary work and as a result of that Cuda is now a functional safe computer and so if you're building a robot Nvidia Coulda yeah okay so so now I wanted to I told you I was going to show you what would we use Omniverse and Cosmos to do in the context of self-driving cars and you know today instead of showing you a whole bunch of uh uh videos of of cars driving on the road I'll show you some of that too um but I want to show you how we use the car
to reconstruct digital twins automatically using Ai and use that capability to train future am models okay let's play it the autonomous vehical Revolution is here building autonomous vehicles like all robots requires three computers Nvidia dgx to train AI models Omniverse to test drive and generate synthetic data and drive agx a supercomputer in the car building safe autonomous vehicles means addressing Edge scenarios but real world data is limited so synthetic data is essential for training the autonomous vehicle data Factory powered by Nvidia Omniverse AI models and Cosmos generates synthetic driving scenarios that enhance training data by
orders of magnitude first omnimap fuses map and geospatial data to construct drivable 3D environments driving scenario variations can be generated from replay Drive logs or AI traffic generators next a neuro reconstruction engine uses autonomous vehicle sensor logs to create High Fidelity 4D simulation environments it replays previous drives in 3D and generates scenario variations to amplify training data finally edify 3DS automatically searches through existing asset libraries or generates new assets to create Sim ready scenes the Omniverse scenarios are used to condition Cosmos to generate massive amounts of photorealistic data reducing the Sim to real Gap and
with text prompts generate near infinite variations of the driving scenario with Cosmos neotron video search the massively scaled synthetic data set combined with recorded drives can be curated to train models nvidia's AI data Factory scales hundreds of drives into billions of effective miles setting the standard for safe and advanced autonomous driving is that incredible we take take thousands of drives and turn them into billions of miles we are going to have mountains of training data for autonomous vehicles of course we still need actual cars on the road of course we will continuously collect data for
as long as we shall live however synthetic data generation using this Multiverse physically based physically grounded capability so that we generate data for training AIS that are physically grounded and accurate and or plausible so that we could have an enormous amount of data to train with the AV industry is here uh this is an incredibly exciting time super super super uh uh excited about the next several years I think you're going to see just as computer Graphics was revolutionized such incredible pace you're going to see the pace of Av development increasing tremendously over the next
several years I I think I think um I I think the next part is is robotics so um robots my [Applause] friends the chat GPT moment for General robotics is just around the corner and in fact all of the enabling technologies that I've been talking about is going to make it possible for us in the next several years to see very rapid breakthroughs surprising breakthroughs in in general robotics now the reason why General robotics is so important is whereas robots with tracks and wheels require special environments to accommodate them there are three robots three robots
in the world that we can make that require no green fields Brown field adaptation is perfect if we if we could possibly build these amazing robots we could deploy them in exactly the world that we've built for ourselves these three robots are one agentic robots and agentic AI because you know they're information workers so long as they could accommodate uh the computers that we have in our offices is going to be great number two self-driving cars and the reason for that is we spent 100 plus years building roads and cities and then number three human
or robots if we have the technology to solve these three this will be the largest technology industry the world's ever seen and so we think that robotics era is just around the corner the critical capability is how to train these robots in the case of humano robots the imitation information is rather hard to collect and the reason for that is uh in the case of car you just drive it we're driving cars all the time in the case of these robots the imitation information the the human demonstration is rather laborious to do and so we
need to come up with a clever way to take hundreds of demonstrations thousands of human demonstrations and somehow use artificial intelligence and Omniverse to synthetically generate millions of synthetically generated motions and from those motions the AI can learn uh how to perform a task let me show you how that's done e e e robotics is arriving powered by Nvidia Isaac group we're going to have mountains of data to train robots with Nvidia Isaac Groot Nvidia Isaac group this is our platform to provide technology platform technology elements to the robotics industry to accelerate the development of
General Robotics and um well I have one more thing that I want to show you none of none of this none of this would be possible if not for uh this incredible project that we started uh about a decade ago go inside the company was called project project digits deep learning GPU intelligence training system digits well before we launched it uh I shrunk at the dgx and to harmonize it with RTX agx ovx and all of the other xes that we have in the company and and um I and and it really re ued uh
djx1 really revolutionized where where is djx1 djx1 revolutionized artificial intelligence the reason why we built it was because we wanted to uh make it possible for researchers and startups to have an out-of-the-box AI supercomputer imagine the way supercomputers were built in the past you really have to uh build your own facility and you have to go build your own infrastructure and really engineer it into existence and so we created a supercomputer for AI for AI development for researchers and and startups that comes literally one out of the box I delivered the first one to a
startup company in 2016 called open Ai and Elon was there and and Ilia susc was there and many of Nvidia Engineers were there and and um uh we we celebrated the arrival of djx1 and obviously uh it revolutionized uh artificial intelligence and Computing um but now artificial intelligence is everywhere it's not just in research and and and startup Labs you know we want artificial intelligence as I mentioned in the beginning of our talk this is now the new way of doing Computing this is the new way of doing software every software engineer every engineer every
creative artist everybody who uses computers today as a tool will need a AI supercomputer and so I just wish I just wish that djx1 was smaller and um you know so so um you know imagine ladies and gentlemen our this is nvidia's latest AI supercomputer and and it's finally called project digits right now and if you have a good name for it uh reach out to us um uh this here's the amazing thing this is an AI supercomputer it runs the entire Nvidia AI stack all of Nvidia software runs on this dgx Cloud runs on
this this sits well somewhere and it's wireless or you know connected to your computer it's even a workstation if you like it to be and you could access it you could you could reach it like uh like a cloud supercomputer and nvidia's AI works on it and um it's based on a a super secret chip that we've been working on called GB 110 the smallest Grace Blackwell that we make and I have well you know what let's show let's show everybody insight isn't this just isn't just it's just so cute and this is the chip
that's inside it is in it is in production this top secret chip uh we did in collaboration the CPU the gray CPU was a uh is built for NVIDIA in collaboration with mediatech uh they're the world's leading s SOC company and they worked with us to build this CPU this CPU so and connect it with chipto chip mvy link to the Blackwell GPU and uh this little this little thing here is in full production uh we're expecting this computer to uh be available uh around May time frame and so it's coming at you uh it's
just incredible what we could do and it's just I think it's you really I was trying to figure out do I need more hands or more pockets all right so so uh imagine this is what it looks like you know who doesn't want one of those and if you if you use PC Mac you know anything because because uh you know it's it's a cloud platform it's a cloud computer platform that sits on your desk you could also use it as a l Linux workstation if you like um if you would like to have double
digits this is what it looks like you know and you you connect it you connect it together uh uh with connectx and it has nickel GPU direct all of that out of the box it's like a supercomputer our entire supercomputing stack uh is available and so Nvidia Project digits [Applause] okay well let me let me let me tell you what I told you I told you that we are in production with three new Blackwells not only is the grace Blackwell supercomputers mvlink 72s in production all over the world we now have three new Blackwell systems
in production one amazing AI foundational M World Foundation model the world's first physical AI Foundation model it's open available to activate the world's industries of Robotics and such and three and three robotics three robots we're working on uh agentic AI uh human or robots and self-driving cars uh it's been an incredible year I want to thank all of you for your partnership uh thank all of you for coming I made you a short video to reflect on last year and look forward to the next year play please [Music] [Music] [Applause] [Music] [Music] [Music] [Music] [Music]
[Music] [Music] [Music] have a great CES everybody Happy New Year here thank you [Music]
Related Videos
NVIDIA CEO Jensen Huang Keynote at CES 2025
1:31:50
NVIDIA CEO Jensen Huang Keynote at CES 2025
NVIDIA
2,798,588 views
The TRUTH About RTX 5000 Pricing. it surprised me
20:17
The TRUTH About RTX 5000 Pricing. it surpr...
Vex
414,116 views
2025 Will Be The Year of AI Agents
10:02
2025 Will Be The Year of AI Agents
You Talk AI
43 views
Chip War, the Race for Semiconductor Supremacy | Full Documentary (2023)
51:26
Chip War, the Race for Semiconductor Supre...
TaiwanPlus Docs
312,551 views
The Interview: From Amazon to Space — Jeff Bezos Talks Innovation, Progress and What’s Next
1:06:10
The Interview: From Amazon to Space — Jeff...
New York Times Events
546,599 views
CES 2025 Las Vegas: Mind-Blowing New tech we saw on Day 1 [4K video]
43:48
CES 2025 Las Vegas: Mind-Blowing New tech ...
The Laughing Lion
55,454 views
US Government to BanTP-Link Devices - Live Hacking of a Chinese WiFi Router
30:31
US Government to BanTP-Link Devices - Live...
Matt Brown
1,284,151 views
China's robot army shows WW3 would kill us all.
14:46
China's robot army shows WW3 would kill us...
Digital Engine
843,760 views
D-Wave CEO responds to Jensen Huang's quantum comments
6:11
D-Wave CEO responds to Jensen Huang's quan...
CNBC Television
26,046 views
How do Graphics Cards Work?  Exploring GPU Architecture
28:30
How do Graphics Cards Work? Exploring GPU...
Branch Education
2,761,269 views
Nvidia CEO Jensen Huang delivers keynote at CES [FULL KEYNOTE]
1:57:41
Nvidia CEO Jensen Huang delivers keynote a...
Yahoo Finance
43,114 views
Рождение и жизнь звезд / Что я знаю
1:07:17
Рождение и жизнь звезд / Что я знаю
ПостНаука
122,006 views
CES 2025 Day 3: Insane Tech from Sony's AFEELA, Displace True Wireless TV, & More
13:00
CES 2025 Day 3: Insane Tech from Sony's AF...
MacRumors
47,388 views
The Problem With Elon Musk
42:46
The Problem With Elon Musk
Johnny Harris
6,139,099 views
AMD's CEO Wants to Chip Away at Nvidia's Lead | The Circuit with Emily Chang
24:02
AMD's CEO Wants to Chip Away at Nvidia's L...
Bloomberg Originals
1,196,000 views
Unlock The ROG Lab - CES 2025 Launch Event | ROG
1:40:14
Unlock The ROG Lab - CES 2025 Launch Event...
ROG Global
769,862 views
China's Innovative Edge - The Impact on Global Markets | Economic Transformation
2:05:37
China's Innovative Edge - The Impact on Gl...
Moconomy
273,296 views
TSMC’s New Arizona Fab! Apple Will Finally Make Advanced Chips In The U.S.
16:26
TSMC’s New Arizona Fab! Apple Will Finally...
CNBC
1,627,113 views
They Let me Game on AMD’s Unreleased MONSTER
12:39
They Let me Game on AMD’s Unreleased MONSTER
Linus Tech Tips
375,976 views
AI Revolution: Jensen Huang And Mukesh D. Ambani Discuss India’s Role At NVIDIA'S AI Summit
1:29:06
AI Revolution: Jensen Huang And Mukesh D. ...
Business Today
37,258 views
Copyright © 2025. Made with ♥ in London by YTScribe.com