NVIDIA CEO Jensen Huang Leaves Everyone SPEECHLESS (Supercut)

316.14k views2816 WordsCopy TextShare
Ticker Symbol: YOU
Highlights of #nvidia ( #nvda stock ) Founder and CEO Jensen Huang speaking at AI Summit India. High...
Video Transcript:
general purpose Computing as we know it has existed for 60 years until now for the last 30 years we've had the benefit of mors law an incredible phenomenon without changing the software the hardware can continue to improve in an architecturally compatible way every single industry has subsequently been built on top of it but we know now that the scaling of CPUs has reached as limit the Free Ride of Moors law has ended we no longer can afford to do nothing in software and expect that our Computing experience will continue to improve that costs will decrease
and continue to spread the benefits of it and to benefit from solving greater and greater challenges we started our company to accelerate software our vision was there are applications that would benefit from acceleration that acceleration benefit has the same qualities as Moors law for applications that were impossible or impractical to perform using general purpose Computing we have the benefits of accelerated Computing to realize that capability for example computer Graphics real-time computer Graphics was made possible because of Nvidia coming into the world and make possible this new processor we call gpus but we felt that longterm
accelerated Computing could be be far far more impactful there is no such magical processor that can accelerate everything in the world because if you could do that you would just call it a CPU you need to reinvent the Computing stack from the algorithms to the architecture underneath and connect it to applications on top in one domain after another domain computer Graphics is a beginning but we've taken this Cuda architecture from one industry after another industry after another industry today we accelerate so many important industries ktho is fundamental to semiconductor manufacturing computational lithography simulation computer Aid
and Engineering Quantum Computing so that we can invent the future of computing with classical Quantum hybrid Computing in each one of these different libraries we're able to accelerate the application 20 30 50 times of course it takes a rewrite of software which is the reason why taken so long in each one of these domains we've had to work with the industry work with our ecosystem software developers and customers in order to accelerate those applications for their domains modulus teaching and AI the laws of physics not just to be able to predict next word but to
be able to predict the next moment in time of fluid dynamics and particle physics and so on so forth and of course one of the most famous application libraries we've ever created called called cdnn made it possible to democratize artificial intelligence as we know it these acceleration libraries Now cover so many different domains that it appears that accelerated Computing is used everywhere but that's simply because we've applied this architecture one domain after another domain that we've covered just about every single industry now accelerated Computing or Cuda has reached the Tipping Point the first thing I
happen of course is how we do software our industry is underpinned by the method by which software is done the way that software was done call It software 1.0 programmers would code algorithms we call functions to run on a computer and we would apply it to input information to predict an output somebody would write python or C or Fortran or Pascal or C++ code algorithms that run on a computer you apply input to it and output is produced very classically the computer model that we understood quite well however that approach of developing software has been
disrupted it is now not coding but machine learning using a computer to study the patterns and relationships of massive amounts of observed data to essentially learn from it the function that predicts it and so we are essentially designing a universal function approximator using machines to learn the expected output that would produce such a function and so going back and forth looking this is software 1.0 with human coding to now software 2.0 using machine learning notice who is writing the software the software is now written by the computer and after you're done training the model you
inference the model you then apply that function now as the input that deep learning model that computer vision model speech understanding model is now an input neural network that goes into the GPU that can now make a prediction given new input unobserved input and we have gone from coding to machine learning from developing software to creating artificial intelligence and from software that prefers to run on CPUs to now neural networks that runs best on gpus this at its core is what happened to our industry in the last 10 years we have now seen the complete
reinvention of the Computing stack the whole technology stack has been reinvented the hardware the way that software is developed and what software can do is now fundamentally different we dedicated ourselves to advance this field and and so this is what we now build initially we were building gpus that fit into a PCI Express card that goes into your PC this is what a GPU looks like today this is Blackwell yeah thank you a massive system designed to study data at an enormous scale so that we could discover patterns and relationships and learn the meaning of
the data this is the Greek breakthrough in the last several years we have now learned the representation or the meaning of words and numbers and images and pixels and videos chemicals proteins amino acids fluid patterns particle physics we have now learned the meaning of so many different types of data we have learned to represent information in so many different modalities not only have we learned the meaning of it we can translate it to another modality so one great example of course is translating English to Hindi translating English large body of text into other English summarization
from pixels to image image recognition from words to pixels image generation from images of videos to words captioning from words to proteins used for drug Discovery from words to chemicals discovering new compounds all because of this one instrument that made it possible for us to study data and enormous scales well I just want to say that in order to build the Blackwell system of course the Blackwell gpus is involved but it takes seven other chips tsmc manufactur all of these chips and they're just doing extraordinary job ramping the Blackwell system Blackwell is in full production
and we're expecting to deliver in volume production in Q4 and so this is basically Blackwell now this is one of the things that's really incredible about the system let me show it to you nothing's easy this morning [Music] this is MV link and it goes across the entire back spine of a rack of gpus and these gpus are all connected 72 dual GPU packages of black worlds 144 gpus connected together so it's one giant GPU if I were to spread out all of the chips to show you what this connect together it's essentially a GPU
so large it' be like this big but it's obviously impossible to build gpus that large so we break it up into the smallest chunks we could which is retical limits and the most Advanced Technologies and we connect it together using MV link this is mvlink backspin you're looking at all of the gpus being connected that's the quantum switch that connects all of these gpus together on top Spectrum X if you would like to have ethernet what connects this together this is connected to this switch and this is one of the most advanced switches the world's
ever built now all of this together represents Blackwell and then it runs the software that's on top the Cuda software CNN software Megatron for training the large language models tensor RT for doing the inference tensor RT llm for doing distributed multi-gpu inference for large language models and then on top on top of that we have two software Stacks one is NVIDIA AI Enterprise that I'll talk about in a second and then the other is Omniverse so this is the Blackwell system this is what Nvidia builds today of course the computation is incredible each rack is
3,000 lb 120 kilow 120,000 watts in each rack the density of computing the highest ever the world's ever known and what we're trying to do is to learn larger and smarter models each year we're increasing the amount of data and amount of the model size each by about a factor of two which means that every single year the computation which is the product of those two has to increase by a factor of four now remember there was a time when the world Mo's law was two times every year and a half or 10 times every
5 years 100 times every 10 years we are now moving technology at a rate of four times every year four times every year over the course of 10 years incredible scaling the second thing that we've discovered recently and this is a very big deal intelligence is not just one shot but intelligence requires thinking and thinking is reasoning and maybe you're doing path planning and maybe you're doing some simulations in your mind you're reflecting on your own answers and so as a result thinking results in higher quality answers and we've now discovered a second scaling law
and this is a SCA scaling law at a time of inference the longer you think the higher quality answer you can produce this is not illogical this is very intuitive to all of us if you were to ask me what's my favorite Indian food I would tell you chicken briani okay and I don't have to think about that very much and I don't have to reason about that I just know it and there are many things that you can ask it like for example what's Nvidia good at Nvidia is good at building AI super computers
nvidia's uh great at building gpus and those are things that you know that that's encoded into your knowledge however there are many things that requires reasoning for example if I had to travel from uh Mumbai to California I I want to do it in the in a way that allows me to enjoy four other cities along the way if I were to to tell it I would like to go from California to Mumbai I would like to do it within uh 3 days and I give it all kinds of constraints about what time I'm willing
to leave and able to leave what hotels I like to stay at so on so forth the people I have to meet the number of permutations of that of course quite high and so the planning of that process coming up with a optimal plan is very very complicated and so that's where thinking reasoning planning comes in and the more you compute the higher quality answer uh you could provide and so we now have two fundamental scaling laws that is driving our technology development first for training and now for inference the number of foundation model makers
has more than doubled since the beginning of Hopper there are more companies that realize that fundamental intelligence is vital to their company and that they have to build Foundation model technology and second the size of the models have increased by 20 30 40x the amount of computation necessary to train these model because of uh the size of the models but also multimodality capability reinforcement learning capability itic data generation capability the amount of data that we use to train these models has really grown tremendously that's one and then the other reason of course is that black
well is also used for generating tokens at incredible speeds and so together all of these factors has led to the demand for black well being incredibly High let's talk about now how we're going to use this technology earlier I told you that we have Blackwell we have all of the libraries acceleration libraries that we were talking about before but on top there are two very important platforms that working on one of them is called Nvidia AI Enterprise and the other is called Nvidia Omniverse and I'll explain each one of them very Qui quickly first Nvidia
AI Enterprise this is a time now where the large language models and the fundamental AI capabilities have reached a level of capabilities we're able to now create what is called agents large language models that understand the data that of course is being presented it could be streaming data could video data language model data it could be data of all kinds the first stage is perception the second is reasoning about given its observations what is the mission and what is the task it has to perform in order to perform that task the agent would break down
that task into steps of other tasks and uh it would reason about what it would take and it would connect with other AI models maybe it's a model that understands how to generate images maybe it's a model model that is able to retrieve AI semantic data from a proprietary database so each one of these large language models are connected to the central reasoning large language model we call agent and so these agents are able to perform all kinds of tasks uh some of them are maybe uh marketing agents some of them are customer service agents
some of them are chip design agents envidia has Chip design agents all over our company helping us design chips and so we're going to have agents that are help helping our employees become super employees these agents or agentic AI models uh augment all of our employees to supercharge them make them more productive now when you think about these agents it's really the way you would bring these agents into your company it's not unlike the way you would onboard uh someone uh who's a new employee you have to give them training curriculum you evaluate them and
so they're evaluation systems and you might guard where if you're accounting agent uh don't do marketing and so each one of these agents are guardrail that entire process we put into essentially an agent life cycle Suite of libraries and so this is what we call Nvidia Nemo we have on the one hand the libraries on the other hand what comes out of the output of it is a API inference microservice we call Nims essentially the this is a factory that builds AIS and Nemo is a suite of libraries that on board and help you operate
the AIS and ultimately your goal is to create a whole bunch of Agents the next generation of it is going to be about producing and delivery of AI and as you know the delivery of software coding and the delivery of AI is fundamentally different but dramatically more impactful insanely more exciting the second part is this what happens after agents now remember every single company has employees but most companies the goal is to build something to produce something to make something could be factories it could be warehouses it could be cars and planes and trains and
uh ships all kinds of things that next generation of AI needs to understand the physical world we call it physical AI in order to create physical AI we need three computers and we created three computers to do so the dgx computer which Blackwell for example is is a reference design an architecture for to create things like d GX computers for training the model that model needs a place to be refined it needs a place to learn it needs the place to apply its physical capability it's robotic capability we call that Omniverse a virtual world that
obeys the laws of physics where robots can learn to be robots and then when you're done with the training of it that AI model could then run in the actual robotic system that robotic system could be a car it could be a robot it could be AV it could be a autonomous moving robot it could be a picking arm uh it could be an entire Factory or an entire Warehouse that's robotic and that computer we call agx Jetson agx dgx for training and then Omniverse for doing the digital twin start locally grow globally right right
that's fantastic okay thank you thank you thank you very much Michelle thank you thank you
Related Videos
Nvidia Founder and CEO Jensen Huang on the AI revolution
29:44
Nvidia Founder and CEO Jensen Huang on the...
Goldman Sachs
126,870 views
Tucker Carlson UNLOADS on Diddy, Kamala, Walz, Kimmel, Rich Girls, Conspiracy Theories, and the CIA!
1:14:02
Tucker Carlson UNLOADS on Diddy, Kamala, W...
Joe Polish
1,318,680 views
BlackRock's Rieder on Trump win: ‘Whole new playbook’ is coming
18:40
BlackRock's Rieder on Trump win: ‘Whole ne...
Yahoo Finance
61,482 views
Michael Beckley on The End of China’s Rise & the Future of Global Order | WORLD KNOWLEDGE FORUM 2024
37:48
Michael Beckley on The End of China’s Rise...
American Enterprise Institute
202,513 views
What Happens When Robots Don’t Need Us Anymore? | Posthuman With Emily Chang
24:52
What Happens When Robots Don’t Need Us Any...
Bloomberg Originals
99,091 views
Jensen Huang Special Address from NVIDIA AI Summit Japan
1:26:51
Jensen Huang Special Address from NVIDIA A...
NVIDIA
3,395,910 views
Will Starship Make It Through? Starship’s Boldest Test Yet!
23:37
Will Starship Make It Through? Starship’s ...
Marcus House
178,748 views
How Solid State Cooling Could Change Everything
16:01
How Solid State Cooling Could Change Every...
Undecided with Matt Ferrell
760,336 views
I'm Buying Palantir Stock (PLTR) After Earnings (Here’s Why)
16:28
I'm Buying Palantir Stock (PLTR) After Ear...
Ticker Symbol: YOU
141,146 views
Ex Google CEO: AI Is Creating Deadly Viruses! If We See This, We Must Turn Off AI!
1:49:37
Ex Google CEO: AI Is Creating Deadly Virus...
The Diary Of A CEO
683,044 views
NVIDIA AI Summit: Fireside Chat between Sh. Mukesh Ambani and Jensen Huang
27:09
NVIDIA AI Summit: Fireside Chat between Sh...
Jio
446,568 views
Top Minds in AI Explain What’s Coming After GPT-4o | EP #130
25:30
Top Minds in AI Explain What’s Coming Afte...
Peter H. Diamandis
114,034 views
NEW Tesla Charges In 8 Min | Elon's Future For Tesla
19:15
NEW Tesla Charges In 8 Min | Elon's Future...
Ryan Shaw
27,044 views
AI Copyright Claimed My Last Video
24:11
AI Copyright Claimed My Last Video
Venus Theory
245,155 views
Chris Langan - The Interview THEY Didn't Want You To See - CTMU [Full Version; Timestamps]
1:58:44
Chris Langan - The Interview THEY Didn't W...
CTMU Radio
6,355,028 views
IT'S OVER! I Can't Stay Quiet on AMD vs Nvidia Stock (NVDA) Any Longer
15:08
IT'S OVER! I Can't Stay Quiet on AMD vs Nv...
Ticker Symbol: YOU
253,104 views
MIT's AI Discovers New Science - "Intelligence Explosion"
11:11
MIT's AI Discovers New Science - "Intellig...
Matthew Berman
101,113 views
Jensen Huang, Founder and CEO of NVIDIA
56:27
Jensen Huang, Founder and CEO of NVIDIA
Stanford Graduate School of Business
1,570,324 views
IT'S OVER! The Shocking Truth Behind Nvidia Partner Super Micro (SMCI Stock)
17:11
IT'S OVER! The Shocking Truth Behind Nvidi...
Ticker Symbol: YOU
153,051 views
Tesla Model 3 TRUE Cost Of Ownership After 5 Years. Do I Still Recommend It?
15:06
Tesla Model 3 TRUE Cost Of Ownership After...
Shelby Church
278,872 views
Copyright © 2025. Made with ♥ in London by YTScribe.com