the race towards artificial super intelligence isn't actually sci-fi anymore it's happening right now and ilos husk one of the co-founders of open AI just dropped some jaw-dropping insights about where we're headed now you're going to want to stick around because what Ilia said might actually redefine everything we know about intelligence humanity and the future of our species now one of the first things that ilot Sasa actually talks about is the current Paradigm that we're in and why it's coming to an end he speaks about the death of pre-training and essentially he says that pre-training as
we know it will unquestionably end and this is important because the current AI Paradigm relies heavily on massive text based pre-training but this statement predicts a fundamental shift in how we're going to be training models and it raises questions about what new models will replace today's techniques and how the AI field will adapt once we exhaust all the readily available data the age of pre-training and the age of pre-training is what we might say the g gpt2 model the gpt3 model the scaling laws and I want to specifically call out my uh former collaborators Alec
Radford also Jared Kaplan Dario mode for really making this work but that led to the age of pre-training and this is what's been the driver of all of progress all the progress that we see today extra large neural networks extraordinar large neural networks trained on huge data sets but pre-training as we know it will unquestionably end pre-training will end why will it end because while computers growing through better Hardware better algorithms and largic clusters right all those things keep increasing your compute all these things keep increasing your compute the data is not growing because we
have but one in internet we have but one internet you could even say you could even go as far as to say the data is the fossil fuel of AI it was like created somehow and now we use it and we've achieved Peak data and there'll be no more we have to deal with the data that we have now it still still let us go quite far but this is there's only one internet and he's right we've basically exhausted all forms of available data now what we're going to need to do is we're going to
need to figure out new things and this is where he talks about three key things which are coming next which I'll break down for you in a second so here I'll take um a bit of Liberty to speculate about what comes next actually I don't need to speculate because many people are speculating too and I'll mention their speculations you may have heard the phrase agents it's common and I'm sure that eventually something will happen but people feel like something agents is the future more concretely but also a little bit vaguely synthetic data but what does
synthetic data mean figuring this out is a big Challenge and I'm sure that different people have all kinds of interesting progress there and an inference time compute or maybe what's been most recently most vividly seen in 01 the one model these are all examples of things of people trying to figure out what to do after pre-training and those are all very good things to do so one of the first things he actually mentioned there was agents now agents are being widely discussed is the next big step in Ai and while current systems like chat gbt
are great at generating responses they aren't truly agentic and this just means that they lack the ability to independently set goals reason about their environment and take meaningful action in the world now he did hint that future systems will be genuinely identic capable of reasoning and acting autonomously in ways we've only seen in science fiction and this basically just means that you know future AI will evolve from being passive tools to become active participants and they're going to be making real decisions they're going to be adapting to new situations and they're going to be collaborating
with humans in meaningful self-directed ways and this is going to present a massively in functionality as agentic AI could perform tasks without needing constant prompts or even supervision and this is important because agentic AI is going to fundamentally change how we interact with technology imagine your personal AI assistant that can proactively manage your tasks conduct complex research or even solve real world problems without being micromanaged now the second point that he spoke about was synthetic data and he said that this is one of the potential solutions for the limitations of real world data sets he
described it as both promising and a challenging area of research and while he didn't go into you know all the technical details he emphasized that figuring out how to create highquality synthetic data could unlock new capabilities for AI now this basically just refers to artificially generated data sets that mimic the properties of real world data and they can be created using simulations generative models or even algorithms designed to replicate natural phenomena and as we hit the limit of the real world data which is of course as he referred to in the beginning the one internet
problem synthetic data could provide new training material for these AI systems the ability to simulate edge cases or rare events for example for self-driving cars or even potentially medical diagnosis this is important because without synthetic data the progress of AI could stagnate due to data scarcity and mastering synthetic data generation we could train on scenarios that don't exist or are too rare to capture in in the world you've got this mammals right all the different mammals then you've got non-human primates it's basically the same thing but then you've got the hominids and to my knowledge
hominids are like close relatives to the humans in evolution like the neand there's a bunch of them like it's called homohabilis maybe there there's a whole bunch and they all here and what's interesting is that they have a different slope on their brain to body scaling exponent so that's pretty cool what that means is that there is a precedent there is an example of biology figuring out some kind of different scaling something clearly is different essentially what we do have here is where he's talking about the graph that illustrates how in biology there is a
known relationship between an animal's body size and its brain size most mammals follow a predictable pattern bigger bodies tend to come with bigger brains in a more or less steady ratio but human ancestors the hominids they actually broke that Trend somehow and evolution produced creatures whose brains grew faster than the expected for their body size and this shift suggests that at some point nature figured out a different scaling rule for intelligence in humans and now by mentioning this sasca is drawing an analogy just as biology found a new way to scale intelligence in human evolution
giving us much bigger and more capable brains relative to our bodies the AI field may need to find new methods to move Beyond its current scaling limits and right now ai depends on huge amounts of data and ever larger neural networks but just as humans evolved a new approach to intelligence AI researchers will need to find equally clever next steps that just don't follow the old scaling patterns which is you know what we're currently doing so right now we have our incredible language models and they unbelievable chat Bots and they can even do things but
they're also kind of strangely unreliable and they get confused when while also having dramatically superhuman performance on evals so it's really unclear how to reconcile this but eventually sooner or later the following will be achieved those systems are actually going to be agentic in a real ways whereas right now the systems are not agents in any meaningful sense just very that might be too strong they're very very slightly agentic just beginning it will actually reason and by the way I want to mention something about reasoning is that a system that reasons the more it reasons
the more unpredictable it becomes the more it reasons the more unpredictable it becomes all the Deep learning that we've been used to is very predictable because if you've been working on replicating human intuition essentially it's like the gut fi if you come back to the 0.1 second reaction time what kind of processing we do in our brains well it's our intuition so we've endowed ouris with some of that intuition but reasoning you're seeing some early signs of that reasoning is unpredictable and one reason to see that is because the chess AIS the really good ones
are unpredictable to the best human chess players so we will have to be dealing with AI systems that are incredibly unpredictable they will understand things from limited data they will not get confused all the things which are really big limitations I'm not saying how by the way and I'm not saying when I'm saying that it will and when all those things will happen together with self-awareness because why not self-awareness is useful it is part your ourselves are parts of our own world models when all those come together we will have systems of radically different qualities
and properties that exist today and of course they will have incredible and amazing capabilities but the kind of issues that come up with systems like this and I'll just leave it as an exercise to um imagine it's very different from what we used to and I would say that it's definitely also impossible to predict the future really all kinds of stuff is possible so this is by far the most interesting part of the talk because this is where alas Sasa actually focuses on where AI research might be headed in the long term and this is
packed with some juicy information so he actually talks about super intelligence and if you remember that's the company that he's actually building right now now he first talks about you know it being a gentic and this is where instead of you know passively waiting for instructions a future super intelligent system might take the initiative and act on its own and it's going to have its goals it's going to have its purposes or sets or you know it's going to redefine itself and it's going to be able to seek out opportunities tools or information to achieve
those specific goals and this is going to be insane because this is a big shift from the current models which just respond when prompted but these models are going to be you know probably even reflecting from the environment and learning from the environment which is something that we haven't really had yet now of course today's a models often give the impression of thinking and reasoning but they mostly pattern matches with a robust step-by-step reasoning process but in contrast super intelligence would be able to genuinely reason and carefully consider problems weigh the different possibilities work out
logical steps and make sound decisions even in complex or unfamiliar situations now the crazy thing about all of this as well is that this AI is going to you know understand us you know current AI basically mimics what we do by producing humanlike text or answers and sometimes it even stumbles when dealing with logical or tricky problems s but a super intelligence is going to have a more profound grasp of Concepts it's not going to know what words tend to go together it's truly going to understand the meanings and the relationships and be able to
interpret everything quite like humans do but probably beyond that now the last thing that I found to be truly you know fascinating was the fact that they speak about you know this thing being self-aware and this is the most speculative quality this means that the AI wouldn't just process information externally it's going to have an internal sense of self its own goals its own reasoning process and its place in the world and self-awareness would let it reflect on its own actions potentially leading to more deliberate improvements more considerations and responsible behavior which are qualities we
often associate with constant beings and essentially this slide is absolutely incredible it paints a picture of a future where AI is not just an advanced tool but actually something much more akin to a new form of intelligence one that acts on its own one that thinks deeply and coherently one that truly understands complex ideas and is going to have a sense of itself which is going to be incredible I have a question for you um about sort of autocorrect um so here is here's the question you mentioned reasoning as being um one of the core
aspects of maybe the modeling in the future and maybe a differentiator um what we saw in some of the poster sessions is that hallucinations in today's models are the way we're analyzing mean maybe you correct me you're the expert on this but the way we're analyzing whether a model is hallucinating today without because we know of the dangers of models not being able to reason that we're using a statistical analysis let's say some amount of standard deviations or whatever away from the mean in the future wouldn't it would do you think that a model given
reasoning will be able to correct itself sort of autocorrect itself and that will be a core feature of future model so that there won't be as many hallucinations because the model will recognize when I maybe that's too esoteric of a question but the model will be able to reason and understand when a Hallucination is occurring does the question make sense yes and the answer is also yes I think what you described is extremely highly plausible yeah I mean you should check I mean for yeah it I wouldn't I wouldn't rule out that it might already
be happening with some of the you know early reasoning models of today I don't know but longer term why not yeah I mean it's part of like Microsoft Word like autoc correct it's a you know it's a it's a core feature yeah I just I mean I think calling it autocorrect is really doing it disservice I think you are when you say autocorrect you evoke like it's far grander than autocorrect but other but you know this point aside the answer is yes thank you and that was one of the most fascinating parts of this interview
because he actually responds to someone that asks about hallucinations and whether or not these models are going to be able to solve them now recently we did get some information from some Alman and some people working at open ey and Microsoft talking about how hallucinations are a very tricky problem to solve but ilk seems to believe that in the future these hallucinations are going to no longer exist which means that this is going to be a remarkable step because one of the main problems that stops AI from being widely used is the reliability problem you
have to understand that some Industries cannot afford hallucination in the kind of software that they're using for example in Planes the failure rate for Bo Parts is like 0.1% which means that if you have an AI that fails at around 98% you can't really deploy that safely so this is going to be something that if it does get solved which ilos is saying that it is going to get solved means that well we're going to be living in a very different future the next part here is once again I say everything is interesting but this
one is genuinely super interesting because this is where we get systems that can do everything a human can do and basically everything the human hasn't even thought about yet this is where Elisa is responding to a question that asks about in distribution generalization and this is just basically where if the model is going to be able to solve problems that it hasn't seen before and this is something that humans do very regularly you know you have maybe 16 to 20 hours of driving lessons but you can realistically easily drive all around the world based on
driving in a small City this is something that you know current autonomous cars can't really do you have to like map out the city you know huge training data this is something that that kind of an issue so if we can actually get AI systems that can generalize outside of their training data that is going to be genuinely game changing because you're not going to need you know huge data you're not going to need large data sets it's going to be a situation where you know you only put it in for 10 hours or whatever
and it's able to look at new scenarios and it's going to be able to generalize out of distribution and those implications are pretty incredible because it means that the AI is going to be able to solve things it's never seen before or never encountered which is just incredible the question should not be answered with yes or no because what does it mean out of distribution generalization what does it mean what does it mean in distribution and what does it mean out of distribution because it's a test of time talk I'll say that long long ago
before people were using deep learning they were using things like string matching and grams for machine translation people were using statistical phrase tables can you imagine they had tens of thousands of code of complexity which was I mean it's it was truly unfathomable and back then generalization meant is it literally not in this the same phrasing as in the data set now we may say well sure my model achieves this high score on um I don't know math competitions but maybe the math maybe some discussion in some Forum on the internet was about the same
ideas and therefore it's memorized well okay you could say maybe maybe it's in distribution maybe it's memorization but I also think that our standards for what counts as generalization have increased really quite substantially dramatically unimaginably if you keep track and so I think the answer is to some degree probably not as well as human beings I think it is true that human beings generalize much better but at the same time they definitely generalize out of distribution to some degree I hope it's a useful topological answer