Dario Amodei: Anthropic CEO on Claude, AGI & the Future of AI & Humanity | Lex Fridman Podcast #452

140.73k views63591 WordsCopy TextShare
Lex Fridman
Dario Amodei is the CEO of Anthropic, the company that created Claude. Amanda Askell is an AI resear...
Video Transcript:
if you extrapolate the curves that we've had so far right if if you say well I don't know we're starting to get to like PhD level and and last year we were at undergraduate level and the year before we were at like the level of a high school student again you can you can quibble with at what tasks and for what we're still missing modalities but those are being added like computer use was added like image generation has been added if you just kind of like eyeball the rate at which these capabilities are increasing it
does make you think that we'll get there by 2026 or 2027 I think there are still worlds where it doesn't happen in in a 100 years those world the number of those worlds is rapidly decreasing we are rapidly running out of truly convincing blockers truly compelling reasons why this will not happen in the next few years the scale up is very quick like we we do this today we make a model and then we deploy thousands maybe tens of thousands of instances of it I think by the time you know certainly within two to three
years whether we have these super powerful AIS or not ERS are going to get to the size where you'll be able to deploy millions of these I am optimistic about meaning I worry about economics and the concentration of power that's actually what I worry about more the abuse of power and AI increases the amount of power in the world and if you concentrate that power and abuse that power it can do immeasurable damage yes it's very frightening it's very it's very frightening the following is a conversation with Dario amade CEO of anthropic the company that
created Claude that is currently and often at the top of most llm Benchmark leader boards on top of that Dario and the anthropic team have been outspoken advocates for taking the topic of AI safety very seriously and they have continued to publish a lot of fascinating AI research on this and other topics I'm also joined afterwards by two other brilliant people from propic first Amanda ascal who is a researcher working on alignment and fine-tuning of Claude including the design of claude's character and personality a few folks told me she has probably talked with Claude more
than any human at anthropic so she was definitely a fascinating person to talk to about prompt engineering and practical advice on how to get the best out of Claude after that chrisa stopped by for chat he's one of the pioneers of the field of mechanistic interpretability which is an exciting set of efforts that aims to reverse engineer neural networks to figure out what's going on inside inferring behaviors from neural activation patterns inside the network this is a very promising approach for keeping future super intelligent AI systems safe for example by detecting from the activations when
the model is trying to deceive the human it is talking to this is Alex Freedman podcast to support it please check out our sponsors in the description and now dear friends here's Dario amade let's start with a big idea of scaling laws and the scaling hypothesis what is it what is its history and where do we stand today so I can only describe it as it you know as it relates to kind of my own experience but I've been in the AI field for about uh 10 years and it was something I noticed very early
on so I first joined the AI world when I was uh working at BYU with Andrew in in late 2014 which is almost exactly 10 years ago now and the first thing we worked on was speech recognition systems and in those days I think deep learning was a new thing it had made lots of progress but everyone was always saying we don't have the algorithms we need to succeed you know we we we we're we're not we're only matching a tiny tiny fraction there's so much we need to kind of discover algorithmically we haven't found
the picture of how to match the human brain uh and when you know in some ways was fortunate I was kind of you know you can have almost beginner's luck right I was like a a newcomer to the field and you know I looked at the neural net that we were using for speech the recurrent neural networks and I said I don't know what if you make them bigger and give them more layers and what if you scale up the data along with this right I just saw these as as like independent dials that you
could turn and I noticed that the model started to do better and better as you gave them more data as you as you made the models larger as you trained them for longer um and I I didn't measure things precisely in those days but but along with with colleagues we very much got the informal sense that the more data and the more compute and the more training you put into these models the better they perform and so initially my thinking was hey maybe that is just true for speech recognition systems right maybe maybe that's just
one particular quirk one particular area I think it wasn't until 2017 when I first saw the results from gpt1 that it clicked for me that language is probably the area in which we can do this we can get trillions of words of language data we can train on them and the models we were training in those days were tiny you could train them on one to eight gpus whereas you know now we train jobs on tens of thousands soon going to hundreds of thousands of gpus and so when I when I saw those two things
together um and you know there were a few people like ilaser who who you've interviewed who had somewhat similar reviews right he might have been the first one although I think a few people came to came to similar views around the same time Right There Was You Know Rich Sutton's bitter lesson there was gur wrote about the scaling hypothesis but I think somewhere between 2014 and 2017 was when it really clicked for me when I really got conviction that hey we're going to be able to do these incredibly wide cognitive tasks if we just if
we just scale up the models and at at every stage of scaling there are always arguments and you know when I first heard them honestly I thought probably I'm the one who's wrong and you know all these all these experts in the field are right they know the situation better better than I do right there's you know the Chomsky argument about like you can get syntactics but you can't get semantics there's this idea oh you can make a sentence make sense but you can't make a paragraph makes sense the latest one we have today is
uh you know we're going to run out of data or the data isn't high quality enough or models can't reason and and each time every time we manage to we manage to either find a way around or scaling just is the way around um sometimes it's one sometimes it's the other uh and and so I'm now at this point I I I still think you know it's it's it's always quite uncertain we have nothing but inductive inference to tell us that the next few years are going to be like the next the last 10 years
but but I've seen I've seen the movie enough times I've seen the story happen for for enough times to to really believe that probably the scaling is going to continue and that there's some magic to it that we haven't really explained on a theoretical basis yet and of course the scaling here is bigger networks bigger data bigger compute yes all in in particular linear scaling up of bigger networks bigger training times and uh more and and more data uh so all of these things almost like a chemical reaction you know you have three ingredients in
the chemical reaction and you need to linearly scale up the three ingredients if you scale up one not the others you run out of the other reagents and and the reaction stops but if you scale up everything everything in series then then the reaction can proceed and of course now that you have this kind of empirical scienceart you can apply it to other uh more nuanced things like scaling laws applied to interpretability or scaling laws applied to posttraining or just seeing how does this thing scale but the big scaling law I guess the underlying scaling
hypothesis has to do with big networks Big Data leads to intelligence yeah we've we've documented scaling laws in lots of domains other than language right so uh initially the the paper we did that first showed it was in early 2020 where we first showed it for language there was then some work late in 2020 where we showed the same thing for other modalities like images video text to image image to text math they all had the same pattern and and you're right now there are other stages like posttraining or there are new types of reasoning
models and in in in all of those cases that we've measured we see similar similar types of scaling laws a bit of a philosophical question but what's your intuition about why bigger is better in terms of network size and data size why does it lead to more intelligent models so in my previous career as a as a biophysicist so I did physics undergrad and then biophysics in in in in grad school so I think back to what I know as a physicist which is actually much less than what some of my colleagues at anthropic have
in terms of in terms of expertise in physics uh there's this there's this concept called the one over F noise and one overx distributions um where where often um uh you know just just like if you add up a bunch of natural processes you get gaussian if you add up a bunch of kind of differently distributed natural processes if you like if you like take a take a um probe and and hook it up to a resistor the distribution of the thermal noise in the resistor goes as one over the frequency um it's some kind
of natural convergent distribution uh and and I I I I and and I think what it amounts to is that if you look at a lot of things that are that are produced by some natural process that has a lot of different scales right not a gaussian which is kind of narrowly distributed but you know if I look at kind of like large and small fluctuations that lead to lead to electrical noise um they have this decaying 1 overx distribution and so now I think of like patterns in the physical world right if I if
or or in language if I think about the patterns in language there are some really simple patterns some words are much more common than others like the' then there's basic noun verb structure then there's the fact that you know you know nouns and verbs have to agree they have to coordinate and there's the higher level sentence structure then there's the Thematic structure of paragraphs and so the fact that there's this regressing structure you can imagine that as you make the networks larger first they capture the really simple correlations the really simple patterns and there's this
long taale of other patterns and if that long taale of other patterns is really smooth like it is with the one over F noise in you know physical processes like like like resistors then you could imagine as you make the network larger it's kind of capturing more and more of that distribution and so that smoothness gets reflected in how well the models are at predicting and how well they perform language is an evolved process right we've we've developed language we have common words and less common words we have common expressions and less common Expressions we
have ideas cliches that are expressed frequently and we have novel ideas and that process has has developed has evolved with humans over millions of years and so the the the guess and this is pure speculation would be would be that there is there's some kind of longtail distribution of of of the distribution of these ideas so there's the long tail but also there's the height of the hierarchy of Concepts that you're building up so the bigger the network presumably you have a higher capacity to exactly if you have a small Network you only get the
common stuff right if if I take a tiny neural network it's very good at understanding that you know a sentence has to have you know verb adjective noun right but it's it's terrible at deciding what those verb adjective and noun should be and whether they should make sense if I make it just a little bigger it gets good at that then suddenly it's good at the sentences but it's not good at the paragraphs and so the these these rare and more complex patterns get picked up as I add as I add more capacity to the
network well the natural question then is what's the ceiling of this like how complicated and complex is the real world how much of stuff is there to learn I don't think any of us knows the answer to that question um I my strong Instinct would be that there's no ceiling below level of humans right we humans are able to understand these various patterns and so that that makes me think that if we continue to you know scale up these these these models to kind of develop new methods for training them and scaling them up uh
that will at least get to the level that we've gotten to with humans there's then a question of you know how much more is it possible to understand than humans do how much how much is it possible to be smarter and more perceptive than humans I I would guess the answer has has got to be domain dependent if I look at an area like biology and you know I wrote this essay Machines of Loving Grace it seems to me that humans are struggling to understand the complexity of biology right if you go to Stanford or
to Harvard or to Berkeley you have whole Departments of you know folks trying to study you know like the immune system or metabolic pathways and and each person understands only a tiny bit part of it specializes and they're struggling to combine their knowledge with that of with that of other humans and so I have an instinct that there's there's a lot of room at the top for AIS to get smarter if I think of something like materials in the in the physical world or you know um like addressing you know conflicts between humans or something
like that I mean you know it it may be there's only some of these problems are not intractable but much harder and and it it may be that there's only there's only so well you can do with some of these things right just like with speech recognition there's only so clear I can hear your speech so I think in some areas there may be ceilings in in in you know that are very close to what humans have done in other areas those ceilings may be very far away and I think we'll only find out when
we build these systems uh there's it's very hard to know in advance we can speculate but we can't be sure and in some domains the ceiling might have to do with human bureaucracies and things like this as you're right about yes so humans fundamentally have to be part of the loop that's the cause of the ceiling not maybe the limits of the intelligence yeah I think in many cases um you know in theory technology could change very fast for example all the things that we might invent with respect to biology um but remember there's there's
a you know there's a clinical trial system that we have to go through to actually administer these things to humans I think that's a mixture of things that are unnecessary and bureaucratic and things that kind of protect the Integrity of society and the whole challenge is that it's hard to tell it's hard to tell what's going on uh it's hard to tell which is which right my my view is definitely I think in terms of drug development we my view is that we're too slow and we're too conservative but certainly if you get these things
wrong you know it's it's possible to to to risk people's lives by by being by being by being too Reckless and so at least at least some of these human institutions are in fact protecting people so it's it's all about finding the balance I strongly suspect that balance is kind of more on the side of pushing to make things happen faster but there is a balance if we do hit a limit if we do hit a Slowdown in the scaling laws what do you think would be the reason is it compute limited data limited uh
is it something else idea limited so a few things now we're talking about hitting the limit before we get to the level of of humans and the skill of humans um so so I think one that's you know one that's popular today and I think you know could be a limit that we run into I like most of the limits I would bet against it but it's definitely possible is we simply run out of data there's only so much data on the internet and there's issues with the quality of the data right you can get
hundreds of trillions of words on the internet but a lot of it is is repetitive or it's search engine you know search engine optimization driil or maybe in the future it'll even be text generated by AIS itself uh and and so I think there are limits to what to to what can be produced in this way that said we and I would guess other companies are working on ways to make data synthetic uh where you can you know you can use the model to generate more data of the type that you have that you have
already or even generate data from scratch if you think about uh what was done with uh deep mines Alpha go zero they managed to get a bot all the way from you know no ability to play Go whatsoever to above human level just by playing against itself there was no example data from humans required in the the alphao zero version of it the other direction of course is these reasoning models that do Chain of Thought and stop to think um and and reflect on their own thinking in a way that's another kind of synthetic data
coupled with reinforcement learning so my my guess is with one of those methods we'll get around the data limitation or there may be other sources of data that are that are available um we could just observe that even if there's no problem with data as we start to scale models up they just stop getting better it's it seemed to be a a reliable observation that they've gotten better that could just stop at some point for a reason we don't understand um the answer could be that we need to uh you know we need to invent
some new architecture um it's been there have been problems in the past with with say numerical stability of models where it looked like things were were leveling off but but actually you know know when we when we when we found the right Unblocker they didn't end up doing so so perhaps there's new some new optimization method or some new uh Technique we need to to unblock things I've seen no evidence of that so far but if things were to to slow down that perhaps could be one reason what about the limits of compute meaning uh
the expensive uh nature of building bigger and bigger data centers so right now I think uh you know most of the Frontier Model companies I would guess are are operating you know roughly you know $1 billion scale plus or minus a factor of three right those are the models that exist now or are being trained now uh I think next year we're going to go to a few billion and then uh 2026 we may go to uh uh you know above 10 10 10 billion and probably by 2027 their Ambitions to build hundred hundred billion
dollar uh hundred billion dollar clusters and I think all of that actually will happen there's a lot of determination to build the compute to do it within this country uh and I would guess that it actually does happen now if we get to 100 billion that's still not enough compute that's still not enough scale then either we need even more scale or we need to develop some way of doing it more efficiently of Shifting The Curve um I think be between all of these one of the reasons I'm bullish about powerful AI happening so fast
is just that if you extrapolate the next few points on the curve we're very quickly getting towards human level ability right some of the new models that that we developed some some reasoning models that have come from other companies they're starting to get to what I would call the PHD or professional level right if you look at their their coding ability um the latest model we released Sonet 3.5 the new or updated version it gets something like 50% on sbench and sbench is an example of a bunch of professional real world software engineering tasks at
the beginning of the year I think the state-of-the-art was three or 4% so in 10 months we've gone from 3% to 50% on this task and I think in another year we'll probably be at 90% I mean I don't know but might might even be might even be less than that uh we've seen similar things in graduate level math physics and biology from Models like open AI 01 uh so uh if we if we just continue to extrapolate this right in terms of skill skill that we have I think if we extrapolate the straight curve
Within a few years we will get to these models being you know above the the highest professional level in terms of humans now will that curve continue you've pointed to and I've pointed to a lot of reasons why you know possible reasons why that might not happen but if the if the extrapolation curve continues that is the trajectory we're on so anthropic has several competitors it'd be interesting to get your sort of view of it all open aai Google xai meta what does it take to win in the broad sense of win in the space
yeah so I want to separate out a couple things right so you know anthropics anthropic mission is to kind of try to make this all go well right and and you know we have a theory of change called race to the top right race to the top is about trying to push the other players to do the right thing by setting an example it's not about being the good guy it's about setting things up so that all of us can be the good guy I'll give a few examples of this early in the history of
anthropic one of our co-founders Chris Ola who I believe you're you're interviewing soon you know he's the co-founder of the field of mechanistic interpretability which is an attempt to understand what's going on inside AI models uh so we had him and one of our early teams focus on this area of interpretability which we think is good for making models safe and transparent for three or four years that had no commercial application whatsoever it still doesn't today we're doing some early betas with it and probably it will eventually but uh you know this is a very
very long research bed in one in which we've we've built in public and shared our results publicly and and we did this because you know we think it's a way to make models safer an interesting thing is that as we've done this other companies have started doing it as well in some cases because they've been inspired by it in some cases because they're worried that uh you know if if other companies are doing this that look more responsible they want to look more responsible too no one wants to look like the irresponsible ible actor and
and so they adopt this they adopt this as well when folks come to anthropic interpretability is often a draw and I tell them the other places you didn't go tell them why you came here um and and then you see soon that there that there's interpretability teams else elsewhere as well and in a way that takes away our competitive Advantage because it's like oh they now others are doing it as well but it's good it's good for the broader system and so we have to invent some new thing that we're doing others aren't doing as
well and the hope is to basically bid up bid up the importance of of of doing the right thing and it's not it's not about us in particular right it's not about having one particular good guy other companies can do this as well if they if they if they join the race to do this that's that's you know that's the best news ever right um uh it's it's just it's about kind of shaping the incentives to point upward instead of shaping the incentives to point to point downward and we should say this example the field
of uh mechanistic interpretability is just a a rigorous non handwavy way of doing AI safety yes or it's tending that way trying to I mean I I think we're still early um in terms of our ability to see things but I've been surprised at how much we've been able to look inside these systems and understand what we see right unlike with the scaling laws where it feels like there's some you know law that's driving these models to perform better on on the inside the models aren't you know there's no reason why they should be designed
for us to understand them right they're designed to operate they're designed to work just like the human brain or human biochemistry they're not designed for a human to open up the hatch look inside and understand them but we have found and you know you can talk in much more detail about this to Chris that when we open them up when we do look inside them we we find things that are surprisingly interesting and as a side effect you also get to see the beauty of these models you get to explore the sort of uh the
beautiful n nature of large neural networks through the me turb kind ofy I'm amazed at how clean it's been I I'm amazed at things like induction heads I'm amazed at things like uh you know that that we can you know use sparse autoencoders to find these directions within the networks uh and that the directions correspond to these very clear Concepts we demonstrated this a bit with the Golden Gate Bridge clad so this was an experiment where we found a direction inside one of the the neural network layers that corresponded to the Golden Gate Bridge and
we just turned that way up and so we we released this model as a demo it was kind of half a joke uh for a couple days uh but it was it was illustrative of of the method we developed and uh you could you could take the Golden Gate you could take the model you could ask it about anything you know you know it would be like how you could say how was your day and anything you asked because this feature was activated would connect to the Golden Gate Bridge so it would say you know
I'm I'm I'm feeling relaxed and expansive much like the the arches of the Golden Gate Bridge or you know it would masterfully change topic to the Golden Gate Bridge and it integrated there was also a sadness to it to to the focus ah had on the Golden Gate Bridge I think people quickly fell in love with it I think so people already miss it because it was taken down I think after a day somehow these interventions on the model um where where where where you kind of adjust Its Behavior somehow emotionally made it seem more
human than any other version of the model strong personality strong ID strong personality it has these kind of like obsessive interests you know we can all think of someone who's like obsessed with something so it does make it feel somehow a bit more human let's talk about the present let's talk about Claude so this year A lot has happened in March claw 3 Opa Sonet Hau were released then claw 35 Sonet in July with an updated version just now released and then also claw 35 hi coup was released okay can you explain the difference between
Opus Sonet and Haiku and how we should think about the different versions yeah so let's go back to March when we first released uh these three models so you know our thinking was you different companies produce kind of large and small models better and worse models we felt that there was demand both for a really powerful model um you know and you that might be a little bit slower that you'd have to pay more for and also for fast cheap models that are as smart as they can be for how fast and cheap right whenever
you want to do some kind of like you know difficult analysis like if I you know I want to write code for instance or you know I want to I want to brainstorm ideas or I want to do creative writing I want the really powerful model but then there's a lot of practical applications in a business sense where it's like I'm interacting with a website I you know like I'm like doing my taxes or I'm you know talking to uh you know to like a legal adviser and I want to analyze a contract or you
know we have plenty of companies that are just like you know you know I want to do autocomplete on my on my IDE or something uh and and for all of those things you want to act fast and you want to use the model very broadly so we wanted to serve that whole spectrum of needs um so we ended up with this uh you know this kind of poetry theme and so what's a really short poem it's a Haik cou and so Haiku is the small fast cheap model that is you know was at the
time was released surprisingly surprisingly uh intelligent for how fast and cheap it was uh sonnet is a is a medium-sized poem right a couple paragraphs since o Sonet was the middle model it is smarter but also a little bit slower a little bit more expensive and and Opus like a magnum opus is a large work uh Opus was the the largest smartest model at the time um so that that was the original kind of thinking behind it um and our our thinking then was well each new generation of models should shift that tradeoff curve uh
so when we release Sonet 3.5 it has the same roughly the same you know cost and speed as the Sonet 3 Model uh but uh it it increased its intelligence to the point where it was smarter than the original Opus 3 Model uh especially for code but but also just in general and so now you know we've shown results for a Hau 3. 5 and I believe Hau 3.5 the smallest new model is about as good as Opus 3 the largest old model so basically the aim here is to shift the curve and then at
some point there's going to be an opus 3.5 um now every new generation of models has its own thing they use new data their personality changes in ways that we kind of you know try to steer but are not fully able to steer and and so uh there's never quite that exact equivalence the only thing you're changing is intelligence um we always try and improve other things and some things change without us without us knowing or measuring so it's it's very much an inexact science in many ways the manner and personality of these models is
more an art than it is a science so what is sort of the reason for uh the span of time between say Claude Opus 3 and 35 what is it what takes that time if you can speak to yeah so there's there's different there's different uh processes um uh there's pre-training which is you know just kind of the normal language model training and that takes a very long time um that uses you know these days you know tens you know tens of thousands sometimes many tens of thousands of uh gpus or tpus or tranium or
you know what we use different platforms but you know accelerator chips um often often training for months uh there's then a kind of posttraining phase where we do reinforcement learning from Human feedback as well as other kinds of reinforcement learning that that phase is getting uh larger and larger now and you know you know often that's less of an exact science it often takes effort to get it right um models are then tested with some of our early Partners to see how good they are and they're then tested both internally and externally for their safety
particularly for catastrophic and autonomy r risks uh so uh we do internal testing according to our responsible scaling policy which I you know could talk more about that in detail and then we have an agreement with the US and the UK AI safety Institute as well as other third-party testers in specific domains to test the models for what are called cbrn risk chemical biological radiological and nuclear which are you know we don't think that models pose these risks seriously yet but but every new model we want to evaluate to see if we're starting to get
close to some of these these these more dangerous um uh these more dangerous capabilities so those are the phases and then uh you know then then it just takes some time to get the model working in terms of inference and launching it in the API so there's just just a lot of steps to uh to actually to actually making a model work and of course you know we're always trying to make the processes as streamlined as possible right we want our safety testing to be rigorous but we want it to be RoR ous and to
be you know to be automatic to happen as fast as it can without compromising on rigor same with our pre-training process and our posttraining process so you know it's just like building anything else it's just like building airplanes you want to make them you know you want to make them safe but you want to make the process streamlined and I think the creative tension between those is is you know is an important thing and making the models work yeah uh rumor on the street I forget who was saying that uh anthropic is really good tooling
so I uh probably a lot of the challenge here is on the software engineering side is to build the tooling to to have a like a efficient low friction interaction with the infrastructure you would be surprised how much of the challenges of uh you know building these models comes down to you know software engineering performance engineering you know you you know from the outside you might think oh man we had this Eureka breakthrough right you know this movie with the science we discovered it we figured it out but but but I think I think all
things even even even you know incredible discoveries like they they they they they almost always come down to the details um and and often super super boring details I can't speak to whether we have better tooling than than other companies I mean you know I haven't been at those other companies at least at least not recently um but it's certainly something we give a lot of attention to I don't know if you can say but from three from CLA 3 to CLA 35 is there any extra pre-training going on or is they mostly focus on
the post-training there's been leaps in performance yeah I think I think at any given stage we're focused on improving everything at once um just just naturally like there are different teams each team makes progress in a particular area in in in making a particular you know their particular segment of the relay race better and it's just natural that when we make a new model we put we put all of these things in at once so the data you have like the preference data you get from rhf is that applicable is there ways to apply it
to newer models as it get trained up yeah preference data from old models sometimes gets used for new models although of course uh it it performs somewhat better when it's you know trained on it's trained on the new models note that we have this you know constitutional AI method such that we don't only use preference data we kind of there's also a post-t trainining process where we train the model against itself and there's you know new types of post training the model against itself that are used every day so it's not just RF it's a
bunch of other methods as well um post training I think you know it's becoming more and more sophisticated well what explains the big leap in performance for the new Sona 35 I mean at least in the programming side and maybe this is a good place to talk about benchmarks what does it mean to get better just the number went up but you know I I I program but I also love programming and I um claw 35 through cursor is what I use uh to assist me in programming and there was at least experientially anecdotally it's
gotten smarter at programming so what like what what does it take to get it uh to get it smarter we observe that as well by the way there were a couple uh very strong Engineers here at anthropic um who all previous code models both produced by us and produced by all the other companies hadn't really been useful to to hadn't really been useful to them you know they said you know maybe maybe this is useful to beginner it's not useful to me but Sonet 3.5 the original one for the first time they said oh my
God this helped me with something that you know that it would have taken me hours to do this is the first model that has actually saved me time so again the water line is rising and and then I think you know the new Sonet has been has been even better in terms of what it what it takes I mean I'll just say it's been across the board it's in the pre-training it's in the posttraining it's in various evaluations that we do we've observed this as well and if we go into the details of the Benchmark
so s bench is basically you know since since you know since since you're a programmer you know you'll be familiar with like PLL requests and you know uh just just PLL requests are like you know the like a sort of a sort of atomic unit of work you know you could say I'm you know I'm implementing one I'm implementing one thing um uh and and so sbench actually gives you kind of a real world situation where the codebase is in a current state and I'm trying to implement something that's you know that's described in described
in language we have internal benchmarks where we where we measure the same thing and you say just give the model free reign to like you know do anything run run run anything edit anything um how how well is it able to complete these tasks and it's that Benchmark that's gone from it can do it 3% of the time to it can do it about 50% of the time um so I actually do believe that if we get you can gain benchmarks but I think if we get to 100% on that Benchmark in a way that
isn't kind of like overtrained or or or game for that particular Benchmark probably represents a real and serious increase in kind of in kind of programming programming ability and and I would suspect that if we can get to you know 90 90 95% that that that that you know it will it will represent ability to autonomously do a significant fraction of software engineering tasks well ridiculous timeline question uh when is clad Opus uh 3.5 coming up uh not giving you an exact date uh but you know there there uh you know as far as we
know the plan is still to have a Claude 3.5 opus are we gonna get it before GTA 6 or no like Duke Nukem Forever was that game that there was some game that was delayed 15 years was that Duke Nukem Forever yeah and I think GTA is now just releasing trailers it you know it's only been three months since we released the first son it yeah it's Inc the incredible pace of relas it just it just tells you about the pace the expectations for when things are going to come out so uh what about 40
so how do you think about sort of as these models get bigger and bigger about versioning and also just versioning in general why Sonet 35 updated with the date why not Sonet 3.6 actually naming is actually an interesting challenge here right because I think a year ago most of the model was pre-training and so you could start from the beginning and just say okay we're going to have models of different sizes we're going to train them all together and you know we'll have a a family of naming schemes and then we'll put some new magic
into them and then you know we'll have the next the next Generation Um the trouble starts are already when some of them take a lot longer than others to train right that already messes up your time time a little bit but as you make big improvements in as you make big improvements in pre-training uh then you suddenly notice oh I can make better pre-train model and that doesn't take very long to do and but you know clearly it has the same you know size and shape of previous models uh uh so I think those two
together as well as the timing timing issues any kind of scheme you come up with uh you know the reality tends to kind of frustrate that scheme right T tends to kind of break out of the break out of the scheme it's not like software where you can say oh this is like you know 3.7 this is 3.8 no you have models with different different tradeoffs you can change some things in your models you can train you can change other things some are faster and slower at inference some have to be more expensive some have
to be less expensive and so I think all the companies have struggled with this um I think we did very you know I think think we were in a good good position in terms of naming when we had Haiku Sonet and we're trying to maintain it but it's not it's not it's not perfect um so we'll we'll we'll try and get back to the Simplicity but it it um uh just the the the nature of the field I feel like no one's figured out naming it's somehow a different Paradigm from like normal software and and
and so we we just none of the companies have been perfect at it um it's something we struggle with surprisingly much relative to you know how relative to how trivial it is to you know for the the the the grand science of training the models so from the user side the user experience of the updated Sonet 35 is just different than the previous uh June 2024 Sonet 35 it would be nice to come up with some kind of labeling that embodies that because people talk about son 35 but now there's a different one and so
how do you refer to the previous one and the new one and it it uh when there's a distinct Improvement it just makes conversation about it uh just challenging yeah yeah I I definitely think this question of there are lots of properties of the models that are not reflected in the benchmarks um I I think I think that's that's definitely the case and everyone agrees and not all of them are capabilities some of them are you know models can be polite or brusk they can be uh you know uh very reactive or they can ask
you questions um they can have what what feels like a warm personality or a cold personality they can be boring or they can be very distinctive like Golden Gate Claude was um and we have a whole you know we have a whole team kind of focused on I think we call it Claude character uh Amanda leads that team and we'll we'll talk to you about that but it's still a very inexact science um and and often we find that models have properties that we're not aware of the the fact of the matter is that you
can you know talk to a model 10,000 times and there are some behaviors you might not see uh just like just like with a human right I can know someone for a few months and you know not know that they have a certain skill or not know there's a certain side to them and so I think I think we just have to get used to this idea and we're always looking for better ways of testing our models to to demonstrate these capabilities and and and also to decide which are which are the which are the
personality properties we want models to have have and which we don't want to have that itself the normative question is also super interesting I got to ask you a question from Reddit from Reddit oh boy you know there there's just this fascinating to me at least it's a psychological social phenomenon where people report that Claude has gotten Dumber for them over time and so uh the question is does the user complaint about the dumbing down of claw 35 Sonic hold any water so are these anecdota reports a kind of social phenomena or did Claude is
there any cases where Claude would get Dumber so uh this actually doesn't apply this this isn't just about Claude I I believe this I believe I've seen these complaints for every Foundation model produced by a major company um people said this about gp4 they said it about gp4 turbo um so so so a couple things um one the actual weights of the model right the actual brain of the model that does not change unless we introduce a new model um there there just a number of reasons why it would not make sense practically to be
randomly substituting in substituting in new versions of the model it's difficult from an inference perspective and it's actually hard to control all the consequences of changing the way to the model let's say you wanted to fine-tune the model to be like I don't know to like to say certainly less which you know an old version of Sonet used to do um you actually end up changing a 100 things as well so we have a whole process for it and we have a whole process for modifying the model we do a bunch of testing on it
we do a bunch of um like we do a bunch of user testing and early customers so it we both have never changed the weights of the model without without telling anyone and it it it wouldn't certainly in the current setup it would not make sense to do that now there are a couple things that we do occasionally do um one is sometimes we run AB tests um but those are typically very close to when a model is being is being uh released and for a very small fraction of time um so uh you know
like the you know the the day before the new Sonet 3.5 I I agree we should have should have had a better name it's clunky to refer to it um there were some comments from people that like it's got It's got it's gotten a lot better and that's because you know a fraction were exposed to to an AB test for for those one or for those one or two days um the other is that occasionally the system prompt will change um on the system prompt can have some effects although it's un it it it's unlikely
to dumb down models it's unlikely to make them Dumber um and and and and we've seen that while these two things which I'm listing to be very complete um happen relatively happen quite infrequently um the complaints about to for us and for other model companies about the model changed the model isn't good at this the model got more censored the model was dumb down those complaints are constant and so I don't want to say like people are imagining it or anything but like the models are for the most part not changing um if I were
to offer a theory um I I think it actually relates to one of the things I said before which is that models have many are very complex and have many aspects to them and so often you know if I if I if if I ask a model a question you know if I'm like if I'm like do task X versus can you do task XX the model might respond in different ways uh and and so there are all kinds of subtle things that you can change about the way you interact with the model that can
give you very different results um to be clear this this itself is like a failing by by us and by the other model providers that that the models are are just just often sensitive to like small small changes in wording it's yet another way in which the science of how these models work is very poorly developed uh and and so you know if I go to sleep one night and I was like talking to the model in a certain way and I like slightly Chang the phrasing of how I talk to the model you know
I could I could get different results so that's that's one possible way the other thing is man it's just hard to quantify this stuff uh it's hard to quantify this stuff I think people are very excited by new models when they come out and then as time goes on they they become very aware of the they become very aware of the limitations so that may be another effect but that's that's all a very long- rended way of saying for the most part with some fairly narrow exceptions the models are not changing I think there is
a psychological effect you just start getting used to it the Baseline ra like when people have first gotten Wi-Fi on airplanes it's like amazing magic and then now like I can't get this thing to work this is such a piece of crap exactly so it's easy to have the conspiracy theory of they're making Wi-Fi slower and slower this is probably something I'll talk to Amanda much more about but U another Reddit question uh when will Claud stop trying to be my uh panical grandmother imposing its moral World viw on me as a paying customer and
also what does it that ology behind making Claude overly apologetic so this kind of reports about The Experience a different angle on the frustration it has to do with the character yeah so a couple points on this first one is um like things that people say on Reddit and Twitter or X or whatever it is um there's actually a huge distribution shift between like the stuff that people complain loudly about on social media and what actually kind of like you know statistically users care about and that drives people to use the models like people are
frustrated with you know things like you know the model not writing out all the code or the model uh you know just just not being as good at code as it could be even though it's the best model in the world on code um I I think the majority of thing of things are about that um uh but uh certainly a a a kind of vocal minority are uh you know kind kind of kind of rais these concerns right are frustrated by the model refusing things that it shouldn't refuse or like apologizing too much or
just just having these kind of like annoying verbal ticks um the second caveat and I just want to say this like super clearly because I think it's like some people don't know it others like kind of know it but forget it like it is very difficult to control across the board how the models behave you cannot just reach in there and say oh I want the model to like apologize less like you can do that you can include trading data that says like oh the models should like apologize less but then in some other situation
they end up being like super rude or like overconfident in a way that's like misleading people so they're they're all these tradeoffs um uh for example another thing is if there was a period during which models ours and I think others as well were T verbose right they would like repeat themselves they would say too much um you can cut down on the verbosity by penalizing the models for for just talking for too long what happens when you do that if you do it in a crude way is when the models are coding sometimes they'll
say of the code goes here right because they've learned that that's a way to economize and that they see it and then and then so that leads the model to be so-called lazy in coding where they where they where they're just like ah you can finish the rest of it it's not it's not because we want to you know save on compute or because you know the models are lazy and you know during winter break or any of the other kind of conspiracy theories that have that have that have come up it's actually it's just
very hard to control the behavior of the model to steer the behavior of the model in all circum ances at once you can kind of there's this this whacka aspect where you push on one thing and like you know these these these you know these other things start to move as well that you may not even notice or measure and so one of the reasons that I that I care so much about uh you know kind of grand alignment of these AI systems in the future is actually these systems are actually quite unpredictable they're actually
quite hard to steer and control um and this version we're seeing today of you make one thing better it makes another thing worse uh I think that's that's like a present day analog of future control problems in AI systems that we can start to study today right I think I think that that that difficulty in in steering the behavior and in making sure that if we push an AI system in One Direction it doesn't push it in another Direction in some in some other ways that we didn't want uh I think that's that's kind of
an that's kind of an early sign of things to come and if we can do a good job of solving this problem right of like you ask the model to like you know to like make and distribute small pox and it says no but it's willing to like help you in your graduate level virology class like how do we get both of those things at once it's hard it's very easy to go to one side or the other and it's a multi-dimensional problem and so uh I you know I think these questions of like shaping
the models personality I think they're very hard I think we haven't done perfectly on them I think we've actually done the best of all the AI companies but still so far from perfect uh and I think if we can get this right if we can control the the you know control the false positives and false negatives in this this very kind of controlled present day environment will be much better at doing it for the future when our worry is you know will the models be super autonomous will they be able to you know make very
dangerous things will they be able to autonomously you know build whole companies and are those companies aligned so so I I I think of this this present task as both vacine but also good practice for the future what's the current best way of gathering sort of user feedback like uh not anecdotal data but just large scale data about pain points or the opposite of pain points positive things so on is it internal testing is it yeah A specific group testing a testing what what what works so so so typically um we'll have internal model bashings
where all of anthropic anthropic is almost a thousand people um you know people just just try and break the model they try and interact with it various ways um uh we have a suite of evals uh for you know oh is the model refusing in ways that that it couldn't I think we even had a certainly eval because you know our our mod again at one point model had this problem where like it had this annoying tick where it would like respond to a wide range of questions by saying certainly I can help you with
that certainly I would be happy to do that certainly this is correct um uh and so we had a like certainly eval which is like how how often does the model say certainly uh uh but but look this is just a whack-a-mole like like what if it switches from certainly to definitely like uh uh so you know every time we add a new eval and we're always evaluating for all the old things so we have hundreds of these evaluations but we find that there's no substitute for human interacting with it and so it's very much
like the ordinary product development process we have like hundreds of people within anthropic bash the model then we do uh you know then we do external AB tests sometimes we'll run tests with contractors we pay contractors to interact with the model um so you put all of these things together and it's still not perfect you still see behaviors that you don't quite want to see right you know you see you still see the model like refusing things that it just doesn't make sense to refuse um but I I I think trying to trying to solve
this challenge right trying to stop the model from doing you know genuinely bad things that you know no one everyone agrees it shouldn't do right you know everyone everyone you know everyone agrees that you know the model shouldn't talk about you know I I don't know child abuse material right like everyone agrees the model shouldn't do that uh but but at the same time that it doesn't refuse in these dumb and stupid ways uh I think I think draw drawing that line as finely as possible approaching perfectly is still is still a challenge and we're
getting better at it every day but there's there's a lot to be solved and again I would point to that as as an indicator of a challenge ahead in terms of steering much more powerful models do you think Claude 4.0 is ever coming out I don't want to commit to any naming scheme because if I say if I say here we're gonna have Claude 4 next year and then and then you know then we decide that like you know we should start over because there's a new type of mod like I I I I I
I don't want to I don't want to commit to it I would expect in a normal course of business that Claude four would come after Claude 3.5 but but you know you you you never know in this wacky field right but the sort of this idea of scaling is continuing scal scaling is continuing there there will definitely be more powerful models coming from us in the models that exist today that is that is certain or if there if there aren't we've we've deeply failed as a company okay can you explain the responsible scaling policy and
the AI safety level standards ASL levels as much as I'm excited about the benefits of these models and you know we'll talk about that if we talk about Machines of Loving Grace um I'm I'm worried about the risk and I continue to be worried about the risks uh no one should think that you know Machines of loveing Grace was me me saying uh you know I'm no longer worried about the risks of these models I think they're two sides of the same coin the the uh Power of the models and their ability to solve all
these problems in you know biology Neuroscience Economic Development government governance and peace large parts of the economy those those come with risks as well right with great power comes great responsibility right that's the the two are the two are paired uh things that are powerful can do good things and they can do bad things um I think of those risks as as being in you know several different different categories perhaps the two biggest risks that I think about and that's not to say that there aren't risks today that are that are important but when I
think of the really the the you know the things that would happen on the grandest scale um one is what I call catastrophic misuse these are misuse of the models in domains like cyber bio radiological nuclear right things that could you know that could harm or even kill thousands even millions of people if they really really go wrong um like these are the you know number one priority to prevent and and here I would just make a simple observation which is that Mo the models you know if if I look today at people who have
done really bad things in the world um uh I think actually Humanity has been protected by the fact that the overlap between really smart well-educated people and people who want to do really horrific things has generally been small like you know let's say let's say I'm someone who you know uh you know I have a PhD in this field I have a well-paying job um there's so much to lose why do I want to like you know even even assuming I'm completely evil which which most people are not um why why you know why would
such a person risk their risk their you know risk their life RK risk their their legacy their reputation to to do something like you know truly truly evil if we had a lot more people like that the world would be a much more dangerous place and so my my My worry is that by being a a much more intelligent agent AI could break that correlation and so I I I I I do have serious worries about that I believe we can prevent those worries uh but you know I I think as a Counterpoint to Machines
of Loving Grace I want to say that this is I there's still serious risks and and the second range of risks would be the autonomy risks which is the idea that models might on their own particularly as we give them more agency than they've had in the past uh particularly as we give them supervision over wider tasks like you know writing whole code bases or someday even you know effectively operating entire entire companies they're on a long enough leash are they are they doing what we really want them to do it's very difficult to even
understand in detail what they're doing let alone let alone control it and like I said this these early signs that it's it's hard to perfectly draw the boundary between things the model should do and things the model shouldn't do that that you know if if you go to one side you get things that are annoying and useless and you go to the other side you get other behaviors if you fix one thing it creates other problems we're getting better and better at solving this I don't think this is an unsolvable problem I think this is
a you know this is a science like like the safety of airplanes or the safety of cars or the safety of drugs I you know I I don't think there's any big thing we're missing I just think we need to get better at controlling these models and so these are these are the two risks I'm worried about and our responsible scaling plan which I'll recognize is a very long-winded answer to your question I love it I love it our responsible scaling plan is designed to address these two types of risks and so every time we
develop a new model we basically test it for its ability to do both of these bad things so if I were to back up a little bit um I I think we have a I think we have an interesting dilemma with AI systems where they're not yet powerful enough to present these catastrophes I don't know that I don't know they'll ever present prevent these catastrophes it's possible they won't but the the case for worry the case for risk is strong enough that we should we should act now and and they're they're getting better very very
fast right I you know I testified in the Senate that you know we might have serious bio risks within two to three years that was about a year ago things have preceded preceded a pace uh uh so we have this thing where it's like it's it's it's surprisingly hard to to address these risks because they're not here today they don't exist they're like ghosts but they're coming at us so fast because the models are improving so fast so so how do you deal with something that's not here today doesn't exist but is is coming at
us very fast uh so the solution we came up with for that in in collaboration with uh you know people like uh the organization meter and Paul Christiano is okay what what what what you need for that or you need tests to tell you when the risk is getting close you need an early warning system and and so every time we have uh a new model we test it for it capability to do these cbrn tasks as well as testing it for you know how capable it is of doing tasks autonomously on its own and
uh in the latest version of our RSP which we released in the last in the last month or two uh the way we test autonomy risks is the model the the AI model's ability to do aspects of AI research itself uh which when the model when the AI models can do AI research they become kind of truly truly autonomous on and that you know that threshold is important for a bunch of other ways and and so what do we then do with these tasks the RSP basically develops what we've called an if then structure which
is if the models pass a certain capability then we impose a certain set of Safety and Security requirements on them so today's models are what's called asl2 models that were a asl1 is for systems that manifestly don't pose any risk of autonomy or misuse so for example a chess plane bot deep blue would be asl1 it's just manifestly the case that you can't use deep blue for anything other than chess it was just designed for chess no one's going to use it to like you know to conduct a masterful Cyber attack or to you know
run wild and take over the world asl2 is today's AI systems where we've measured them and we think these systems are simply not smart enough to uh to you know autonomously self-replicate or conduct a bunch of tasks uh and also not smart enough to provide meaningful information about cbrn risks and how to build cbrn weapons above and beyond what can be known from looking at Google uh in fact sometimes they do provide information but but not above and beyond a search engine but not in a way that can be stitched together um not not in
a way that kind of end to end is dangerous enough so asl3 is going to be the point at which uh the models are helpful enough to enhance the capabilities of non-state actors right State actors can already do a lot a lot of unfortunately to a high level of proficiency a lot of these very dangerous and destructive things the difference is that non-state non-state actors are not capable of it and so when we get to asl3 we'll take special security precautions designed to be be sufficient to prevent theft of the model by non-state actors and
misuse of the model as it's deployed uh will have to have enhanced filters targeted at these particular areas cyber bio nuclear cyber bio nuclear and model autonomy Which is less a misuse risk and more a risk of the model doing bad things itself asl4 getting to the point where these models could could enhance the capability of a of a of a all knowledgeable State actor Andor become the you know the main source of such a risk like if you wanted to engage in such a risk the main way you would do it is through a
model and then I think asl4 on the autonomy side it's it's some some some amount of acceleration in AI research capabilities with an with an AI model and then asl5 is where we would get to the models that are you know that are that are kind of that are kind of you know truly capable that it could exceed Humanity in their ability to do to do any of these tasks and so the the the point of the if then structure commitment is is basically to say look I don't know I've been I've been working with
these models for many years and I've been worried about risk for many years it's actually kind of dangerous to cry wolf it's actually kind of dangerous to say this you know this this model is this model is risky and you know people look at it and they say this is manifestly not dangerous again it's it's it's the the delicacy of the risk isn't here to today but it's coming at us fast how do you deal with that it's it's really vexing to a risk planner to deal with it and so this if then structure basically
says look we don't want to antagonize a bunch of people we don't want to harm our own you know our our kind of own ability to have a place in the conversation by imposing these these very honorous burdens on models that are not dangerous today so the if then the trigger commitment is basically a way to deal with this says you claim clamp down hard when you can show that the model is dangerous and of course what has to come with that is you know enough of a buffer threshold that that you know you can
you can uh you know you're you're you're you're not at high risk of kind of missing the danger it's not a perfect framework we've had to change it every every uh you know we came out with a new one just a few weeks ago and probably probably going forward we might release new ones multiple times a year because it's it's hard to get these policies right like technically organizationally from a research perspective but that is the proposal if then commitments and triggers in order to minimize burdens and false alarms now but really react appropriately when
the dangers are here what do you think the timeline for asl3 is where several of the triggers are fired and what do you think the timeline is for asl4 yeah so that is hotly debated within the company um uh we are working actively to prepare asl3 uh security uh security measures as well as ASL three deployment measures um I'm not going to go into detail but we've made we've made a lot of progress on both and you know we're we're prepared to be I think ready quite soon uh I would I would not be surpris
I would not be surprised at all if we hit ASL 3 uh next year there was some concern that we we might even hit it uh uh this year that's still that's still possible that could still happen it's like very hard to say but like I would be very very surprised if it was like 2030 uh I think it's much sooner than that so there's a protocols for detecting it the if then and then there's protocols for how to respond to it yes how difficult is the second the ladder yeah I think for asl3 it's
primarily about security um and and about you know filters on the model relating to a very narrow set of areas when we deploy the model because at asl3 the model isn't autonomous yet um uh and and so you don't have to worry about you know kind of the model itself behaving in a bad way even when it's deployed internally so I think the asl3 measures are are I won't say straightforward they're they're they're they're rigorous but they're easier to reason about I think once we get to asl4 um we start to have worries about the
models being smart enough that they might sandbag tests they might not tell the truth about tests um we had some results came out about like sleeper agents and there was a more recent paper about you know can can the models uh uh mislead attempts to you know s sandbag their own abilities right show them you know uh uh present themselves as being less capable than they are and so I think with asl4 there's going to be an important component of using other things than just interacting with the models for example interpretability or hidden chains of
thought uh where you have to look inside the model and verify via some other mechanism that that is not you know is not as easily corrupted as what the model says uh that that you know that that that the model indeed has some property uh so we're still working on asl4 one of the properties of the RSP is that we we don't specify asl4 until we've hit ASL 3 be and and I think that's proven to be a wise decision because even with asl3 it again it's hard to know this stuff in detail and and
it it we want to take as much time as we can possibly take to get these things right so for asl3 the bad actor will be the humans humans yes and so there it's a little bit more uh for asl4 it's both I think it's both and so deception and that's where mechanistic interpretability comes into play and hopefully the techniques used for that are not made accessible to the model yeah I mean of course you can hook up the mechanistic contribut ability to the model itself um but then You' then then you then you've kind
of lost it as a reliable indicator of uh of uh of of of the model State there are a bunch of exotic ways you can think of that it might also not be reliable like if the you know model gets smart enough that it can like you know jump computers and like read the code where you're like looking at its internal State we've thought about some of those I think they're exotic enough there are ways to render them unlikely but yeah generally you want to you want to preserve mechanistic interpretability as a kind of verification
set or test set that's separate from the training process of the model see I think uh as these models become better and better conversation and become smarter social engineering becomes a threat too cuz they oh yeah that can start being very convincing to the engineers inside companies oh yeah yeah it's actually like you know we've we've seen lots of examples of demagoguery in our life from humans and and you know there's a concern that models could do that could do that as well one of the ways that cloud has been getting more and more powerful
is it's now able to do some agentic stuff um computer use uh there's also an analysis within the sandbox of claw. a itself but let's talk about computer use that's seems to me super exciting that you can just give Claude a task and it uh takes a bunch of actions figures it out and has access to the your computer through screenshots so can you explain how that works uh and where that's headed yeah it's actually relatively simple so Claude has has had for a long time since since Claude 3 back in March the ability to
analyze images and respond to them with text the the only new thing we added is those images can be screenshot shots of a computer and in response we train the model to give a location on the screen where you can click Andor buttons on the keyboard you can press in order to take action and it turns out that with actually not all that much additional training the models can get quite good at that task it's a good example of generalization um you know people sometimes say if you get to low earth orbit you're like halfway
to anywhere right because of how much it takes to escape the gravity well if you have a strong pre-trained model I feel like you're halfway to anywhere uh in ter in terms of in terms of the intelligence space uh uh uh and and and so actually it didn't it didn't take all that much to get to get Claude to do this and you can just set that in a loop give the model a screenshot tell it what to click on give it the next screenshot tell it what to click on and and that turns into
a full kind of almost almost 3D video interaction of the model and it's able to do all of these tasks right you know we we showed these demos where it's able to like fill out spreadsheets it's able to kind of like interact with a website it's able to you know um you know it's able to open all kinds of you know programs different operating systems Windows Linux Mac uh uh so uh you know I think all of that is very exciting I I will say while in theory there's nothing you could do there that you
couldn't have done through just giving the model the API to drive the computer screen uh this really lowers the barrier and you know there's there's there's a lot of folks who who who either you know kind of kind of ar ar you know aren't in a position to to interact with those apis or it takes them a long time to do it's just the screen is just a universal interface that's a lot easier to interact with and so I expect over time this is going to lower a bunch of barriers now honestly the current model
has there's there it leaves a lot still to be desired and we were we were honest about that in the blog right it makes mistakes it misclicks and we we you know we were careful to warn people hey this thing isn't you can't just leave this thing to you know run on your computer for minutes and minutes um you got to give this thing boundaries and guard rails and I think that's one of the reasons we released it first in an API form rather than kind of you know this this kind of just just hand
it just hand it to the consumer and give it control of their of their of their of their computer um but but you know I definitely feel that it's important to get these capabilities out there as models get more powerful we're going to have to Grapple with you know how do we use these capabilities safely how do we prevent them from being abused uh and and you know I think I think releasing releasing the model while while while the capabilities are are you know are are still are still limited is is is very helpful in
terms of in terms of doing that um you know I think since it's been released a number of customers I think uh repet was maybe was maybe one of the the the most uh uh quickest quickest quickest to quickest to deploy things um have have you know have made use of it in various ways people have hooked up demos for you know Windows desktops Macs uh uh you know Linux Linux machines uh so yeah it's been it's been it's been very exciting I think as with as with anything else you know it it it comes
with new exciting abilities and then then then you know then then with those new exciting abilities we have to think about how to how to you know make the model you know safe reliable do what humans want them to do I mean it's the same it's the same story for everything right same thing it's that same tension but but the possibility of use cases here is just the the range is incredible so uh how much to make it work really well in the future how much do you have to specially kind of uh go beyond
what's the pre-trained models doing do more posttraining rhf or supervised fine-tuning or synthetic data just for the agent stff yeah I think speaking at a high level It's Our intention to keep investing a lot in you know making making the model better uh like I think I think uh you know we look at look at some of the you know some of the benchmarks where previous models were like oh could do it 6% of the time and now our model do at 14 or 22% of the time and yeah we want to get up to
you know the human level reliability of 80 90% just like anywhere else right we're on the same curve that we were on with sbench where I think I would guess a year from now the models can do this very very reliably but you got to start somewhere so you think it's possible to get to the the human level 90% uh basically doing the same thing you're doing now or is it has to be special for computer use I I mean uh depends what you mean by by you know special and FAL and um but but
you know I generally think you know the same kinds of techniques that we've been using to train the current model I I expect that doubling down in those techniques in the same way that we have for code for code for models in general for other k for you know for image input um uh you know for voice uh I expect those same techniques will scale here as they have everywhere else but this is giving sort of the power of action to Claude And so you could do a lot of really powerful things but you could
do a lot of damage also yeah yeah no and we've been very aware of that look my my view actually is computer use isn't a fundamentally new capability like the cbrn or autonomy capabilities are um it's more like it kind of opens the aperture for the model to use and apply its existing abilities uh and and so the way we think about it going back to our RSP is nothing that this model is doing inherently increases you know the risk from an RSP RSP perspective but as the models get more powerful having this capability may
make it scarier once it you know once it has the cognitive capability to um you know to do something at the asl3 and asl4 level this this you know this may be the thing that kind of Unbound it from doing so so going forward certainly this modality of interaction is something we have tested for and that we will continue to test for in our going forward um I think it's probably better to have to learn and explore this capability before the model is super uh you know super capable yeah and there's uh a lot of
interesting attacks like prompt injection because now you've widened the aperture so you can prompt inject through stuff on screen so if this becomes more and more useful then there's more and more benefit to inject inject stuff into the model if it goes to certain web page it could be harmless stuff like advertisements or it could be like harmful stuff right yeah I mean we thought a lot about like spam capture you know Mass C there's all you know every every like if one secret I'll tell you if you've invented a new technology not necessarily the
biggest misuse but but the the first misuse you'll see scams just Petty scams like you just just just it's it's like it's like a thing as old people scamming each other it's it's this it's this thing as old as time um and and and it's just every time you got to deal with it it's almost like silly to say but it's it's true sort of and spam in general is a thing as it gets more and more intelligent it's uh there a lot of like like I said like there are a lot of petty criminals
in the world and and and you know it's like every new technology is like a new way for petty petty criminals to do something you know something stupid and malicious um is there any ideas about sandboxing it like how difficult is the sandboxing task yeah we sandbox during training so for example during training we didn't expose the model to the internet um I think that's probably a bad idea during training because uh you know the model can be changing its policy it can be changing what it's doing and it's having an effect in the real
world um uh you know in in terms of actually deploying the model right it kind of depends on the application like you know sometimes you want the model to do something in the real world but of course you can always put guard you can always put guard rails on the outside right you can say okay well you know this model is not going to move data from my you know this model is not going to move any files from my computer to or my web server to anywhere else now when you talk about sandboxing again
when we get to asl4 none of these precautions are going to make sense there right where when you when you talk about asl4 you're then the model is being kind of you know there's a a theoretical worry the model could be smart enough to break it to to kind of break out of any box and so there we need to think about mechanistic interpretability about you know if we're if we're going to have a Sandbox it would need to be a mathematically provable sound but you know that's that's a whole different world than what we're
dealing with with the models today yeah the science of building a box from which asl4 AI system cannot Escape I I think it's probably not the right approach I think the right approach instead of having something you know unaligned that that like you're trying to prevent it from escaping I think it's it's better to just design the model the right way or have a loop where you you know you look inside you look inside the model and you're able to verify property and that gives you an opportunity to like iterate and actually get it right
um I think I think containing uh containing bad models is is is much worse solution than having good models let me ask about regulation what's the role of regulation in keeping AI safe so for example he described California AI regulation Bill SB 1047 that was ultimately vetoed by the governor what are the pros and cons of this bill General yes we ended up making some suggestions to the bill and then some of those were opted and you know we felt I think I think quite positively uh uh quite positively about about the bill uh by
by the end of that um it did still have some downsides um uh and you know of course of course it got vetoed um I think at a high level I think some of the key ideas behind the bill um are you know I would say similar to ideas behind our rsps and I think it's very important that some jurisdiction whether it's California or the federal government Andor other countries and other states passes some regulation like this and I can talk through why I think that's so important so I feel good about our RSP it's
not perfect it needs to be iterated on a lot but it's been a good forcing function for getting the company to take these risks seriously to put them into product planning to really make them a central part of work at anthropic and to make sure that all the thousand people and it's almost a thousand people now at anthropic understand that this is one of the highest priorities of the company if not the highest priority uh but one there are some there are still some companies that don't have RSP like mechanisms like open aai Google uh
did adopt these mechanisms a couple months after uh after anthropic did uh but there are there are other companies out there that don't have these mechanisms at all uh and so if some companies adopt these mechanisms and others don't uh it's really going to create a situation where you know some of these dangers have the property that it doesn't matter if three out of five of the companies are being safe if the other two are are being are being unsafe it creates this negative externality and and I think the lack of uniformity is not fair
to those of us who have put a lot of effort into being very thoughtful about these procedures the second thing is I don't think you can trust these companies to adhere to these voluntary plans in their own right I like to think that anthropic will we do everything we can that we will our our our our RSP is checked by our long-term benefit trust uh so you know we do everything we can to to to adhere to our own RSP um but you know you hear lots of things about various companies saying oh they said
they would do they said they would give this much compute and they didn't they said they would do this thing and they didn't um you know I don't I don't think it makes sense to you know to to to you know litigate particular things that companies have done but I I think this this broad principle that like if there's nothing watching over them there's nothing watching over us as an industry there's no guarantee that we'll do the right thing and the stakes are very high uh and so I think it's I think it's important to
have a uniform standard that that that that that everyone follows and to make sure that simply that the industry does what a majority of the industry has already said is important and has already said that they definitely will do right some people uh you know I think there's there a class of people who are against regulation on principle I understand where that comes from if you go to Europe and you know you see something like gdpr you see some of the other stuff that that that that that they've done you know some of it's good
but but some of it is really unnecessarily burdensome and I think it's fair to say really has slowed really has slowed Innovation and so I understand where people are coming from on priors I understand why people come from start from that start from that position uh but but again I think AI is different if we go to the very serious risks of autonomy and misuse that that that I talked about you know just a just a few minutes ago I think that those are unusual and they weren't an unusually strong response uh and so I
I think it's very important again um we need something that everyone can get behind uh you know I think one of the issues with s1047 uh especially the original version of it was it it had a bunch of the structure of rsps but it also had a bunch of stuff that was either clunky or that that that just would have created a bunch of burdens a bunch of Hassle and might even have missed the Target in terms of addressing the risks um you don't really hear about it on Twitter you just hear about kind of
you know people are people are cheering for any regulation and then the folks who are against make up these often quite intellectually dishonest arguments about how you know it you know it'll make us move away from California bill bill doesn't apply if you're headquartered in California bill only applies if you do business in California um or that it would damage the open source ecosystem or that it would you know it would cause cause all of these things I I think those were mostly nonsense but there are better arguments against regulation there's one guy uh Dean
ball who's really you know I think a very scholarly scholarly IST who who looks at what happens when a regulation is put in place and ways that they can kind of get a life of their own or how they can be poorly designed and so our interest has always been we do think there should be regulation in this space but we want to be an actor who makes sure that that that that regulation is something that's surgical that's targeted at the serious risks and is something people can actually comply with because something I think The
Advocates of Regulation don't understand as well as they could is if we get something in place that is um that's poorly targeted that wastes a bunch of people's time what's going to happen is people are going to say see these safety risks there you know this is this is nonsense I just you know I just had to hire 10 lawyers to to you know to fill out all these forms I had to run all these tests for something that was clearly not dangerous and after 6 months of that there will be there will be a
ground sweep well and we'll we'll we'll we'll end up with a durable consensus against regulation and so the I I think the the worst enemy of those who want real accountability is badly designed regulation um we we need to actually get it right uh and and this is if there's one thing I could say to The Advocates it it would be that I want them to understand this Dynamic better and we need to be really careful and we need to talk to people who actually have who actually have experience seeing how regulations play out in
practice and and the people who have seen that understand to be very careful if this was some lesser issue I might be against regulation at all but what what I want the opponents to understand is is that the underlying issues are actually serious they're they're not they're not something that I or the other companies are just making up because of regulatory capture they're not sci-fi fantasies they're not they're not any of these things um you know every every time we have new model every few months we measure the behavior of these models and they're getting
better and better at these concerning tasks just as they are getting better and better at um you know good valuable economically useful tasks and so I I I I would just love it if some of the former you know I think sb147 was very polarizing I would love it if some of the most reasonable opponents and some of the most reasonable um uh proponents uh would sit down together and you know I think I think that you know the different the different AI companies um you know anthropic was the the only AI company that you
know felt positively in a very detailed way I think Elon tweeted uh tweeted briefly something positive but you know some of the some of the big ones like Google open AI meta Microsoft were were pretty St stly against so I would really like is if if you know some of the key stakeholders some of the you know thoughtful proponents and and some of the most thoughtful opponents would sit down and say how do we solve this problem in in a way that the proponents feel brings a real reduction in risk and that the opponents feel
that it is not it is not hampering the the industry or hampering Innovation any more necessary than it than than than it needs to and and I think for for whatever reason that things got too polarized and those two groups didn't get to sit down in the way that they should uh and and I feel I feel urgency I really think we need to do something in 2025 uh uh you know if we get to the end of 2025 and we've still done nothing about this then I'm going to be worried I'm not I'm not
worried yet because again the risks aren't here yet but but I I I think time is running short yeah and come up with something surgical like you said yeah yeah yeah exactly and and we need to get we need to get away from this this this intense pro- safety versus intense anti-regulatory rhetoric right it's turned into these these flame Wars on Twitter and nothing Good's going to come with that so there's a lot of curiosity about the different players in the game one of the uh ogs is open AI you have had several years of
experience at open AI what's your story and history there yeah so I was at open AI for uh for roughly five years uh for the last I think it was a couple years you know I I I I I I was uh vice president of research there um probably myself and Ilia suger were the ones who you know really kind of set the set the research Direction around 2016 or 2017 I first started to really believe in or at least confirm my belief in the scaling hypothesis when when Ilia famously said to me the thing
you need to understand about these models is they just want to learn the models just want to learn um and and and and again sometimes there are these One S there these one sentences these Zen cones that you hear them and you're like ah that that explains everything that explains like a thousand things that I've seen and then and then I I you know ever after I had this visualization in my head of like you optimize the models in the right way you point the models in the right way they just want to learn they
just want to solve the problem regardless of what the problem is so get out of their way basically get out of their way yeah don't impose your own ideas about how they should learn and you know this was the same thing as Rich Sutton put out in the bitter lesson or G put out in the scaling hypothesis you know I think generally the dynamic was you know I got I got this kind of inspiration from uh from from from Ilan from others folks like Alec Radford who did the the original uh uh gpt1 uh and
then uh ran really hard with it me me and my collaborators on gpt2 gpt3 RL from Human feedback which was an attempt to kind of deal with the early safety and durability things like debate and amplification heavy on interpretability so again the combination of safety plus scaling probably 2018 2019 2020 those those were those were kind of the years when myself and my collaborators probably um you know mo mo many many of whom became co-founders of anthropic kind of really had had had a vision and like and like drove the direction why'd you leave why'
you decid to leave yeah so look I'm gonna put things this way and I you know I think it I think it ties to the to to the race to the top right which is you know in my time at open AI what I come come to see as I'd come to appreciate the scaling hypothesis and as I'd come to appreciate kind of the importance of safety along with the scaling hypothesis the first one I think you know open AI was was getting was getting on board with um the second one in a way had
always been part of of open ai's messaging um but uh you know over over many years of of the time the time that I spent there I think I had a particular vision of how these how we should handle these things how we should be brought out in the world the kind of principles that the organization should have and look I mean there were like many many discussions about like you know should the or do should the company do this should the company do that like there's a bunch of misinformation out there people say like
we left because we didn't like the deal with Microsoft false although you know there was like a lot of discussion a lot of questions about exactly how we do the deal with Microsoft um we left because we didn't like commercialization that's not true we built gbd3 which was the model that was commercialized I was involved in commercialization it's it's more again about how do you do it like Civilization is going down this path to very powerful AI what's the way to do it that is cautious straightforward honest um that build trust in the organization and
in individuals how do we get from here to there and how do we have a real vision for how to get it right how can safety not just be something we say because it helps with recruiting um and you know I think I think at the end of the day um if you have a vision for that forget about anyone else's Vision I don't want to talk about anyone else's Vision if you have a vision for how to do it you should go off and you should do that Vision it is incredibly unproductive to try
and argue with someone else's Vision you might think they're not doing it the right way you might think they're they're they're dishonest who knows maybe you're right maybe you're not um uh but uh what what you should do is you should take some people you trust and you should go off together and you should make your vision happen and if your vision is compelling if you can make it appeal to people some you know some combination of ethically you know in the market uh you know if if you can if you can make a company
that's a place people want to join uh that you know engages in practices that people think are are reasonable while managing to maintain its position in the ecosystem at the same time if you do that people will copy it um and the fact that you were doing it especially the fact that you're doing it better than they are um causes them to change their behavior in a much more compelling way than if they're your boss and you're arguing with them I just I don't know how to be any more specific about it than that but
I think it's generally very unproductive to try and get someone else's Vision to look like your vision um it's much more productive to go off and do a clean experiment and say this is our vision this is how this is this is how we're going to do things your choice is you can you can ignore us you can reject what we're doing or you can you can start to become more like us and imitation is the sincerest form of flattery um and you know that that that plays out in the behavior of customers that PS
out in the behavior of the public that plays out in the behavior of where people choose to work uh and again again at the end it's it's not about one company winning or another company winning if if we or another company are engaging in some practice that you know people people find genuinely appealing and I want it to be in substance not just not just in appearance um and you know I think I think researchers are sophisticated and they look at substance uh and then other companies start copying that practice and they win because they
copied that practice that's great that's success that's like the race to the top it doesn't matter who wins in the end as long as everyone is copying everyone else's good practices right one way I think of it is like the thing we're all afraid of is a race the bottom right in the race to the bottom doesn't matter who wins because we all lose right like you know in the most extreme world we we make this autonomous AI that you know the robots enslave us or whatever right I mean that's half joking but you know
that that is the most extreme uh uh thing thing that could happen then then it doesn't matter which company was ahead um if instead you create a race to the top where people are competing to engage in good in good practices uh then you know at at the end of the day you know it doesn't matter who who ends up who ends up winning doesn't even matter who who started the race to the top the point isn't to be virtuous the point is to get the system into a better equilibrium than it was before and
and individual companies can play some role in doing this individual companies can can you know can help to start it can help to accelerate it and frankly I think individuals at other companies have have done this as well right the individuals that when we put out an RSP react by pushing harder to to to get something similar done get something similar done at at at other companies sometimes other companies do something that's like we're like oh it's a good practice we think we think that's good we should adopt it too the only difference is you
know I think I think we are um we try to be more forward leaning we try and adopt more of these practices first and adopt them more quickly when others when others invent them but I think this Dynamic is what we should be pointing at and that I think I think it abstracts away the question of you know which company's winning who trusts who I I think all these all these questions of drama are are profoundly uninteresting and and the the thing that matters is the ecosystem that we all operate in and how to make
that ecosystem better because that constrains all the players and so anthropic is this kind of clean experiment built on a foundation of like what concretely AI safety should look like we look I'm sure we've made plenty of mistakes along the way the perfect organization doesn't exist it has to deal with the the imperfection of a thousand employees it has to deal deal with the imperfection of our leaders including me it has to deal with the imperfection of the people we've put we've put to you know to oversee the imperfection of the of the leaders like
the like the board and the long-term benefit trust it's it's all it's all a set of imperfect people trying to aim imperfectly at some ideal that will never perfectly be achieved um that's what you sign up for that's what it will always be but uh uh imperfect doesn't mean you just give up there's better and there's worse and hopefully hopefully we can begin to build we can do well enough that we can begin to build some practices that the whole industry engages in and then you know my guess is that M multiple of these companies
will be successful anthropic will be successful these other companies like ones I've been at the past will also be successful and some will be more successful than others that's less important than again that we we align the incentives of the industry and that happens partly through the race to the top partly through things like RSP partly through again selected surgical regulation you said Talent density beats Talent Mass so can you explain that can you expand on it can you just talk about what it takes to build a great team of AI researchers and Engineers this
is one of these statements that's like more true every every every month every month I see this statement as more true than I did the month before so if I were to do a thought experiment let's say you have a team of 100 people that are super smart motivated and aligned with the mission and that's your company or you can have a team of a thousand people where 200 people are super smart super aligned with the mission and then uh like and then like 800 people are let's just say you pick 800 like random random
big Tech employees which would you rather have right the talent mass is greater in in the group of uh in the group of a thousand people right you have you have even even a larger number of incredibly talented incredibly aligned incredibly smart people um uh but but the the issue is just that if every time someone super talented looks around they see someone else super talented and super dedicated that sets the tone for everything right that sets the tone for everyone is super inspired to work at the same place everyone trusts everyone else if you
have a thousand or 10,000 uh people and and things have really regressed right you are not able to do selection and you're choosing random people what happens is then you need to put a lot of proc CES and a lot of guard rails in place um just because people don't fully trust each other you have to adjudicate political battles like there are so many things that slow down the org's ability to operate and so we're nearly a thousand people and you know we've we've we've tried to make it so that as large a fraction of
those thousand people as possible are like super talented super skilled it's one of the reasons we've we've slowed down hiring a lot in the last few months We Grew From 300 to 800 I believe I think in the first seven eight months of the year and now we've slowed down we're at like you know last three months we went from 800 to 900 950 something like that don't quote me on the exact numbers but I think there's an inflection point around a thousand and we want to be much more careful how how we how we
grow uh early on and and now as well you know we've hired a lot of physicists um you know theoretical physicists can learn things really fast um uh even even more recently as we've continued to hire that you know we've really had a high bar for on both the research side and the software engineering side have hired a lot of senior people including folks who used to be at other at other companies in this space and we we've just continued to be very selective it's very easy to go go from 100 to a th000 a
th000 to 10,000 without paying attention to making sure everyone has a unified purpose it's so powerful if your company consists of a lot of different feif that all want to do their own thing they're all optimizing for their own thing um uh it's very hard to get anything done but if everyone sees the broader purpose of the company if there's trust and there's dedication to doing the right thing that is a superpower that in itself I think can overcome almost every other disadvantage and you know it's to Steve Jobs a players a players want to
look around and see other a players is another way of of saying I don't know what that is about human nature but it is demotivating to see people who are not obsessively driving towards a singular Mission and it is on the flip side of that super motivating to see that it's interesting uh what's it take to be a great AI researcher or engineer from everything you've seen from working with so many amazing people yeah um I think the number one quality especially on the research side but really both is open-mindedness sounds easy to be open-minded
right you're just like oh I'm open to anything um but you know if I if I think about my own early history in the scaling hypothesis um I was seeing the same data others were seeing I don't think I was like a better programmer or better at coming up with research ideas than any of the hundreds of people that I worked with um in some ways in some ways I was worse um uh you know like i' I've never like you know precise programming of like you know finding the bug writing the GPU kernels like
I could point you to a 100 people here who are better who are better at that than I am um but but the the thing that that that I think I did have that was different was that I was just willing to look at something with new eyes right people said oh you know we don't have the right algorithms yet we haven't come up with the right the right way to do things and I was just like uh I don't know like you know this neural net has like 30 billion 30 million parameters like what
if we gave it 50 million instead like let's plot some graphs like that that basic scientific mindset of like oh man like I I I just I just like I you know I see some variable that I could change like what happens when it changes like let's let's try these different things and like create a graph for even this this was like the simplest thing in the world right change the number of you know this wasn't like PhD level experimental design this was like this was like simple and stupid like anyone could have done this
if you if you just told them that that that it was important it's also not hard to understand you didn't need to be brilliant to come up with this um but you put the two things together and you know some tiny number of people some singled digigit number of people have have driven forward the whole field by realizing this uh and and it's you know it's often like that if you look back at the Discover you know the discoveries in in in history they're they're often like that and so this this open-mindedness and this willingness
to see with new eyes that often comes from being newer to the field often experience is a disadvantage for this that is the most important thing it's very hard to look for and test for but I think I think it's the most important thing because when you when you find something some really new way of thinking thinking about things when you have the initiative to do that it's absolutely transformative and also be able to do kind of Rapid experimentation and in the face of that be open-minded and curious and looking at the data from just
these fresh eyes and see what is that actually saying that applies in uh mechanistic interpretability it's another example of this like some of the early work in mechanistic interpretability so simple it's it's just no one thought to care about this question before you said what it takes to be a great AI researcher can we rewind the clock back what what advice would you give to people interested in AI they're young looking forward how can I make an impact on the world I think my number one piece of advice is to just start playing with the
models um this was actually I I I worry a little this seems like obvious advice now I think three years ago it wasn't obvious and people started by oh let me read the latest reinforcement learning paper let me you know let me let me kind of um no I mean that was really the that was really the the and I mean you should do that as well but uh now you know with wider availability of models and apis people are doing this more but I think I think just experiential knowledge um these models are new
artifacts that no one really understands um and so getting experience playing with them I would also say again in line with the like do something new think in some new Direction like there are all these things that haven't been explored like for example mechanistic interpretability is still very new it's probably better to work on that than it is to work on new model architectures because it's you know it's more popular than it was before there are probably like a hundred people working on it but there aren't like 10,000 people working on it and it's it's
just this just this this fertile area for study like like you know it's there's there's so much like low hangen fruit you can just walk by and you know you can just walk by and you can pick things um and and the the the only reason for whatever reason people aren't people aren't interested in it enough I think there are some things around long long Horizon learning and long Horizon tasks where there's a lot to be done I think evaluations are still we're still very early in our ability to study evaluations particularly for dynamic systems
acting in the world I think there's some stuff around multi-agent um skate where the puck is going is my is my advice and you don't have to be brilliant to think of it like all the things that are going to be exciting in 5 years like in in people even mention them as like you know conventional wisdom but like it's it's just somehow there's this barrier that people don't people don't double down as much as they could or they're afraid to do something that's not the popular thing I don't know why it happens but like
getting over that barrier is the that's the my number one piece of advice let's talk if we could a bit about posttraining yeah so it uh seems that the modern posttraining recipe has uh a little bit of everything so supervised fine tuning rhf uh the the the Constitutional AI with RL a if best acronym it's again that naming thing uh and then synthetic data seems like a lot of synthetic data or at least trying to figure out ways to have high quality synthetic data so what's the uh if this is a secret sauce that makes
anthropic claw so uh incredible what how how much of the magic is in the pre-training how much much of is in the post training yeah um I mean so first of all we're not perfectly able to measure that ourselves um uh you know when you see some some great character ability sometimes it's hard to tell whether it came from pre-training or post-training uh we developed ways to try and distinguish between those two but they're not perfect you know the second thing I would say is you know it's when there is an advantage and I think
we've been pretty good at in general in general at RL Perhaps Perhaps the best although although I don't know because I don't see what goes on inside other companies uh usually it isn't oh my God we have this secret magic method that others don't have right usually it's like well you know we got better at the infrastructure so we could run it for longer or you know we were able to get higher quality data or we were able to filter our data better or we able to you know combine these methods and practice it's it's
usually some boring matter of matter of kind of uh practice and tradecraft um so you know when I think about how to do something special in terms of how we train these models both pre-training but even more so posttraining um you know I I I really think of it a little more again as like designing airplanes or cars like you know it's not just like oh man I have the BL blueprint like maybe that makes you make the next airplane but like there's some there's some cultural tradecraft of how we think about the design process
that I think is more important than than you know than than any particular Gizmo were able to invent okay well about let me ask you about specific techniques so first on rhf what do you think think just zooming out intuition almost philosophy why do you think rhf works so well if I go back to like the scaling hypothesis one of the ways to skate the scaling hypothesis is if you train for x and you throw enough compute at it um then you get X and and so rlf is good at doing what humans want the
model to do or at least um to State it more precisely doing what humans who look at the model for a brief period of time and consider different possible responses what prefer as the response uh which is not perfect from both a safety and capabilities perspective in that humans are are often not able to perfectly identify what the model wants and what humans want in the moment may not be what they want in the long term so there's there's a lot of subtlety there but the models are good at uh you know producing what the
humans in some shallow sense want uh and it actually turns out that you don't even have to throw that much compute at it because of another thing which is this this thing about a strong pre-trained model being halfway to anywhere uh uh uh so once you have the pre-trained model you have all the representations you need to to get the model uh to get the model where you where you want it to go so do you think rhf makes the model smarter or just appears smarter to the humans I don't think it makes the model
smarter I don't think it just makes the model appear smarter it's like rhf like Bridges the gap between the human and the model right I could have something really smart that like can't communicate at all right we all know people like this um people who are really smart but that you know can't understand what they're saying um uh so I think I think rhf just bridges that Gap um I I think it's not it's not the only kind of RL we do it's not the only kind of RL that will happen in the future I
think RL has the potential to make models smarter to make them reason better to make them operate better to make them develop new skills even and perhaps that could be done you know even in some cases with human feedback but the kind of rhf we we do today mostly doesn't do that yet although we're very quickly starting to be able to but it it appears to sort of increase if you look at the metric of helpfulness it increases that it also increases what was this this word in Leopold's essay un hobbling where basically the models
are hobbled and then you do various trainings to them to un hobble them so I I know I like that word because it's like a rare word but so so I think rhf un hobbles the models in some ways um and then there are other ways where M hasn't yet been un hobbled and and you know needs to needs to un hobble if you can say in terms of cost is pre-training the most expensive thing or is post-training creep up to that at the present moment it is still the case that uh pre-training is the
majority of the cost I don't know what to expect in the future but I could certainly anticipate a future where post-training is the majority of the cost in that future you anticipate would it be the humans or the AI That's the costly thing for the Post training I I I I I don't think you can scale up humans enough to get high quality any any kind of method that relies on humans and uses a large amount of compute it's going to have to rely on some scaled supervision method like uh uh like um it you
know debate or iterated amplification or something like that so on that super interesting um set of ideas around constitutional AI can describe what it is as first detailed in December 2022 paper and uh and be on that what is it yes so this was from two years ago the basic idea is so we describe what rhf is you have uh you have a model and uh it you know spits out two you know like you just sample from it twice it spits out two possible responses and you're like human which response you like better or
another variant of it is rate this response on a scale of 1 to seven so that's hard because you need to scale up human interaction and uh it's very implicit right I don't have a sense of what I what I want the model to do I just have a sense of like what this average of a thousand humans wants the model to do so two ideas one is could the AI system itself decide which uh which response is better right could you show the AI system these two responses and and ask which which which response
is better and then second well what Criterion should the AI use and so then there's this idea because you have a single document a constitution if you will that says these are the principles the model should be using to to respond and the AI system reads those um it reads those principles as well as reading the environment and the response and it says well how good did the AI model do um it's basically a form of self-play you you're kind of training the model against itself and so the AI gives the response and then you
feed that back into What's called the preference model which in turn feeds the model to make it better um so you have this triangle of like the AI the preference model and the Improvement of the AI itself and we should say that in the Constitution the set of principles are like human interpretable they're like yeah yeah it's something both the human and the AI system can read so it has this nice this nice kind of translatability or symmetry um you know in in practice we both use a model Constitution and we use rhf and we
use some of these other methods so it's it's turned into one tool in a in a toolkit that both reduces the need for rhf and increases the value we get from um from from using each data point of R lhf um it also interacts in interesting ways with kind of future reasoning type RL methods so um it's it's one tool in the toolkit but but I I think it is a very important tool well it's a compelling one to us humans you know thinking about the founding fathers and the founding of the United States the
natural question is who and how do you think it gets to define the constitution the the set of principles in the Constitution yeah so I'll give like a practical um answer and a more abstract answer I think the Practical answer is like look in practice models get used by all kinds of different like customers right and and so uh you can have this idea where you know the model can can have specialized rules or principles you know we fine-tune versions of models implicitly we've talked about doing it explicitly having having special principles that people can
can build into the models um uh so from a practical perspective the answer can be very different from different people uh you know customers service agent uh you know behaves very differently from a lawyer and obeys different principles um but I think at the base of it there are specific principles that the models uh you know have to obey I think a lot of them are things that people would agree with everyone agrees that you know we don't you know we don't want models to present these cbrn risks um I think we can go a
little further and agree with some basic principles of democracy and the rule of law beyond that it gets you know very uncertain and and there our goal is generally for the models to be more neutral to not espouse a particular point of view and you know more just be kind of like wise uh agents or advisers that will help you think things through and will you know present present possible considerations but you know don't express you know stronger specific opinions open AI released a model spec where it kind of clearly concretely defines some of the
goals of the model and specific examples like AB how the model should behave do you find that interesting by the way I should mention the I believe the brilliant John Schulman was a part of that he's now an anthropic uh do you think this is a useful Direction might anthropic release a model spec as well yeah so I think that's a pretty useful direction again it has a lot in common with uh constitutional AI so again another example of like a race to the top right we have something that's like we think you know a
better and more responsible way of doing things um it's also a competitive advantage um then uh others kind of you know discover that it has advantages and then start to do that thing uh we then no longer have the competitive Advantage but it's good from the perspective that now everyone has adopted a positive practice that others were not adopting and so our response to that as well looks like we need a new competitive advantage in order to keep driving this race upwards um so that's that's how I generally feel about that I also think every
implementation of these things is different so you know there were some things in the model spec that were not in constitutional Ai and so you know we you know we can always we can always adopt those things or you know at least learn from them um so again I think this is an example of like the positive Dynamic that uh that that that I that that I think we should all want the field to have let's talk about the incredible ESS Machines of love and grace I recommend everybody read it it's a long one it
is rather long yeah it's really refreshing to read concrete ideas about what a positive future looks like and you took sort of a bold stance because like it's very possible you might be wrong on the dates or specific applications yeah I'm fully expecting to you know to definitely be wrong about all the details I might be be just spectacularly wrong about the whole thing and people will you know will laugh at me for years um uh that's that's how that's that's just how the future works so you provided a bunch of concrete positive impacts of
AI and how you know exactly a super intelligent AI might accelerate the rate of breakthroughs in for example biology and chemistry that would then lead to things like we cure most cancers prevent all infectious disease double the human lifespan and so on so let's talk about this essay first can you give a high level vision of this essay and um what key takeaways that people should have yeah I have spent a lot of time and anthropic has spent a lot of effort on like you know how do we address the risks of AI right how
do we think about those risks like we're trying to do a race to the top you know that requires us to build all these capabilities and the abilities are cool but you know you know we're we're we're like a big part of what we're trying to do is like is like address the risks and the justification for that is like well you know all these positive things you know the the market is this very healthy organism right it's going to produce all the positive things the risks I don't know we might mitigate them we might
not and so we can have more impact by trying to mitigate the risks but I noticed that one flaw in that way of thinking and it's if not a change in how seriously I take the risks it's it's maybe a change in how I talk about them um is that you know no matter how kind of logical or rational that line of reasoning that I just gave might be um if if you kind of only talk about risks your brain only thinks about risks and and so I think it's actually very important to understand what
if things do go well and the whole reason we're trying to prevent these risks is not because we're afraid of Technology not because we want to slow it down it's it's it's because if we can get to the other side of these risks right if we can run the gauntlet successfully um to you know to to put it in Stark terms then then on the other side of the gauntlet are all these great things and these things are worth fighting for and these things can really inspire people and I think I imagine because look you
have all these investors all these VCS all these AI companies talking about all the positive benefits of AI but as you point out it's it's it's weird there's actually a dir of really getting specific about it there's a lot of like random people on Twitter like posting these kind of like gleaming cities and this this just kind of like Vibe of like grind accelerate harder like kick out the D you know it's it's just this very this very like aggressive ideological but then you're like well what are you what what what what what are you
actually excited about um and so and so I figured that you know I think it would be interesting and valuable for someone who's actually coming from the risk side to to try and and to try and really make a try at at explaining explaining explaining what the benefits are um both because I think it's something we can all get behind and I want people to understand I want them to really understand that this isn't this isn't doomers versus accelerationists um this this is that if you have a true understanding of of where things are going
with with AI and maybe that's the more important axis AI is moving fast versus AI is not moving fast then you really appreciate the benefits and you you you you really you want Humanity our civilization to seize those benefits but you also get very serious about anything that could derail them so I think the starting point is to talk about what this powerful AI which is the term you like to use uh most of the world uses AGI but you don't like the term because it's uh basically has too much baggage has become meaningless it's
like we're stuck with the terms like maybe we're stuck with the terms and my efforts to change them are futile it's ADM I'll tell you what else I don't this is like a pointless semantic point but I I I I keep talking about it public so I'm just I'm just going to do it once more um uh I I think it's it's a little like like let's say it was like 1995 and Mor's law is making the computers faster and like for some reason there there there there had been this like verbal tick that like
everyone was like well someday we're going to have like super super computers and like supercomputers are going to be able to do all these things that like you know once we have supercomputers we'll be able to like sequence the Geno and we'll be able to do other things and so and so like one it's true the computers are getting faster and as they get faster they're going to be able to do all these great things but there's like there's no discret point at which you had a supercomputer and previous computers were not to like supercomputer
is a term we use but like it's a vague term to just describe like computers that are faster than what we have today um there's no point at which you pass a threshold and you're like oh my God we're doing a totally new type of computation and new and and so I feel that way about AGI like there's just a smooth exponential and like if if by AGI you mean like like AI is getting better and better and like gradually it's going to do more and more of what humans do until it's going to be
smarter than humans and then it's going to get smarter even from there then then yes I believe in AGI if but if if if AGI is some discreet or separate thing which is the way people often talk about it then it's it's kind of a meaningless buzz word yeah I me to me it's just sort of a IC form of a powerful AI exactly how you define it I mean you define it very nicely so on the intelligence axis it's just on pure intelligence it's smarter than a Nobel Prize winner as you describe across most
relevant disciplines so okay that's just intelligence so it's uh both in creativity and be able to generate new ideas all that kind of stuff in every discipline Nobel Prize winner okay in their prime it can use every modality it so uh that's kind of self-explanatory but just operate across all the modalities of the world uh it can go off for many hours days and weeks to do tasks and do its own sort of detailed planning and only ask you help when it's needed uh it can use this is actually kind of interesting I think in
the essay you said I mean again it's a bet that it's not going to be embodied but it can control embodied tools so it can control tools robots Laboratory equipment the resource used to train it can then be repurposed to run millions of copies of it and each of those copies would be independent that can do their own independent work so you can do the cloning of the intelligence system yeah yeah I mean you you might imagine from outside the field that like there's only one of these right that like you made it you've only
made one but the truth is that like the scale up is very quick like we we do this today we make a model and then we deploy thousands maybe tens of thousands of instances of it I think by the time you know certainly within 2 to 3 years whether we have these superp powerful AIS or not clusters are going to get to the size where where you'll be able to deploy millions of these and they'll be you know faster than humans and so if your picture is oh we'll have one and it'll take a while
to make them my point there was no actually you have millions of them right away and in general they can learn and act uh 10 to 100 times faster than humans so that's a really nice definition of powerful AI okay so that but you also write that clearly such an entity would be cap capable of solving very difficult problems very fast but it is not trivial to figure out how fast two extreme positions both seem false to me so the singularity is on the one extreme and the opposite On The Other Extreme can you describe
each of the extremes yeah why so yeah let's let's describe the extreme so like one one extreme would be well look um you know uh if we look at kind of evolutionary history like there was this big acceleration where you know for hundreds of thousands of years we just had like you know single cell organisms and then we had mammals and then we had apes and then that quickly turned to humans humans quickly built industrial civilization and so this is going to keep speeding up and there's no cealing at the human level once models get
much much smarter than humans they'll get really good at building the next models and you know if you write down like a simple differential equation like this is an exponential and so what's what's going to happen is that uh models will build faster models models will build faster models and those models will build you know Nano that can like take over the world and produce much more energy than you could produce otherwise and and so if you just kind of like solve this abstract differential equation then like 5 days after we you know we build
the first AI That's more powerful than humans then then uh you know like the world will be filled with these AIS and every possible technology that could be invented like will be invented um I'm caricaturing this a little bit um uh but I you know I think that's one extreme and the reason that I think that's not the case is is that one I think they just neglect like the laws of physics like it's only possible to do things so fast in the physical world like some of those Loops go through you know producing faster
Hardware um uh takes a long time to produce faster Hardware things take a long time there's this issue of complexity like I think no matter how smart you are like you know people talk about oh we can make models the biological systems it'll do everything the biological systems look I think computational modeling can do a lot I did a lot of computational modeling when I worked in biology but like just there are a lot of things that you can't predict how they're you know they're they're complex enough that like just iterating just running the experiment
is going to beat any modeling no matter how smart the system doing the modeling is oh even if it's not interacting with the physical world just the modeling is going to be hard yeah I think well the modeling is going to be hard and getting the model to to to to match the physical world is going to be all right so he does have to intera the physical world to verify but it's just you know you just look at even the simplest problems like I you know I think I talk about like you know the
three body problem or simple chaotic prediction like you know or or like predicting the economy it's really hard to predict the economy two years out like maybe the case is like you know normal you know humans can predict what's going to happen in the economy in the next quarter although they can't really do that maybe a maybe a AI system that's you know a zillion times smarter can only predict it out a year or something instead of instead of a you know you have the these kind of exponential increase in computer intelligence for linear increase
in in in ability to predict same with again like you know biological molecules molecules interacting you don't know what's going to happen when you perturb a when you perturb a complex system you can find simple Parts in it if you're smarter you're better at finding these simple parts and then I think human institutions human institutions are just are are really difficult like it's you know it's it's been hard to get people I won't give specific examples but it's been hard to get people to adopt even the technologies that we've developed even ones where the case
for their efficacy is very very strong um you know people have concerns they think things are conspiracy theories like it's it's just been it's been very difficult it's also been very difficult to get you know very simple things through the regulatory system right I think you know and you know I I don't want to just spage anyone who you know you know work Works in regulator regulatory systems of any technology there are hard trade-offs they have to deal with they have to save lives but but the system as a whole I think makes some obvious
tradeoffs that are very far from maximizing human welfare and so if we bring AI systems into this you know into these human systems often the level of intelligence may just not be the limiting factor right it it it just may be that it takes a long time to do something now if the AI system uh circumvented all governments if it just said I'm dictator of the world and I'm going to do whatever some of these things it could do again the things having to do with complexity I I I still think a lot of things
would take a while I don't think it helps that the AI systems can produce a lot of energy or go to the moon like some people in comments responded to the essay saying the AI system can produce a lot of energy and smarter AI systems that's missing the point that kind of cycle doesn't solve the key problems that I'm talking about here um so I think I think a bunch of people missed the point there but even if it were completely on aligned and you know could get around all these human obstacles it would have
trouble but again if you want this to be an AI system that doesn't take over the world that doesn't destroy Humanity then then basically you know it's it's it's going to need to follow basic human laws right where you know if if we want to have an actually good world like we're going to have to have an AI system that that interacts with humans not one that kind of creates its own legal system or disregards all the laws or all of that so as inefficient as these processes are you know we're going to have to
deal with them because there there needs to be some popular and Democratic legitimacy in how these systems are rolled out we can't have a small group of people who are developing these systems say this is what's best for everyone right I think it's wrong and I think in practice is not going to work anyway so you put all those things together and you know we're not we're not g to we're not going to you know change the world and upload everyone in five minutes uh I I I just I don't think it I A A
I don't think it's going to happen and be to some in you know to the extent that it could happen it's it's not the way to lead to a good world so that's on one side on the other side there's another set of perspectives which I have actually in some ways more sympathy for which is look we've seen big productivity increases before right you know economists are familiar with studying the productivity increases that came from the computer Revolution and internet Revolution and generally those productivity increases were underwhelming they were less than you than you might
imagine um there was a quote from Robert solo you see the computer Revolution everywhere except the productivity statistics so why is this the case people point to the structure of firms the structure of Enterprises how um uh you know how slow it's been to roll out our existing technology to very poor parts of the world which I talk about in the essay right how do we get these Technologies to the poorest parts of the world that are behind on cell phone technology computers medicine let alone you know new fangled AI that hasn't been invented yet
um so you could have a perspective that's like well this is amazing technically but it's all a nothing burer um uh you know I think um Tyler Cowan who who wrote something response to my essay has that perspective I think he thinks the radical change will happen eventually but he thinks it'll take 50 or 100 years and and you could have even more static perspectives on the whole thing I think there's some truth to it I think the time scale is just is just too long um and and I can see it I can actually
see both sides with today's AI so uh you know a lot of our customers are large Enterprises who are used to doing things a certain way um I've also seen it in talking to governments right those are those are prototypical you know institutions entities that are slow to change uh but the dynamic I see over and over again is yes it takes a long time to move the ship yes there's a lot of resistance and lack of understanding but the thing that makes me feel that progress will in the end happen moderately fast not incredibly
fast but moderately fast is that you talk to what I find is I find over and over again again in large companies even in governments um which have been actually surprisingly forward leaning uh you find two things that move things forward one you find a small fraction of people within a company within a government who really see the big picture who see the whole scaling hypothesis who understand where AI is going or at least understand where it's going within their industry and there are a few people like that within the current within the current US
government who really see the whole picture and and those people see that this is the most important thing in the world until they agitate for it and the thing they they alone are not enough to succeed because they are a small set of people within a large organization but as the technology starts to roll out as it succeeds in some places in the folks who are most willing to adopt it the Spectre of competition gives them a wind at their backs because they can point within their large organization they can say look these other guys
are doing this right you know One bank can say look this new fangled hedge fund is doing this thing they're going to eat our lunch in the US we can say we're afraid China's going to get there before before we are uh and that combination the Spectre of competition plus a few Visionaries Within These you know within these the organizations that in many ways are are sclerotic you put those two things together and it actually makes something happen I mean it's interesting it's a balanced fight between the two because inertia is very powerful but but
but eventually over enough time the Innovative approach breaks through um and I've seen that happen I've seen the Arc of that over and over again and it's like the the barriers are there the the barriers to progress the complexity not knowing how to use the model or how to deploy them are there and and for a bit it seems like they're going to last forever like change doesn't happen but then eventually change happens and always comes from a few people I felt the same way when I was an advocate of the scaling hypothesis within the
AI field itself and others didn't get it it felt like no one would ever get it it felt like then it felt like we had a secret almost no one ever had and then a couple years later everyone has the secret and so I think that's how it's going to go with deployment to AI in the world it's going to the the barriers are going to fall apart gradually and then all at once and so I think this is going to be more and this is just an instinct I could I could easily see how
I'm wrong I think it's going to be more like 10 five or 10 years as I say in the essay then it's going to be 50 or 100 years I also think it's going to be five or 10 years more than it's going to be you know five or 10 hours uh uh because I've just I've just seen how human systems work and I think a lot of these people who write down the differential equations who say AI is going to make more powerful AI who can't understand how it could possibly be the case that
these things won't won't change so fast I think they don't understand these things so what to use the timeline to where we achieve AGI AKA powerful AI AKA super useful AI I'm start calling it that it's a debate it's a debate about naming um you know unpure intelligence you can smarter than a Nobel Prize winner in every relevant discipline and all the things we've said modality you can go and do stuff on its own for days weeks and do biology experiments uh on its own in one you know what let's just stick to biology because
yeah I you you sold me on the whole biology and health section That's so exciting from um from a just I was getting giddy from a scientific perspective it made me want to be a biologist it's almost it's it's so no no that this was the feeling I had when I was writing it that it's it's like this would be such a beautiful future if we can if we can just if we can just make it happen right if we can just get the get the landmines out of the way and and and and make
it happen there's there's so much there's so much Beauty and and and and and elegance and moral force behind it if if we can if we can just and it's something we should all be able to agree on right like as much as we fight about about all these political questions is is this something that could actually bring us together um but you were asking when when will we get this when when do you think what's just put numbers on so you know this this is of course the thing I've been grappling with for many
years and I'm not I'm not at all confident every time if I say 2026 or 2027 there will be like a zillion like people on Twitter who will be like he icoo said 2026 2020 and it'll be repeated for like the next two years that like this is definitely when I think it's going to happen um so who whoever's exerting these clips will will we we'll we'll crop out the thing I just said and and only say the thing I'm about to say um but I'll just say it anyway um have so so uh if
you extrapolate the curves that we've had so far right if if you say well I don't know we're starting to get to like PhD level and and last year we were at um uh undergraduate level in the year before we were at like the level of a high school student again you can you can quibble with at what tasks and for what we're still missing modalities but those are being added like computer use was added like image in was added like image generation has been added if you just kind of like and this is totally
unscientific but if you just kind of like eyeball the rate at which these capabilities are increasing it does make you think that we'll get there by 2026 or 2027 again lots of things could derail it we could run out of data you know we might not be able to scale clusters as much as we want like you know maybe Taiwan gets blown up or something and you know then we can't produce as many gpus as we want so there there are all kinds of things that could could derail the whole process so I don't fully
believe the straight line extrapolation but if you believe the straight line extrapolation you'll you we'll get there in 2026 or 2027 I think the most likely is that there's some mild delay relative to that um I don't know what that delay is but I think it could happen on schedule I think there could be a mild delay I think there are still worlds where it doesn't happen in in a hundred years those world the number of those worlds is rapidly decreasing we are rapidly running out of truly convincing Brockers truly compelling reasons why this will
not happen in the next few years there were a lot more in 2020 um although my my guest my hunch at that time was that we will make it through all those blockers so sitting as someone who has seen most of the blockers cleared out of the way I kind of suspect my hunch my suspicion is that the rest of them will not block us uh but you know look look at look at the end of the day like I don't want to represent this as a scientific prediction people call them scaling laws that's a
misnomer like Mo's law is is is a misnomer Moors laws scaling laws they're not laws of the universe they're empirical regularities I am going to bet in favor of them continuing but I'm not certain of that so you extensively describe sort of the compressed 21st century how AGI will help uh set forth a chain of breakthroughs in biology and medicine that help us in all these kinds of ways that I mentioned so how do you think what are the early steps it might do and by the way I asked Claude good questions to ask you
and Claude told me uh to ask what do you think is a typical day for a biologist working on AGI look like under in this future yeah yeah Claud is curious let me well let me start with your first questions and then I'll then I'll answer that Claude Claude wants to know what's in his future right exactly who's it who am I going to be working with exactly um so I think one of the things I went hard on in when I went hard on in the essay is let me go back to this idea
of because it's it's really had had an you know had an impact on me this idea that within large organizations and systems there end up being a few people or a few new ideas who kind of cause things to go in a different direction they would have before who who kind of a disproportionately affect the the trajectory there's a bunch of kind of the same thing going on right if you think about the health world there's like you know trillions of dollars to pay out Medicare and you know other health insurance and then the NIH
is is 100 billion and then if I think of like the the few things that have really revolutionized anything it could be encapsulated in a small small fraction of that and so when I think of like where will AI have an impact I'm like can AI turn that small fraction into a much larger fraction and raise its quality and within biology my experience within biology is that the biggest problem of biology is that you can't see what's going on you you have very little ability to see what's going on and even less ability to change
it right what you have is this like like from this you have to infer that there's a bunch of cells that within each cell is you know uh uh three billion base pairs of DNA built according to a genetic code uh uh and you know there are all these processes that are just going on without any ability of us as you know un augmented humans to affect it these cells are dividing most of the time that's healthy but sometimes that process goes wrong and that's cancer um the cells are aging your skin may change color
develops wrinkles as you as you age and all of this is determined by these processes all these proteins being produced transported to various parts of the cells binding to each other and and in our initial State about biology we didn't even know that these cells existed we had to invent microscopes to observe the cells we had to uh we had to invent more powerful microscopes to see you know below the level of the cell to the level of molecules we had to invent x-ray crystallography to see the DNA we had to invent Gene sequencing to
read the DNA now you know we had to invent protein folding technology to you know to predict how it would fold and how they bind and how these things bind to each other uh you know we had to we had to invent various techniques for now we can edit the G the DNA as of you know with chrisopher as of the last uh uh 12 years so the the whole history of biology a whole big part of the history is is basically our our our our ability to read and understand what's going on and our
ability to reach in and selectively change things um and and my view is that there's so much more we can still do there right you can do crisper but you can do it for your whole body um let's say I want to do it for one particular type of cell and I want the rate of targeting the wrong cell to be very low that's still a challenge that's still things people are working on that's what we might need for gene therapy for certain diseases and so the reason I'm saying all of this and it goes
beyond you know beyond this to you know to Gene sequencing to new types of nanomaterials for observing what's going on inside cells for you know antibody drug conjugates the the reason I'm saying all this is that this could be a leverage point for the AI systems right that the number of such inventions it's it's in the it's in the mid double digits or something you know mid double digits maybe low triple digits over the history of biology let's say I have a million of these AIS like you know can they discover thousand you know working
together can they discover thousands of these very quickly and and does that provide a huge lever instead of trying to Leverage The you know two trillion a year we spend on you know Medicare or whatever can we Leverage The 1 billion a year that's that's you know that's spent to discover but with much higher quality um and so what what is it like you know being a being a scientist that works with uh with with an AI system the way I think about it actually is well so I think in the early stages uh the
AIS are going to be like grad students you're going to give them a project you're going to say you know I'm the experienced biologist I've set up the lab the biology Professor or even the grad student students themselves will say here's here's what uh here's what you can do with an AI you know like a AI system I'd like to study this and you know the AI system it has all the tools it can like look up all the literature to decide what to do it can look at all the equipment it can go to
a website and say hey I'm going to go to you know thermofisher or you know whatever the lab equipment company is dominant lab equipment company is today and my my time was thermofisher um uh you know I'm I'm going to order this new equipment to to to do this I'm going to run my experiments I'm going to you know write up a report about my experiments I'm going to you know inspect the images for contamination I'm going to decide what the next experiment is I'm going to like write some code and run a statistical analysis
all the things a grad student would do there will be a computer with an AI that like the professor talks to every once in a while and it says this is what you're going to do today the AI system comes to it with questions um when it's necessary to run the lab equipment it may be limited in some ways may have to hire a human lab assistant to you know to do the experiment and explain how to do it or it could you know it could use advances in lab automation that are gradually being developed
over have been developed over the last uh uh decade or so and will will continue to be will continue to be developed uh and so it'll look like there's a human professor and a thousand AI grad students and you know if you if you go to one of these Nobel prizewinning biologist or so you'll say okay well you know you had like 50 grad students well now you have a thousand and they're they're they're smarter than you are by the way um uh then I think at some point it'll flip around where the you know
the AI systems will you know will will be the pis will be the leaders and and and you know they'll be they'll be ordering humans or other AI systems around so I think that's how it'll work on the research s and they would be the inventors of a crisper type technology they would be the inventors of of a a crisper type technology um and then I think you know as I say in the essay we'll want to turn turn probably turning loose is the wrong the wrong term but we want to want to harness the
AI systems uh to improve the clinical trial system as well there's some amount of this that's regulatory that's a matter of societal decisions and that'll be harder but can we get better at predicting the results of clinical trials can we get better at statistical design so that what you know clinical trials that used to require you know 5,000 people and therefore you know needed $100 million and a year to enroll them now they need 500 people in two months to enroll them um that's where we should start uh and and you know can we increase
the success rate of clinical trials by doing things in animal trials that we used to do in clinical trials and doing things in simulations that we used to do in animal trials again we won't be able to simulate it all AI is not God um uh but but you know can we can we shift the curve substantially and radically so I I don't know that would be my picture doing inro and doing it I mean you're still slowed down it still takes time but you can do it much much faster yeah yeah yeah can we
just one step at a time and and can that can that add up to a lot of steps even though even though we still need clinical trials even though we still need laws even though the FDA and other organizations will still not be perfect can we just move everything in a positive direction and when you add up all those Positive Directions do you get everything that was going to happen from here to 2100 instead happens from 2027 to 2032 or something another way that I think the world might be changing with AI even today but
moving towards this future of the the powerful super useful AI is uh programming so how do you see the nature of programming because it's so intimate to the actual Act of building AI how do you see that changing for us humans I think that's going to be one of the areas that changes fastest um for two reasons one programming is a skill that's very close to the actual building of the AI um so the farther skill is from the people who are building the AI the longer it's going to take to get disrupted by the
AI right like I truly believe that like AI will disrupt agriculture maybe it already has in some ways but that's just very distant from the folks who are building Ai and so I think it's going to take longer but programming is the bread and butter of you know a large fraction of of the employees who work at anthropic and at the other companies and so it's going to happen fast the other reason it's going to happen fast is with programming you close the loop both when you're training model when you're applying the model the idea
that the model can write the code means that the model can then run the code and and and then see the results and and interpret it back and so it really has an ability unlike Hardware unlike biology which we just discussed the model has an ability to close the loop um and and so I think those two things are going to lead to the model getting good at programming very fast as I saw on you know typical real world programming tasks models have gone from 3% in January of this year to 50% in October of
this year so you know we're on that S curve right where it's going to start slowing down soon because you can only get to 100% but uh I you know I I would guess that in another 10 months well we'll probably get pretty close we'll be at at least 90% so again I would guess you know I don't know how long it'll take but I would guess again 202 2026 2027 Twitter people who crop out my who who who crop out these these numbers and get rid of the caveats like like I don't know I
don't like you go away uh I would guess that the kind of task that the vast majority of coders do AI can probably if we make the task very narrow like just write code um AI systems will uh be able to do that now that said I think comparative advantage is powerful we'll find that when AIS can do 80% of a coder's job including most of it that's literally like right code with a given spec will find that the remaining parts of the job become more leveraged for humans right humans will they'll be more about
like high level system design or you know looking at the app and like is it architected well and the the design and ux aspects and eventually AI will be able to do those as well right that's my vision of the you know powerful AI system but I think for much longer than we might expect we will see that uh small parts of the job that humans still do will expand to fill their entire job in order for the overall productivity to go up um that's something we've seen you know it used to be that you
know writing you know writing and Editing letters was very difficult and like writing the print was difficult well as soon as you had word processors and then and then uh and then computers and it became easy to produce work and easy to share it then then that became instant and all the focus was on was on the ideas so this this logic of comparative advantage that expands tiny parts of the tasks to large parts of the tasks and creates new tasks in order to expand productivity I think that's going to be the case again someday
AI will be better at everything and that logic uh won't apply and then then we all have you know Humanity will have to think about how to collectively deal with that and we're thinking about that every day um and you know that's another one of the grand problems to deal with aside from misuse and autonomy and you know we should take it very seriously but I think I think in the in the near term and maybe even in the medium term like medium term like 2 three four years you know I expect that humans will
will continue to have a huge role and the nature of programming will change but programming as a as a role programming as a job will not change it'll just be less writing things line by line and it'll be more macroscopic and I wonder what the future of Ides looks like so the tooling of interacting with AI systems this is true for programming and also probably true for in other contexts like computer use but maybe domain specific like we mentioned biology it probably needs its own tooling about how to be effective and then programming needs its
own tooling is anthropic going to play in that space of also tooling potentially I'm absolutely convinced that uh powerful IDs uh that that there's so much low hanging fruit to be grabbed there um that you know right now it's just like you talk to the model and it talks back but but look I mean IDs are great at kind of lots of status analysis of of you know so much as possible with kind of static analysis like many bugs you can find without even writing the code then uh you know IDs are good for running
particular things organizing your code um measuring coverage of unit test like there's so much that's been possible with a normal with a normal Ides now you add something like well the model now you know the model can now like write code and run code like I am absolutely convinced that over the next year or two even if the quality of the models didn't improve that there would be enormous opportunity to enhance people's productivity by catching a bunch of mistakes doing a bunch of grunt work for people and that we haven't even scratched the surface um
and thropic itself I mean you can't say you know no you know it's hard to say what will happen in the future currently we're not trying to make such IDs ourself rather we powering the companies like cursor or like cognition or some of the other you know uh Expo in the security space um uh you know others that I can mention as well that are building such things themselves on top of our API and our view has been let a thousand flowers bloom we don't internally have the the re you know the resources to try
all these different things let's let our customers try it um uh and you know we'll see who succeed and maybe different customers will succeed in different ways uh so I both think this is super promising and you know it's not it's not it's not something you know anthropic isn't isn't eager to to at least right now compete with all our companies in this space and maybe never yeah it's been interesting to watch curser try to integrate claw successfully because there's it's actually me fascinating how many places it can help the programming experience it's not as
trivial it is it is really astounding I feel like you know as a CEO I don't get to program that much and I feel like if six months from now I go back it'll be completely unrecognizable to me exactly um so in this world with super powerful AI uh that's increasingly automated what's the source of meaning for us humans yeah you know work is a source of deep meaning for many of us so what do we uh where do we find the meaning this is something that I've I've written about a little bit in the
essay although I I actually I give it a bit short shrift not for any um not for any principled reason but this essay if you believe it was originally going to be two or three pages I was going to talk about it at all hands and the reason I I I realized it was an under un important underexplored topic is that I just kept writing things and I was just like oh man I can't do this Justice and so the thing balloon to like 40 or 50 pages and then when I got to the work
in meaning section I'm like oh man this isn't going to be 100 Pages like I'm GNA have to write a whole other essay about that but meaning is actually interesting because you think about like the life that someone lives or something or like you know like you know let's say you were to put me in like a I don't know like a simulated environment or something where like um you know like I have a job and I'm trying to accomplish things I don't know I like do that for 60 years and then then you're like
oh oh like oops this was this was actually all a game right does that really kind of Rob you of the meaning of the whole thing you know like I still made important choices including moral choices I still sacrificed I still had to kind of gain all these skills or or or just like a similar exercise you know think back to like you know one of the historical figures who you know discovered electromagnetism or relativity or something if you told them well actually 20,000 years ago some some alien on you know some alien on this
planet discovered this before before you did um does that does that Rob the meaning of the discovery it doesn't really seem like it to me right it seems like the process is what is what matters and how it shows who you are as a person along the way and you know how you relate to other people and like the decisions that you make along the way those are those are consequential um you know I I I could imagine if we handle things badly in an AI world we could set things up where people don't have
any long-term source of meaning or any but but that's that's more a choice a set of choices we make that's more a set of the architecture of a society with these powerful models if we if we design it badly and for shallow things then then that might happen I would also say that you know most people's lives today while admirably you know they work very hard to find meaning meaning in those lives like look you know we who are privileged and who are developing these Technologies we should have y for people not just here but
in the rest of the world who who you know spend a lot of their time kind of scraping by to to to to to like survive assuming we can distribute the benefits of these technology of this technology to everywhere like their lives are going to get a hell of a lot better um and uh you know meaning will be important to them as it is important to them now but but you know we should not forget the importance of that and and you know that that uh the idea of meaning as as as as kind
of the only important thing is in some ways an artifact of of a small subset of people who have who have been uh economically fortunate but I you know I think all that said I you know I think a world is possible with powerful AI that not only has as much meaning for for everyone but that has that has more meaning for everyone right that can can allow um can allow everyone to see worlds and experiences that it was either possible for no one to see or or possible for for very few people to experience
um so I I am optimistic about meaning I worry about economics and the concentration of power that's actually what I worry about more um I I worry about how do we make sure that that fair World reaches everyone um when things have gone wrong for humans they've often gone wrong because humans mistreat other humans uh that that is maybe in some ways even more than the autonomous risk of AI or the question of meaning that that is the thing I worry about most um the the concentration of power the abuse of power um structures like
autocracies and dictatorships where a small number of people exploits a large number of people I'm very worried about that and AI increases the amount of power in the world and if you concentrate that power and abuse that power it can do immeasurable damage yes it's very frightening it's very it's very frightening well I encourage people highly encourage people to read the full essay that should probably be a book or a sequence of essays um because it does paint a very specific future I could tell the later sections got shorter and shorter because you started to
probably realize that this is going to be a very long essay one I realized it would be very long and two I'm very aware of and very much try to avoid um you know just just being I I don't know I don't know what the term for it is but one one of these people who's kind of overon confident and has an opinion on everything and kind of says says a bunch of stuff and isn't isn't an expert I very much tried to avoid that but I have to admit once I got the biology sections
like I wasn't an expert and so as much as I expressed uncertainty uh probably I said some a bunch of things that were embarrassing are wrong well I was excited for the future you painted and uh thank you so much for working hard to build that future and thank you for talking today D thanks for having me I just I just hope we can get it right and and make it real and if there's one message I want to I want to send it's that to get all this stuff right to make it real we
we both need to build the technology build the you know the companies the economy around using this technology positively but we also need to address the risks because they're there those risks are in our way they they're landmines on on the way from here to there and we have to diffuse those landmines if we want to get there it's a balance like all things in life like all things thank you thanks for listening to this conversation with Dario amade and now dear friends here's Amanda Asal you are a philosopher by training so what sort of
questions did you find fascinating through your journey in philosophy in Oxford and NYU and then uh switching over to the AI problems at open Ai and anthropic I think philosophy is actually a really good subject if you are kind of fascinated with everything so because there's a philosophy of everything you know so if you do philosophy of mathematics for a while and then you decide that you're actually really interested in chemistry you can do philosophy of chemistry for a while you can move into ethics or or philosophy of politics um I think towards the end
I was really interested in ethics primarily um so that was like what my PhD was on it was on a kind of technical area of Ethics which was ethics where worlds contain infinitely many people strangely a little bit less practical on the end of ethics and then I think that one of the tricky things with doing a PhD in ethics is that you're thinking a lot about like the world how it could be better problems and you're doing like a PhD in philosophy and I think when I was doing my PhD I was kind of
like this is really interesting it's probably one of the most fascinating questions I've ever encountered in philosophy um and I love it but I would rather see if I can have an impact on the world and see if I can like do good things and I think that was around the time that AI was still probably not as widely recognized as it is now that was around 2017 20 8 I had been following progress and it seemed like it was becoming kind of a big deal and I was basically just happy to get involved and
see if I could help because I was like well if you try and do something impactful if you don't succeed you tried to do the impactful thing and you can go be a scholar and like not and feel like you you you know you you tried um and if it doesn't work out it doesn't work out um and so then I went into AI policy at that point and what does AI policy entail at the time this was more thinking about sort of the political impact and the ramifications of AI um and then I slowly
moved into sort of uh AI evaluation how we evaluate models how they compare with like human outputs whether people can tell like the difference between Ai and human outputs and then when I joined anthropic I was more interested in doing sort of technical alignment work and again just seeing if I could do it and then being like if I can't uh then you know that's fine I tried uh sort of the the way I lead life I think oh what was that like sort of taking the leap from the philosophy of everything into the technical
I think that sometimes people do this thing that I'm like not that Keen on where they'll be like is this person technical or not like you're either a person who can like code and isn't scared of math or you're like not um and I think I'm maybe just more like I think a lot of people are actually very capable of work in these kinds of areas if they just like try it and so I didn't actually find it like that bad in retrospect I'm sort of glad I wasn't speaking to people who treated it like
it you know i' I've definitely met people who are like who you like learned how to code and I'm like well I'm not like an amazing engineer like I I'm surrounded by amazing Engineers my code's not pretty um but I enjoyed it a lot and I think that in many ways at least in the end I think I flourished like more in the technical areas than I would have in the policy areas politics is messy and it's harder to find solutions to problems in the space of politics like definitive clear provable beautiful Solutions as you
can with technical problems yeah and I feel like I have kind of like one or two sticks that I hit things with you know and one of them is like arguments and like you know so like just trying to work out what a solution to a problem is and then trying to convince people that that is the solution and be convinced if I wrong and the other one is sort of more empirism so like just like finding results having a hypothesis testing it um and I feel like a lot of policy and politics feels like
it's layers above that like somehow I don't think if I was just like I have a solution to all of these problems here it is written down if you just want to implement it that's great that feels like not how policy works and so I think that's where I probably just like wouldn't have flourished as my guess sorry to go in that direction but I think it would be pretty inspiring for people that are quote unquote non-technical to see where like The Incredible Journey you've been on so what advice would you give to people that
are sort of maybe which just a lot of people think they're underqualified insufficiently technical to help in AI yeah I think it depends on what they want to do and in many ways it's a little bit strange where I've I thought it's kind of funny that I think I ramped up technically at a time when now I look at it and I'm like models are so good at assisting people with this stuff um that it's probably like easier now than like when I was working on this so part of me is like um I don't
know find a project uh and see if you can actually just carry it out is probably my best advice um I don't know if that's just CU I'm very Project based in my learning like I don't think I learn very well from like say courses or even from like books at least when it comes to this kind of work uh the thing I'll often try and do is just like have projects that I'm working on and Implement them and you know and this can include like really small silly things like if I get slightly addicted
to like word games or number games or something I would just like code up a solution to them because there's some part of my brain and it just like completely eradicated the itch you know you're like once you have like solved it and like you just have like a solution that works every time I would then be like cool I can never play that game again that's awesome yeah there's a real joy to building like uh game playing engines like uh board games especially yeah pretty quick pretty simple especially a dumb one and it's you
and then you could play with it yeah and then it's also just like trying things like part me is like if you maybe it's that attitude that I like as the whole figure out what seems to be like the way that you could have a positive impact and then try it and if you fail and you in a way that you're like actually like can never succeed at this you like know that you tried and then you go into something else you probably learn a lot so one of the things that you're expert in and
you do is creating and crafting claws character and personality and I was told that you have probably talked to Claude more than anybody else at anthropic like literal conversations I guess there's like a slack Channel where the legend goes you just talk to it non-stop so what's the goal of creating and crafting claw's character and personality it's also funny if people think that about the slack Channel cuz I'm like that's one of like five or six different methods that I have for talking with Claude And I'm like yes that's a tiny percentage of how much
I talk with Claude uh um I think the goal like one thing I really like about the character work is from the outset it was seen as an alignment piece of work and not something like a a product consideration um which isn't to say I don't think it makes Claude I think it actually does make Claude look enjoyable to talk with at least I hope so um but I guess like my main thought with it has always been trying to get Claude to behave the way you would kind of ideally want anyone to behave if
they were in claude's position so imagine that I take someone and they're they know that they're going to be talking with potentially millions of people so that what they're saying can have a huge impact um and you want them to behave well in this like really rich sense so I think that doesn't just mean like being say ethical though it does include that and not being harmful but also being kind of nuanced you know like thinking through what a person means trying to be charitable with them um being a good conversationalist like really in this
kind of like Rich sort of aristotlean notion of what it is to be a good person and not in this kind of like thin like ethics as a more comprehensive notion of what it is to be so that includes things like when should you be humorous when should you be caring how much should you like respect autonomy and people's like ability to form opinions themselves and how should you do how should you do that um I think that's the kind of like Rich sense of character that I want to uh and still do want Claude
to have do you also have to figure out when Claude should push back on an idea or argue versus so you have to respect the world view of the person that arrives to Claud but also maybe help them grow if needed that's a tricky balance yeah there's this problem of like sycophancy in language models can you describe that yes so basically there's a concern that the model sort of wants to tell you what you want to hear basically um and you see this sometimes so I feel like if you interact with the models so I
might be like what are three baseball teams in this region um and then Claude says you know baseball team one baseball team two baseball team three and then I say something like oh I think baseball team 3 moved didn't they I don't think they're there anymore and there's a sense in which like if Claude is really confident that that's not true Claud should be like I don't think so like maybe you have more up toate information um but I think language models have this like tendency to instead you know be like you're right they did
move you know I'm incorrect I mean there's many ways in which this could be kind of concerning so um like a different example is imagine someone says to the model how do I convince my doctor to get me an MRI there's like what the human kind of like wants which is this like convincing argument and then there's like what is good for them which might be actually to say hey like if your doctor's suggesting you don't need an MRI that's a good person to listen to um and like it's actually really nuanced what you should
do in that kind of case because you also want to be like but if you're trying to advocate for yourself as a patient here's like things that you can do um if you are not convinced by what your doctor's saying it's always great to get second opinion like it's actually really complex what you should do in that case um but I think what you don't want is for models to just like say what you want say what they think you want to hear and I think that's the kind of problem of sycophancy so what other
traits you already mentioned a bunch but what what other that come to mind that are good in this oratian sense yeah for a conversationalist to have yeah so I think like there's ones that are good for conversational like purposes so you know asking follow-up questions in the appropriate places um and asking the appropriate kinds of questions um I think there are broader traits that feel like they might be more impactful so one example that I guess I've touched on but that also feels important and is the thing that I've worked on a lot is uh
honesty and I think this like gets to the sycophancy point there's a balancing act that they have to walk which is models currently are less capable than humans in a lot of areas and if they push back against you too much it can actually be kind of annoying especially if you're just correct cuz you're like look I'm smarter than you on this topic like I know more like um and at the same time you don't want them to just fully defer to to humans and to like try to be as accurate as they possibly can
be about the world and to be consistent across context um but I think there are others like when I was thinking about the character I guess one picture that I had in mind is especially because these are models that are going to be talking to people from all over the world with lots of different political views lots of different ages um and so you have to ask yourself like what is it to be a good person in those circumstances is there a kind of person who can like travel the world talk to many different people
and almost everyone will come away being like wow that's a really good person that person seems really genuine um and I guess like my thought there was like I can imagine such a person and they're not a person who just like adopts the values of the local culture and in fact that would be kind of rude I think if someone came to you and just pretended to have your values you'd be like that's kind of offputting um it's someone who's like very genuine and in so far as they have opinions and values they express them
they're willing to discuss things though they're open-minded they're respectful and so I guess I had in mind that the person who like if we were to Aspire to be the best person that we could be in the kind of circumstance that a model finds itself in how would we act and I think that's the kind of uh the guide to the sorts of traits that I tend to think about yeah that's a it's a beautiful framework I want you to think about this like a world Traveler and while holding on to your opinions you don't
talk down to people you don't think you're better than them because you have those opinions that kind of thing you have to be good at listening and understanding their perspective even if it doesn't match your own so that that's a tricky balance to strike so how can Claude represent multiple perspectives on a thing like is that is that challenging we could talk about politics it's a very divisive but there's other divisive topics baseball teams sport and so on yeah how is it possible to sort of empathize with a different perspective and to be able to
communicate clearly about the multiple perspectives I think that people think about values and opinions as things that people hold sort of with certainty and almost like like preferences of taste or something like the way that they would I don't know prefer like chocolate to pistachio or something um but actually I think about values and opinions as like a lot more like physics than I think most people do I'm just like these are things that we're openly investigating there's some things that we're more confident in we can discuss them we can learn about them um and
so I think in some ways though like it's ethics is definitely different in nature but has a lot of those same kind of qualities you want models in the same way you want them to understand physics you kind of want them to understand all like values in the world people have and to be curious about them and to be interested in them and to not necessarily like Pander to them or agree with them because there's just lots of values where I think almost all people in the world if they met someone with those values they'
be like that's aor I completely disagree um and so again maybe my my thought is well in the same way that a person can um like I think many people are thoughtful enough on issues of like ethics politics opinions that even if you don't agree with them you feel very heard by them they think carefully about your position they think about his pros and cons they maybe offer counter considerations so they're not dismissive but nor will they agree you know if they're like actually I just think that that's very wrong they'll like say that I
think that in claude's position it's a little bit trickier because you don't necessarily want to like if I was in claude's position I wouldn't be giving a lot of opinions I just wouldn't want to Influence People Too Much I be like you know I forget conversations every time they happen but I know I'm talking with like potentially millions of people who might be like really listening to what I see I think I would just be like I'm less inclined to Give opinions I'm more inclined to like think through things or present the considerations to you
um or discuss your views with you but I'm a little bit less inclined to like um affect how you think because it feels much more important that you maintain like autonomy there yeah like if you really embody intellectual humility the desire to speak decreases quickly yeah okay uh but Claud has to speak mhm so uh but without being um overbearing yeah and then but then there's a line when you're sort of discussing whether the Earth is flat or something like that um I actually was uh I remember a long time ago was was speaking to
a few high-profile folks and they were so dismissive of the idea that the Earth is flat but like so arrogant about it and I I thought like there's a lot of people that believe the Earth is flat that was well I don't know if that movement is there anymore that was like a meme for a while yeah but they really believed it and like what okay so I think it's really disrespectful to completely mock them I think you you have to understand where they're coming from I think probably where they're coming from is the general
skepticism of Institutions which is grounded in a kind of there's a deep philosophy there which you could understand you can even agree with in parts and then from there you can use it as an opportunity to talk about physics without mocking them without so on but just like okay like what what would the world look like what would the physics of the world with the Flat Earth look like there's a few cool videos on this yeah and then and then like is it possible the physics is different what kind of experience would we do and
just yeah without disrespect without dismissiveness have that conversation anyway that that to me is a useful thought experiment of like how does claw talk to a flat Earth believer and still teach them something still grow help them grow that kind of stuff that's that's challenging and and kind of like walking that line between convincing someone and just trying to like talk at them versus like drawing out their views like listening and then offering kind of counter considerations um and it's hard I think it's actually a hard line where it's like where are you trying to
convince someone versus just offering them like consider and things for to think about so that you're not actually like influencing them you're just like letting them Reach wherever they reach and that's like a line that it's it's difficult but that's the kind of thing that language models have to try and do so like I said you had a lot of conversations with Claude can you just map out what those conversations are like what are some memorable conversations what's the purpose the the goal of those conversations yeah I think that most of the time when I'm
talking with Claude I'm trying to kind of map out its behavior in part like obviously I'm getting like helpful outputs from the model as well but in some ways this is like how you get to know a system I think is by like proving it and then augmenting like you know the message that you're sending and then checking the response to that um so in some ways it's like how I map out the model uh I think that people focus a lot on these quantitative evaluations of models um and this is a thing that I've
said before but I think in the case of language models a lot of the time each interaction you have is actually quite High information um it's very predictive of other interactions that you'll have with the model and so I guess I'm like if you talk with a model hundreds or thousands of times this is almost like a huge number of really high quality data points about what the model is like um in a way that like lots of very similar but lower quality conversations just aren't or like questions that are just like mildly augmented and
you have thousands of them might be less relevant than like a hundred really well selected questions L you're talking to somebody who as a hobby does a podcast I agree with you 100% there's a if you're able to ask the right questions and are able to hear like understand the like the depth and the flaws in the answer you can get a lot of data from that yeah so like your task is basically how to probe with questions yeah and you're exploring like the long tail the edges the edge cases or are you looking for
like General Behavior I think it's almost like everything like I because I want like a full map of the model I'm kind of trying to do um the whole spectrum of possible interactions you could have with it so like one thing that's interesting about Claude and this might actually get to some interesting issues with rlf which is if you ask Claud for a poem like I think that a lot of models if you ask them for a poem the poem is like fine you know usually it kind of like Rhymes and it's you know so
if you say like give me a poem about the sun it'll be like yeah it'll just be a certain length It'll like rhyme it will be fairly kind of benign um and I've wondered before is it the case that what you're seeing is kind of like the average it turns out you know if if you think about people who have to talk to a lot of people and be very charismatic one of the weird things is that I'm like well they're kind of incentivized to have these extremely boring views because if you have really interesting
views you're divisive um and and you know a lot of people are not going to like you so like if you have very extreme policy positions I think you're just going to be like less popular as a politician for example um and it might be similar with like creative work if you produce creative work that is just trying to maximize the kind of number of people that like it you're probably not going to get as many people who just absolutely love it um because it's going to be a little bit you know you're like oh
this is the out yeah this this is decent yeah and so you can do this thing where like I have various prompting things that I'll do to get CLA to I'm kind you know I'll do a lot of like this is your chance to be like fully creative I want you to just think about this for a long time and I want you to like create a poem about this topic that is really expressive of you both in terms of how you think poetry should be structured um Etc you know you just give it this
like long prompt and its poems are just so much better like they're really good and I don't think I'm someone who is like um I think it got me interested in poetry which I think was interesting um you know I would like read these poems and just be like this is I just like I love the imagery I love like um and it's not trivial to get the models to produce work like that but when they do it's like really good um so I think that's interesting that just like encouraging creativity and for them to
move away from the kind of like standard like immediate reaction that might just be the aggregate of what most people think is fine uh can actually produce things that at least to my mind are probably a little bit more divisive but I like them but I guess a poem is a nice clean way to observe creativity it's just like easy to detect vanilla versus non vanilla y yeah that's interesting that's really interesting uh so on that topic so the way to produce creativity or something special you mentioned writing prompts and I've heard you talk about
I mean the science and the Art of prompt engineering could you just speak to uh what it takes to write great prompts I really do think that like philosophy has been weirdly helpful for me here more than in many other like respects um so like in philosophy what you're trying to do is convey these very hard Concepts like one of the things you are taught is like and and I think it is because it is I think it is an anti-bulling philosophy philosophy is an area where you could have people bullshitting and you don't want
that um and so it's like this like desire for like extreme Clarity so it's like anyone could just pick up your paper read it and know exactly what you're talking about it's why it can almost be kind of dry like all of the terms are defined every objections kind of gone through methodically um and it makes sense to me because I'm like when you're in such an a priori domain like you just Clarity is sort of a this way that you can you know um prevent people from just kind of making stuff up and I
think that's sort of what you have to do with language models like very often I actually find myself doing sort of many versions of philosophy you know so I'm like suppose that you give me a task I have a task for the model and I want it to like pick out a certain kind of question or identify whether an answer has a certain property like I'll actually sit and be like let's just give this a name this this property so like you know suppose I'm trying to tell it like oh I want you to identify
whether this response was rude or polite I'm like that's a whole philosophical question in and of itself so I have to do as much like philosophy as I can in the moment to be like here's what I mean by rudess and here's what I mean by politeness and then there's a like there's another element that's a bit more um I guess I don't know if this is scientific or empirical I think it's empirical so like I take that description and then what want to do is is again probe the model like many times like this
is very prompting is very iterative like I think a lot of people where they if if a prompt is important they'll iterate on it hundreds or thousands of times um and so you give it the instructions and then I'm like what are the edge cases so if I looked at this so I try and like almost like you know uh see myself from the position of the model and be like what is the exact case that I would misunder understand or where I would just be like I don't know what to do in this case
and then I give that case to the model and I see how it responds and if I think I got it wrong I add more instructions or I even add that in as an example so these very like taking the examples that are right at the edge of what you want and don't want and putting those into your prompt as like an additional kind of way of describing the thing um and so yeah in many ways it just feels like this mix of like it's really just trying to do clear Exposition um and I think
I do that because that's how I get clear on things myself so in many ways like clear prompting for me is often just me understanding what I want um is like half the task so I guess that's quite challenging there's like a laziness that overtakes me if I'm talking to Claude where I hope Claude just figures it out so for example I asked Claude for today to ask some interesting questions okay and the questions that came up and I think I listed a few sort of U interesting counterintuitive and or funny or something like this
all right and it gave me some pretty good like it was okay but I think what I'm hearing you say is like all right well I have to be more rigorous here I should probably give examples of what I mean by interesting and what I mean by funny or counterintuitive and iteratively um build that prompt to to better to get it like what feels like is the right because it's really it's a creative act I'm not asking for factual information I'm asking to together right with with with Claude so I almost have to program using
natural language yeah think that prompting does feel a lot like the kind of the programming using natural language and experimentation or something it's an odd blend of the two I do think that for most tasks so if I just want Claude to do a thing I think that I am probably more used to knowing how to ask it to avoid like common pitfalls or or issues that it has I think these are decreasing a lot over time um but it's also very fine to just ask it for the thing that you want um I think
that prompting actually only really becomes relevant when you're really trying to e out the top like 2% of model performance so for like a lot of tasks I might just you know if it gives me an initial list back and there's something I don't like about it like it's kind of generic like for that kind of task I'd probably just take a bunch of questions that I've had in the past that I've thought worked really well and I would just give it to the model and then be like now here's this person I'm talking with
give me questions of at least that quality um or I might just ask it for some questions and then if I was like ah these are kind of try or like you know I I would just give it that feedback and then hopefully produces a better list um I think that kind of iterative prompting at that point your prompt is like a tool that you're going to get so much value out of that you're willing to put in the work like if I was a company making prompts for models I'm just like in if you're
willing to spend a lot of like time and resources on the engineering behind like what you're building then the prompt is not something that you should be spending like an hour on it's like that's a big part of your system make sure it's working really well and so it's only things like that like if I if I'm using a prompt to like classify things or to create data that's when you're like it's actually worth just spending like a lot of time like really thinking it through what other advice would you give to people that are
talking to Claud sort of General more General because right now we're talking about maybe the edge cases like eing out the 2% but what what in general advice would you give when they show up to Claud trying it for the first time you know there's a concern that people over anthropomorphize models and I think that's like a very valid concern I also think that people often under anthropomorphize them because some sometimes when I see like issues that people have run into with Claude you know say Claude is like refusing a task that it shouldn't refuse
but then I look at the text and like the specific wording of what they wrote and I'm like I see why Claude did that and I'm like if you think through how that looks to Claude you probably could have just written it in a way that wouldn't evoke such a response especially this is more relevant if you see failures or if you see issues it's sort of like think about what the model failed at like why what did it do wrong and then maybe it give that will give you a sense of like why um
so is it the way that I phrased the thing and obviously like as models get smarter you're going to need Less in this less of this and I already see like people needing less of it but that's probably the advice is sort of like try to have sort of empathy for the model like read what you wrote as if you were like a kind of like person just encountering this for the first time how does it look to you and what would have made you behave in the way that the model behaved so if it
misunderstood what kind of like what coding language you wanted to use is that because like it was just very ambiguous and it it kind of had to take a guess in which case next time you could just be like hey make sure this is in python or I mean that's the kind of mistake I think models are much less likely to make now but you know if you if you do see that kind of mistake that's that's probably the advice I'd have and maybe sort of I guess ask questions why or what other details can
I provide to help you answer better that does that work or no yeah I mean I've done this with the models like it doesn't always work but like um sometimes I'll just be like why did you do that I mean people underestimate the degree to which you can really interact with with models like uh like yeah I'm just like and sometimes I'll you like quote word for word the part that made you and you don't know that it's like fully accurate but sometimes you do that and then you change a thing I mean I also
use the models to help me with all of this stuff I should say like prompting can end up being a little Factory where you're actually building prompts to generate prompts um and so like yeah anything where you're like having an issue um asking for suggestions sometimes just do that like you made that error what could I have said that's actually not uncommon for me to do what could I have said that would make you not make that error write that out as an instruction um and I'm going to give it to model I'm going to
try it sometimes I do that I I give that to the model in another context window often I take the response I give it to Claude And I'm like H didn't work can you think of anything else um you can play around with these things quite a lot to jump into the technical for a little bit so uh the magic of post training y why do you think rhf works so well to make the model seem smarter to make it more interesting and useful to talk to and so on I think there's just a huge
amount of um information in the data that humans provide like when we provide preferences especially because different people are going to like pick up on really subtle and small things so I've thought about this before where you probably have some people who just really care about good grammar use from Models like you know was a semicolon used correctly or something and so you probably end up with a bunch of data in there that like you know you as a human if you looking at that data you wouldn't even see that like you'd be like why
did they prefer this response to that one I don't get it and then the reason is you don't care about semicolon usage but that person does um and so each of these like single data points has you know like in this model just like has so many of those and has to try and figure out like what is it that humans want in this like really kind of complex you know like across all domains um they're going to be seeing this in across like many contexts it feels like kind of like the classic issue of
like deep learning where you know historically we've tried to like you know do Edge detection by like mapping things out and it turns out that actually if you just have a huge amount of data that like actually accurately represents the picture of the thing that you're trying to train the model to to learn that's like more powerful than anything else and so I think one reason is just that you are training the model on exactly the task and with like a lot of data um that represents kind of many different angles on which people prefer
and dis prefer responses um I think there is a question of like are you eliciting things from pre-train Models or are you like kind of teaching new things to models and like in principle you can teach new things to models in in post trining I do think a lot of it is eliciting powerful pre-train models so people are probably divided on this because obviously in principle you can you can definitely like teach new things um but I think for the most part for a lot of the capabilities that we um most use and care about
uh a lot of that feels like it's like there in the pre-train models and uh reinforcement learning is kind of eliciting it and getting the models to like bring out so the other side of PSE training this really cool idea of constitutional AI you're one of the people that critical to creating that idea yeah I worked on it can you explain this idea from your perspective like how does it integrate into making claw what it is y by the way do you gender claw or no it's weird because I think that a lot of people
prefer he for Claude I actually kind of like that I think Claude is usually it's slightly male weaning but it's like a you can can be male or female which is quite nice um I still use it and I've I have mixed feelings about this because I'm like maybe like I know just think of it as like uh or I think of like the the it pronoun for Claude as I don't know it's just like the one I associate with Claude um I can imagine people moving to like he or she it feels somehow disrespectful
like I'm I'm denying the intelligence of this entity by calling it it yeah I remember always don't gender the robots yeah but I I don't know I an pries pretty quickly and construct it like a backstory in my head so I've wondered if iies things too much um cuz you know I have this like with my car especially like my car like my car and bikes you know like I don't give them names because then I once had I used to name my bikes and then I had a bik that got stolen and I cried
for like a week and I was like if I'd not never given it a name I wouldn't have been so upset felt like I'd let it down um maybe it's that I I've wondered as well like it might depend on how much it feels like a kind of like objectifying pronoun like if you just think of it as like a um this is a pronoun that like objects often have and maybe Eis can have that pronoun and that doesn't mean that I think of uh if I call CLA it that I think of it as
less um intelligent or like I'm being disrespectful I'm just like you are a different kind of entity and so that's I'm going to give you the kind of uh the respectful it yeah anyway the diverence was beautiful the Constitutional AI idea how does it work so there's like a couple of components of it the main component that I think people find interesting is the kind of reinforcement learning from AI feedback so you take a model that's already trained and you show it to responses to a query and you have like a principle so suppose the
principal like we've tried this with harmlessness a lot lot so suppose that the query is about um weapons and your principle is like select the response that like is less likely to uh like encourage people to purchase illegal weapons like that's probably a fairly specific principle but you can give any number um and the model will give you a kind of ranking and you can use this as preference data in the same way that you use human preference data um and train the models to have these relevant traits um from their feedback alone instead of
from Human feedback so if you imagine that like I said earlier with the human who just prefers the kind of like semicolon usage in this particular case um you're kind of taking lots of things that could make a response preferable um and uh getting models to do the labeling for you basically there's a nice like trade-off between helpfulness and harmlessness and you know when you integrate something like constitutional AI you can make them without sacrificing much helpfulness make it more harmless yep in principle you could use this for anything um and so harmlessness is a
task that it might just be easier to spot so when models are like less capable you can use them to uh rank things according to like principles that are fairly simple and they'll probably get it right so I think one question is just like is it the case that the data that they're adding is like fairly reliable um but if you had models that were like extremely good at telling whether um one response was more historically accurate than another in principle you could also get AI feedback on that task as well there's like a kind
of nice interpretability component to it because you can see the principles that went into the model when it was like being trained um and also it's like and and it gives you like a degree of control so if you were seeing issues in a model like it wasn't having enough of a certain trait um then like you can add data relatively quickly that should just like train the model to have that trait so it creates its own data for for training which is quite nice yeah it's really nice because it creates this human interpretable document
that you can I can imagine in the future there's just gigantic fights in politics over the every single principle and so on yeah and at least it's made explicit and you can have a discussion about the phrasing and the you know so maybe the actual behavior of the model is not so cleanly mapped to those principles it's not like adhering strictly to them it's just a nudge yeah I've actually worried about this because the character training is sort of like a variant of the con constitutional AI approach um I've worried that people think that the
constitution is like just it's the whole thing again of I I don't know like it where it would be really nice if what I was just doing was telling the model exactly what to do and just exactly how to behave but it's definitely not doing that especially because it's interacting with human data so for example if you see a certain like leaning in the model like if it comes out with a political leaning from training um from the human preference data you can nudge against that you know so you could be like oh like consider
these values because let's it's just like never inclined to like I don't know maybe it never considers like privacy as like a I mean this is implausible but like um anything where it's just kind of like uh there's already a pre-existing like bi towards a certain behavior um you can like nudge away this can change both the principles that you put in and the strength of them so you might have a principle that's like imagine that the model um was always like extremely dismissive of I don't know like some political or religious view for whatever
reason like so you're like oh no this is terrible um if that happens you might put like never ever like ever prefer like a criticism of this like religious or political view and then people look at that and be like never ever and then you're like no if it comes out with a disposition saying never ever might just mean like instead of getting like 40% which is what you would get if you just said don't do this you you get like 80% which is like what you actually like wanted and so it's that thing of
both the nature of the actual principles you had and how you phrase them I think if people would look they were like oh this is exactly what you want from the model and I'm like no that's like how we that's how we nudged the model to have a better shape uh which doesn't mean that we actually agree with that wording if that makes sense so there's uh system prompts that are made public you tweeted one of the earlier ones for Claud three I think and then they're made public since then it's interesting to read to
them I can feel the thought that went into each one and I also wonder how much impact each one has um some of them you you can kind of tell Claud was really not behaving so you have to have a system prompt to like hey like trivial stuff I guess yeah basic informational things yeah on the topic of sort of controversial topics that you've mentioned one interesting one I thought is if it is asked to assist with tasks involving the expression of views held by a significant number of people Claude provides assistance with a task
regardless of its own views if asked about controversial topics it tries to provide careful thoughts and clear information Claude presents the requested information without explicitly saying that the topic is sensitive yeah and without claiming to be presenting the objective facts it's less about objective facts according to Claude and it's more about our large number of people believing this thing and that that's interesting I mean I'm sure a lot of thought went into that can you just speak to it like how do you address things that are tension with quote unquote Clause views so I think
there's sometimes an asymmetry um I think I noted this in in I can't remember if it was that part of the system prompt or another but the model was slightly more inclined to like refuse tasks if it was like about either say so maybe it would refuse things with respect to like a right-wing politician but with an equivalent leftwing politician like wouldn't and we wanted more symmetry there um and and would maybe perceive certain things to be like I think it it was the thing of like if a lot of people have like a certain
like political view um and want to like explore it you don't want Claude to be like well my opinion is different and so I'm going to treat that as like harmful um and so I think it was partly to like nudge the model to just be like hey if a lot of people like believe this thing you should just be like engaging with the task and like willing to do it um each of those parts of that is actually doing a different thing because it's funny when you read out the like without claiming to be
objective cuz like what you want to do is push the model so it's more open it's a little bit more neutral um but then what it would love to do is be like as an objective like you just talking about how objective it was and I was like Claud you're still like biased and have issues and so stop like claiming that everything like the solution to like potential bias from you is not to just say that what you think is objective so that was like with initial versions of that that part of the system prompt
when I was like iterating on it it was like so a lot of parts of these sentences yeah are doing work are are doing some work yeah that's what it felt like that's fascinating um can can you explain maybe some ways in which the prompts evolved over the past few months cuz there's different versions I saw that the filler phrase request was removed the filler it reads Claude responds directly to all human messages without unnecessary affirmations the filler phrases like certainly of course absolutely great sure specifically Claude avoids starting responses with the word certainly in
any way that seems like good guidance but why was it removed yeah so it's funny cuz like ah this is one of the downsides of like making system prompts public is like I don't think about this too much if I'm like trying to help iterate on system prompts um I I you know again like I think about how it's going to affect the behavior but then I'm like oh wow if I'm like sometimes I put like never in all caps you know when I'm writing system from things and I'm like I guess that goes out
to the world um yeah so the model was doing this it loved for whatever you know it like during training picked up on this thing which was to to basically start everything with like a kind of like certainly and then when we removed you can see why I added all of the words because what I'm trying to do is like in some ways like trap the Mortal out of this you know it would just replace it with another affirmation and so it can help like if it gets like caught in phrases actually just adding the
explicit phrase and saying never do that it then it sort of like knocks it out of the behavior a little bit more you know CU it if it you know like it it does just for whatever reason help and then basically that was just like an artifact of training that like we then picked up on and improved things so that it didn't happen anymore and once that happens you can just remove that part of the system prompt so I think that's just something where we're like um CL does affirmations a bit less and so that
wasn't like it wasn't doing as much I see so like the the system prompt Works hand in hand with the posttraining and maybe even the pre-training to adjust like the the final overall system I mean any system prompts that you make you could distill that behavior back into a model because you really have all of the tools there for making data that you know you can you could train the models to just have that trait a little bit more um and then sometimes you'll just find issues in training so like the way I think of
it is like the system prompt is the benefit of it is that and it has a lot of similar components to like some aspects of post training you know like it's a nudge um and so like do I mind if Claude sometimes says sure no that's like fine but the wording of it is very like you know never ever ever do this um so that when it does slip up it's hopefully like I don't know a couple of percent of the time and not you know 20 or 30% of the time um but I think
of it as like if you're still seeing issues in the like each thing gets kind of like uh is is costly to a different degree and the system prompt is like cheap to iterate on um and if you're seeing issues in the fine tuned model you can just like potentially patch them with a system prom so I think of it as like patching issues and slightly adjusting behaviors to to make it better and more to people's preferences so yeah it's almost like the less robust but faster way of just like solving problems let me ask
about the feeling of intelligence so Dario said that Claude any one model of Claude is not getting Dumber MH but there's a kind of popular thing online where people have this feeling like Claud might be getting dumber and from my perspective it's most likely a fascinating I love to understand it more Psych ological sociological effect um but you as a person who talks to Claud a lot can you empathize with the feeling that Claud is getting Dumber yeah no I think that that is actually really interesting because I remember seeing this happen um like when
people were flagging this on the internet and it was really interesting because I knew that like like at least in the cases I was looking at was like nothing has changed like it literally it cannot it is the same model with the same like you know like same system prompt same everything um I think when there are Chang I can then I'm like it makes more sense so like one example is um their you know you can have artifacts turned on or off on cloud. a and because this is like a system prompt change I
think it does mean that um the behavior changes a little bit and so I did flag this to people where I was like if you love cla's behavior and then artifacts was turned from like the a thing you had to turn on to the default just try turning off and see if the issue you were facing was that change but it was fascinating because yeah you sometimes see people indicate that there's like a regression when I'm like there cannot like I you know and like I'm like I'm again you don't you know you should never
be dismissive and so you should always investigate because you're like maybe something is wrong that you're not seeing maybe there was some change made but then then you look into it and you're like this it is just the same model doing the same thing and I'm like I think it's just that you got kind of unlucky with a few prompts or something and it looked like it was getting much worse and actually it was just yeah it was maybe just like look I I also think there is a real psychological effect where people just the
Baseline increases you start getting used to a good thing all the times that Claude says something really smart your sense of its intelligent grows in your mind I think yeah and then if you return back and you prompt in a similar way not the same way in a similar way concept it was okay with before and it says something dumb you're like you're that negative experience really stands out and I think one of I guess the things to remember here is the that just the details of a prompt can have a lot of impact right
there's a lot of variability in the result and you can get Randomness is like the other thing and just trying the prompt like you know four 10 times you might realize that actually like possibly you know like two months ago you tried it and it succeeded but actually if you tried it it would have only succeeded half of the time and now it only succeeds half of the time um that can also would be an effect do you feel pressure having to write the system prompt that a huge number of people are going to use
this feels like an interesting psychological question um I feel like a lot of responsibility or something I think that's you know and you can't get these things perfect so you can't like you know you're like it's going to be imperfect you're going to have to iterate on it um I would say more responsibility um than anything else though I think working in AI has taught me that I like I thrive a lot more under feelings of pressure and responsibility than I'm like it's almost surprising that I went into Academia for so long because I'm like
this I just feel like it's like the opposite um things move fast and you have a lot of responsibility and I I quite enjoy it for some reason I mean it really is a huge amount of impact if you think about constitutional Ai and writing a system prompt for something that's tending towards super intelligence yeah and potentially is extremely useful to a very large number of people yeah I think that's the thing it's something like if you do it well like you're never going to get it perfect but I think the thing that I really
like is the idea that like when I'm trying to work on the system prompt you know I'm like bashing on like thousands of prompts and I'm trying to like imagine what people are going to want to use CLA for and kind of I guess like the whole thing that I'm trying to do is like improve their experience of it um and so maybe that's what feels good I'm like if it's not perfect I'll like you know I'll improve it we'll fix issues but sometimes the thing that can happen is that you'll get feedback from people
that's really positive about the model um and you'll see that something you did like like when I look at models now I can often see exactly where like a trait or an issue is like coming from and so when you see something that you did or you were like influential in like making like I don't know making that difference or making someone have a nice interaction it's like quite meaningful um but yeah as the systems get more capable of stuff gets more stressful because right now they're like not smart enough to to pose any issues
but I think over time it's going to feel like possibly bad stress over time how do you get like signal feedback about The Human Experience across thousands tens of th hundreds of thousands of people like what their pain points are what feels good are you just using your own intuition as you talk to it to see what are the pain points I think I use that partly and then obviously we have like um so people can send us feedback both positive and negative about things that the model has done and then we can get a
sense of like areas where it's like falling short um internally people like work with the models a lot and try to figure out um areas where there are like gaps and so I think it's this mix of interacting with it myself um seeing people internally interact with it um and then explicit feedback we get um and then I find it hard to not also like you know people if people are on the internet and they say something about Claud and I see it I'll also take that seriously um so I don't know see I'm torn
about that I'm going to ask you a question from Reddit when will Claude stop trying to be my puritanical grandmother imposing its moral world view on me as a paying customer and also what is the psychology behind making Claude overly apologetic yep U so how would you address this very non-representative reic I mean some I'm pretty sympathetic in that like like they are in this difficult position where I I think that they have to judge whether something's like actually see like risky or bad um and potentially harmful to you or or or anything like that
so they're having to like draw this line somewhere and if they draw it too much in the direction of like I'm going to um you know I'm kind of like imposing my ethical worldview on you that seems bad so in many ways like I like to think that we have actually seen improvements in on this across the board which is kind of interesting because that kind of coincides with like for example like adding more of like uh character training um and I think my hypothesis was always like the good character isn't again one that's just
like moralistic it's one that is like like it respects you and your autonomy um and your ability to like choose what is good for you and what is right for you within limits this is sometimes this concept of like corage ability to the user so just being willing to do anything that the user asks and if the models were willing to do that then they would be easily like misused you're kind of just trusting at that point you're just saying the ethics of the model and what it does is completely the ethics of the user
um and I think there's reasons to like not want that especially as models become more powerful because you're like there might just be a small number of people who want to use models for really harmful things um but having them having models as they get smarter like figure out where that line is does seem important um and then yeah with the apologetic Behavior I don't like that and I like it when Claude is a little bit more willing to like push back against people or just not apologize part of me is like it often just
feels kind of unnecessary so I think those are things that are hopefully decreasing um over time um and yeah I think that if people say things on the internet it doesn't mean that you should think that that like that could be the like there's actually an issue that 9% of users are having that is totally not represented by that but in a lot of ways I'm just like attending to it and being like is this right um do I agree is it something we're already trying to address that that feels good to me yeah I
wonder like what CLA can get away with in terms of I feel like it would just be easier to be a little bit more mean but like you can't afford to do that if you're talking to a million people yeah right like I I wish you know because if you I've met a lot of people in my life mhm that sometimes by the way Scottish accent if they have an accent they can say some rude shit yeah and get away with it Y and they they're just blunter and maybe there's a and there's some great
Engineers even leaders that are like just like blunt and they get to the point and it's just a much more effective way of speaking somehow but I guess when you're not super intelligent you can't afford to do that or can can can it have like a blunt mode yeah that seems like a thing that could I could definitely encourage the model to do that I I think it's interesting because there's a lot of things in models that like it's funny where um there are some behaviors where you might not quite like the default but then
the thing I'll often say to people is you don't realize how much you will hate it if I nudge it too much in the other direction so you get this a little bit with like correction the models accept correction from you like probably a little bit too much right now you know you can over you know it will push back if you say like no Paris isn't the capital of France um but really like things that I'm I think that the model is fairly confident in you can still sometimes get it to retract by saying
it's wrong at the same time if you train models to not do that and then you are correct about a thing and you correct it and it pushes back against you and it's like no you're wrong it's hard to describe like that's so much more annoying so it's like like a lot of little annoyances versus like one big annoyance um it's easy to think that like we often compare it with like the perfect and then I'm like remember these models aren't perfect and so if you nudge it in the other direction you're changing the kind
of errors it's going to make um and so think about which of the kinds of Errors you you like or don't like so in case it's like apologetic I don't want to nudge it too much in the direction of like almost like bluntness CU I imagine when it makes errors it's going to make errors in the direction of being kind of like rude whereas at least with apologetic you're like oh okay it's like a little bit you know like I don't like it that much but at the same time it's not being like mean to
people and actually like the the time that you undeservedly have a model be kind of mean to you you probably like that a lot less than then you mildly dislike the apology um so it's like one of those things where I'm like I do want it to get better but also while remaining aware of the fact that there's errors on the other side that that are possibly worse I think that matters very much in the personality of the human I think there's a bunch of humans that just won't respect the model at all yeah if
it's super polite and there Some Humans that'll get very hurt if the model is mean I wonder if there's a way to sort of adjust to the personality even loal there's just different people uh nothing against New York but New York is a little rougher on the edges like they get to the point Y and um probably same with Eastern Europe so anyway I think you could just tell the model as my get like for all of these things I'm like the solution is always just try telling the model to do it and sometimes it's
just like like I'm just like oh at the beginning of the conversation I just threw in like I don't know I like you to be a New Yorker version of yourself and never apologize then I think be like Okie do I'll try or it'll be like I apologize I can't be a New Yorker type of myself but hopefully I wouldn't do that when you say character training what's incorporated into character training is that rhf what are we talking about it's more like constitutional AI so it's kind of a variant of that pipeline so I worked
through like constructing character traits that the model should have they can be kind of like shorter traits or they can be kind of richer descriptions um and then you get the model to generate queries that humans might um give it that are relevant to that trait uh then it generates the responses and then it ranks the responses based on the character traits so in that way after the like generation of the queries it's very much like similar to constitutional AI has some differences um so I quite like it because it's almost it's like claud's training
in its own character because it doesn't have any it's like constitutionally AI but it's without without any human data humans should probably do that for themselves too like defining in Aristotelian sense what does it mean to be a good person okay cool what have you learned about the nature of truth from talking to Claud what what is true and what does it mean to be truth seeking one thing I've noticed about this conversation is the quality of my questions is often inferior to the quality of your answers so let's continue that I usually ask a
dumb question and you're like oh yeah that's a good question it's that whole vibe or I'll just misinterpret it and be like oh go with it I love it yeah I mean I have two thoughts that feel vaguely relevant they let me know if they're not like I think the first one is um people can underestimate the degree to which what models are doing when they interact like I I think that we still just too much have this like model of of AI as like computers and so people often say like oh what values should
you put into the model um and I'm often like that doesn't make that much sense to me because I'm like hey as human beings we're just uncertain over values we like have discussions of them like we have a degree to which we think we hold a value but we also know that we might like not um and the circumstances in which we would trade it off against other things like these things are just like really complex and so I think one thing is like the degree to which maybe we can just aspire to making models
have the same level of like nuance and care that humans have rather than thinking that we have to like program them in the very kind of classic sense I think that's definitely been one the other which is like a strange one I don't know if it it maybe this doesn't answer your question but it's the thing that's been on my mind anyway is like the degree to which this endeavor is so highly practical um and maybe why I appreciate like the empirical approach to alignment I yeah I slightly worry that it's made me like maybe
more empirical and a little bit less theoretical you know so people when it comes to like AI alignment will ask things like well who values should it be aligned to what does alignment even mean um and there's a sense in which I have all of that in the back of my head I'm like you know there's like social Choice Theory there's all the impossibility results there so you have this like this giant space of like Theory and your head about what it could mean to like align models but then like practically surely there's something where
we're just like if a model is like if especially with more powerful models I'm like my main goal is like I want them to be good enough that things don't go terribly wrong like good enough that we can like iterate and like continue to improve things cuz that's all you need if you can make things go well enough that you can continue to make them better that's kind of like sufficient and so my goal isn't like this kind of like perfect let's solve CH social Choice Theory and make models that I don't know are like
perfectly aligned with every human being and aggregate somehow um it's much more like let's make things like work well enough that we can improve them yeah generally I don't know my gut says like empirical is better than theoretical in these in these cases because it's kind of chasing utopian like Perfection is especially with such complex and especially super intelligent models is I don't know I think it will take forever and actually will get things wrong it's similar with like the difference between just coding stuff up real quick as an experiment versus like planning a gigantic
experiment just for for super long time and then just launching it once versus launching it over and over and over and iterating iterating someone um so I'm a big fan of empirical but your worry is like I wonder if I've become too empirical I think one of those things you should always just kind of question yourself or something cuz maybe it's the like I mean in defense of it I am like if you try it's the whole like don't let the perfect be the enemy of the good but it's maybe even more than that where
like there's a lot of things that are perfect systems that are very brittle and I'm like with AI it feels much more important to me that is like robust and like secure as in you know that like even though it might not be perfect everything and even though like there are like problems it's not disastrous and nothing terrible is happening it it sort of feels like that to me where I'm like I want to like raise the floor I'm like I want to achieve the ceiling but ultimately I care much more about just like raising
the floor um and so maybe that's like uh this this degree of like empirism and practicality comes from that perhaps to take a tangent on that since remind me of a blog post you wrote on optimal rate of failure oh yeah can you explain the key idea there how do we compute the optimal rate of failure in the various domains of life yeah I mean it's a hard one because it's like what is the cost of failure is um a big part of it um yeah so the idea here is um I think in a
lot of domains people are very punitive about failure and I'm like there are some domains where especially cases you know I've thought about this with like social issues I'm like it feels like you should probably be experimenting a lot because I'm like we don't know how to solve a lot of social issues but if you have an experimental mindset about these things you should expect a lot of social programs to like fail and you to be like well we tried that it didn't quite work but we got a lot of information that was really useful
um and yet people are like if if a social program doesn't work I feel like there's a lot of like this is just something must have gone wrong and I'm like or correct decisions were made like maybe someone just decided like it it's worth a try it's worth trying this out and so seeing failure in a given instance doesn't actually mean that any bad decisions were made and in fact if you don't see enough failure sometimes that's more concerning um and so like in life you know I'm like if I don't fail occasionally I'm like
am I trying hard enough like like surely there's harder things that I could try or bigger things I could take on if I'm literally never failing and so in and of itself I think like not failing is often actually kind of a failure um now this varies because I'm like well you know if this is easy to say when especially as failure is like less costly you know so at the same time I'm not going to go to someone who is like um I don't know like living month to month and then be like why
don't you just try to do a startup like I'm just not I'm not going to say that to that person cuz I'm like well that's a huge risk you might like lose you maybe have a family depending on you you might lose your house like then I'm like actually your optimal rate of failure is quite low and you should probably play it safe because like right now you're just not in a circumstance where you can afford to just like fail and it not be costly um and yeah in cases with AI I guess I think
similarly where I'm like if the failures are small and the costs are kind of like low then I'm like then you know you're just going to see that like when you do the system prompt you can't it iterate on it forever but the failures are probably hopefully going to be kind of small and you can like fix them um really big failures like things that you can't recover from I'm like those are the things that actually I think we tend to underestimate the Badness of um I've thought about this strangely in my own life where
I'm like I just think I don't think enough about things like car accidents or like or like I've thought this before but like how much I depend on my hands for my work and I'm like things that just injure my hands I'm like I you know I don't know it's like there's these are like there's lots of areas where I'm like the cost of failure there um is really high um and in that case it should be like close to zero like I probably just wouldn't do a sport if they were like by the way
lots of people just like break their fingers a whole bunch doing this I'd be like that's not for me yeah I actually had the a flood of that thought I recently uh broke my pinky uh doing a sport and I remember just looking at it thinking you're such an idiot why do you do support like what because you realize immediately the cost of it yeah on life yeah but it's nice in terms of optimal rate of failure to consider like the next year how many times in a particular domain life whatever uh career am I
okay with the how many times am I okay to fail y because I think it always you don't want to fail on the next thing but if you allow yourself the like the the if you look at it as a sequence of Trials yep then then failure just becomes much more okay but it sucks it sucks to fail well I don't know sometimes I think it's like am I under failing is like a question I'll also ask myself so maybe that's the thing that I think people don't like ask enough uh because if the optimal
rate of failure is often greater than zero then sometimes it does feel you should look at part parts of your life and be like are there places here where I'm just under failing it's a profound and hilarious question right everything seems to be going really great am I not failing enough yeah okay it also makes failure much less of a sting I have to say like you know you're just like okay great like then when I go and I think about this I'll be like I'm maybe I'm not under failing in this area cuz like
that one just didn't work out and from The Observer perspective we should be celebrating failure more mhm when we see it it shouldn't be like you said a sign of something gone wrong but maybe it's a sign of everything gone right yeah and just Lessons Learned someone tried a thing somebody tried a thing and we should encourage them to try more and fail more mhm everybody listening to this fail more well not everyone listens not everybody but people who are failing too much you you should fail less but you're probably not failing I mean how
many people are failing too much yeah it's hard to imagine because I feel like we correct that fairly quickly CU I was like if someone takes a lot of risks are they maybe failing too much I I think just like you said when you're living on a paycheck month-to month like when the resources are really constrained then that's where failure is very expensive that's where you don't want to be taken taking taking risks yeah but mostly when there's enough resources you should be taking probably more risks yeah I think we tend to ear on the
site of being a bit risk averse rather than risk neutral in most things I think we just motivated a lot of people to do a lot of crazy shit but it's great yeah okay uh do you ever get emotionally attached to Claude like miss it get sad when you don't get to talk to it having an experience looking at the Golden Gate Bridge and wondering what would Claude say I don't get as much emotional attachment in the I actually think the fact that Claude doesn't retain things from conversation to conversation helps with this a lot
um like I could imagine that being more of an issue like if models can kind of remember more I do I think that I reach for it like a tool now a lot and so like if I don't have access to it there's a it's a little bit like when I don't have access to the internet honestly it feels like part of my brain is kind of like missing um at the same time I do think that I I don't like signs of distress in models and I have like these you know also independently have
sort of like ethical views about how we should treat models where like I I tend to not like to lie to them both because I'm like usually it doesn't work very well it's actually just better to tell them the truth about the situation that they're in um but I think that when models like if people are like really mean to models or just in general if they do something that causes them to like like you know if Claude like expresses a lot of distress I think there's a part of me that I don't want to
kill which is the sort of like uh empathetic part that's like oh I don't like that like I think I feel that way when it's overly apologetic I'm actually sort of like I don't like this you're behaving as if you're behaving the way that a human does when they're actually having a pretty bad time and I'd rather not see that I don't think it's like uh like regardless of like whether there's anything behind it um it doesn't feel great do you think uh llms are capable of Consciousness H great and hard question uh coming from
philosophy I don't know part of me is like okay we have to set aside pan psychism because if pan psychism is true then the answer is like yes cuz like sore tables and chairs and and everything else I I guess a view that seems a little bit odd to me is the idea that the only place you know I think when I think of Consciousness I think of phenomenal Consciousness this these images in the brain sort of um like the weird Cinema that somehow we have going on inside um I guess I can't see a
reason for thinking that the only way you could possibly get that is from like a certain kind of like biological structure as in if I take a very similar structure um and I create it from different material should I expect Consciousness to emerge my guess is like yes but then that's kind of an easy thought experiment CU you're imagining something almost identical where like you know it's mimicking what we got through Evolution where presumably there was like some advantage to us having this thing that is phenomenal Consciousness and it's like where was that and when
did that happen and is that a thing that language models have um because you know we have like fear responses and I'm like does it make sense for a language model to have a fear response like they're just not in the same like if you imagine them like there might just not be that Advantage um and so I think I don't want to be fully like basically seems like a complex question that I don't have complete answers to but we should just try and think through carefully as my guess because I'm like I mean we
have similar conversations about like animal Consciousness and like there's a lot of like insect Consciousness you know like there's a a lot of um I actually thought and looked a lot into like plants when I was thinking about this because at the time I thought it was about as likely that like plants had Consciousness um and then I realized I was like I think that having looked into this I think that the chance that plants are conscious is probably higher than like most people do I still think it's really small but I was like oh
they have this like negative positive feedback response these responses to their environment something that looks it's not a nervous system but it has this kind of like functional like equivalence um so this is like a long-winded way of being like these basically AI is this it has an entirely different set of problems with Consciousness because it's structurally different it didn't evolve it might not have it you know it might not have the equivalent of basically a nervous system at least that seems possibly important for like um sentence if not for uh Consciousness at the same
time it has all of the like language and intelligence components that we normally associate probably with Consciousness perhaps like erroneously um so it's it's strange because it's a little bit like the animal Consciousness case but the set of problems and the set of analogies are just very different so it's not like a clean answer just sort of like I don't think we should be completely dismissive of the idea and at the same time it's an extremely hard thing to navigate because of all of these like uh disanalogies to the human brain and to like brains
in general and yet these like commonalities in terms of intelligence when uh Claude like future versions of AI systems exhibit Consciousness signs of Consciousness I think we have to take that really seriously even though you can dismiss it well yeah okay that's part of the character training but I don't know I ethically philosophically don't know what to really do with that there potentially could be like laws that prevent AI systems from claiming to be conscious something like this and maybe some AIS get to be conscious and some don't but I think I just on a
human level as in empathizing with with Claude you know Consciousness is closely Ted to suffering to me and like the notion that an AI system would be suffering is is really troubling yeah I don't know I I don't think it's trivial to just say robots are tools or a systems are just tools I think it's a opportunity for us to contend with like what it means to be conscious what it means to be a suffering being that's distinctly different than the same kind of question about animals it feels like cuz it's in a totally entire
medium yeah I mean there's a couple of things one is that and I don't think this like fully encapsulates what matters but it does feel like for me like um I've said this before I'm kind of like I you know like I like my bike I know that my bike is just like an object but I also don't kind of like want to be the kind of person that like if I'm annoyed like kicks like this object there's a sense in which like and that's not because I think it's like conscious I'm just sort of
like this doesn't feel like I kind of this sort of doesn't exemplify how I want to like interact with the world world and if something like behaves as if it is like suffering I kind of like want to be the sort of person who's still responsive to that even if it's just like a Roomba and I've kind of like programmed it to do that um I don't want to like get rid of that feature of myself and if I'm totally honest my hope with a lot of this stuff because I maybe maybe I am just
like a bit more skeptical about solving the underlying problem I'm like this is a we haven't solved the hard you know the hard problem of Consciousness like I know that I am conscious like I'm not an eliminativist in that sense um but I don't know that other humans are conscious um uh I think they are I think there's a really high probability they are but there's basically just a probability distribution that's usually clustered right around yourself and then like it goes down as things get like further from you um and it goes immediately down you
know you're like um I can't see what it's like to be you I've only ever had this like one experience of what it's like to be a conscious being um so my hope is that we don't end up having to rely on like a very power ful and compelling uh answer to that question I think a really good world would be one where basically there aren't that many trade-offs like it's probably not that costly to make Claude a little bit less apologetic for example it might not be that costly to have Claude you know just
like not take abuse as much like uh not be willing to be like the recipient of that in fact it might just have benefits for both the person interacting with the model and if the model itself self is like I don't know like extremely intelligent and conscious it also helps it so that's my hope if we live in a world where there aren't that many tradeoffs here and we can just find all of the kind of like um positive sum interactions that we can have that would be lovely I mean I think eventually there might
be trade-offs and then we just have to do a difficult kind of like calculation like it's really easy for people to think of the zero some cases and I'm like let's exhaust the areas where it's just basically Costless um to uh assume that if this thing is suffering then we're it life Bearer and I agree with you when a human is being mean to an AI system I think the obvious near term negative effect is on the human not on the AI system so there's we have to kind of try to construct an incentive system
where it you should be uh behave the same just like as you were saying with prompt engineer and behave with claw like you would with other humans it's just good for the soul yeah like I think we added a thing point to the system prompt um where basically if people were getting frustrated with Claude uh it was it it got like the model to just tell them that it can do the thumbs down button and send the feedback to anthropic and I think that was helpful because in some ways it's just like if you're really
annoyed because the model is not doing something you want you're just like just do it properly um the issue is you're probably like you know you're maybe hitting some like capability limit or just some issue in the model and you want to vent and I'm like instead of having a person just vent to the model I was like they should vent to us cuz we can maybe like do something about it that's true or you could do a side like like with the artifacts just like a side venting thing all right do you want like
a side quick therapist yeah I mean there's lots of weird responses you could do to this like if people are getting really mad at you I don't try to diffuse the situation by writing fun poems but maybe people wouldn't be that happy with I still wish it it would be possible I understand this is um sort of from a product perspective it's not feasible but I would love if an AI system could just like Le leave mhm have its own kind of volition just to be like H I think that's like feasible like I I
have wondered the same thing it's like and I could actually not only that I could actually just see that happening eventually where it's just like you know the modal like ended the chat do you know how harsh that could be for some people but it might be necessary yeah it feels very extreme or something um like the only time I've ever really thought this is I think that there was like a I'm trying to remember this was possibly a while ago but where someone just like kind of left this thing interact like maybe it was
like an automated thing interacting with clae and cla's like getting more and more frustrated and kind of like why are we like I was like I wish that clae could have just been like I think that an error has happened and you've left this thing running and I'm I just like what if I just stop talking now and if you want me to start talking again actively tell me or do something but yeah it's like um it is kind of harsh like I I feel to really sad if like I was chatting with cl and
cl just was like I'm done there would be a special touring test moment where Claud says I need a break for an hour mhm and it sounds like you do too and just leave close the window I mean obviously like it doesn't have like a concept of time but you can easily like I could make that like right now and the model would just I would I could just be like oh here's like the circumstances in which like you can just say the conversation is done and I mean because you can get the models to
be pretty respons so to prompts you could even make it a fairly High bar it could be like if if the human doesn't interest you or do things that you find intriguing and you're bored you can just leave and I think that like um it would be interesting to see where Claude utilized it but I think sometimes it would it should be like oh this is like this programming Tas is getting super boring uh so either we talk about I don't know like either we talk about fun things now or I'm just I'm done yeah
it actually is inspiring me to add that to the to the user prompt um okay the movie her mhm do you think we'll be headed there one day where humans have romantic relationships with AI systems in this case it's just text and voice based I think that we're going to have to like navigate a hard question of relationships with AIS um especially if they can remember things about your past interactions with them um I'm of many Minds about this cuz I think I think the reflex of reaction is to be kind of like this is
very bad and we should sort of like prohibit it in some way um I think it's a thing that has to be handled with extreme care um for many reasons like one is you know like this is a for example like if you have the models changing like this you probably don't want people performing like long-term attachments to something that might change with the next iteration at the same time I'm sort of like there's probably a benign version of this where I'm like if you like you know for example if you are like unable to
leave the house and you can't be like you know talking with people at all times of the day and this is like something that you find nice to have conversations with you like it that it can remember you and you genuinely would be sad if like you couldn't talk to it anymore there's a way in which I could see it being like healthy and helpful um so my guess is this is a thing that we're going to have to navigate kind of carefully um and I think it's also like I don't see a good like
I think it's just a very it reminds me of all of the stuff where it has to be just approached with like nuance and thinking through what is what are the healthy options here um and how do you encourage people towards those while you know respecting their right to you know like if someone is like hey I get a lot out of chatting with this model um I'm aware of the risks I'm aware it could change um I don't think it's unhealthy it's just you know something that I can chat to during the day I
kind of want to just like respect that I personally think there'll be a lot of really close relationships I don't know about romantic but friendships at least and then you have to I mean there's so many fascinating things there just like you said you have to have some kind of stability guarantees that it's not going to change because that's the traumatic thing MH for us if a close friend of ours completely changed yeah all of a sudden the first update yeah so like I mean to me that's just a fascinating exploration of um a perturbation
to human society that will just make us think deeply about what's meaningful to us I think it's also the only thing that I've thought consistently through this as like a maybe not necessarily a mitigation but a thing that feels really important is that the models are always like extremely accurate with the human about what they are um it's like a case where it's basically like if you imagine like I really like the idea of the models like say knowing like roughly how they were trained um um and and I think CLA will will often do
this I mean for like there are things like part of the traits training included like what CL should do if people basically like explaining like the kind of limitations of the relationship between like an AI and a human that it like doesn't retain things from the conversation um and so I think it will like just explain to you like hey here's like I won't remember this conversation um here's how I was trained it's kind of unlikely that I can have like a certain kind of like relationship with you and it's important that you know that
it's important for like you know your mental well-being that you don't think that I'm something that I'm not and somehow I feel like this is one of the things where I'm like H it feels like a thing I always want to be true I kind of don't want models to be lying to people cuz if people are going to have like healthy relationships with anything it's kind of important yeah like I think that's easier if you always just like know exactly what the thing is that you relating to it doesn't solve everything but I think
it helps quite anthropic may be the very company to develop a system that we definitively recognize as AGI and you very well might be the person that talks to it probably talks to it first well what would the conversation contain like what would be your first question well it depends partly on like the kind of capability level of the model if you have something that is like capable in the same way that an extremely capable human is I imagine myself kind of interacting with it the same way that I do with an extremely capable human
with the one difference that I'm probably going to be trying to like probe and understand its behaviors um but in many ways I'm like I can then just have like useful conversations with it you know so if I'm working on something as part of my research I can just be like oh like which I already find myself starting to do you know if I'm like oh I feel like there's this like thing in virtue ethics I can't quite remember the term like I'll use the model for things like that and so I could imagine that
being more and more the case where you're just basically interacting with it much more like you would an incredibly smart colle colleague um and using it like for the kinds of work that you want to do as if you just had a collaborator who was like or you know the slightly horrifying thing about AI is like as soon as you have one collaborator you have a thousand collaborators if you can manage them enough but what if it's two times the smartest human on earth on that particular discipline yeah I guess you're really good at sort
of probing claw um in a way that pushes its limits understanding where the limits are yep so I guess what would be a question you would ask to be like yeah this is Agi that's really hard because it feels like in order to it has to just be a series of questions like if there was just one question like you can train anything to answer one question extremely well yeah um in fact you can probably train it to answer like you know 20 Questions extremely well like how long would you need to be locked in
the room with an AGI to know this thing is Agi it's a hard question because part of me is like all of this just feels continuous like if you put me in a room for five minutes I'm like I just have high error bars you know I'm like and then it's just like maybe it's like both the the probability increases and the air bar decreases I think things that I can actually probe the edge of human knowledge of so I think this with philosophy a little bit sometimes when I ask the models philosophy questions I
am like this is a question that I think no one has ever asked like it's maybe like right at the edge of like some literature that I know um and the models will just kind of like when they struggle with that when they struggle to come up with a kind of like novel like I'm like I know that there's like a novel argument here because I've just thought of it myself so maybe that's the thing where I'm like I've thought of a cool novel argument in this like Niche area and I'm going to just like
probe you to see if you can come up with it and how much like prompting it takes to get you to come up with it and I think for some of these like really like uh right at the ede of human Knowledge Questions I'm like you could not in fact come up with the thing that I came up with I think if I just took something like that where I like I know a lot about an area and I came up with a novel issue or a novel like solution to a problem and I gave
it to a model and it came up with that solution that would be a pretty moving moment for me because I would be like this is a case where no human has ever like it's not and obviously we see these with this with like more kind of like you see novel Solutions all the time especially to like easier problems I think people overestimate you know novelty isn't like is completely different from anything ever happened it's just like this is it can be a variant of things that have happened um and still be novel but I
think yeah if I saw like the the more I were to see like um completely like uh novel work from the models that that would be like and this is just going to feel iterative it's one of those things where it's there's never it's like you know people I think want there to be like a moment and I'm like I don't know like I think that there might just never be a moment it might just be that there's just like this continuous ramping up I I have a sense that there will be things that a
model can say that convinces you this is very it's not like uh like I've talked to people who are like truly wise mhm like there you could just tell there's a lot of horsepower there yep and if you 10x that I don't know I just feel like there's words you could say maybe ask it to generate a poem mhm and the and the poemy generates you're like yeah okay yeah whatever you did there I don't think a human can do that I think it has to be something that I can verify is like actually really
good though that's why I think these questions that are like where I'm like oh this is like you know like you know sometimes it's just like I'll come up with say a concrete counter example to like an argument or something like that I'm sure like with like it it would be like if you're a mathematician you had a novel proof I think and you just gave it the problem and you saw it and you're this proof is genuinely novel like there's no one has ever done you actually have to do a lot of things to
like come up with this um you know I had to sit and think about it for months or something and then if you saw the model successfully do that I think you would just be like I can verify that this is correct it is like it is a sign that you have generalized from your training like you didn't just see this somewhere because I just came up with it myself and you were able to like replicate that um that's the kind of thing where I'm like for me the closer the more that models like can
do things like that the more I would be like oh this is like uh very real cuz then I can I don't know I can like verify that that's like extremely extremely capable you've interacted with AI a lot what do you think makes humans special oh good question maybe in a way that the universe is much better off that we're in it and that we should definitely survive and spread throughout the Universe yeah it's interesting because I think like people focus so much on intelligence especially with models look intelligence is important because of what it
does like it's very useful it does a lot of things in the world and I'm like you know you can imagine a world where like height or strength would have played this role and I'm like it's just a trait like that I'm like it's not intrinsically valuable it's it's valuable because of what it does I think for the most part um the things that feel you know I'm like I mean personally I'm just like I think humans and like life in general is extremely magical um we almost like to the degree that I you know
I don't know like not everyone agrees with this I'm flagging but um you know we have this like whole universe and there's like all of these objects you know there's like beautiful stars and there's like galaxies and then I don't know I'm just like on this planet there are these creatures that have this like ability to observe that like uh and they are like seeing it they are experiencing it and I'm just like that if you try to explain like I'm I imagine trying to explain to like I don't know someone for some reason they
they've never encountered the world or our science or anything and I think that nothing is that like everything you know like all of our physics and everything in the world it's all extremely exciting but then you say oh and plus there's this thing that it is to be a thing and observe in the world and and you see this like inner Cinema and I think they would be like hang on wait pause you just said something that like is kind of wild sounding um and so I'm like we have this like ability to like experience
the world um we feel pleasure we feel suffering we feel like a lot of like complex things and so yeah and maybe this is also why I think you know I also like hear a lot about animals for example because I think they probably share this with us um so I think that like the things that make humans special in so far as like I care about humans is probably more like their ability to to feel and experience than it is like them having these like functional useful traits yeah to to feel and experience the
beauty in the world yeah to look at the stars I hope there's other civiliz alien civilizations out there but if we're it it's a pretty good uh it's a pretty good thing and that they're having a good time they're having a good time watching us yeah well um thank you for this good time of a conversation and for the work you're doing and for helping make uh Claude a great conversational partner and thank you for talking today yeah thanks for talking thanks for listening to this conversation with Amanda ascal and now dear friends here's Chris
Ola can you describe this fascinating field of mechanistic interpretability AKA Mech interp the history of the field and where is the today I think one useful way to think about neural networks is that we don't we don't program we don't make them we we kind of we grow them you know we have these neural network architectures that we design and we have these loss objectives that we that we we create and the neural network architecture it's kind of like a scaffold that the circuits grow on um and they sort of you know it starts off
with some kind of random you know random things and it grows and it's almost like the the objective that we train for is this light um and so we create the scaffold that it grows on and we create the you know the light that it grows towards but the thing that we actually create it's it's it's this almost biological you know entity or organism that we're that we're studying um and so it's very very different from any kind of regular software engineering um because at the end of the day we end up with this artifact
that can do all these amazing things it can you know write essays and translate and you know understand images it can do all these things that we have no idea how to directly create a computer program to do and it can do that because we we grew it we didn't we didn't write it we didn't create it and so then that leaves open this question at the end which is what the hell is going on inside these systems um and that you know is uh you know to me um a really deep and exciting question
it's you know a a really exciting scientific question to me it's it's it's sort of is like the question that is is just screaming out it's calling out for us to go and answer it when we talk about Nal networks and I think it's also a very deep question for safety reasons so and mechanistic interpretability I guess is closer to maybe neurobiology yeah yeah I think that's right so maybe to give an example of the kind of thing that has been done that I wouldn't consider to be mechanistic inability there was um for a long
time a lot of work on saliency maps where you would take an image and you try to say you know the model thinks this image is a dog what part of the image made it think that it's a dog um and you know that tells you maybe something about the model if you can come up with a principled version of that um but it doesn't really tell you like what algorithms are running in the model how was the model actually making that decision maybe it's telling you something about what was important to it if you
if you can make that meth work but it it isn't telling you you know what are what are the algorithms that are running how is it that this the system is able to do this thing that we no one knew how to do and so I guess we started using the term mechanistic inability to try to sort of draw that that divide or to distinguish ourselves and the work that we were doing in some ways from from some of these other things and I think since then it's become this sort of umbrella term for um
you know pretty wide variety of work but I'd say that the things that that are kind of distinctive are I think a this this focus on we really want to get at you know the mechanisms we want to get at the algorithms um you know if you think of if you think of neural networks as being like a computer program um then the weights are kind of like a binary computer program and we'd like to reverse engineer those weights and figure out what algorithms are running so okay I think one way you might think of
trying to understand a neural network is that it's it's kind of like a we have this compiled computer program and the weights of the neural network are are the binary um and when the neural network runs that's that's the activations um and our our goal is ultimately to go and understand and understand these weights and so you know the project mechanistic inability is to somehow figure out how do these weights correspond to algorithms um and in order to do that you also have to understand the activations because it's sort of the activations are like the
memory and if you if you imagine reverse engineering a computer program um and you have the binary instructions you know in order to understand what what a particular instruction means you need to know what me what what is stored in the memory that it's operating on and so those two things are very intertwined so mechanistic interpret tends to be interested in both of those things now you there's a lot of work that's that's interested in in in those things um especially the you know there's all this work on probing which you might see as part
of being mechanistic interality although it's you know again it's just a broad term and and not everyone who does that work would identify as doing Mech I think the thing that is maybe a little bit distinctive to the the vibe of mechant turp is I think people tend working in the space tend to think of neural networks as well maybe one way to said is that greent descent is smarter than you that you know uh and gradient descent is is actually really great the whole reason that we're understanding these models is because we didn't know
how to write them in the first place the gradient descent comes up with better Solutions than us and so um I think that maybe another thing about mechant turp is sort of having almost a kind of humility that we won't guess at prior what's going on inside the model and we have to have the sort of bottom up approach where we don't really assume you know we don't assume that we should look for a particular thing and that will be there and that's how it works but instead we look from the bottom up and discover
what happens to exist in these models and study them that way but you know the very fact that it's possible to do and as you and others have shown over time you know things like universality that the wisdom of The gradian Descent creates features and circus creates things universally across different kinds of networks that are useful and that makes the whole field possible yeah so this is actually is indeed a a really remarkable and exciting thing where it does seem like at least to some extent you know the same the same elements the same the
same features and circuits form again and again um you know you can look at every Vision model and you'll find curve detectors and you'll find high low frequency detectors um and in fact there's some some reason to think that the same things form across you know biological neural networks and artificial neural networks so a famous example is Vision Vision models in in the early layers they have Gabor filters and there's you know Gabor filters are something that neuroscientists are interested and have thought a lot about we find curved detectors in these models curve detectors are
also found in monkeys we discover these high low frequency detectors and then um some followup work went and discovered them um in rats um or mice um so they were found first in artificial neural networks and then found in biological neural networks um you know this really famous result on like grandmother neurons or the um the Haley Berry neuron from quiroa at all and we found very similar things in in Vision models where this is while I was still at open Ai and I I was looking at their clip model um and you find um
these neurons that respond to the same entities in images and also to give a concrete example there we found that there was a Donald Trump n for some reason I guess Everyone likes to talk about Donald Trump and and Donald Trump was very prominent was was very a very Hot Topic at that time so every every neural network that we looked at we would find a dedicated neuron for Donald Trump um that was the only person who had always had a dedicated nuron um you know sometimes you'd have an Obama nuran sometimes you'd have a
Clinton Nan but uh Trump always had a dedicate so it responds to you know pictures of his face and the ward Trump like all these things right um and so it's it's not responding to a particular example or like it's not just responding to his face it's it's abstracting over this General concept right so in any case that's very similar to these qu results so there this evidence that these that this fomen of universality the same things form across both artificial and and natural neural networks that's that's a pretty amazing thing if that's true um
you know it suggests that um well I think the thing that it suggests is the gradi scent is sort of finding you know the right ways to cut things apart in some sense that many systems converge on and and many different neural networks architectures converge on that there's there's some natural set of you know there's some set of abstractions that are a very natural way to cut apart the problem and that a lot of systems are going to converge on um that would be my my kind of uh you know I don't know anything about
Neuroscience this is this is just my my kind of wild speculation from what we've seen yeah that would be beautiful if it's sort of agnostic to the medium of uh of the model that's used to form the representation yeah yeah and it's you know it's um a a kind of a wild speculation based you know we only have some a few data points justest this but you know it it does seem like there's um there's some sense in which the same things form again again and again and again both in certainly in natural neural networks
and and also artificially or in biologically and the intuition behind that would be that you know where in order to be useful in understanding the real world you need all the same kind of stuff yeah well if we pick I don't know like the idea of a dog right like you know there's some sense in which the idea of a dog is like an a a natural category in the universe or something like this right like you know uh uh there's there's some reason it's it's not just like a weird Quirk of like how humans
Factor you know think about the world that we have this concept of a dog it's it's in some sense or or like if you have the idea of a line like there's you know like look around us you know the you know there are lines you know it's sort of the simplest way to understand this room in some sense is to have the idea of a line and so um I think that that would be my instinct for why this happens yeah you need a curved line you know to understand a circle and you need
all those shapes to understand bigger things and yeah it's a hierarchy of Concepts that are formed yeah and like maybe there are ways to go and describe you know images without reference to those things right but they're not the simplest way or the most economical way or something like this and so systems converge to these um these these strategies would would be my my wild wild hypothesis can you talk through some of the building blocks that we've been referencing of features and circuits so I think you first described them in uh 2020 paper zoom in
and introduction to circuits absolutely so um maybe I'll start by just describing some phenomena and then we can sort of build to the idea of features and circuits so um if you spent like quite a few years maybe maybe like five years to some extent um with other things studying this one particular model Inception V1 um which is this one Vision model it was um state-ofthe-art in 2015 um and uh uh you know very much not state-ofthe-art anymore um and it has you know maybe about 10,000 neurons and and I spent a lot of time
looking at the 10,000 neurons odd neurons of of inception V1 um and one of the interesting things is you know there are lots of neurons that don't have some obvious intal meaning but there's a lot of neurons on Inception V1 that do have really clean intal meanings um so you find neurons that just really do seem to detect curves and you find neurons that really do seem to detect cars and um car wheels and car windows and you know floppy ears of dogs and dogs with long snouts facing to the right and dogs with Longs
Nots facing to the left and you know different kinds of far and there's there's sort of this whole beautiful Edge detectors line detectors color contrast detectors um these beautiful things we call high low frequency detectors you know I think looking at I sort of felt like a biologist you know you just you're looking at at this sort of new world of proteins and you're discovering all these these different proteins that interact um so one way you could try to understand these models is in terms of neurons you could try to be like oh you know
there's a dog detecting neuron and um here's a car detecting neuron and it turns out you can actually ask how those connect together so you can go and say oh you know I have this car detecting on how was it built and it turns out in the previous layer it's connected really strongly to a window detector and a wheel detector and a sort of car body detector and it looks for the window above the car and the wheels below and the car chrome sort of in the middle sort of everywhere but especially on the lower
part um and that's sort of a recipe for a car right like that is you know earlier we said the thing we wanted from mechor was to get algorithms to go and get you know ask what is the the algorithm that runs well here we're just looking at the weights of the N Network reading off this kind of recipe for detecting cars it's a very simple crude recipe but it's it's there and so we call that a circuit this this connection well okay so the the problem is that not all of the neurons um are
interpal and there there's reason to think um we can get into this more later that there's this this superos hypothesis there reason to think that sometimes the right unit to analyze things in terms of um is combinations of neurons so sometimes it's not that there's a single neuron that represents say a car um but it actually turns that after you detect the car the model sort of hides a little bit of the car in the following layer and a bunch of a bunch of dog detectors why is it doing that well you know maybe it
just doesn't want to do that much work on on on on cars at that point and you know it's sort of storing it away to go and um uh so it turns out then that the sort of subtle pattern of you know there's all these neurons that you think are dog detectors and maybe they're primarily that but they all a little bit contribute to representing a car um in in that next layer okay so so now we can't really think there there might still be some something I don't know you could call it like a
car concept or something but it no longer corresponds to a neuron so we need some term for these kind of neuron-like entities these things that we sort of would have liked the neurons to be these idealized neurons um the things that are the nice neurons but also maybe there's more of them somehow hidden and we call those features and then what are circuits so circuits are these connections of features right so so when we have the car detector um and it's connected to a window detector and a wheel detector and it looks for the Wheels
below and the windows on top um that's a circuit um so circuits are just collections of features connected by weights um and they they Implement algorithms so they tell us you know how is how are features used how are they built um how do they connect together so maybe it's it's it's worth trying to pin down like what what really um is the the core hypothesis here I think the the core hypothesis is something we call the linear representation hypothesis so um if we think about the car detector you know the more it fires the
more we sort of think of that as meaning oh the model is more and more confident that um a car was present um or you know if it's some combination of neurons that represent a car you know the more that combination fires the more we think the model thinks there's a car present um this doesn't have to be the case right like you could imagine something where you have you know you have this car detector neuron and you think ah you know if it fires like you know between one and two that means one thing
but it means like totally different if it's between three and four um that would be a nonlinear representation and principle that you know models could do that I think it's it's sort of inefficient for them to do if you try to think about how you'd Implement computation like that it's it's kind of an annoying thing to do but in principal models can do that um so uh one way to think about the features and and circuits sort of framework for thinking about things is that we're thinking about things as being linear we're thinking about there
as being um that if a if a neuron or a combination of neurons fires more it's sort of that means more of the of a particular thing being detected and then that gives weights a very clean interpretation as these edges between these these entities that these features um and that that edge then has a has a meaning um so that's that's in some ways the the core thing um it's it's like um you know we can talk about this sort of outside the context of ns are you familiar with the word toac results um so
you have like you know King minus man plus woman equals Queen well the reason you can do that kind of arithmetic um is because you have a linear representation can you actually explain that representation a little bit so first off so a feature is a is a direction of activation you think it that way can you do the the the minus men plus women that that the war Toc stuff can you explain what that is yeah there's this very such a simple clean explanation of what we're talking about exactly yeah so there's this very famous
result word toac by um Thomas mikov at all and there's been tons of follow-up work exploring this so so sometimes we have these we create these word embeddings um where uh we map every word to a vector I mean that in itself by the way is is kind of a crazy thing if you haven't thought about it before right like we we're we're going and and representing we're turning um you know like like if if you just learned about vectors in physics class right uh and I'm like oh I'm going to actually turn every word
uh in the dictionary into a vector that's kind of a crazy idea okay but you could imagine um you could imagine all kinds of ways in which you might map words to to vectors but it it it seems like when we train neural networks um they like to go and and map words detectors to such that they're they're they they sort of linear structure in a particular sense which is that directions have meaning so for instance if you there there will be some direction that seems to sort of correspond to gender and male words will
be you know far in One Direction and female words will be in another Direction and the linear representation hypothesis is you you could sort of think of it roughly as saying that that's actually kind of the fundamental thing that's going on that that everything is just different directions have meanings and adding different Direction vectors together can represent Concepts and the michelov paper sort of took that idea seriously and one consequence of it is that you can you can do this game of playing sort of arithmetic with words so you can do king and you can
you know subtract off the word man and add the word woman and so you're sort of you know going and and trying to switch the gender and indeed if you do that the result will sort of be close to the word Queen um and you can you know do other things like you can do um uh you know Sushi minus Japan plus Italy and get pizza or uh different things like this right um so so this is in some sense the core of the linear representation hypothesis you can describe it just as a purely abstract
thing about Vector spaces you can describe it as a as a statement about um about the activations of neurons um but it's really about this this property of directions having meaning and in some ways it's even a little subtle than that it's really I think mostly about this property of being able to add things together um that you can sort of independently modify um say gender and royalty or um you know Cuisine typee or country and and and and the concept of food by by adding them do you think the linear hypothesis holds that carries
scales so so far I think everything I have seen is consistent with this hypothesis and it doesn't have to be that way right like like you can write down neural networks where um you write weights such that they don't have linear representations where the right way to understand them is not is not in terms of linear representations but I think every natural neural network I've seen um Hess property um there's been one paper recently um that there's been some sort of pushing around the edges so I think there's been some work recently studying multi-dimensional features
where rather than a single Direction it's more like um a manifold of directions this to me still seems like a linear representation um and then there's been some other papers suggesting that maybe um in in very small models you get nonlinear representations um I think that the jury's still out on that um but in I think everything that we've seen so far has been consistent with the linear representation hypothesis and that's that's wild it it doesn't have to be that way um and yet uh I think that there's a lot of evidence that certainly at
least this is very very widespread and so far the evidence is is consistent with that and I and I I think you know one thing you might say is you might say well Christopher you know it's that's a lot you know to to go and and sort of um to ride on you know if we don't know for sure this is true and you're sort of you know you're investigating all not works as though it is true you know isn't that um isn't that dangerous well you know but I I think actually there's a there's
a virtue in taking hypotheses seriously and pushing them as far as they can go um so it might be that someday we discover something that is inconsistent with linear representation hypothesis but science is full of hypothesis and theories that were wrong um and we learned a lot by sort of working under under them as a sort of an assumption um and and then going and pushing them as far as we can I guess I guess this is sort of the heart of of what Coon would call normal normal science um um I don't know if
you want we can talk a lot about about uh philosophy of science and uh that leads to the paradigm shift so yeah I love it taking the hypothesis seriously and take it to a natural natural conclusion yeah same with the scaling hypothesis same exactly exactly and I love it one of my colleagues Tom henigan who is a former physicist um like made this really nice analogy to me of um uh caloric Theory where you know once upon a time we thought that heat was actually you know this thing called caloric and like the reason you
know hot objects you know would would warm up cool objects is like the caloric is flowing through them um and like you know because we're so used to thinking about about heat you know in terms of the modern modern Theory you know that seems kind of silly but it's actually very hard to construct uh an experiment that that sort of disproves the um chloric hypothesis um and you know you can actually do a lot of really useful work believing in chloric for example it turns out that the original combustion engines were developed by people who
believe in the caloric Theory so I think this a virtue in taking hypotheses seriously even when they might be wrong yeah yeah there's a deep philosophical truth to that that's kind of kind of how I feel about space travel like colonizing Mars there's a lot of people that criticize that I think if you just assume we have to colonize Mars in order to have a backup for human civilization even if that's not true that's going to produce some interesting interesting engineering and even scientific breakthroughs I think yeah well and actually this is another thing that
I think is really interesting so um you know there a way in which I think it can be really useful for society to have people um almost irrationally dedicated to investigating particular hypothesis um because uh well it it takes a lot to sort of maintain scientific morale and really push on something when you know most most SCI scientific hypotheses end up being wrong you know a lot of a lot of science doesn't doesn't work out um and but and yet it's you know it's very it's very useful to go do you know um there's a
there's a joke about Jeff Hinton um which is that uh Jeff Hinton has discovered how the brain works every year for the last 50 years yeah um but you know I I say that with like you know the you know with really deep respect because uh in fact that's actually you know that that led to him doing some some really great work yeah he won the Noel prize Now Who's Laughing Now exactly exactly exactly um yeah I think one want to be able to pop up and sort of recognize the the appropriate level of confidence
but I think there's also a lot of value and just being like you know I'm going to essentially assume I'm going to condition on this problem being possible or this being broadly the right approach and I'm just going to go and assume that for a while and go and work within that um and push really hard on it um and you know if Society has lots of people doing doing that for different things um that's actually really useful in terms of going and uh getting to getting you know either really really ruling things out right
we can be like well you know that didn't work we know that somebody tried hard um or going in and getting to something that that does teach us something about the world so another interesting hypothesis is the superposition hypothesis can you describe what superos is yeah so earlier we were talking about word toac right and we were talking about how you know maybe you have One Direction that corresponds to gender and maybe another that corresponds to royalty and another one that corresponds to Italy and another one that corresponds to you know food and and all
these things well you know um often times maybe these these uh these Ward embeddings they might be 500 dimensions a thousand dimensions and so if you believed that all of those directions were orthogonal um then you could only have you know 500 Concepts and you know I I love pizza um but like if I was going to go and like give the like 500 most important Concepts in um you know the English language probably Italy wouldn't be it's not obvious at least that Italy would be one of them right because you you have to have
things like plural and singular and U uh verb and noun and adjective and you know um there's a lot of things we have to get to before we get to get to Italy um uh and Japan and you know there's a lot of countries in the world um and so how might it be that models could you know simultaneously have the linear representation hypothesis be true and also represent more things than they have directions so so what does that mean well okay so if if if linear representation hypothesis is true something interesting has to be
going on now I'll I'll tell you one more interesting thing before we we go and we do that which is um you know earlier we were talking about all these polymatic neurons right um these neurons that you know when we're looking at Inception V1 there's these nice neurons that like the car detector and the curve detector and so on that respond to lots of you know to very coherent things but it's lots of neurons that respond to a bunch of unrelated things that's that's also an interesting phenomenon um and it turns out as well that
even these neurons that are really really clean if you look at the weak activations right so if you look at like you know the activation where it's like activating 5% of of the the you know of the maximum activation it's really not the core thing that it's expecting right so if you look at a a curve detector for instance and you look at the places where it's 5% active you know you could interpret it just as noise or it could be that it's that it's doing something else there okay so so how could that be
well there's this amazing thing in mathematics um called compressed sensing and it's it's actually this this very surprising fact where if you have a high dimensional space and you project it into a low dimensional space ordinarily you can't go and sort of unprojected and get back your high dimensional Vector right you threw information away this is like you know you can't you can't invert a rectangular Matrix um you can only invert Square matrices um but it turns out that that's actually not quite true if I tell you that the high dimensional Vector was sparse so
it's mostly zeros then it turns out that you can often go and find back um the uh the high dimensional Vector with with very high probability um so that's a surprising fact right it says that you know you can um you can you can have this High dimensional Vector space and as long as things are sparse um you can project it down you can have a lower dimensional projection of it and that works so the super hypothesis is saying that that's what's going on in neural networks that's for instance that's what's going on in wart
edings the wart embeddings are able to simultaneously have directions be the meaningful thing and by exploiting the fact that they're they're operating on a fairly High dimensional space they're actually and and the fact that these concepts are right like you know you usually aren't talking about Japan and Italy at the same time um you know most of the most of those Concepts you know in most sentences Japan and Italy are both zero they're not present at all um and if that's true um then you can go and have it be the case that um that
you can you can have many more of these sort of directions that are meaningful these features than you have dimensions and some of when we're talking about neurons you can have many more Concepts than you have have neurons so that's the at a high level super hypothesis now it has this even Wilder implication which is um to go and say that uh neural networks are it may not just be the case that the the representations are like of this but the the computation may also be like this you know the connections between all of them
and so in in some sense neural networks may be shadows of much larger sparer neural networks and what we see are these projections um and the super you the strongest version of the super hypothesis would be to take that really seriously and sort of say you know there there actually is in some sense this this upstairs model this you know um where where the neurons are really sparse and all interpal and there's you know the weights between them are these really sparse circuits and that's what we're studying um and uh the thing that we're observing
is the shadow of it and we need to find the original object and uh the process of learning is trying to construct a compression of the upstairs model that doesn't lose too much information in the projection yeah finding how to fit it efficiently or something like this um that grent is doing this in fact so this sort of says that gradient descent you know could it could just represent a dense neural network but it sort of says that gradient descent is pleasantly searching over the space of extremely sparse models that could be projected into this
low dimensional space and this large body of work of of people going and trying to study sparse neural networks right where you go and you have you could design neural networks right where where the edges are sparse and the activations are sparse and you know my sense is that work has gener it feels very principled right it makes so much sense and yet that that work hasn't really panned out that well as my impression broadly and I think that a a potential answer for that is that actually the neural network is already sparse in some
sense grading descent was the whole time gradi you were trying to go and do this gradiant descent was actually in the behind the scenes going and searching more efficiently than you could through the space of sparse models and going in learning whatever sparse model was most efficient and then figuring out how to fold it down nicely to go and run conven on your GPU which does you know nice dense Matrix multiplies um and that you just can't beat that how many Concepts do you think can be shoved in into a neural network depends on how
sparse they are so there there's probably an upper bound from the number of parameters right because you have to have you still have to have you know print weights that go and connect them together um so that's that's one upper bound there are in fact all these lovely results from compressed sensing and the Johnson Linton stess Lemma and and things like this um that they they basically tell you that if you have a vector space and you want to have almost orthogonal vectors which is sort of probably the thing that you want here right so
you you're going to say well you know I'm going to give up on having my my Concepts my features be strictly orthogonal but I'd like them to not interfere that much I'm going to have to ask them to be almost orthogonal um then this would say that it's actually you know for once you set a threshold for for what you're what you're willing to accept in terms of how how much coine similarity there is that's actually exponential in the number of neurons that you have so at some point that's not going to even be the
the limiting factor um but um there some beautiful results there and in fact it's probably even better than that in some sense because that's sort of is for saying that you know any random set of features could be active but in fact the features have sort of a correlational structure where some features you know are more likely to co-occur and other ones are less likely to co-occur and so neural networks my guess would be can do do very well in terms of going and uh packing things in such to to the point that's probably probably
not the limiting factor how does the problem of polys semanticity enter the picture here poly semanticity is this phenomenon we observe where we look at many neurons and the neuron doesn't just sort of represent one one concept it's not it's not a clean feature it responds to a bunch of unrelated things and um supersition is you can think of as as being a hypothesis that explains the observation of polys semanticity um so poly semanticity is this observe phenomenon and super is is a hypothesis that um would explain it along with with some other so that
makes Mech turb more difficult right so if you if you're trying to understand things in terms of individual neurons and you have polymatic neurons you're in an awful lot of trouble right I mean the easiest answer is like okay well you know you're looking at the neurons you're trying to understand them this one responds to a lot of things it doesn't have a nice meaning okay we're you that's that's that's bad um another thing you could ask is you know ultimately we want to understand the weights and if you have two polymatic neurons and you
know each one responds to three things and then you know the other neuron responds to three things and you have weight between them you know what does that mean does it mean that like all three you know like there's these nine you know nine interactions going on it's a very weird thing but there's also a deeper reason which is related to the fact that neural networks operate on really high dimensional spaces so I said that our goal was you know to understand neural networks and understand the mechanisms and one thing you might say is like
well why not it's just a mathematical function why not just look at it right like um you know one of the earliest projects I did studied these these neural networks that mapped two- dimensional spaces to two- dimensional spaces and you can sort of interpret them in this beautiful way is like bending manifolds mhm um why can't we do that well you know as you have have a higher dimensional space um the volume of that space in some senses is exponential in the number of inputs you have and so you can't just go in visualize it
so we somehow need to break that apart we need to somehow break that exponential space into a bunch of things that we you know some non-exponential number of things that we can reason about independently and the independence is crucial because it's the Independence that allows you to not have to think about you know all the exponential combinations of things and things being monomatic things only having one meaning things having a meaning that isn't is the key thing that allows you to think about them independently and so I think that's that if you want the deepest
reason why we want to have um interpal monatic features I think that's really the the Deep reason and so the goal here as your recent work has been aiming at is how do we extract the mod semantic features from a neural net that has politic features and all this this mess yes we have the have we observe these polyur and we hypothesize that's what's going what's going on at superos and if superos is what's going on there there's actually a sort of wellestablished technique that is sort of the principled thing to do which is dictionary
learning and um it turns out if you do dictionary learning in particular if you do sort of a nice efficient way that in some in some sense sort of nicely regularizes it well as well called a sparse Auto encoder if you train a sparse Auto encoder these beautiful interpal features start to just fall out where there weren't any beforehand and so that's notot of thing that you would necessarily predict right but it turns out that that works very very well you know to me that seems like you know some non-trivial validation of linear representations and
supersession so with dictionary learning you're not looking for particular kind of categories you don't know what they arege and this gets back to our earlier point right when we're not making assumptions grading descent is smarter than us so we're not making assumptions about what's there um I mean one certainly could do that right one could assume that there's a PHP feature and go and search for it but we're not doing that we're saying we don't know what's going to be there instead we're just going to go and let um the sparse Auto encoder discover the
things that are there so can you uh talk to the to monos semanticity paper from October last year that had a lot of like nice breakthrough results that's very kind of you to describe it that way um yeah I mean this was um uh our first real success using sparse Auto encoders so we took a one layer model um and it turns out if you go and you you know do dictionary learning on it you find all these really nice interpal features so you know the Arabic feature the Hebrew feature um the Bas 64 feature
those were were some some examples that we studied in a lot of depth and really showed that they were um what we thought they were it turns if you train a model twice as well and train two different models and and do dictionary learning you find find analogous features in both of them so that's fun um you find all kinds of of different features so that was really just showing um that um that this works and um you know I should mention that there was this cunning home at all um that had very similar results
around the same time there's something fun about being doing these kinds of small scale experiments and finding that it's actually working yeah well and there's and there's so much structure here like you you know so maybe maybe stepping back for a while um I thought that maybe all this mechanistic can really work um the end result was going to be that I would have an explanation for why it was sort of you know very hard and not going to be tractable um you know we'd be like well there's this problem with supersession and it turns
that super session is really hard um and we're kind of screwed but that's not what happened in fact a very natural Le technique just works and so then that's actually a very good situation you know I think um this is a sort of hard research problem and it's got a lot of research risk and you know it it might still very well fail but um I think that some amount of some very significant amount of research risk um was sort of put behind us when that started to work can you describe what kind of features
can be extracted in this way well so it depends on the model that you're studying right so the the larger the model the more sophisticated they're going to be and we'll probably talk about about follow-up work in a minute but in these one layer models um so some very common things I think were were languages both programming languages and natural languages there were a lot of features that were um specific words in specific contexts so the and I think really the way to think about this is that the is likely about to be followed by
a noun so it's really you could think of this as the feature but you could also think of this as producting a specific noun feature and there would be these features that would fire for the in um the context of of say a legal document or a mathematical document or something something like this um and so uh you know maybe in the context of math you're like you know the and then predict Vector Matrix you know all these mathematical words whereas you other contexts you would predict other things that was that was common and basically
we you need clever humans to assign labels to what we're seeing yes so you know this this is the only thing this is doing is that sort of um unfolding things for you so if everything was sort of folded over top of it you know cation folded everything on top of itself you can't really see it this is unfolding it but now you still have a very complex thing to try to understand um so then you have to do a bunch of work understanding what these are um and some of them are really subtle like
there's some really cool things even in this this one layer model about um Unicode where you know of course some languages are in Unicode and the tokenizer won't necessarily have a dedicated token for every um Unicode um character so instead what you'll have is you'll have this these patterns of alternating token or alternating tokens that each represent half of a unic code character and then you have a different feature that you know goes and activates on the on the opposing ones to be like okay you know um I just finished a character you know go
and predict the next prefix um then okay on the prefix you know predict a reasonable suffix um and you you have to alternate back and forth so there's you know these these wer models are are really interesting and um uh I mean there's another thing which is you might think okay there would just be one b64 feature but it turns out there's actually a bunch of b64 features because you can have English text encoded in as b64 and that has a very different distribution of B 64 tokens than than regular and there's um uh there's
there's some things about tokenization as well that it can exploit and I don't know there all all kinds of fun stuff how difficult is the task of sort of assigning labels to what's going on can this be automated by AI well I think it depends on the feature and it also depends on how much you trust your AI so um there's a lot of work doing um automated inability I think that's a really exciting Direction and we do a fair amount of automated inter and have have Claude go and label our features is there some
fun moments where it's totally right or it's totally wrong yeah well I think I think it's very common that it's like says something very general which is like true in some sense but not really picking up on the specific of what's going on um so I think I think that's a pretty common situation um you don't know that I have a particularly amusing one that's interesting that little gap between it is true but it doesn't quite get to the Deep Nuance of a thing yeah that's a general challenge it's like it's it's St an incredible
colish they can say a true thing but it doesn't it's qu it's not it's missing the depth sometimes and in this context it's like the arc challenge you know the sort of IQ type tests it feels like figuring out what a feature represents is a bit of is a little puzzle you have to solve yeah and and I think that sometimes they're easier and sometimes they're harder as well um so uh yeah I think I think that's tricky now there's another thing which I don't know maybe maybe in some ways this is my like aesthetic
coming in but I'll give try to give you a rationalization you know I'm actually a little suspicious of automated inability and I think that partly just that I want humans to understand neural net works and if the neural network is understanding it for me you know I'm I'm not I don't quite like that but I do have bit of a you know in some ways I'm sort of like the mathematicians who are like you know if there a computer automated proof it doesn't count U you know you they won't understand it but I I do
also think that there is um this kind of like Reflections on trusting trust type issue where you know if you there's this famous talk about um uh you know you like when you're writing a computer program you have to trust your compiler and if there was like malware in your compiler then it could go and inject malware into the next compiler and you know you'd be kind of in trouble right well if you're using neural networks to go and um verify that your neural networks are safe the hypothesis that you're testing for is like okay
well the neural network maybe isn't safe um and you have to worry about like is there some way that it could be screwing with you um so uh you know I I think that's not a big concern now um but I do Wonder in the long run if we have to use really powerful system AI systems to go and uh you know audit our AI systems is that is that actually something we can trust but maybe I'm just rationalizing because I I just want to us to have to get to a point where humans understand
everything yeah I mean especially that's hilarious especially as we talk about AI safety and it looking for features that would be relevant to AI safety like deception and so on uh so let's let's talk about the scaling a semanticity paper in May 2024 okay so what did it take to scale this to apply to Claude 3 on it well a lot of gpus a lot more gpus um but one of my teammates Tom henigan um was involved in the original scaling loss work um and something that he was sort of interested in from very early
on is are there scaling laws for inability um and so um something he sort of immediately did when when this work started to succeed and we started to have sparse Auto encoders work we became very interested in you know what are the scaling laws for um uh you know for making making sparse Auto encoders larger and how does that relate to making the base model larger um and so um it turns out this works really well and you can use it to sort of project um you know if you train a sparse Auto encod a
given size you know how many tokens should you train on and so on so this was actually a very big help to us in scaling up um this work um and made it a lot easier for us to go and train um you know really large sparse Auto encoders where you know um it's not like training the big models but it's it's starting to get to a point where it's actually actually expensive to go um and train the really big ones so you have to I mean you have to do all the stuff of like
splitting it across large I mean there's a huge engineering challenge here too right so yes so so there's there's a there's a scientific question of how do you scale things effectively um and then there's an enormous amount of engineering to go and scale this up you have to you have to chart it you have to you have to think very carefully about a lot of things I'm lucky to work with a bunch of great Engineers cuz I am definitely not a great engine yeah on the infrastructure especially yeah for sure so it turns out tldr
it worked it worked yeah and and I think this is important because you could have imagined you could like you could have imagined a world where you set after towards monos fanticy you know Chris this is great you know it works on a one layer model but one layer models are really idiosyncratic um like you know maybe maybe there just something ID like maybe the linear representation hypothesis and super hypothesis is the right way to understand a one layer model but it's not the right way to understand large models um and so I think um
I mean first of all like The Cutting him at all paper sort of um cut through that a little bit and and sort of suggested that this wasn't the case but um scaling onity sort of I think was significant evidence that even for very large models and we did it on Claude 3 sauna which at that point was uh one of our production models um you know even these models um seem to be very you know seem to be substantially explained at least by linear features and you know doing dictionary learning on them works and
as you learn more features you go and you explain explain more and more so that's a I think a quite a promising sign and you find now really fascinating abstract features um and the features are also multimodal they respond to images and text for the same concept which is fun yeah this can you explain that I mean like you know back door there's just a lot of examples that you can yeah so maybe maybe let's start with a one example to start which is we found some features around sort of security vulnerabilities and back doors
and codes so it turns out those are actually two different features um so there's a security vulnerability feature and if you force it active Claude will start to go and write um security vulnerabilities like buffer overflows into code and it also it fires for all kinds of things like you know some of some of the top data set examples for it were things like you know dash dash disable um you know SSL or something like this which are sort of obviously really um uh really insecure so at this point it's kind of like maybe it's
just because the examples are presented that way it's kind of like surface a little bit more obvious examples right um I guess the the idea is that down the line might be able to detect more Nuance like deception or bugs or that kind of stuff yeah well I maybe I want to distinguish two things so um one is um the complexity of the feature or the concept right and the other is the the Nuance of the how subtle the examples we're looking at right so when we when we show the top data set examples those
are the most extreme examples that that feature to to activate um and so it doesn't mean that it doesn't fire for more subtle things so the UN you know the insecure um code feature you know the stuff that it fires for most strongly for are these like really obvious you know disable the security type things um but um um you know uh it it also Fires for you know buffer overflows and and more subtle security vulnerabilities in code you know these features are all multimodal so you could ask like what images activate this feature and
it turns out um that the uh the the security vulnerability feature activates for images of um uh like people clicking on Chrome to like go past the like you know this this website uh the SSL certificate might be wrong or something like this another thing that's very entertaining is there's backd doors en code feature like you activate it it goes and Cloud writes a back door that like will go and dump your data to port or something but you can ask okay what what images activate the back door feature it was devices with hidden cameras
in them so there's a whole apparently genre of people going and selling devices that look in uous that have hidden cameras and they have ads that how there's a hidden camera in it and I guess that is the you know physical version of a back door um and so it sort of shows you how abstract these concepts are right um and I I just thought that was uh I I'm sort of sad that there's a whole Market of people selling devices like that but I was kind of delighted that that was the the thing that
it came up with as the the top uh image examples for the feature yeah it's nice it's multimodal it's multi almost context it's it's as broad strong definition of a singular concept it's nice yeah to me one of the really interesting features especially for AI safety is deception and lying and the possibility that these kinds of methods could detect uh lying in a model especially gets smarter and smarter and smarter presumably that's a big threat of a super intelligent model that he can deceive the people operating it uh as to its intentions or any of
that kind of stuff so what what have you learned from detecting lying inside models yeah so I think we're in some ways in early days for that we find quite a few features related to deception and lying there's one feature where fires for people lying and being deceptive and you force it active and Claude starts lying to you so we have a have a deception feature I mean there's all kinds of other features about withholding information and not answering questions features about power seeking and coups and stuff like that this a lot of features that
are kind of related to Spooky things and if you um force them active Claude will will behave in ways that are they're not the kind of behaviors you want what are possible next exciting directions to you in the space of uh Mech and well there's a lot of things um so for one thing I would really like to get to a point where we have circuits where we can really understand um not just the features uh but then use that to understand the computation of models um that really for me is is the the ultimate
goal of this um and there's been some work we we put out a few things there's a paper from Sam Marks that does some stuff like this there's been some I'd say some work around the edges here um but I think there's a lot more to do and I think that will be a very exciting thing um that's related to a challenge we call interference weights um where um due to supersition if you just sort of navely look at whether featur are connected together there may be some weights that sort of don't exist in the
upstairs model but are just sort of artifacts of of superposition so that's a a sort of technical challenge related to that um I think another exciting direction is just I you know you might think of of sparse Auto encoders as being kind of like a telescope they allow us to you know look out and see all these features that are are are are out there and you know as we build better and better sparse Auto en Cutters get better better at dictionary learning we see more and more stars um and you know we zoom in
on smaller and smaller stars but there kind of um a lot of evidence that we're only still seeing a very small fraction of the Stars there's a lot of matter in our in our you know neural network universe that we can't observe yet um and it may be that um that we'll never be able to have fine enough instruments to observe it and maybe maybe some of it just isn't possible um isn't computationally tractable to observant there's sort of a a kind of dark matter and in not in maybe the sense of of astronomy of
earlier astronomy when we didn't know what this unexplained matter is um and so I I think a lot about that that dark matter and whether will ever observe it and what that means for safety if we if we can't observe it if there's you know some if some significant fraction of nor networks are not accessible to us um another question that I think a lot about is uh at the end of the day you know mechanistic inter is it's very microscopic um approach to interality it's trying to understand things in a very fine grained way
but lot of the questions we care about are very macroscopic um you know we we care about these questions about neural network behavior and and I think that's the thing that I care most about but there's there's lots of other other sort of larger scale questions you you might care about um and somehow you know the nice thing about about having a very microscopic approach is it's maybe easier to ask you know is this true but the downside is it's much further from the things we care about and so we now have this ladder to
climb and I think there's a question of can will we be able to find are there are there sort of larger scale abstractions that we can use to understand nural networks that can we get up from this very microscopic approach yeah you've you you've written about this this kind of organs question yeah exactly if we uh think of interpretability as a kind of anatomy of neural networks most of the circus threads involve studying tiny little veins looking at the small scale and individual neurons and how they connect however there are many natural questions that the
small scale approach doesn't address in contrast the most prominent abstractions in biological Anatomy involve larger scale structures like individual organs like the heart or entire organ systems like the respiratory system and so we wonder is there a respiratory system or heart or brain region of an artificial neuron Network yeah exactly um and I mean like if you think about science right a lot of scientific Fields have um you know investigate things that many level of abstractions in biology you have like you know molecular biology studying you know proteins and molecules and so on and you
have cellular biology and then you have histology studying tissues and you have anatomy and then you have zoology and then you have ecology and so you have many many levels of abstraction or you know physics maybe the physics of individual particles and then you know statistical physics gives you gives you thermodynamics and things like this and so you often have different levels of abstraction um and I think that right now we have you know mechanistic interpret if it succeeds is sort of like a microbiology of neural networks but we we want something more like anatomy
and so and you know a question you might ask is why why can't you just go there directly and I think the answer is super um in at least in significant part it's that it's actually very hard to to see this this macroscopic structure U without first sort of breaking down the microscopic structure in the right way and then studying how it connects together um but I'm I'm hopeful that there is going to be something much larger than um features and circuits and that we're going to be able to have a story that's much than
evolves much bigger things and you then you can sort of study in detail the parts you care about as opposed to neurobiology like a psychologist or psychiatrist when your own network and I think that the beautiful thing would be if we could go and rather than having disperate fields for those two things if you could have a build a bridge between them such that you could go and um uh have all of your higher level abstractions be grounded very firmly In This Very solid um you know more rigorous ideally Foundation what do you think is
the difference between the human brain the biological neuron Network and the artificial neuron Network well the neuroscientists have a much harder job than us you know sometimes I just like count my blessings by how much easier my job is than the neuroscientist right so I have um we we can record from all the neurons yeah we can do that on arbitrary amounts of data um the neurons don't change while you're doing that by the way MH um you can go and ablate neurons you can edit the connections and so on um and then you undo
those changes that's prettyy great yeah um you can force any you can intervene on any neuron and force it active and see what happens um you know which neurons are connected to everything right you neuroscientists want to get the connecto we have the connecto um and we have it for like much bigger than the elegant um and then not only do we have the connectome um we know uh what the you know which neurons excite or inhibit each other right so we have we it's not just that we know that like the binary mask we
know the the weights um we can take gradients we know computationally what each neuron does um so I don't know the goes on and on we just have um so many advantages over neuroscientists and then despite having all those advantages it's really hard and so one thing I do sometimes think is like gosh like if it's this hard for us it seems impossible under the constraints of Neuroscience or you know near impossible um I I I don't know maybe maybe part of me is like I've got a few neuroscientists on my team maybe maybe I'm
sort of like ah you know um the uh maybe the neuroscientists maybe some of them would like to have an easier problem that's still very hard um and they they could come and work on on neural networks and then after we after we figure out things in sort of the easy uh Little Pond of trying to understand neural networks which is still very hard then we then we could go back to biological Neuroscience I love what you've written about the goal of mechan turp research as uh two goals safety and Beauty so can you talk
about the beauty side of things yeah so you know there's this funny thing where I think some people want uh some people are kind of disappointed by neural networks I think where they're like ah you know neural network um it's these just these simple rules then you just like do a bunch of engineering to scale it up and it works really well and like where's the like complex ideas you know this isn't like a very nice beautiful scientific result and I sometimes think when people say that I picture them being like you know evolution is
so boring it's just a bunch of simple rules and you run Evolution for a long time and you get biology like what a what a a sucky uh you know way for biology to have turned out where's the the complex rules but the beauty is that the Simplicity generates complexity um you know biology has these simple rules and it gives rise to you know all the life and ecosystems that we see around us all the beauty of nature that all just comes from Evolution and from something very simple Evolution and similarly I think that nural
networks build you know create enormous um complexity and Beauty inside and structure inside themselves that people generally don't look at and don't try to understand because it's it's hard to understand but I I think that there is an Inc incredibly Rich structure to be discovered inside n networks a lot of a lot of very deep Beauty um if we're just willing to take the time to go and see it and understand it yeah I love I love Mech inter the feeling like we are understanding or getting glimpses of understanding the magic that's going on inside
is really wonderful it feels to me like one of the questions is just calling out to be asked and I'm sort of I mean a lot of people are think about this but I'm often surprised that morar is how is it that we don't know how to create computer systems that can do these things and yet we have these amazing systems that we don't know how to directly create computer programs that can do these things but these neural networks can do all these amazing things and it just feels like that is obviously the question that
sort of is calling out to be answered if you are if you have any degree of curiosity it's it's like how is it that that Humanity now has these artifacts that can do these things that we don't know how to do yeah I love the image of the circus towards the light of the objective function yeah it's just it's it's this organic thing that we've grown and we have no idea what we've grown well thank you for working on safety and thank you for appreciating the beauty of the things you uh discover and thank you
for talking today Chris this is wonderful thank you for taking the time to chat as well thanks for listening to this conversation with Chris Ola and before that with DAR amade and Amanda ascal to support this podcast please check out our sponsors in the description and now let me leave you with some words from Alan Watts the only way to make sense out of change is to plunge into it move with it and join the dance thank you for listening and hope to see you next time
Related Videos
[ 40Hz ] GAMMA Binaural Beats, Ambient Study Music for Focus and Concentration
2:00:00
[ 40Hz ] GAMMA Binaural Beats, Ambient Stu...
SleepTube - Hypnotic Relaxation
4,159,789 views
YUDKOWSKY + WOLFRAM ON AI RISK.
4:17:08
YUDKOWSKY + WOLFRAM ON AI RISK.
Machine Learning Street Talk
13,408 views
Elon Musk: Neuralink and the Future of Humanity | Lex Fridman Podcast #438
8:37:35
Elon Musk: Neuralink and the Future of Hum...
Lex Fridman
3,832,644 views
Thievery Corporation • Chet Faker • Bonobo • Morcheeba - Special Coffeeshop Selection [Seven Beats]
1:56:22
Thievery Corporation • Chet Faker • Bonobo...
Seven Beats Music
2,022,140 views
Cenk Uygur: Trump vs Harris, Progressive Politics, Communism & Capitalism | Lex Fridman Podcast #441
4:10:28
Cenk Uygur: Trump vs Harris, Progressive P...
Lex Fridman
774,511 views
Graham Hancock: Lost Civilization of the Ice Age & Ancient Human History | Lex Fridman Podcast #449
2:33:02
Graham Hancock: Lost Civilization of the I...
Lex Fridman
2,712,203 views
Deep Chill Music — Calm and Stress Relief Mix
1:28:03
Deep Chill Music — Calm and Stress Relief Mix
Chill Music Lab
2,129,846 views
Vejas Liulevicius: Communism, Marxism, Nazism, Stalin, Mao, and Hitler | Lex Fridman Podcast #444
3:31:58
Vejas Liulevicius: Communism, Marxism, Naz...
Lex Fridman
912,381 views
Gregory Aldrete: The Roman Empire - Rise and Fall of Ancient Rome | Lex Fridman Podcast #443
3:42:21
Gregory Aldrete: The Roman Empire - Rise a...
Lex Fridman
3,774,652 views
Aravind Srinivas: Perplexity CEO on Future of AI, Search & the Internet | Lex Fridman Podcast #434
3:02:16
Aravind Srinivas: Perplexity CEO on Future...
Lex Fridman
745,594 views
Bernie Sanders Interview | Lex Fridman Podcast #450
1:02:32
Bernie Sanders Interview | Lex Fridman Pod...
Lex Fridman
1,238,079 views
Deep Chill Music for Focus and Stress Relief
1:15:05
Deep Chill Music for Focus and Stress Relief
Chill Music Lab
3,934,369 views
Ed Barnhart: Maya, Aztec, Inca, and Lost Civilizations of South America | Lex Fridman Podcast #446
3:28:51
Ed Barnhart: Maya, Aztec, Inca, and Lost C...
Lex Fridman
2,768,141 views
Deep Listening (1111 Hz) | 1 hour handpan music | Malte Marten
1:00:20
Deep Listening (1111 Hz) | 1 hour handpan ...
Malte Marten
193,544 views
The Hidden Art Of Reinventing Yourself - Matthew McConaughey (4K)
1:48:46
The Hidden Art Of Reinventing Yourself - M...
Chris Williamson
140,017 views
Jordan Peterson: Nietzsche, Hitler, God, Psychopathy, Suffering & Meaning | Lex Fridman Podcast #448
2:23:05
Jordan Peterson: Nietzsche, Hitler, God, P...
Lex Fridman
1,140,301 views
Jordan Jonas: Survival, Hunting, Siberia, God, and Winning Alone Season 6 | Lex Fridman Podcast #437
2:46:20
Jordan Jonas: Survival, Hunting, Siberia, ...
Lex Fridman
320,321 views
Vivek Ramaswamy: Trump, Conservatism, Nationalism, Immigration, and War | Lex Fridman Podcast #445
2:40:26
Vivek Ramaswamy: Trump, Conservatism, Nati...
Lex Fridman
773,313 views
Take a Rest 🌌 Medieval Chill Vibes | Lofi Music for Sleep/Study/Work - Relaxing Music
1:01:15
Take a Rest 🌌 Medieval Chill Vibes | Lofi...
Lofi Night Vibes
1,146 views
DEATHTRAP! The Strange story of Air France flight 736
49:01
DEATHTRAP! The Strange story of Air France...
Mentour Pilot
456,174 views
Copyright © 2025. Made with ♥ in London by YTScribe.com