Anthropic's Chilling 18-Month Warning: AI Apocalypse in 18 Months

40.13k views3494 WordsCopy TextShare
TheAIGRID
Prepare for AGI with me - https://www.skool.com/postagiprepardness 🐤 Follow Me on Twitter https://...
Video Transcript:
so a crazy article has come out from anthropic that I completely missed basically anthropic is daating that we have 18 months just 18 before something catastrophic happens within AI now currently the date for this article not the date of recording this video was the 31st of October 2024 which means 18 months from today April the 30th 2026 now that is actually a lifetime away when you think about it in terms of AI terms because that is a few months away and essentially a few months in AI we all know how crazy a few months can be so you can see here it basically says that look increasingly powerful AI systems have the potential to accelerate scientific progress unlock new Medical Treatments and of course grow the economy but along with the remarkable new capabilities of these AI come significant risks and governments should urgently take action on AI policy in the next 18 months the window for proactive risk prevention is closing fast basically what they're trying to say here is that the way how air models have been rapidly advancing we might be losing time to get that window down where we can actively Implement guard rails to actually prevent AI getting out of control and if you want to know how crazy AI development just has gone in the past couple of days take a look at this where it actually looks at how we've gone from gpt1 all the way from GPT 4 now to this Innovation area where we're going to be having scaling breakthroughs and even potentially super intelligent gpt1 which was nothing special to uh to 01 which is sort of PhD level in math and computer science and then 01 I think kicks off this new phase of you know we sort of think about as an innovation era which is basically now until super intelligence we'll see if that's 6 years or maybe even a bit less than that um the Hallmark I think here is that you know we're spending $200 billion on the models we probably can't spend a lot more than that you know probably can't spend you know 200 trillion on the models so there's only limited amounts of scaling left in terms of orders of magnitude and so we need corresponding uh Innovations to sort of come alongside uh definitely Advanced reasoning and test time compute uh is one of those uh and we think there's probably a few more handful of others that will get us you know to now the article continues to state that narrowly targeted regulation can allow us to get the best of both worlds realizing the benefits of AI while mitigating the risks and of course dragging our feet might lead to the worst of Both Worlds poorly designed knee-jerk regulation that hampers progress while also failing to be effective at preventing risks basically what they're stating here is that and I do believe this I'm not just reading this off the script I actually genuinely believe this is that if we do this really slowly what's going to happen is that there's going to be knee-jerk regulation which basically just means that something will occur first like the tragedy will occur first and then the regulation will be built around such tragedy so that it cannot occur again and I don't think we want to do regulation like that we would rather prevent the issue from ever happening then for us to wait for a catastrophe and then of course for us to then construct the regulation around that now they also talk about how what's going on inside AI companies is pretty crazy so it says in the last year and this is where they talk about urgency AI systems have grown dramatically better at math graduate level reasoning and computer coding along with many other capabilities and inside AI companies we see continued progress on as yet undisclosed systems and results now essentially what they mean by this is that this means that they are looking at systems that have not been publicly announced or revealed and this could include proprietary AI models or Technologies you know like Advanced AI systems that are in development but kept confidential for competitive ethical or security reasons or they could be talking about breakthrough results which is where they have significant progress or discoveries in AI capabilities that these companies are not ready to share with the public and of course possibly because they are being tested or because was revealing them could pose certain risks so this is where they're stating that these advancers offer many positive applications but progress in these same broad capabilities also brings with it the potential for destructive applications either from the misuse of AI in domains such as cyber security or biology or from The Accidental or autonomous behavior of the system itself now they're basically stating here that like these advances that we're going to see with these future models are offering all of these insane applications but the thing is is that because we're currently building these General models that are able to do quite a lot of different things than they generalized to other categories the problem is is that this progress also goes to those areas for misuse so basically what they're stating here is that you know the broad capabilities also brings with it the potential for destructive applications either from the misuse of AI in domains such as cyber security or biology or from The Accidental or autonomous behavior of the AI system itself so that last line there is a little bit more chilling considering the fact that you know 2025 2026 is going to be the year apparently for advanced AI agents so the craziest thing about this all okay is that they state that in the realm of cyber capabilities models have rapidly Advanced on a broad range of coding tasks and cyber offense evaluations and take a look at this guys because this when I read this I was like whoa this is actually pretty crazy when we actually think about it so it says on the software engineering bench software engineering task models have improved from being able to solve 1. 9 6% of a test set of reward coding problems which was clawed to in October 2023 all the way up to now it can do 40 9% just in basically a year I think because it says claw 2 was October 2023 and then claw 3. 5 Sonic in October 2024 is 49% guys so can we imagine what October 2025 looks like inwe that's probably going to be close to 90% guys and that is going to be pretty in it's like incredible I can't I can't even speak um and it also says internally our Frontier team has found that current models can already assist on a broad range of cyber offense related tasks and we expect that the next generation of models which will be able to plan over long multi-step tasks will be even more effective so that's of course something that we know when you give these models the ability to think for such a long time they're able to come up with a lot more coherent responses and are able to find solutions to problems that you know they otherwise weren't able to so I think I'm not going to lie guys this is going to be something that's really really crazy because if we're seeing that rate of progress on the software engineering progress Benchmark in just one year we can really start to think about okay what happens the year after that the year after that I mean it's really going to get to that point where things get incredible I mean this is why they're literally stating that look within 18 months we're going to have really Advanced AI systems and if we're not able to actually fix that right now and Implement these C rails then things are going to be able to go off the rails so overall what we're looking at is a situation that basically tells us that these systems are going to be even better than these current ones and it's probably going to happen a lot faster now they basically talk about here that they were also looking at the potential for cbrn which is chemical biological radiological and nuclear misuses and the UK AI safety Institute actually tested a range of models from industry actors including anthropic and they concluded that these models can be used to obtain expert level knowledge about biology and chemistry and for several models replies to science questions were on par with those given by PhD level experts so for those of you who are unaware of what I'm saying guys they've got a situation on their hands where these models the problem is is that pretty much every model has some kind of jailbreak I'm not sure how you're going to get these models to not say certain things literally every time a model is released I see someone on Twitter called plyy and he goes model jailbroken model pwned and basically there doesn't seem to be a real proper solution to the issue that is jailbreaking and they're basically saying that look models can be used to obtain expert level knowledge about biology and chemistry and of course the knowledge that they're giving them isn't like gbt 3.
5 level knowledge it isn't just complete jgun it's actually the kind of knowledge that is really good and is given by PhD level expert so when we actually look at where we are now the models that they've been testing we have to think about this in the next 3 to 5 years because these models are going to get so smart that if you have that model available to the average person we know that this is going to be some kind of issue you can see here it says about a year ago we warned that Frontier models might post real risks in the Cyber and cbrn domains within 2 to 3 years based on the progress described above we believe we are now substantially closer to such risks surgical careful regulation will soon be needed so basically what they're stting here is that look when you've got all of this progress the swe bench going so crazy we need to think about the fact that this is going to be some kind of exponential Improvement and we need to be prepared for this kind of risk of course you can see how AI systems have progressed dramatically in their understanding of the Sciences in the last year the widely used Benchmark the GP QA saw scores on its hardest section grow from 38. 8% when it was was released in November 2023 to 59. 4% in June 2024 and to 77.
3% in September which is opening I 01 and the crazy thing about all of this is that human experts score 81.
Related Videos
Bill Gates on possibility, AI, and humanity
1:04:38
Bill Gates on possibility, AI, and humanity
Reid Hoffman
218,802 views
The Thrill and Threat of Mind Hacking | Posthuman with Emily Chang
24:58
The Thrill and Threat of Mind Hacking | Po...
Bloomberg Originals
80,003 views
The Investing & Crypto Expert: "We Only Have 6 Years Until Everything Changes!" - Raoul Pal
2:13:05
The Investing & Crypto Expert: "We Only Ha...
The Diary Of A CEO
2,315,684 views
"Zero Doubt" Non-Human Intelligence on Earth - Col. Karl Nell & Alex Klokus | SALT iConnections NY
21:36
"Zero Doubt" Non-Human Intelligence on Ear...
SALT
1,271,798 views
The Race to Harness Quantum Computing's Mind-Bending Power | The Future With Hannah Fry
24:02
The Race to Harness Quantum Computing's Mi...
Bloomberg Originals
3,338,207 views
the sham legacy of Richard Feynman
2:48:11
the sham legacy of Richard Feynman
Angela Collier
280,428 views
Connor Leahy on Why Humanity Risks Extinction from AGI
1:58:51
Connor Leahy on Why Humanity Risks Extinct...
Future of Life Institute
7,243 views
Vertical AI Agents Could Be 10X Bigger Than SaaS
42:13
Vertical AI Agents Could Be 10X Bigger Tha...
Y Combinator
233,775 views
Mo Gawdat on AI: The Future of AI and How It Will Shape Our World
47:41
Mo Gawdat on AI: The Future of AI and How ...
Mo Gawdat
297,092 views
Eric Schmidt DROPS BOMBSHELL: China DOMINATES AI!
10:06
Eric Schmidt DROPS BOMBSHELL: China DOMINA...
TheAIGRID
78,824 views
The Most Misunderstood Concept in Physics
27:15
The Most Misunderstood Concept in Physics
Veritasium
16,983,010 views
Ex Google CEO: AI Is Creating Deadly Viruses! If We See This, We Must Turn Off AI!
1:49:37
Ex Google CEO: AI Is Creating Deadly Virus...
The Diary Of A CEO
1,518,254 views
Why this top AI guru thinks we might be in extinction level trouble | The InnerView
26:31
Why this top AI guru thinks we might be in...
TRT World
1,071,208 views
The Savings Expert: Are You Under 45? You Won't Get A Pension! Don't Buy A House! - Jaspreet Singh
2:28:47
The Savings Expert: Are You Under 45? You ...
The Diary Of A CEO
2,538,278 views
Dominic Cummings: is AI already in control?
1:47:50
Dominic Cummings: is AI already in control?
The Spectator
64,547 views
Something Strange Happens When You Follow Einstein's Math
37:03
Something Strange Happens When You Follow ...
Veritasium
17,302,200 views
Top Minds in AI Explain What’s Coming After GPT-4o | EP #130
25:30
Top Minds in AI Explain What’s Coming Afte...
Peter H. Diamandis
532,620 views
$100B Founder Breaks Down The Biggest AI Business Opportunities For 2025
1:28:15
$100B Founder Breaks Down The Biggest AI B...
My First Million
78,223 views
Eric Schmidt - The Risks and Opportunities of an AI Future | Prof G Conversations
34:07
Eric Schmidt - The Risks and Opportunities...
The Prof G Show – Scott Galloway
72,648 views
What Is an AI Anyway? | Mustafa Suleyman | TED
22:02
What Is an AI Anyway? | Mustafa Suleyman |...
TED
1,839,484 views
Copyright © 2025. Made with ♥ in London by YTScribe.com