Discussion about this post

User's avatar
Jelle Donders's avatar

The effort to engage with the ideas is still appreciated, but I think this largely argues with strawmen of AI risk arguments. Would love to respond if I had more time.

Two quick things:

1) Recursive self improvement only becomes a dominant force once the last few human bottlenecks are automated.

2) Agentic AI that takes the human out of the loop will outcompete safe, responsible systems that don't. An AGI CEO or AGI military decision making system would decimate adversaries with humans in charge. Also, at some point we'll stop understanding why an AI comes to their conclusion. Even if it explains it to us, we wouldn't vet for it. So even if AI stays in an advisory role, companies and countries effectively still have to do what the advisor tells them or their adversaries will outcompete them.

Expand full comment
Matt's avatar

Love this whole piece. It strikes at the core of what ultimately drove me away from LessWrong style thinking after many years steeped in it (including hosting an ACX meetup).

Software engineering is a process of find bug, debug it, uncover next bug, debug it, etc. Hard takeoff is based on the idea that we will suddenly break out of that paradigm and AI will self-debug better than humans can do. That simply doesn't resemble the way software development has ever worked or will ever work. We get faster and faster at debugging, we handle and abstract away bug-prone patterns, but there's always another bug to be hit when venturing into completely unexplored territory.

I would extend the Waymo analogy to note that the pattern of thinking around hard takeoff right now is falling into the exact same trap we once fell into with self-driving cars: looking at the rate of change for the first (and easiest) 80% of the problem and assuming it will continue into the last 20%, where the nastiest edge cases lie.

Before we get anything resembling truly human-level agentic AGI we will probably go through that same process, where we gradually hit more and more exotic edge cases. But as with self-driving cars, it only takes one edge case to make the whole system spin out of control in a way that makes it unusable. That's a major impediment for current-gen agents, and while I think we'll whittle away at the problem with each advance in capability, the idea that we'll hit some unexpectedly critical threshold and it'll abruptly disappear is a form of magical thinking.

Expand full comment
61 more comments...

No posts