Helping the world prepare for extremely powerful AI @open_phil (views my own), writer and editor of Planned Obsolescence newsletter.

Berkeley, CA
Joined October 2017
My take: 1. We don't have an AI agent that can fully automate R&D. 2. We could soon. 3. This agent would have *enormously* bigger impacts than AI products have had so far. 4. This doesn't require a "paradigm shift," just the same corporate R&D that took us from GPT-2 to o3.
beginning to dawn on me that we’ll probably never see a day when the first AGI is revealed. it won’t be like the moon landing or even the announcement of the iphone. just a steady climb up the capability ladder, with no consensus on where the threshold lies or when we cross it
Ajeya Cotra retweeted
Yeah "no growth in per capita income until the Industrial Revolution" reflects Malthusian population dynamics not a lack of technological progress — human capabilities grew steadily, which was reflected in population growth.
I realize this view of history is extremely common these days, but I think ppl underestimate the technological and cultural progress between 100k ya and the rise of cities 5k ya, and again from there to the height of the Bronze Age, and then to the classical era, and so forth.
9
13
1
180
Ajeya Cotra retweeted
Suppose it became clear that AIs were very capable and egregiously misaligned (e.g., AIs have repeatedly tried to escape and then create bioweapons). What should we (companies/goverments/etc) do? I'm most interested in the views of regulation/government-involvement skeptics.
Ajeya Cotra retweeted
🌱⚠️ weeds-ey but important milestone ⚠️🌱 This is a first concrete example of the kind of analysis, reporting, and accountability that we’re aiming for as part of our Responsible Scaling Policy commitments on misalignment.
Where is my LLM Google Doc integration that makes comments and tracks in suggestions? Can @snewmanpv come back for one last job or something?
3
1
1
37
Ajeya Cotra retweeted
Economist a right to point out that Baumol's effect will be very important when AI can automate some tasks but not others. But when AI+robots fully automates ALL physical and cognitive tasks, you can massively increase output on all tasks needed for manufacturing and technological progress. And so you can quickly develop huge numbers of incredibly high-tech robots. There are questions about how much bottlenecks and gathering data and producing new machinery and mining will slow this process down. But the best analyzes I've seen suggest that even with today's physical technology, we could double the amount of physical capital every year or quicker. The model that Seb cites assumes that AI will never automate all tasks! See this post on the industrial explosion for more. x.com/TomDavidsonX/status/19…
Here's a great paper by Nobel winner Philippe Aghion (and Benjamin F. Jones and Charles I. Jones) on AI and economic growth. The key takeaway is that because of Baumol's cost disease, even if 99% of the economy is fully automated and infinitely productive, the overall growth rate will be dragged down and determined by the progress we can make in that final 1% of essential, difficult tasks. And this logic still applies *even* in a world with AGIs that can automate *every* task a human can do. In this world, the "hard to improve" tasks would no longer be human-centric ones, but physics-centric ones. The economy's growth rate stops being a function of how fast/well the AGI can "think" and starts being a function of how fast it can manipulate the physical world. Essentially, post-AGI does not necessarily mean post-scarcity: the entire cost and value of the economy becomes concentrated in the physically constrained tasks: generating energy, mining resources, manufacturing goods, transportation and so on. nber.org/system/files/workin…
1
3
27
Ajeya Cotra retweeted
Re "is AI slowing down" - For sure the edge has been taken off the AGI-by-2027 predictions... but claims that progress is stalling also seem pretty off to me. I continue to expect AI will get a lot better over the next 5-15 years, and that will be a wild ride. More:
We're in a strange spot right now with AI. The anti-AI crowd believes progress has halted and are doing a victory lap. Insiders at all labs maintain advancement continues at pace. Only one of these versions of reality will survive the new year. Gemini 3 is very close now.
Ajeya Cotra retweeted
I don’t have much to add to the bubble discussion, but the “this time is different” argument is, in part, based on the sincere belief of many at the AI labs that there is a race to superintelligence & the winner gets,.. everything. It is a key dynamic that is not discussed much
This tweet is unavailable
Ajeya Cotra retweeted
1. This is more about their perceived timelines than whether they're AGI-pilled (clearly yes) 2. What matters re: valuations is perceptions relative to the market. I thought the market was slow to recognize AI potential before. Not sure if erring in the opposite direction now.
I promise you that 'openai is secretly not agi-pilled' is a bad take if you believe it, I'd be excited to take the opposite side from you in a wide variety of financial transactions
2
3
91
Ajeya Cotra retweeted
Introducing MALT: a dataset of natural and prompted transcripts where agents behave in ways that threaten evaluation integrity (like generalized reward hacking & sandbagging) on HCAST & RE-Bench. For general use, we're releasing a split of MALT samples collected on public tasks.
TBC, Noah's analysis does assume that companies don't develop true drop-in replacements for human workers before their loans come due, which many bulls at those companies disagree with. If loans come due in the next 2y though, I think they're likely too optimistic.
9
I'm interested in more detailed work on how much profit AI companies will capture and whether there will be a bust in the next 2y. Appreciated the emphasis that "market crash" != "low-impact technology"
This tweet is unavailable
4
5
108
Ajeya Cotra retweeted
I'm just one person, and my programming needs are somewhat unusual (building various kinds of statistical forecasting models). But I'm just not seeing the consistent productivity gains from LLMs that I would have expected if you'd asked me 6 months ago.
Ajeya Cotra retweeted
wild to to see this from the dallas fed
Ajeya Cotra retweeted
The Thinking Machine Tinker API was a moderate update towards thinking it would be possible to secure models against insider threat, because it might allow a lot of ML research to be done without direct model access.
2
1
26
If you're a generalist EA who's working on AI safety or policy because it's the most important problem, I'd consider switching to biosecurity. AI is more important but bio seems a lot more neglected and there's straightforward object level work that could help a lot.
Convention wisdom is that bioweapons are humanity's greatest weakness – 100x cheaper to make than to defend against. Andrew Snyder-Beattie thinks conventional wisdom is likely wrong. He has a plan cheap enough to do without government. Useful even in worst case scenarios like mirror bacteria. Effective enough to save most people. In one of my all-time fav interviews he lays out a low-tech 4-step approach developed by his research team at Open Philanthropy, to fix a problem most have thought unsolvable. ASB is hiring for many roles in this project from logistics to biotech to manufacturing, and has $100s millions to deploy. Enjoy, links below! 2:10 How bad it could get 9:19 The worst-case scenario: mirror bacteria 18:14 Why low-tech 25:30 Prevention 31:21 The “4 pillars” plan 33:09 ASB is hiring now to make this happen 35:11 Everyone was wrong: biorisks are defence dominant 40:23 Pillar 1: Lungs 55:53 Pillar 2: Biohardening 1:15:19 Pillar 3: Detection 1:28:40 Pillar 4: The wrench hypothesis 1:40:12 The plan's biggest weaknesses 1:44:44 Would chaos make this impossible to pull off? 1:51:50 Would rogue AI make bioweapons? 1:57:57 We can feed the world even if all the plants die 2:07:03 Could a bioweapon make the Earth uninhabitable? 2:09:35 What ASB is hiring for 2:30:27 How to protect yourself and your family (On the 80,000 Hours Podcast, available anywhere you get podcasts.)
8
18
2
253
Ajeya Cotra retweeted
I mostly agree with this post: AI isn't plateauing, trend extrapolation is useful, and substantial economic impacts seem soon. However, trends don't imply huge economic impacts in 2026 and naive extrapolations suggest full automation of software engineering is ~5 years away.
As a researcher at a frontier lab I’m often surprised by how unaware of current AI progress public discussions are. I wrote a post to summarize studies of recent progress, and what we should expect in the next 1-2 years: julian.ac/blog/2025/09/27/fa…
6
10
1
150
Ajeya Cotra retweeted
Some new theoretical economics papers looking at the implications of AGI. These two papers argue that a true AGI-level AI (equivalent to a human genius), if achieved, would eventually displace most human labor and reduce the economic value of remaining human work to near-zero.
Ajeya Cotra retweeted
I'm excited that, this year, interpretability finally works well enough to be practically useful in the real world! We found that, with enough effort into dataset construction, simple linear probes are cheap, real-time, token level hallucination detectors and beat baselines
Imagine if ChatGPT highlighted every word it wasn't sure about. We built a streaming hallucination detector that flags hallucinations in real-time.
22
117
6
1,634
Ajeya Cotra retweeted
We’re hosting a historic hackathon with @METR_Evals, inspired by their latest paper that measured the real-world impact of AI coding tools. Here's how it works: 🤖 Half of participants will build with AI tools 👩‍💻 Half of participants will build without AI tools Judging is blind and scored on speed, precision, and quality. $15K prize + credits from @openai, @anthropicai, and @raindrop_ai. Speakers and Judges include @swyx, @bfioca, @shyamalanadkat, @lioronai, and more. 📍 SF | Sept 6 | In-person only | Application in comments Who will win: Man or Machine? Comment / RT with you pick for a chance to win $1000 Factory credits.