AI policy researcher, @lfschiavo wife guy, fan of cute animals and sci-fi, Substack writer, stealth-ish non-profit co-founder

San Francisco, CA
Joined February 2013
New blog post (link in next tweet) - "Standards, Incentives, and Evidence: The Frontier AI Governance Triad." I explain the basic building blocks of making sure that frontier AI systems are safe and secure.
7
7
1
102
Trying not to think about how Vince Gilligan said some wrong things about AI + just focus on the fact that all his work is great and his new show is supposed to be great *fingers in ears* la la la la
5
1
14
The actual phasing out part is hard in part due to that earlier misalignment, but there is an info asymmetry btwn companies + users here as well as abundant evidence of harmful uses such that deferring entirely to users’ preferences is, again, predictably causing harm
4
2
1
47
Not sure exactly what he meant ofc but basically my take here is - reasoning models blow away earlier models on basically every safety + capability dimension, and generally we just get better at alignment over time, so old models predictably increase various harms
3
67
Roon was right btw
14
5
1
178
Some wild content over at at @Gossip_Goblin
1
1
11
Why would people want to allow the human experience to be displaced by computers? As a creation of man, AI will not be divorced from the flaws of human nature; indeed, it is more likely to magnify those flaws. This is not safe; it is dangerous.
“In reality, it is not so” is Ilya-coded (complimentary)
The world needs honest and courageous entrepreneurs and communicators who care for the common good. We sometimes hear the saying: “Business is business!” In reality, it is not so. No one is absorbed by an organization to the point of becoming a mere cog or a simple function. Nor can there be true humanism without a critical sense, without the courage to ask questions: "Where are we going? For whom and for what are we working? How are we making the world a better place?"
31
(Antitrust, sensitive IP etc. are sometimes cited as reasons not to do this but IMO these are not the real barriers. Solvable if there is a desire to do do, easy excuses to disengage otherwise)
2
AI companies sometimes cooperate on safety but this is slowed by people disliking each other + some companies’ whole identity being “we’re better than them in XYZ ways.” May be more promising long term for companies to trust their AIs to negotiate with the other companies’ AIs.
4
8
Miles Brundage retweeted
How can we give defenders of critical infrastructure an advantage over attackers? @Miles_Brundage proposes Operation Patchlight: a national effort to use each jump in frontier AI capabilities to find and patch critical open-source code vulnerabilities before these AI capabilities diffuse to attackers. Attacks on critical infrastructure are already a big problem. 61% of US hospitals report ransomware affecting clinical care, with 17% saying it caused serious patient harm. It takes 491 days on average to apply critical security updates to hospital equipment, and cyber incidents in healthcare cost over $15 billion in 2023 alone. But it could get much worse. Future agentic AI systems may constitute advanced persistent threats, scaling the effective workforce of cyberattackers by orders of magnitude. The proposal addresses a market failure. Attackers can make millions from a single ransomware campaign, while open-source code maintainers (whose libraries underpin ~70% of commercially used software) remain chronically under-resourced. The plan: for the US government and AI labs to jointly fund AI-powered vulnerability discovery, plus more funds to give every hospital and power plant administrator an always-on AI security assistant that helps implement patches and security improvements. As AI systems become more capable, both the need for this project will become more acute and its feasibility more apparent, as future AI systems become capable of both cyberattacks and protecting infrastructure against them. Even large investments into AI-powered cyberdefense could prove to be extremely cost-effective for the US government. Read the online essay: ifp.org/operation-patchlight…
🚀The Launch Sequence book debut is in 11 days! Start the countdown: every day until then, I’ll post at least one short summary on each of the ideas in the book. Then we’ll start shipping the books to Congress. More details in-thread (1/3).
If you’re a hotel manager and you decide to take down the sign in the bathroom that says “we care about oceans and stuff. Don’t make us wash your towels too much,” do you get kicked out of the Hotel Illuminati or something?
2
13
Miles Brundage retweeted
I made a list of forecasts of the impact of AI on economic growth over the next decade. A few observations... (🧵): tecunningham.github.io/posts…
Miles Brundage retweeted
The UK’s @AISecurityInst is one of the rare successes of government. I wrote a piece for @Samfr substack on why and what lessons it gives us for how to make government work. There are five key lessons. A thread ⬇️
Miles Brundage retweeted
🚀 Hello, Kimi K2 Thinking! The Open-Source Thinking Agent Model is here. 🔹 SOTA on HLE (44.9%) and BrowseComp (60.2%) 🔹 Executes up to 200 – 300 sequential tool calls without human interference 🔹 Excels in reasoning, agentic search, and coding 🔹 256K context window Built as a thinking agent, K2 Thinking marks our latest efforts in test-time scaling — scaling both thinking tokens and tool-calling turns. K2 Thinking is now live on kimi.com in chat mode, with full agentic mode coming soon. It is also accessible via API. 🔌 API is live: platform.moonshot.ai 🔗 Tech blog: moonshotai.github.io/Kimi-K2… 🔗 Weights & code: huggingface.co/moonshotai
Gov't gets 10%, Trump himself gets least 200 tokens in the Model Spec
2
19
That too
Trump is going to demand half of OpenAI, you know
2
7
(also wasn't the restructure supposed to mean no more complaining from OpenAI about fundraising being hard?)
2
7
The American AI industry has many problems but a lack of eager investors is not one of them
1
7
😬
Replying to @shiringhaffary
Friar also hinted the idea of US gov't taking a role in AI financing, such as a gov't "backstop," or "guarantee," but stopped short of explaining how that would work
1
3
45