Science and speculation
Will we agree about AI risks in time?
Science is a social institution like the judicial system. Just like courts make you throw out perfectly good Bayesian evidence like whether the defendant seems like a shifty guy to you or rumors you heard about his whereabouts,1 the scientific process has epistemically conservative norms about what kinds of evidence and arguments are admissible. This means that the process of science gets to the truth much slower than a perfectly rational Bayesian should, and indeed much slower than many particular farseeing individuals actually did. Svante Arrhenius made the first estimate of global warming from human carbon dioxide emissions back in 1896, probably about eighty years before something like a “scientific consensus” emerged and ninety four years before the first IPCC Assessment Report.2
But they say if you want to go fast go alone, if you want to go far go together. At its best, the point of the scientific process is to add bricks to a solid foundation of facts that we can establish “beyond a reasonable doubt,” that we can all agree on regardless of our priors because the weight of the evidence we’ve built up is enough to overwhelm those differences. This is a good piece of social technology to have. We’re never going to agree on subtle questions of priors, and it’s great to have a process that lets us at least agree about where the giant likelihood ratios lie. And like our adversarial justice system, scientific institutions are to a significant extent designed around the fact that we all have biases and conflicts of interest — personal virtue isn’t as verifiable or scalable as specific norms like preregistration, reproducibility, and peer review.
This is why, when AI Snake Oil (now known as AI As a Normal Technology) wrote AI existential risk probabilities are too unreliable to inform policy two years ago, I was much more sympathetic to it than most people who are as concerned as I am about existential risks from AI. While I strongly disagree with the authors on the object level, my guess is that loosening the norms around how much policymaking should be based on science is much more likely to lead to vaccine bans and other destructive crackpottery than thoughtfully crafted AI safety regulation.3
The arguments about superintelligence and the Singularity are speculative. “Speculative” should not be a pejorative term meaning “obviously wrong and dismissible” (Arrhenius’ arguments about CO2 emissions were speculative), but it does make sense that society draws a normative distinction between speculative and scientifically established claims.
Unfortunately, these conservative norms have a serious chance of getting us killed in the case of AI. The pace of change in the field is vastly outstripping our ability to develop a well-grounded scientific picture of it. Just last year, METR generated a lot of buzz with an RCT showing that early 2025 AI tools actually seemed to make open source developers a little slower at their work tasks. By the time my colleagues started looking at late 2025 AI tools, the experiment design broke because participants started dropping out of the study or changing what they worked on because they were so afraid of being randomized into the no-AI arm for their normal tasks.
Still, the project of scientific research on models’ capabilities and risks doesn’t seem doomed. We could have lived in a world where superintelligence truly came as a bolt from the blue, like a lot of the speculative writing from ten or fifteen years ago contemplated. The world we find ourselves in is extremely far from that. We used to debate questions like whether people would make AI agents at all, or whether they would be connected to the internet, or whether they would be allowed to run on their own without a human in the loop. We know the answers to all those questions now. We have seen thousands of public examples of AI agents lying, cheating, and working around safeguards to try to achieve their task. With every new model there are new researchers who take catastrophic risks seriously. And precisely because we are so far from a bulletproof consensus on anything and the field moves at such a breakneck pace, there is so much grist for research and testing and sensemaking. The scientific fruit is on the floor.
We almost certainly won’t be able to develop an evidence base about AI risks anywhere near as robust as what we have for greenhouse gases causing climate change or tobacco causing lung cancer. But science is a process made out of humans, and a “scientific consensus” is ultimately a set of facts the relevant scientists managed to agree about. This develops through some messy combination of building up the foundation of established facts through shared experience and solid experiments, making logical arguments on top of established facts, arguing with one another to establish the right frames to think about things, observing whose predictions turn out to be more right, and much more. I think we do have a shot at building a real-life scientific consensus robust enough to motivate serious technical standards before it’s too late. AI is going to keep improving at a breakneck pace, and in some ways that makes this an easier problem than tobacco or climate change. Scientific consensus, like everything else, could be dragged along at the speed of AI.
See ChatGPT on hearsay.
See ChatGPT on the history of climate change, especially the table at the end.
The same goes for courts — there are many cases where courts make the wrong object-level judgment, including many where the conservative epistemic norms of the system are directly responsible for the error. But the alternative to a conservative, adversarial judicial system is probably not a speedy and reliable Bayesian justice system but rather more latitude for the state to punish people who are politically inconvenient.


> I think we do have a shot at building a real-life scientific consensus robust enough to motivate serious technical standards before it’s too late.
That's the big question, isn't it? How large do you think the shot is?
Conversely, what is your unconditional forecast of human extinction before 2040?
I just watched The AI Doc last night with my sister who is visiting from across the country. It was my third viewing of the film. She liked the film and thought it was informative for someone like her who doesn't know much about AI. But I got the sense that even after watching it she is still part of the 99+% of Americans who wouldn't name AI when asked what is the most important problem facing America today.
I thought I had a decent understanding of why most Americans are concerned about AI, but very few name AI as *the most important* problem, but after this latest watch-through I'm doubting whether I understand why this is the case.
Is it just the lack of scientific consensus about the risks? AI risk has been a mainstream topic for three years now since Hinton left Google and began speaking out, so people have heard how large many leading experts think the risk is. Do they just remain skeptical that the risk is anywhere near as high as those experts say due to the lack of consensus?
Stuart Russell frequently points out in interviews that the annual risk of human extinction humanity is on track to take on with the development of AI over the next decade or two is about a *million times* higher than is acceptable given acceptable risk thresholds for meltdowns in new nuclear power plants that national and international regulatory bodies have set. Even if most Americans intuitively only partially believe e.g. Hinton and Bengio's 10-50% AI existential risk estimates (say by putting only 10% weight on their estimates), this still results in a 1-5% AI risk estimate, which I would think is more than high enough to make AI the most important problem facing the country today. Yet 99+% of people don't think AI is the most important problem and I'm not sure why. And I'm not sure what would lead them to start thinking it is. And I'm skeptical that we will get the political will necessary for sufficiently "serious technical standards" and regulation to get created and passed until a lot more people start considering AI risk to be the most important problem facing society today.
So this is why I think that people like you clearly voicing what your unconditional forecast of human extinction before 2040 is might be useful -- it communicates the gravity of the situation in a way that I would think would cause many people to update to the view that AI is the most important problem, even if they only partially believe you. But I could easily be wrong--maybe this is not at all what is needed for people to update. Maybe nothing short of scientific consensus will be enough.