Now we describe how our model works in more detail, including: written descriptions of how the model works, the equations forming the model, the reasoning behind our modeling choices, and how the most important parameters of the model are set. We'll start with how the coding time horizon progression works, then discuss how we model improvements in AI efficiency and how automation affects the model, then describe how we model intelligence explosions. You can use the model diagram or table of contents on the right to navigate. If you'd instead like to see the forecasts produced by our model, go to the forecast page.
The final section Recap of the Model is a succinct summary of the whole model, which some readers may prefer and which can serve as a useful reference. It contains little motivation or explanation, and mainly consists of the formulas.
Now we will describe how our model works in more detail, including: written descriptions of how the model works, the equations forming the model, the reasoning behind our modeling choices, and how the most important parameters of the model are set. We'll start with how the coding time horizon progression works, then discuss how we model improvements in AI efficiency and how automation affects the model, then describe how we model intelligence explosions. If you'd instead like to see the forecasts produced by our model, go to the forecast page.
The final section Recap of the Model is a succinct summary of the whole model, which some readers may prefer and which can serve as a useful reference. It contains little motivation or explanation, and mainly consists of the formulas.
We've significantly upgraded our timelines and takeoff model! We’ve found it useful for clarifying and informing our thinking, and we hope others do too. We plan to continue thinking about these topics, and update our models and forecasts accordingly.
The future is very hard to predict. We don't think this model, or any other model, should be trusted completely. The model takes into account what we think are the most important dynamics and factors, but it doesn't take into account everything. Also, only some of the parameter values in the model are grounded in empirical data; the rest are intuitive guesses. If you disagree with our guesses, you can change them above.
Nevertheless, we think that modeling work is important. Our overall view is the result of weighing many considerations, factors, arguments, etc.; a model is a way to do this transparently and explicitly, as opposed to implicitly and all in our head. By reading about our model, you can come to understand why we have the views we do, what arguments and trends seem most important to us, etc.
The future is uncertain, but we shouldn’t just wait for it to arrive. If we try to predict what will happen, if we pay attention to the trends and extrapolate them, if we build models of the underlying dynamics, then we'll have a better sense of what is likely, and we'll be less unprepared for what happens. We’ll also be able to better incorporate future empirical data into our forecasts.
In fact, the improvements we’ve made to this model as compared to our timelines model at the time of AI 2027 (Apr 2025), have resulted in a roughly 2-4 year shift in our median for full coding automation. The plurality of this has come from improving our modeling of AI R&D automation, but other factors such as revising parameter estimates have also contributed. These modeling improvements have resulted in a larger change in our views than the new empirical evidence that we’ve observed. You can read more about the shift in our blog post.
In our blog post, we give a brief survey of other methods for forecasting AI timelines and takeoff speeds, including other formal models such as FTM and GATE but also simpler methods like revenue extrapolation and intuition: you can read that here.
You can find our forecasts here. These are based on simulating the model lots of times, with each simulation resampling from our parameter distributions. We also display our overall views after adjusting for out-of-model considerations.
With Eli's parameter estimates, the model gives a median of late 2030 for full automation of coding (which we refer to as the Automated Coder milestone or AC), with 19% probability by the end of 2027. With Daniel’s parameter estimates, the median for AC is late 2029.
In our results analysis, we investigate which parameters have the greatest impact on the model’s behavior. We also discuss the correlation between shorter timelines and faster takeoff speeds.
Our blog post discusses how these predictions compare to the predictions of our previous model from AI 2027.
In terms of how how much they affect Eli’s all-things-considered views (i.e. Eli’s overall forecast of what will happen, taking into account factors outside of the model), the top known limitations are:
Not explicitly modeling data as an input to AI progress. Our model implicitly assumes now that any data progress is proportional to algorithmic progress. But data in practice could be either more or less bottlenecking. Along with accounting for unknown unknowns, this limitation pushes Eli’s median timelines back by 2 years.
Not modeling automation of hardware R&D, hardware production, and general economic automation. We aren’t modeling these, and while they have longer lead times than software R&D, a year might be enough for them to make a substantial difference. This limitation makes Eli place more probability on fast takeoffs from AC to ASI than the model does, especially increasing the probability of <3 year takeoffs (from ~43% to ~60%).
Meanwhile, Daniel increases his uncertainty somewhat in both directions in an attempt to account for model limitations and unknown unknowns and makes his takeoff somewhat faster for reasons similar to Eli’s.
A more thorough discussion of this model’s limitations can be found here, and discussion of how they affect our all-things-considered views is here.
We start with a high-level description of how our model can intuitively be divided into 3 stages. The next section contains the actual explanation of the model, and accompanies the arrowed diagram on the right.
Our model's primary output is the trajectory of AIs' abilities to automate and accelerate AI software R&D. But although the same formulas are used in Stages 1, 2, and 3, new dynamics emerge at certain milestones (Automated Coder, Superhuman AI Researcher), and so these milestones delineate natural stages.

Stage 1 predicts when coding in the AGI project will be fully automated. This stage centrally involves extrapolating the METR-HRS coding time horizon study.
Milestone endpoint: Automated Coder (AC). An AC can fully automate an AGI project's coding work, replacing the project’s entire software engineering staff.
The main drivers in Stage 1 are:
Besides coding, we track one other type of skill that is needed to automate AI software R&D: research taste. While automating coding makes an AI project faster at implementing experiments, automating research taste makes the project better at setting research directions, selecting experiments, and learning from experiments.
Stage 2 predicts how quickly we will go from an Automated Coder (AC) to a Superhuman AI researcher (SAR), an AI with research taste matching the top human researcher.
Milestone endpoint: Superhuman AI Researcher (SAR): A SAR can fully automate AI R&D.
The main drivers of how quickly Stage 2 goes is:
Finally we model how quickly AIs are able to self-improve once AI R&D is fully automated and humans are obsolete. The endpoint of Stage 3 is asymptoting at the limits of intelligence.
The primary milestones we track in Stage 3 are:
In our simulations, we see a wide variety of outcomes ranging from a weeks-long takeoff from SAR to ASI, to a fizzling out of the intelligence explosion requiring further increases in compute to get to ASI.
To achieve a fast takeoff, there usually needs to be a feedback loop such that each successive doubling of AI capabilities takes less time than the last. In the fastest takeoffs, this is usually possible via a taste-only singularity, i.e. the doublings would get faster solely from improvements in research taste (and not increases in compute or improvements in coding). Whether a taste-only singularity occurs depends on which of the following dominates:
Continued improvements in coding automation matter less and less, as the project gets bottlenecked by their limited supply of experiment compute.
We've significantly upgraded our timelines and takeoff model! We’ve found it useful for clarifying and informing our thinking, and we hope others do too. We plan to continue thinking about these topics, and update our models and forecasts accordingly.
The future is very hard to predict. We don't think this model, or any other model, should be trusted completely. The model takes into account what we think are the most important dynamics and factors, but it doesn't take into account everything. Also, only some of the parameter values in the model are grounded in empirical data; the rest are intuitive guesses. If you disagree with our guesses, you can change them above.
Nevertheless, we think that modeling work is important. Our overall view is the result of weighing many considerations, factors, arguments, etc.; a model is a way to do this transparently and explicitly, as opposed to implicitly and all in our head. By reading about our model, you can come to understand why we have the views we do, what arguments and trends seem most important to us, etc.
The future is uncertain, but we shouldn’t just wait for it to arrive. If we try to predict what will happen, if we pay attention to the trends and extrapolate them, if we build models of the underlying dynamics, then we'll have a better sense of what is likely, and we'll be less unprepared for what happens. We’ll also be able to better incorporate future empirical data into our forecasts.
In fact, the improvements we’ve made to this model as compared to our timelines model at the time of AI 2027 (Apr 2025), have resulted in a roughly 2-4 year shift in our median for full coding automation. The plurality of this has come from improving our modeling of AI R&D automation, but other factors such as revising parameter estimates have also contributed. These modeling improvements have resulted in a larger change in our views than the new empirical evidence that we’ve observed. You can read more about the shift in our blog post.
In our blog post, we give a brief survey of other methods for forecasting AI timelines and takeoff speeds, including other formal models such as FTM and GATE but also simpler methods like revenue extrapolation and intuition: you can read that here.
You can find our forecasts here. These are based on simulating the model lots of times, with each simulation resampling from our parameter distributions. We also display our overall views after adjusting for out-of-model considerations.
With Eli's parameter estimates, the model gives a median of late 2030 for full automation of coding (which we refer to as the Automated Coder milestone or AC), with 19% probability by the end of 2027. With Daniel’s parameter estimates, the median for AC is late 2029.
In our results analysis, we investigate which parameters have the greatest impact on the model’s behavior. We also discuss the correlation between shorter timelines and faster takeoff speeds.
Our blog post discusses how these predictions compare to the predictions of our previous model from AI 2027.
In terms of how how much they affect Eli’s all-things-considered views (i.e. Eli’s overall forecast of what will happen, taking into account factors outside of the model), the top known limitations are:
Not explicitly modeling data as an input to AI progress. Our model implicitly assumes now that any data progress is proportional to algorithmic progress. But data in practice could be either more or less bottlenecking. Along with accounting for unknown unknowns, this limitation pushes Eli’s median timelines back by 2 years.
Not modeling automation of hardware R&D, hardware production, and general economic automation. We aren’t modeling these, and while they have longer lead times than software R&D, a year might be enough for them to make a substantial difference. This limitation makes Eli place more probability on fast takeoffs from AC to ASI than the model does, especially increasing the probability of <3 year takeoffs (from ~43% to ~60%).
Meanwhile, Daniel increases his uncertainty somewhat in both directions in an attempt to account for model limitations and unknown unknowns and makes his takeoff somewhat faster for reasons similar to Eli’s.
A more thorough discussion of this model’s limitations can be found here, and discussion of how they affect our all-things-considered views is here.
We start with a high-level description of how our model can intuitively be divided into 3 stages. The next section contains the actual explanation of the model, and accompanies the arrowed diagram on the right.
Our model's primary output is the trajectory of AIs' abilities to automate and accelerate AI software R&D. But although the same formulas are used in Stages 1, 2, and 3, new dynamics emerge at certain milestones (Automated Coder, Superhuman AI Researcher), and so these milestones delineate natural stages.

Stage 1 predicts when coding in the AGI project will be fully automated. This stage centrally involves extrapolating the METR-HRS coding time horizon study.
Milestone endpoint: Automated Coder (AC). An AC can fully automate an AGI project's coding work, replacing the project’s entire software engineering staff.
The main drivers in Stage 1 are:
Besides coding, we track one other type of skill that is needed to automate AI software R&D: research taste. While automating coding makes an AI project faster at implementing experiments, automating research taste makes the project better at setting research directions, selecting experiments, and learning from experiments.
Stage 2 predicts how quickly we will go from an Automated Coder (AC) to a Superhuman AI researcher (SAR), an AI with research taste matching the top human researcher.
Milestone endpoint: Superhuman AI Researcher (SAR): A SAR can fully automate AI R&D.
The main drivers of how quickly Stage 2 goes is:
Finally we model how quickly AIs are able to self-improve once AI R&D is fully automated and humans are obsolete. The endpoint of Stage 3 is asymptoting at the limits of intelligence.
The primary milestones we track in Stage 3 are:
In our simulations, we see a wide variety of outcomes ranging from a weeks-long takeoff from SAR to ASI, to a fizzling out of the intelligence explosion requiring further increases in compute to get to ASI.
To achieve a fast takeoff, there usually needs to be a feedback loop such that each successive doubling of AI capabilities takes less time than the last. In the fastest takeoffs, this is usually possible via a taste-only singularity, i.e. the doublings would get faster solely from improvements in research taste (and not increases in compute or improvements in coding). Whether a taste-only singularity occurs depends on which of the following dominates:
Continued improvements in coding automation matter less and less, as the project gets bottlenecked by their limited supply of experiment compute.