Cross-posted to LessWrong https://www.lesswrong.com/posts/teD4xjwoeWc4LyRAD/what-role-should-evolutionary-analogies-play-in
This is based on a research project that I’ve been working on under the supervision of Dr. Vael Gates. This work is funded by the Long-Term Future Fund. I would like to thank Vael Gates for their support and supervision, and Tilman Räuker for their valuable feedback.
Epistemic status: exploratory, fairly high uncertainty. My background is in experimental physics, and I only have limited experience with biology or AI.
The motivation for this project was a question from Technical AGI Safety Research Outside AI by Richard Ngo.
What are the most compelling arguments for and against discontinuous versus continuous takeoffs? In particular, how should we think about the analogy from human evolution, and the scalability of intelligence with compute?
In short, it seems that many people are quite confused about the use of evolution as an analogy in forecasting AI takeoff speeds (if not, then I certainly am!) and I hope to clarify things through this post. In particular, I’m focusing mostly on “how should we think about the analogy from human evolution”, although I think this itself is a big enough question that I haven’t given a complete answer to it. I’ve made an effort to accurately represent existing arguments and stay up to date with new ones, but at the end of the day these are still my interpretations of them. Comments and feedback are very welcome, and all mistakes are my own.
Table of Contents
- 0 Overview
- 1 Evolutionary arguments for takeoff speeds
- 1.1 Hominid variation
- 1.2 Changing selection pressures
- 1.3 Cumulative cultural evolution
- 1.4 Brains and compute
- 1.5 Biological anchors
- 1.6 One algorithm and Whole Brain Emulation
- 1.7 Anthropic effects
- 1.8 Drivers of intelligence explosions
- 1.9 Summary
- 2 Evaluating analogies and disanalogies
- 2.1 A non-exhaustive list of analogies and disanalogies
- Analogy: Timescales
- Analogy: Necessary compute
- Analogy: Cumulative cultural evolution
- Analogy: Intelligence is non-hard
- Analogy: Thresholds
- Analogy: Number of neurons
- Analogy: Structure of the learning system
- Disanalogy: Training environments
- Disanalogy: Optimisation objectives
- Disanalogy: Changing selection pressures
- Disanalogy: Data constraints
- Disanalogy: Energy costs
- Disanalogy: Drivers of intelligence explosions
- 2.2 Discussion
- 2.3 Summary
- 2.1 A non-exhaustive list of analogies and disanalogies
- 3 Further work
- 4 Conclusions
- People use evolution as both an analogy and a disanalogy in evolutionary arguments for TAI takeoff speeds
- Many of these analogies and disanalogies are potentially quite weak and could be investigated further
- We can think about these analogies and disanalogies in terms of a framework in three levels, based on Steven Byrnes’ post:
- the “outer algorithm” level: level of an algorithm that searches for an “inner algorithm”, like evolution
- the “inner algorithm” level: the level of the “human intelligence algorithm”
- effects level: analogies/disanalogies to the effects of the inner+outer algorithm
- Many uncertainties remain, like how large of a role these analogies/disanalogies should play, and how we can define appropriate measures of growth for informed forecasts and AI governance
- I think we should look a little more into the foundations of these analogies, e.g. evolution of modularity, macroevolution, threshold effects and traits, etc.
Many arguments for AI takeoff speeds are based on analogies to evolution. Given that evolution has given rise to general intelligence, this seems like a reasonable thing to do. But how far can we extend this analogy? What are the important analogies that make this a valid comparison, and where does the analogy break down? This seems like an important question to answer, especially if many of our forecasts are based on evolutionary priors.
In writing this post, I hope to get us a few steps closer to answering these questions. The first section focuses on clarifying the current state of the debate around AI takeoff speeds - the most common arguments, and my personal thoughts.
The second section is a collection of what I think are the analogies and disanalogies between AI and evolution, in terms of takeoffs. These were compiled from posts and comments that I’ve seen on LessWrong, and some are based on my own impressions.
I believe that our current understanding of these topics is insufficient, and that more work is needed, either to steelman existing arguments or find better ones (see 0.2 Motivation). The third section thus contains some suggestions for further research directions, with the caveat that I’m unsure about the importance of this area relative to other areas of AI safety. On the upside, it seems likely that these problems can be tackled by those without a technical background in AI.
There are several reasons and motivations for why I decided to write this post. First, I’m worried that many evolution-based arguments about takeoff speeds might have flimsy premises, and I hope to increase discussion about this. I’ve also seen similar concerns echoed in several places. Second, the debate about takeoff speeds is spread over multiple blogs, papers, posts, and comments. One problem with the debate being so spread out is that different people are using different definitions for the same terms, and so I figured it might help to gather everything I could find in a single place.
I think there are also other reasons to believe that work understanding the connections between AI and evolution can be useful. For one, it can help improve our understanding of related questions, e.g. “How are AI takeoff scenarios likely to arise? How should we prepare for them? How much should we trust evolutionary priors in forecasts about TAI timelines?”
I also believe that we need stronger arguments before making political or research decisions. For instance, takeoffs that happen over multiple years could require different strategic considerations, and your views on this topic could influence which AI safety work you think is most important. Ben Garfinkel also pointed out that it can be risky to make large commitments on the basis of several blog posts, and I broadly agree with this.
0.3.1 Takeoff of what?
If we’re looking at the question of AI takeoffs from a policy decision-making perspective, plausibly the main consideration should be the timelines for transformative AI (TAI). Holden Karnofsky defines this as AI that is “powerful enough to bring us into a new, qualitatively different future”, and operationalises this in several ways for specificity. When talking about AI takeoffs, what I therefore mean is “the time period leading up to TAI” under this definition.
I think this should be distinguished from artificial general intelligence (AGI), by which I mean an AI system that is able to exhibit intelligent behaviour in many domains. Contrast this to “narrow intelligence”, which is only able to behave intelligently on a limited number of tasks, particularly in novel domains .
A lot of previous discussion about AI takeoffs has used AGI and TAI quite interchangeably - I think this is a reasonable position because general intelligence likely offers abilities like the ability “to tackle entirely new domains”, thus making it plausible that an AGI would be transformative. However, I don’t want to rule out the possibility that TAI can arise with a fairly “narrow” AI, so in this post I’m most interested in the takeoff of TAI, which may or may not necessarily be AGI. For the sake of clarity, I’ll try to mention whether I’m referring to TAI generally or AGI specifically.
0.3.2 Takeoff speeds
A more disputed definition is how to measure takeoff speeds - this was discussed more extensively by Matthew Barnett in Distinguishing definitions of takeoff. The definition that I’ll use is based on the discontinuous progress investigation by AI Impacts, again based on decision-relevance.
- Continuous takeoff: (the development of TAI) follows a trajectory that is in line with what one would have expected by extrapolating from past progress
- Discontinuous takeoff: (the development of TAI) follows a trajectory that is very far above what one would have expected by extrapolation of past progress
I also like these definitions because it suggests some kind of a qualitative difference between before and after we reach TAI. Other definitions that are solely based on times like “fast takeoffs happen over several months, slow takeoffs happen over several years” to me sound somewhat arbitrary, but perhaps I’ve failed to consider something. As others have mentioned, calling takeoffs “fast” or “slow” can also be confusing because “slow” takeoffs happen over several years, which is very fast to most people . This raises the question of why the debate has mostly revolved around takeoffs that take “2 years vs 2 months”, rather than “20 years vs 2 years” .
In practice, it can be quite hard to know what trajectory we should be paying attention to, and so operationalising the definitions of continuous and discontinuous takeoff can be challenging. One trajectory we could observe is the growth of world economic output, because it seems highly likely that a powerful AI would have large effects on economic productivity (especially if we’re comparing it to technological development during the industrial revolution).
Even if we’re observing a quantitative measure of takeoff speed like economic output, it’s still hard to tell when exactly we’re satisfying the criteria for continuous takeoff - e.g. are we deviating from existing trends because of a discontinuity, or is this just random noise? One proxy that we can use to determine a discontinuity is one that Paul Christiano gives in Takeoff Speeds. Specifically, a continuous takeoff  is a “complete 4 year interval in which world output doubles, before the first 1 year interval in which world output doubles.” This captures our rough intuition as to what we should see from a discontinuity - that the growth is fast relative to what we might ordinarily expect, based on the overall time that the drivers of the growth have been present.
While I think these definitions are broadly useful, I also believe that there are potentially some complications with these economic operationalisations of takeoff speeds. These are discussed a bit more in the section on analogies and disanalogies between AI and evolution.
0.4 Some assumptions about evolution
I mentioned my concerns about the “analogy to evolution” being rather vague. Part of this is due to the “analogy” being sloppy, but I suspect this is also due to “evolution” being a somewhat nebulous term. What exactly about evolution are we referring to? In the interests of concreteness, here are a few things about evolution that I’m assuming or relying upon for later parts of the post:
- Evolution is myopic: In general, it’s very hard to discern global trends in evolution, or if there is a well-defined notion of “evolutionary progress”. Evolution is often characterised as “blind”, and for the purposes of this post I think of evolution as a myopic search algorithm over possible organisms.
- Evolution is optimising for reproductive fitness.
- Evolution led to the development of general intelligence: One possible dispute here is whether or not humans are “general” enough in their intelligence (see for example François Chollet’s essay on intelligence explosion), but following the definition from above, I definitely count humans as being generally intelligent.
My understanding is that these views are fairly widely held and that stronger claims about evolution could also be made, but I don’t wish to make too many assumptions that many people might disagree about. In listing out common arguments about evolution and takeoff speeds, some extra assumptions are made in each argument, but I’m not necessarily claiming that these assumptions are true (though some of them may well be true). Here is a non-exhaustive list of such assumptions:
- Takeoff of human intelligence was continuous/discontinuous: for instance, Yudkowsky claims that the growth of human intelligence was discontinuous over evolutionary timescales (see 1.1 Hominid variation)
- Evolution has a long-term goal.
- Evolution has always been optimising for intelligence, or relative capabilities.
Evolution is often relatively inefficient at optimising for a particular goal, compared to other possible algorithms (for instance, it’s a lot more efficient to directly build a CNN, rather than evolve eyes to see and a ventral visual cortex to recognise objects)
1 Evolutionary arguments for takeoff speeds
There have been quite a few arguments for takeoff speeds based on analogies to evolution. These arguments often follow the rough structure of:
- Evolution has property X
- AI takeoffs are sufficiently analogous to evolution
- Thus, AI takeoffs are likely to have a property similar to X
The focus of this section is to give an overview of arguments that (somewhat) follow this framework, and the next section tries to clarify where point 2 holds and where it breaks down. If there’s some argument that you think I’ve missed or poorly represented, please do leave a comment or reach out to me personally. For the most part, I’ll not be commenting on whether these arguments make sense to me in this section.
Note: I wrote the below before the discussion between Yudkowsky and Christiano was posted onto LessWrong.
1.1 Hominid variation
The classic argument for AI takeoffs is sometimes known as the “hominid variation” argument, an example of which is in Eliezer Yudkowsky’s post, Surprised by Brains. This argues that a continuous takeoff is more likely, by drawing analogies from the scaling of intelligence in humans.
The basic argument is that evolution caused discontinuous progress in intelligence while it was (instrumentally) optimising for it, and AI is likely to do the same. A very similar argument is given in Intelligence Explosion Economics, where Yudkowsky argues that cognitive investment (e.g. greater brain compute, improved cognitive algorithms - not necessarily intelligence directly) leads to increased marginal returns in fitness.
Here’s my attempt at breaking down this argument more carefully:
- Claim: Evolution induced discontinuous progress in “success” 
- Humans are vastly more successful than other hominids, because human intelligence is far superior
- On evolutionary timescales, humans and other hominids are very close together
- Evolution was optimising for intelligence (instrumentally, while it was optimising for fitness) throughout evolutionary history
- Intelligence is a key factor in the relative “success” of humans, as compared to other hominids
- Conclusion: Evolution must have induced discontinuous progress in intelligence
- Claim: Hence, scaling of AI is likely to be discontinuous
- Evolution must have induced discontinuous progress in intelligence
- Intelligence scaling in AI is sufficiently similar to intelligence scaling in evolution
- Conclusion: Intelligence scaling in AI is likely to be discontinuous
1.2 Changing selection pressures
One notable response to the hominid variation argument is given by Paul Christiano, in Arguments about fast takeoff. My current understanding of the argument is that it boils down to two main disagreements:
- While it's true that humans are vastly more successful than other hominids, it’s not clear that this is because our intelligence is far superior
- Evolution wasn’t optimising for intelligence for all of evolutionary history; the selection pressures from evolution are constantly changing, and the selection pressure for intelligence was weak until very recently
To visualise this, consider the graph of intelligence growth starting from the point when the selection pressures were in the direction of increased intelligence, then the growth was actually continuous. Thus it was a change in selection pressures that caused a seemingly discontinuous takeoff (over evolutionary timescales), as opposed to evolution stumbling upon some “secret sauce” or crossing some intelligence threshold.
If evolution had been optimising for human-like success, then the takeoff would have appeared continuous even on longer timescales. Finally, Christiano argues that a similar “changing selection pressures” effect isn’t going to happen in AI, because humans are always going to be optimising AI systems for their usefulness.
1.3 Cumulative cultural evolution
This argument is proposed by calebo in Musings on Cumulative Cultural Evolution and AI, and can be thought of as pushback against the “changing selection pressures” argument discussed earlier. One of the main counterarguments given by Christiano was the possibility that the relative evolutionary success of humans was not due to intelligence.
The cumulative cultural evolution argument pushes back against this by referring to a conceptual model of human social learning. A highly simplified view of this says that the right mixture of social and asocial learning is necessary for fast development. Before that, we may get into an “overhang”, e.g. where we have undergone a large amount of asocial learning, such that when we start doing social learning, development happens very quickly. The argument implies that we might need to consider a similar possibility in the development of TAI, and suggests an update towards a discontinuous takeoff.
1.3.1 My opinions on this debate
I think this argument raises quite a few additional questions, some of which I suspect already have answers but I’m unaware of:
- What evidence is there that the increase in human intelligence was due to changing selection pressures?
- If the sudden growth in human intelligence (relative to all of evolutionary history) was indeed due to changing selection pressures, when exactly did these selection pressures change? What is the evidence for this?
- In evolutionary history, to what extent did human capabilities increase because of increased individual intelligence, as opposed to other reasons?
- It seems that the analogous property of “fitness” is “usefulness”, in the sense that this is what both the evolutionary and AI algorithms are optimising for respectively. What then is the analogous property of “selection pressures”? Couldn’t it also be the case that optimising for usefulness also leads to some change in the equivalent of “selection pressures” in AI?
My personal impression is that this debate can’t be settled without the above questions being answered, because they are quite fundamental to the aforementioned arguments.
1.4 Brains and compute
This argument from evolution is somewhat different from the hominid variation argument because it involves a quantitative comparison to what evolution was able to achieve with brains (albeit a very rough comparison). This works by asking the question, “how much was the brain able to do with X compute?”, then using the answer to this to answer, “how much do we expect AI systems to be able to do with Y compute?”.
An example of this is given by Scott Alexander in Neurons and Intelligence: A Birdbrained Perspective, and another investigation by AI Impacts. If we look into our evolutionary history we observe a kind of “intelligence ladder”. On the 80k podcast, Paul Christiano describes it this way (emphasis mine):
“A better way to compare [the abilities of AI relative to humans] is to look at what evolution was able to do with varying amounts of compute. If you look at what each order of magnitude buys you in nature, you’re going from insects to small fish to lizards to rats to crows to primates to humans. Each of those is one order of magnitude, roughly…”
This argument thus gives a rough prior for what we might expect from the continued development of advanced AI systems.
|Neuron count [9,10]||Organism||Capabilities|
Basic memory and understanding
Theory of mind, complex reasoning, etc.
|~5 x 103||Jellyfish|
|2.5 x 105||Ants|
|1.6 x 107||Frog|
|1.31 x 108||Star-nosed mole|
|3.10 x 108||Pigeon|
|2.25 x 109||Dog|
|9.6 x 109||Brown bear|
|2.8 x 1010||Chimpanzee|
|8.6 x 1010 (1015 connections )||Humans|
Table 1: Using neuron count as a rough prior for the capabilities of agents, by comparing with organisms with different levels of intelligence. The capabilities column is very high in uncertainty and should be taken with a grain of salt.
Scott Alexander also further states that we should expect abilities to “scale linearly-ish with available computing power”, which is especially alarming given current compute trends, and if the scaling hypothesis is true. Looking at the intelligence ladder also suggests that we compare typical human performance with technologies relative to things found in biology or evolution. Moreover, similar arguments can be made by drawing comparisons with the energy efficiency of the brain, rather than their overall capabilities (e.g. because this is more measurable).
I think there’s a tremendous amount of uncertainty in what exactly should go into the rightmost column of table 1, and that one could raise many questions about it. For instance, “to what extent are these capabilities due to intelligence, rather than other factors?” Another question might be whether the listed organisms might exhibit different behaviours in environments that differ from their natural habitat. This raises an issue of differences between observed behaviours and actual capabilities “out of distribution”. For this reason I’ve mostly only listed very vague capabilities that I think are reasonably likely to generalise to different living environments.
One could also claim that neuron counts are just the wrong approach altogether – for instance, Table 1 doesn’t include “African Elephant”, with 2.5 x 1011 neurons below humans, but most would agree that humans are smarter than elephants. This raises the question of how much trust we should place in this analogy.
These arguments don’t explicitly argue for whether takeoff is likely to be continuous or discontinuous; they instead provide a prior for what we should expect in the case of AI.
1.4.1 Side note: Dog-level AI
In the introduction, I mentioned that I didn’t want to rule out the possibility of TAI arising from sub-superintelligence, or perhaps even a fairly narrow intelligence. One way to frame this is to compare things with “human-level AI”, for instance by talking about a “dog-level AI”. But what does a “dog-level AI” actually mean? In the 2018/19 Overview of Technical AI Alignment episode on the AI Alignment Podcast, Rohin Shah and Buck Shlegeris define this to mean “a neural net that if put in a dog’s body replacing its brain would do about as well as a dog”. They point out that such a system could plausibly become superhuman at many things in other environments - I think this suggests that the capabilities in table 1 are not upper bounds by any means, and potentially might say very little about how powerful “dog-level” systems actually are. So these capabilities really should be taken with a grain of salt!
1.5 Biological anchors
A close relative of the previous argument is to look at the total compute done over evolution and to use this as a prior. The best example of this is probably Ajeya Cotra’s Draft report on AI timelines, which uses biological features (“anchors”) to develop a quantitative forecast of TAI timelines. This report wasn’t looking at AI takeoff speeds specifically, but I still think that this analogy is an important one.
The forecast uses six biological anchors, and the one that we’re most interested in is the evolution anchor, describing the total compute over evolution from earliest animals to modern humans. With this anchor, we’re treating all of evolution as a single overarching algorithm, and asking how much compute it has taken to get to human-level intelligence . Cotra gives a value of about 1041 floating point operations (FLOPs) for this, a number derived from Joseph Carlsmith’s report, How Much Computational Power Does It Take to Match the Human Brain?. Cotra assigns 10% weight to the evolution anchor in the resulting forecast.
Another intuition we might have is that evolution has done a significant amount of “pretraining” via neural architecture search, yielding the biological neural networks (i.e. our brains) that we’re familiar with today. Given that humans have hand-engineered artificial neural network architectures like RNNs, we might expect that the appropriate benchmark is the amount of compute required to “train” a person’s brain over the course of their lifetime. This describes the lifetime anchor, which predicts 1024 FLOPs to reach TAI. Cotra discusses this in much more detail in the report, including why this anchor seems quite unlikely to be correct. As of December 2021, the most computationally expensive systems require around 1024 FLOPs for the final training run (see the Parameter, Compute and Data Trends in Machine Learning project for more details on compute calculations).
We can think of the evolution anchor as an upper bound for how much compute we would need to reach TAI. Given that evolution is myopic and inefficient (e.g. for yielding intelligence), and it was able to yield general intelligence despite this, then it seems extraordinarily likely that we’ll reach AGI with the same amount of compute as over all of evolution - especially if our algorithms are more efficient. If we accept the results of the report as being broadly accurate, then we get a very rough lower bound for how slow a takeoff can be before reaching TAI. I think that a better understanding of these anchors plausibly leads to better understanding of the ways in which we might build AGI.
Of course, one could reasonably worry that the results are severe overestimates (or at least the upper bounds are so high that they aren’t helpful), see for instance Yudkowsky’s critique of the report.
1.6 One algorithm and Whole Brain Emulation
Another argument that one could make is that intelligence is generally described using a single simple algorithm. In such a scenario, stumbling across this algorithm could lead to a sharp discontinuity in progress.
My impression is that the main sources of disagreement about this argument are:
- Whether or not such a simple algorithm exists
- Whether or not we would be able to find such an algorithm if it does exist
One reason to suppose that such an algorithm exists is to use evolution as a proof of concept (i.e. “Evolution was able to do this, so why shouldn’t we be able to?”). If evolution led to human intelligence through a myopic search process that aims to optimise for fitness, then it surely must not be that hard to stumble upon this algorithm in AI research.
In particular, if humans are able to emulate human brains sufficiently well, then whole brain emulation (WBE) would provide a relatively clear pathway to yielding general intelligence. This plausibly makes a discontinuous takeoff more likely, although whether or not this happens depends on how easy brain emulation actually is. A key hypothesis that a lot of WBE hinges on is that we don’t need to “understand” the brain in order to emulate it, potentially making the task significantly easier (see for instance page 8 of Whole Brain Emulation: A Roadmap). Such a system may be more intelligent than existing systems for several hardware-related reasons (e.g. due to much faster computational speed, see Table 2), but fundamentally still operates in the same fashion. We might thus expect a system built via WBE to be limited in how “superintelligent” it may be.
1.7 Anthropic effects
One reason to be skeptical about the “one algorithm” argument is that observer-selection effects may conceal the difficulty in discovering an “algorithm for intelligence”.
Suppose there is a universe where finding this “intelligence algorithm” is very hard, and evolution was unsuccessful in discovering it. In this case, there would tautologically be no observer around to make the observation that “finding intelligence is very hard”. At the same time, in a universe where finding the intelligence algorithm is very hard but evolution successfully stumbled upon it, the observers would conclude that “the intelligence algorithm can’t be that hard to find, because even blind evolution was able to happen upon it”. The takeaway is that an observer sees intelligent life arise on their planet *regardless of how hard it is to find an algorithm for intelligence. *
This argument is described in more detail in Shulman and Bostrom’s How Hard is Artificial Intelligence? Evolutionary Arguments and Selection Effects. Importantly, the authors also consider how different theories of observation-selection effects can be used to counter this objection to the “one algorithm” argument. This relies on (definitions from the LW wiki):
- The Self-Sampling Assumption (SSA): “all other things equal, an observer should reason as if they are randomly selected from the set of all actually existent observers (past, present and future) in their reference class”.
- The Self-Indication Assumption (SIA): “all other things equal, an observer should reason as if they are randomly selected from the set of all possible observers”.
The SIA strongly favours universes where there are more observers with experiences indistinguishable from ours. It thus implies that we should place more credence in universes where intelligence is “not hard” to come by, as compared to universes where intelligence is hard, in favour of the “one algorithm” argument .
At the same time, evidence from convergent evolution (the independent evolution of similar features in different taxa) is less susceptible to anthropic effects from the SSA. Thus in a universe like ours where multiple independent lineages lead to complex learning and memory (or other signs of intelligence), we can infer that discovering an intelligence algorithm is boundedly hard .
Thus depending on which framework of anthropic reasoning you use (based on SSA and SIA), Shulman and Bostrom argue that the “one algorithm” argument is still plausible, despite observer-selection effects.
1.8 Drivers of intelligence explosions
The arguments in this section are in favour of discontinuous takeoff, and are mostly disanalogies to things that happen in evolution (more on this in section 2). In particular, they are of the form:
- Evolution has a certain property X
- But AI has [or more powerful AI systems would have] a property Y that drives intelligence growth much faster than it did in evolution
- Thus we should expect a discontinuous takeoff (or even an intelligence explosion)
I’ve collected a list of such arguments in Table 2 - these are mostly arguments first popularised around the time when Superintelligence was published or earlier.
|Generally cannot duplicate intelligent agents exactly nor quickly||Easy duplication (and orders of magnitude faster), much less computationally expensive than training from scratch|
|Rate of compute is partially constrained by speed of neural signals||Electric signals can be transmitted millions of times faster than in biological neurons|
|Serial depth of the human brain is limited to about 100 steps (see Connectionist Models and Their Properties)||Powerful AI systems could handle much more compute without parallelisation|
|Neural network size is constrained by biological properties, like skull size||AI models could get much larger and there is nothing in principle that stops artificial neural networks from being significantly larger than say human brains|
|Cannot “modify source code” very easily (whether of oneself or another organism)||Recursive improvement could lead to a feedback loop that speeds up growth|
Table 2: (Non-exhaustive) Comparing proposed drivers of intelligence growth in evolution and AI - the disanalogous properties of the drivers between evolution and AI are used to argue for a discontinuous takeoff, or even an intelligence explosion.
More discussion of this can be found in AGI Safety from first principles: Superintelligence, and in Plenty of Room Above Us.
In this section I discussed a few of what I think are the most prominent arguments relating to takeoff speeds, that draw an analogy to evolution. I suspect that I’ve missed some arguments, but as far as I’m aware the ones listed have been the most prominent over the past few years.
2 Evaluating analogies and disanalogies
The first part of this section gives a big list of analogies and disanalogies, and the second considers plausible frameworks for thinking about them. I attempt to relate these analogies to the arguments from section 1, and also try to draw some preliminary conclusions.
Some posts like Against evolution as an analogy for how humans will create AGI lay out reasons for why using evolution as an analogy can be problematic from the perspective of neuroscience. In this post I’ll be taking a different approach - I’m hoping to facilitate discussions that clarify existing arguments, rather than to push a particular point of view about whether or not evolution is a good analogy.
By “analogy” I’m referring mostly to the definition of analogy in Daniel Kokotajlo’s post on “outside views”. I don’t think any of the arguments from the previous section are rigorous enough to make them reference class forecasts.
There are several different kinds of “analogies to evolution”:
- Analogies about the plausibility of TAI development
- Analogies about how AGI might be built
- Analogies about when AGI might arrive
- Analogies about the influence of TAI
2.1 A non-exhaustive list of analogies and disanalogies
For each of these, I’ve characterised them as either an analogy or a disanalogy. I think it’s very likely that people will disagree with my classification, and that I’ll change my mind with more information. I’ve given each of these categorisations a certainty score that reflects my confidence in whether something is an analogy or a disanalogy (the higher the score, the more certain I am with the classification). I’d love to hear examples that support or contradict the claims here, or suggestions for further analogies/disanalogies.
The approach that I’ve taken is to go for breadth rather than depth - so I’ve also included some comparisons that I think are quite weak but are things that one “might want to consider”, or analogies that people have made in the past. I hope that this will serve as a useful preliminary outline of the analogies and disanalogies that can be better formalised in further work.
Certainty score: 4/5
Example(s): Surprised by Brains
According to this analogy, over evolutionary timescales, the development of human level intelligence was dramatically fast. If we compare this with the timescales of AI development, we should expect something similar (i.e. a discontinuous takeoff).
I think you can split these analogies to timescales into two types:
- Analogies to intelligence scaling in evolution
- Analogies to capability scaling in evolution
This class of analogies looks at how intelligence scaled in evolution, to say something about how intelligence will scale in AI. One example of this type of analogy is the hominid variation argument in section 1.1.
These analogies often come with a hidden implication of “if intelligence scales in a particular way, AI development is also likely to scale this way on some measure of capability (e.g. GDP)”. This is a crucial (and I think very plausible) assumption - as mentioned previously, from a decision-making perspective what we’re interested in is the effects of such a system. So although these analogies have the benefit of being relatively direct (because we’re comparing a directly analogous quantity in evolution and AI, namely intelligence), we also run into the problem of “intelligence” being nebulous and hard to measure. Furthermore, it can be hard to tell how exactly this translates into societal impacts.
I use quite a different definition of “analogy” compared to Steven Byrnes in Against evolution as an analogy for how humans will create AGI - this refers to something specific and satisfies three key properties, which I’ll not describe here. I’m not sure that many existing arguments from evolution fit the definition of “analogy” that Byrnes uses - instead I just count anything that draws any kind of similarity to the development of capabilities or intelligence in evolution as an “analogy”. Granted, this is quite vague, but I think this is sufficient because the goal of this post is more to gather arguments that relate to evolution, rather than argue a particular point.
Byrnes instead argues against what I would consider to be one particular analogy for the development of AGI, where an “outer algorithm” (e.g. evolution) does the majority of the work in improving an “inner algorithm” (e.g. the algorithm in the human brain - i.e. the AGI itself).
The second class of analogies is separated for precisely this reason - there can be a difference between the effects of scaling intelligence, and the general scaling of capabilities (for example, see the changing selection pressures argument). This looks at how capabilities scaled in evolution, to say something about how capabilities will scale in AI.
This is less direct of an analogy because it’s more ambiguous what exactly it is we’re comparing; having much greater capabilities than your competitors can mean very different things in an evolutionary environment and in the modern world today.
- Why should it be valid to make such a comparison? What is the justification for the internal working mechanisms being sufficiently similar? (Or if they are importantly different, how?)
Analogy: Necessary compute
Certainty score: 4/5
Example(s): Ajeya Cotra’s Draft Report on AI timelines, section on Brains and Compute
(See 1.5 Biological anchors for more details.)
This is a direct analogy to the amount of compute required for intelligence to arise in evolution. It informs possible quantitative priors that we can use in forecasts for AI takeoff scenarios, and are arguably most informative as providing upper bounds for when to expect TAI (and thus lower bounds for how slow a continuous takeoff can be).
Possible weaknesses with this analogy include:
- It’s not clear what numbers we should be comparing with: How should we compare brain FLOPs and computer FLOPs? What measures should we use, if neuron counts are sufficient to characterise intelligence?
- It’s not clear how useful the bounds provided by this analogy are: one could argue that they significantly overestimate how long it would take to reach TAI - e.g. due to major paradigmatic changes in AI research
Analogy: Cumulative cultural evolution
Certainty score: 3/5
Example(s): Musings on Cumulative Cultural Evolution and AI
(See 1.3 Cumulative cultural evolution for more details.)
This is an analogy used to argue for discontinuous takeoff via a model of the discontinuous growth mechanisms in evolution, in turn implying the plausibility of similar mechanisms in TAI development. We could arguably also claim that there have been some small discontinuities in certain domains like Go.
Possible weaknesses with this analogy include:
- It’s not clear how much agreement there is on the proposed model of cultural evolution - i.e. other models which don’t imply a discontinuous takeoff also exist
- One could question whether or not intelligence can only arise through the right combination of social and asocial learning, and how well such a model describes the bottlenecks for TAI development currently
Analogy: Intelligence is non-hard
Certainty score: 4/5
Example(s): Hard Takeoff
(See 1.6 One algorithm and Whole Brain Emulation for more details)
This analogy claims that because even blind evolution was able to discover an algorithm for general intelligence, finding the right algorithm for general intelligence in TAI is boundedly difficult. While it’s not clear how easy or hard intelligence really is, although it plausibly does seem reasonably non-hard.
Possible weaknesses with this analogy:
- One could object to this claim by pointing to observer-selection effects (see 1.7 Anthropic effects). Maybe we really did just get really lucky!
- Perhaps the space of algorithms that we’re searching through via AI research is sufficiently different from the space of algorithms that evolution searched through
- If it’s possible that “narrow intelligences” could be transformative, then drawing analogy with how easy it was to yield general intelligence in evolution may not be very sound
Certainty score: 4/5
Example(s): Hard Takeoff, Surprised by Brains
This analogy is between evolutionary thresholds and thresholds that hold back the development of AI. For instance, one might imagine that above some threshold intelligence, agents are able to undergo recursive self-improvement, leading to discontinuous takeoff. This threshold is often taken to be around “human-level”. An example of such an argument is given by Yudkowsky in Hard Takeoff:
“...observed evolutionary history - the discontinuity between humans, and chimps who share 95% of our DNA - lightly suggests a critical threshold built into the capabilities that we think of as ‘general intelligence’, a machine that becomes far more powerful once the last gear is added.”
In evolution, thresholds can arise when a small change in one variable past a certain threshold value leads to a large change in another variable. Examples of these include threshold traits that explain why guinea pigs develop different numbers of digits. But these generally only describe small scale changes, rather than very large changes in traits.
Possible weaknesses with this analogy:
- It’s not clear how strong threshold effects are on the scale of macroevolution; perhaps we shouldn’t be surprised that some small threshold effects can happen on small scales in complex dynamical systems, but this doesn’t say much about the likelihood of larger scale thresholds 
- It’s not a priori clear why we should expect this to be analogous in AI development - what properties might cause threshold effects to arise?
Analogy: Number of neurons
Certainty score: 4/5
Example(s): 80k podcast with Paul Christiano
A rough gauge we can use for the capabilities of AI systems of different sizes is by looking at the capabilities of organisms with brains of differing numbers of neurons (see 1.4 Brains and compute).
Note that this doesn’t refer to the number of connections per neuron - in chapter 1 of Deep Learning, Goodfellow et al. point out that artificial neural networks have had close to the same number of connections per neuron as in the human brain for many years, but we’re still orders of magnitude away from having the same number of neurons.
A related suggestion is to look at the ratio of content to architecture size, rather than just the architecture size - see Source code size vs learned model size in ML and in humans?
- Perhaps we should be looking at the connections in the neocortex, rather than just the number of neurons
- The number of neurons isn’t sufficient to tell us about intelligent capabilities - e.g. elephant brains have more neurons than human brains
Analogy: Structure of the learning system
Certainty score: 2/5
Example(s): My computational framework for the brain
Evolution designed human brains with a particular structure, and perhaps there are important features of this that need to be mimicked when developing AI in order for the resulting systems to be generally intelligent.
A natural comparison that we might make is based on the architectures and modularity of learning systems. Different parts of the brain have different functions, and modern ML systems often consist of several models designed for different purposes. As a simple example, a CNN has a head that learns a feature map, which is fed as an input vector into a fully connected neural net for classification (Richard Ngo calls this type of modularity architectural modularity, and distinguishes this from emergent modularity arising in a neural net from training). Some people argue that this modular structure makes it easier to evolve - parts can be added or removed from the system without affecting the function of other modules, although others contest this claim.
Other related factors one might want to think about are:
- The extent to which processing is done in parallel in the brain vs in ML (which could depend on how parallelisable the algorithm is, what hardware is being used etc.)
My impression is that there’s a fair bit of disagreement on how much artificial neural networks would need to resemble the brain in order to develop AGI. More discussion on this in The Brain as a Universal Learning Machine and Jeff Hawkins on neuromorphic AGI within 20 years.
- It’s hard to tell what features of the brain are the most crucial for developing general intelligence - one hypothesis is that the cortex is key for developing intelligence, but not all animals have a neocortex
Disanalogy: Training environments
Certainty score: 2/5
Example(s): AGI safety from first principles: Superintelligence
In natural selection, populations become adapted to existing environments, not to future ones. This increases fitness relative to the context within which selection is happening. Perhaps AI training environments are going to be more conducive to a fast takeoff, because these are hand-designed by AI researchers rather than purely based on the current state of the natural environment.
Some might counter that training of an AI in a virtual environment is insufficient; general intelligence can only arise in an environment that resembles the real biological one. However, this is potentially a pessimistic view, especially given that many features of biological environments can be simulated.
Richard Ngo previously also discussed some ways in which different environments could bottleneck AGI development.
- Current understanding of which features of the environment are most important for developing intelligence (or capabilities) is not very good
- How important is it that, plausibly, more intelligent agents are able to modify their environments more? How much were humans able to change the environment, in fitness-relevant ways? Should we look for something similar in AI?
Disanalogy: Order of training
Certainty score: 1/5
Example(s): Analogies and General Priors on Intelligence
In Analogies and General Priors on Intelligence, riceissa and Sammy Martin point out the order of learned tasks is different for AI, as compared with human evolution:
“Language, for instance, was evolved very late along the human lineage, while AI systems have been trained to deal with language from much earlier in their relative development. It is unknown how differences such as this would affect the difficulty-landscape of developing intelligence.”
This is perhaps related to the analogy mentioned by Stuart Russell in Human Compatible. Specifically, Russell argues that there are certain milestones, like the ability to plan and understand language, before we can reach AGI. Possible disanalogies here include “how fast we’ll be able to reach all of these milestones” and “whether there is a fixed order in which the milestones are reached”.
A proxy for how fast AI takeoff occurs (if it does) would thus be something along the lines of, “how fast do you think we’ll be able to reach all of these milestones”?
- This doesn’t seem to have very much information value (at least the order doesn’t matter match, although my intuition is that Stuart Russell’s mention of milestones may be more helpful)
Disanalogy: Optimisation objectives
Certainty score: 3/5
Example(s): AGI safety from first principles: Superintelligence, Takeoff speeds
Current AI systems tend to be trained to maximise performance on specific tasks (e.g. image classification), leading to relatively narrow capabilities. On the other hand, organisms in evolution need to deal with multiple different tasks (e.g. reasoning, image recognition, planning), and thus need to be more well-rounded to survive.
But how exactly should we think of the “optimisation objectives” for AI systems? If we’re thinking about a single task like classification using a CNN, then the objective might be to “minimise the loss function”. If we think more broadly, then one could also argue that the objectives for AI systems are to be useful to humans (by design, if we assume that the objective functions have been specified correctly). On the other hand, evolution is optimising for fitness.
- This might not hold up if the scaling hypothesis is true. For instance, GPT-3 demonstrates some degree of generalisation to tasks on which it was never trained, like simple arithmetic
- How do we know that there aren’t any sub-objectives that agents may be optimising for?
Disanalogy: Changing selection pressures
Certainty score: 4/5
Example(s): Takeoff speeds
Paul Christiano argues that selection pressures won’t change in AI (since AI systems are being designed for usefulness to humans), but have constantly been changing throughout evolutionary history, and so a discontinuous takeoff in AI is less likely.
By selection pressures, I’m referring to a driver of natural selection that favours particular traits, like “high social intelligence”. One potential cause for these are environmental effects - fitness is defined relatively to the current environment, and so changes in the environment could change what natural selection (or a more general search algorithm) preferentially selects for. This could be induced by agents that modify the environment they are in; depending on how much control agents have in modifying the environment, they may be able to control the selection pressures in the way that they “desire” (similar to how humans build houses for shelter).
- Just because AI research is ultimately being driven by usefulness, this doesn’t mean that there won’t be the equivalent of changing selection pressures in AI that could lead to a discontinuous takeoff
- What do we mean by “usefulness”? Doesn’t what we consider to be useful change over time?
Disanalogy: Data constraints
Certainty score: 2/5
Example(s): the scaling “inconsistency”: openAI’s new insight (and Scaling Laws for Autoregressive Generative Modeling)
This disanalogy is between how much data is available for “training” in evolution as opposed to AI. Currently, ML systems are partially bottlenecked by a lack of labelled data, rather than data in general, which I’ll term as a lack of “useful data”. A corollary of this is that we might expect language models to eventually become data constrained. In evolution on the other hand, there is no shortage of useful data to train on - organisms gather information by interacting with the environment.
- What counts as “useful data”?
- Perhaps the more appropriate comparison here is between RL and evolution, removing the bottleneck from having a large enough labelled dataset
- How likely is it that supervised learning will get bottlenecked from there being not enough useful data? Isn’t the amount of useful data just going to continue increasing dramatically?
- Should we be more worried about sample efficiency than data constraints? This could be especially important given recent breakthroughs, e.g. EfficientZero.
Disanalogy: Energy costs
Certainty score: 3/5
Example(s): Training a single AI model can emit as much carbon as five cars in their lifetimes
The brain seems to be very energy efficient, and machine learning models take a ton of energy consumption to train . For instance, our brains are able to run on about 20W, which is comparable with the power of my desk lamp. In contrast, training of neural networks has led to increasing concerns about the carbon footprint of ML. Perhaps a feature to pay attention to is how the energy costs and efficiency of AI systems changes over time.
This could be important because of tradeoffs that exist in the evolutionary environment. On the one hand, having a larger brain generally yields cognitive benefits that increase odds of survival and maximising fitness. On the other hand, having larger brains requires more energy costs, reducing energy that can be devoted toward reproduction and development, thus reducing fitness .
- Comparisons are often done between energy consumption of the human brain vs energy required for neural network training. Perhaps this is the wrong comparison, and we should be comparing with inference compute instead, which is significantly lower
- TAI could be developed even with a very energy inefficient model, so this comparison may not be very useful
Disanalogy: Drivers of intelligence explosions
Uncertainty score: 5/5
Example(s): AGI Safety from first principles
(See 1.8 Drivers of intelligence explosions for more details.) This class of disanalogies argues that there are important ways in which AI development is different from evolutionary growth, and that this changes the picture completely.
- Is there any evidence for us to believe that these mechanisms (e.g. recursive self-improvement) are likely to kick-in before reaching TAI? How much credence should we place on something like this? How do we know that this is likely to happen?
Based on the above, it seems like there are many different considerations that might plausibly need to be considered. In order for analogies to evolution to be useful in helping us draw conclusions, I think it would help to understand which properties need to be analogous and why.
I think now is a good time to revisit the original question that inspired this post: “How should we think about the analogy to evolution?”
2.2.1 A possible framework: levels of analysis
The list of analogies and disanalogies is quite messy as it stands, and overall I’m very unsure about how to orient towards them.
One possibility is to follow an approach that is a modified version of Marr’s Levels of Analysis (credits to Vael for suggesting this). According to Kraft and Griffiths:
"Marr (1982) famously argued that any information processing system can be analyzed at three levels, that of (1) the computational problem the system is solving; (2) the algorithm the system uses to solve that problem; and (3) how that algorithm is implemented in the “physical hardware” of the system.”
An example of this that I really like is proposed by Steven Byrnes, that separates the development of general intelligence into an “outer algorithm” and an “inner algorithm”.
|Outer algorithm (runs an automated search for the inner algorithm)||Evolution searching over the space of possible cognitive architectures and algorithms||Humans searching over the space of AI system architectures and algorithms|
|Inner algorithm (the general intelligence)||The human brain algorithm||E.g. SGD|
Table 3: Steven Byrnes’ description of the analogy to evolution involves breaking down the process of AGI development into two parts - an inner algorithm (for the trained intelligent agent) and an outer algorithm which searches for the inner algorithm.
I think this way of viewing the list of analogies and disanalogies leads to a decent classification of the internal algorithmic aspects, but some analogies are also about the effects and capabilities of them. To include these other analogies, I suggest one other category, the “effects”, yielding three levels:
- Inner algorithm level: analogies/disanalogies relating to the inner algorithm (e.g. narrowness of optimisation objectives)
- Outer algorithm level: analogies/disanalogies relating to the outer algorithm (e.g. changing selection pressures)
- Effects: analogies/disanalogies that aren’t about the workings of the inner+outer algorithm, but are about the observed effects (e.g. level of intelligence)
|Necessary compute||~1e41 FLOPs is an upper bound||Currently systems ~1e24 FLOPs|
|Cumulative cultural evolution||Discontinuous jumps are plausible via overhangs||Some small (?) discontinuities e.g. AlphaGo|
|Intelligence is non-hard||Even blind evolution discovered general intelligence|
|Thresholds||Threshold traits, theory of saltation|
|Number of neurons||Human brain: ~1e11 neurons, ~1e15 synapses||Current ML models: 1e13 parameters|
|Structure of the learning system||E.g. modularity in the brain||E.g. modular ML system architectures|
|Training environments||Certain features of evolutionary environments are needed for intelligence||Maybe not?|
|Order of training||Learned cognitive skills in a particular order||Need not follow the same order|
|Optimisation objectives||Wider range of incentives||Narrower; specific tasks|
|Changing selection pressures||Selection pressures change all the time||Always optimising for usefulness|
|Data constraints||No shortage of data||Lack labelled data|
|Energy costs||Very energy efficient brain||ML systems not as efficient yet|
|Drivers of intelligence explosion (e.g. recursive improvement)||Limited recursive improvement due to biological constraints||AIs go from the inner to the outer - recursive improvement|
Table 4: Summary of the analogies, disanalogies, and which level of analysis they fall into.
At the end of the day, I think it’s also plausible that trying to frame these analogies/disanalogies just isn’t that useful yet. I’m worried that this is a bad framing that turns out to be harmful in the long-term, so I’d appreciate any thoughts on this .
It seems very reasonable to me that we would have to think of each level of analysis differently. Processes at the inner algorithm level are probably going to be quite different, for instance. Furthermore, because ultimately what we’re interested in is the effect-level consideration of TAI, it seems that the inner algorithm level factors are less important (e.g. just because the environments are similar is far from sufficient evidence that TAI will develop the same way as evolution).
I expect that the analogies and disanalogies will change over time, some becoming more analogous and some less so, perhaps due to different approaches becoming more dominant in ML. I don’t have a strong opinion about which things might become more or less analogous.
2.2.2 Should we use evolution as an analogy?
So, should we even be using these analogies and disanalogies?
We should probably be more confident in the conclusions derived from these comparisons if they are used as plausibility arguments, as opposed to purely quantitative measures, simply because a plausibility claim is weaker. This of course depends on how the quantitative measure is being done, and what it is being used for (e.g. I believe bioanchors can still be a useful approach as a rough proxy or for getting an upper bound for when we might reach TAI).
I believe a lot of the value of these analogies could come from their information value (e.g. in helping us understand the mechanisms that could lead to different takeoff scenarios). As an example, understanding the role of modular systems in evolution could end up being very useful if it is indeed a large driver in the evolution of human intelligence.
Also, if we’re looking at intelligence specifically, what other options do we have? I think we shouldn’t be too quick to dismiss analogies to evolution, depending on the use case.
Overall, I think that the way in which analogies and disanalogies to evolution are being used is pretty reasonable. Personally, I think that using evolution as an analogy makes sense and is a good idea, as long as it remains exactly that - an analogy. My impression is that we would need a lot more evidence or understanding to be able to use evolution as evidence in a more rigorous fashion.
2.2.3 Relation to other arguments about takeoff speeds
Perhaps you’re thinking that most of these analogies just aren’t going to be very significant compared to other arguments about takeoffs in general, like the scaling hypothesis or concerns about an AI overhang.
Given the uncertainties regarding evolution and the many possible disanalogies, I think existing arguments that include both evolutionary and other considerations make a lot of sense. But how much exactly? I don’t have a strong opinion on this, but I think a large source of variance here is how convincing you think the disanalogies are (especially the drivers of intelligence of explosions).
Overall I’m unsure about more precise proportions as to how strongly to weigh evolutionary arguments against other ones (e.g. economic arguments) in forecasts.
The analogies and disanalogies that I mentioned previously are nowhere near exhaustive. Looking at things overall, I think it’s potentially useful or impactful to investigate which of these analogies or disanalogies are the most important and why.
3 Further work
In this section, I list some brief thoughts on what might potentially be useful questions to answer. I don’t have a good grasp of how tractable these are, or how likely it is that they’ll be highly useful for furthering understanding. I leave these considerations as an exercise for the reader.
3.1 Clarifying definitions
3.1.1 Measures of growth
This is an example of what I think is an ambiguity that often arises when talking about takeoffs, that I’m hoping to get some clarity on. Specifically, I think the choice of measure is quite important when extrapolating previous progress to determine whether a takeoff is continuous or not.
One concern with the definition that I used previously, global economic output (see 0.3 Definitions), is the possibility that global economic output grows continuously, while another measure (like intelligence) grows discontinuously. Consider the example of LED efficiency:
Figure 4: [Image source] Graphs showing the change in light source efficiency over time. The graph on the right shows the same graph as on the left, but with two rough curves highlighted - a continuous curve that represents continuous growth in lighting efficiency, and a discontinuous curve that represents growth in LED efficiency since first invention.
We can observe several things:
- LEDs were first produced around 1960, and suddenly underwent drastic efficiency improvements in the 1990s
- If we choose to look at “lighting efficiency”, then we’re more likely to claim a continuous takeoff
- If we choose to look at “LED efficiency”, then we’re more likely to claim a discontinuous takeoff
This illustrates some of what I think are existing difficulties with defining discontinuities. What exactly should we be measuring if we’re going to be talking about takeoff speeds?
Another problem is using economic growth as the operationalisation of “increased capabilities”. Some people like Will MacAskill think that it’s plausible that we won’t get an economic “takeoff” at all; instead, we may hit diminishing returns. This would be problematic if economic growth is the main measure that we’re using. Daniel Kokotajlo also raises some concern about this in Against GDP as a metric for AI timelines and takeoff speeds. This could mean that purely using economic output as a measure of takeoff speed is insufficient, and determining the continuity of progress could require observing multiple different trends simultaneously (economic growth, compute efficiency), some of which are the above analogies/disanalogies.
Another confusion that I had when working on this project is how we came to the definitions of “continuous” and “discontinuous” takeoff in the first place. However, since I’m viewing things with an eye towards political decision-making, I’m under the impression that the continuous/discontinuous distinction might not make as large of a difference as I’d thought (see 0.2 Motivation). More discussion of this can be found at Soft takeoff can still lead to decisive strategic advantage, and the subsequent review.
I’m interested in better understanding exactly how slow the takeoff would need to be to make a big difference in our ability to respond to increasingly rapid AI development, and having terms that distinguish between these scenarios. In short, currently we’ve been asking, “if takeoff happens at a rate X, then will society be able to do Y in time?” I want to ask, “if we want to be able to do Y in time, what X is necessary?”.
3.2 Clarifying understanding
If we temporarily ignore arguments about evolution altogether, my impression is that most people seem to think that a continuous takeoff is more likely (under Paul Christiano’s definition of “slow takeoff”). I’m unsure about why exactly this is the case, and I’m curious to hear what people think about this - is this because the evolutionary arguments in favour of a discontinuous takeoff, like the hominid variation argument, seem largely unconvincing?
In general, I’d be curious to get a better idea of what the most relevant takeoff measures are, and the main drivers of takeoff speeds (as well as the role of evolutionary analogies in that). In particular, do you think that the role of evolutionary analogies are relatively small compared to other arguments about takeoff speeds? If so, why?
I also think it would be good to look into specific cases of these analogies or disanalogies and steelmanning or critiquing them. Some topics in evolution that I think it could be valuable to look into include:
- Evolution of modularity
- Macroevolution, especially understanding whether or not there are global trends over evolutionary history
- Threshold effects in evolution
- Different theories of the evolution of general intelligence (e.g. cultural learning)
I wrote this post to raise and address some uncertainties around using evolution as an analogy for the development of TAI, and potentially encourage new work in this area that benefits political decision-making. In particular, I focus on arguments about the takeoff speeds of TAI, although the conclusions may apply more generally.
The post split into three main sections:
- Reviewing key arguments involving analogies or disanalogies to evolution
- A list of the analogies and disanalogies, and evaluating them
- Possible avenues for further work
Reviewing key arguments
- The hominid variation argument: (discontinuous takeoff) because intelligence scaled discontinuously in evolution, it seems plausible that something similar could happen in AI
- Changing selection pressures: (continuous takeoff) we won’t get something similar, because the discontinuity in evolution arose from a change in selection pressures - this won’t happen in AI because humans always design AI systems for usefulness
- Cumulative cultural evolution: (discontinuous takeoff) the discontinuity in evolution might not just be due to selection pressures - perhaps it’s due to a social learning/asocial learning overhang instead
- Brains and compute: (unsure) a rough proxy for how quickly AI capabilities will scale is to look at the capabilities of organisms with different numbers of neurons, and comparing that to the neuron count of current systems
- Bioanchors: (unsure) we can place upper bounds on how much compute is needed for TAI, forecast when we might expect TAI using 2020 deep learning [and put a weak lower bound on how slow takeoff will be]
- One algorithm and WBE: (discontinuous takeoff) building AGI can’t be that hard, because blind evolution was able to do it
- Anthropic effects: (discontinuous takeoff) one might object to the “one algorithm” argument because of observation-selection effects, but the SIA and SSA frameworks at least partially dispel such concerns
- Drivers of intelligence explosions: (discontinuous takeoff) AI is different from evolution in crucial ways, like the ability to do recursive improvement quickly
A non-exhaustive list of analogies and disanalogies
- Timescales: (analogy) intelligence and capability scaled in fashion in evolution, so we should expect them to scale in fashion in AI
- Necessary compute: (analogy) how much compute we need to reach TAI is vaguely similar to (and upper bounded by) how much compute was needed to develop general intelligence over evolution
- Cumulative cultural evolution: (analogy) one model of the evolutionary development of intelligence suggests a way in which growth could be discontinuous, so something similar should happen in AI
- Intelligence is non-hard: (analogy) it can’t be that hard to find the “intelligence algorithm” in AI because blind evolution found it – and when we stumble across it, discontinuous takeoff occurs!
- Thresholds: (analogy) akin to how small changes in one variable can lead to large changes in another (threshold traits), once we cross some threshold, intelligence can grow really fast
- Number of neurons: (analogy) a gauge for the capabilities is AI systems is to look at how many neurons they have, and what organisms with that many neurons are able to do
- Structure of the learning system: (analogy) human brains have a particular abstracted structure, and we should expect AGI to have a similar high-level one
- Training environment: (disanalogy) fitness is defined relative to the environment, and training environments can be different between AI and evolution in important ways
- Order of training: (disanalogy) in evolution, we developed the ability to do certain tasks before others - we probably shouldn’t expect the same order of “learned tasks” in future AI development
- Optimisation objectives: (disanalogy) AI systems are currently trained to perform well on specific tasks, whereas organisms in evolutionary environments have to deal with multiple different tasks, leading to a broader range of abilities
- Changing selection pressures: (disanalogy) unlike evolution, selection pressures won’t change in AI, because AI is always optimised for usefulness
- Data constraints: (disanalogy) ML systems are partially bottlenecked by a lack of labelled data, but evolutionary “agents” can just collect data by interacting with the environment
- Energy costs: (disanalogy) the brain is much more energy efficient that ML systems
- Drivers of intelligence explosions: (disanalogy) you don’t get things like extremely fast recursive self-improvement in evolution
These can be analysed by thinking of them as existing at three different levels:
- Outer algorithm level: searches for an “inner algorithm”, e.g. evolution
- Inner algorithm level: e.g. the human intelligence algorithm
- Effects: analogies/disanalogies to the observed effects of the inner+outer algorithm
Overall I’m uncertain about how large of a role these analogies and disanalogies should play when thinking about takeoff speeds, although I believe that they can be quite useful for information value and as a rough estimate of what to expect from TAI.
Other things (in addition to the role of analogies/disanalogies to evolution) that I think remain unclear include:
- How do we define appropriate measures of growth?
- Should we be using different definitions of continuous and discontinuous takeoff?
- What are the main reasons why people believe continuous takeoff is more likely?
Some things I think it would help to look into:
- Evolution of modularity
- Macroevolution and global trends in evolution
- Threshold effects and traits
- The evolution of general intelligence
I’d love to see some discussion on these topics, including things that I got wrong and other stuff that you think should be included!
 To make things more complicated, the line between the domain-specific and domain-general intelligence can be somewhat fuzzy. For instance, what counts as a single “domain”?
 Then again, “continuous takeoff” isn’t a very mathematically sound term either, but I’m not sure how to contest that!
 Specifically I think it would be useful to quantify how likely a “20 year takeoff” would be compared to a “2 year one”. My impression is that several arguments for why the 20 year takeoff is unlikely already exist (e.g. in Intelligence Explosion Microeconomics).
 This is referred to as a “slow takeoff” rather than a “continuous takeoff” in the original article.
 I’m not very sure what kind of “success” Yudkowsky is referring to, so I’ve put this in quotes. Presumably this means capabilities, i.e. the ability to achieve one’s goals, which seems very similar to the definition of power in the formalisation of instrumental convergence.
 I’m somewhat unsure what exactly I should be referring to for the “start of evolutionary history”, so this is admittedly vague. One possibility is to treat the evolution of the first neurons (around 500 million years ago) as a starting point, similar to the approach used in Draft report on AI timelines.
 I think of this as being like applying a linear transformation to the time axis such that evolutionary history gets contracted down into several decades. But is such an approach valid?
 One could counter this point by saying although it looks continuous in evolution, this would be discontinuous in AI development e.g. due to recursive self improvement (see 2 Evaluating analogies and disanalogies).
 Perhaps the important consideration is cortical neuron count, rather than neuron count in general. The numbers here only represent neuron counts in general.
 Source: Wikipedia page List of animals by number of neurons. Note that I haven’t bothered to look into the primary sources individually.
 By contrast, state of the art ML systems have used up to 1013 parameters.
 The report also describes other bioanchors, such as the lifetime anchor, i.e. “the computation done by a child’s brain over the course of growing to be an adult”. These other anchors are many orders of magnitude less in the required compute, so the compute predicted solely by the evolution anchor could be a significant overestimate.
 This is very much oversimplified - a more complete consideration depends on things like the plausibility of the Great Filter. I don’t feel like I have a good understanding of this, and will defer to Shulman and Bostrom’s paper.
 Of course, the evidence for this may be unclear. For instance, it’s possible that species that display convergent evolution only do so because of some shared obscure reason in their biologies. Shulman and Bostrom also list several other caveats to this argument.
 In terms of macroevolution, one could point to the theory of saltation, which postulates radical mutational changes that radically alter the phenotype of a species. This is in stark contrast with the theory of gradualism, which instead argues that phenotypic changes in organisms only happen continuously.
 For more details about the efficiency of AI systems, see OpenAI’s AI and Efficiency.
 More discussion by Burkart et al in Future directions for studying the evolution of general intelligence.
 Here’s another framework for thinking about the analogies and disanalogies that I briefly considered, with three categories: (1) analogies/disanalogies for how we will build AGI, (2) analogies/disanalogies for how soon or quickly we should expect AGI, and (3) analogies/disanalogies for how soon or quickly we should expect something that is as transformative as AGI.