Summary of Situational Awareness - The Decade Ahead

By OscarD🔸 @ 2024-06-08T11:29 (+140)

Original by Leopold Aschenbrenner, this summary is not commissioned or endorsed by him.

Short Summary

I. From GPT-4 to AGI: Counting the OOMs

Past AI progress

Training data limitations

Will training data be a limitation to continuing language model scaling?

Trend extrapolations

The modal year of AGI is soon

The modal year in which AGI arrives is in the late 2020s, even if you think the median is after that.

II. From AGI to Superintelligence: the Intelligence Explosion

The basic intelligence explosion case

Objections and responses

The power of superintelligence

Much of the first AGI’s efforts will likely be directed towards further AI research. But once superintelligences exist, they will likely also spend considerable subjective time on other fields. This could look like compressing the equivalent of all of 20th-century technological progress into a few years.

III The Challenges

IIIa. Racing to the Trillion-Dollar Cluster

AI will increasingly require huge amounts of capital expenditure on data centres, energy, and semiconductor fabs. Based on naive (but plausible) trend extrapolation, we can estimate some numbers for the latest single training run in each year:

YearOOMsH100s EquivalentCostPowerPower Reference Class
2022~GPT-4 cluster~10k~$500M~10 MW~10,000 average homes
~2024+1 OOM~100k$billions~100 MW~100,000 homes
~2026+2 OOMs~1M$10s of billions~1 GWThe Hoover Dam, or a large nuclear reactor
~2028+3 OOMs~10M$100s of billions~10 GWA small/medium US state
~2030+4 OOMs~100M$1T+~100 GW>20% of US electricity

IIIb. Lock Down the Labs: Security for AGI

The power of espionage

AI secrets can probably be easily stolen by state actors.

Securing model weights

“An AI model is just a large file of numbers on a server. This can be stolen. All it takes an adversary to match your trillions of dollars and your smartest minds and your decades of work is to steal this file.”

Protecting algorithmic insights

Necessary steps for improved security

IIIc. Superalignment

Controlling AGI, and thereafter superintelligence, is an unsolved but solvable technical problem.

IIId. The Free World Must Prevail

“Our generation too easily takes for granted that we live in peace and freedom. And those who herald the age of AGI in SF too often ignore the elephant in the room: superintelligence is a matter of national security, and the United States must win.”

Objection: China is well behind on AI research, and chip export controls mean they will stay well behind on hardware. So we don’t need to worry much about the US losing its lead.

It would be very bad if China controls superintelligence, as this could lead to a stable totalitarian regime lasting indefinitely, with none of the even limited checks and balances of current regimes.

Superintelligence will likely invent new weapons that continue the historical trend of greatly reducing the cost of destruction (e.g. very effective and controllable bioweapons, innumerable tiny drones armed with poison darts, presumably things we have not and cannot think of).

IV. The Project

Much of AI governance work focuses on regulations and standards for AI companies to follow. But this misses the bigger picture, that the USG will not let superintelligence be developed privately, and will nationalise AGI companies.

V. Parting Thoughts

Many people do not accept the centrality of AGI (“Deep learning is hitting a wall!”, “just another tech bubble!”). But there are two other untenable positions:

Core tenets of AGI realism:

The story in this essay is only one way things could go. But I (Leopold) think it is the modal outcome for this decade.

Responses to Situational Awareness

I (Oscar) will try to compile some of the notable responses to Leopold’s report here, feel free to suggest other things I should add to this list, or point me to a better existing list.

  1. ^

     There is more uncertainty because over the intervening four years OpenAI, along with other leading AI companies, began releasing less information publicly about their training runs.


Yannick_Muehlhaeuser @ 2024-06-10T16:27 (+27)

One response I think is worth reading. 

Toby Tremlett @ 2024-06-12T12:36 (+2)

Seems like a helpful addition to the debate-- have you considered link-posting it (in full)? 

trevor1 @ 2024-06-10T05:06 (+5)

One thing I really liked about it was the title: "situational awareness". I think that phrase is very well-put given the situation, and I got pretty good results from it in conversations which were about AI but not Leopold's paper.

I also found "does this stem from the pursuit of situational awareness" or "how can I further improve situational awareness" to be helpful questions to ask myself every now and then, but I haven't been trying this for very long and might get tired of it eventually (or maybe they will become reflexive automatic instincts which stick and activate when I'd want them to activate; we'll see).

jacquesthibs @ 2024-06-11T14:43 (+3)

GPT-2 was trained in 2019 with an estimated 4e21 FLOP, and GPT-4 was trained in 2023 with an estimated 8e24 to 4e25 FLOP.

Correction: GPT-2 was trained in 2018 but partially released in February 2019. Similarly, GPT-4 was trained in 2022 but released in 2023.

OscarD @ 2024-06-11T20:13 (+1)

Thanks, fixed. I was basing this off of Table 1 (page 20) in the original but I suppose Leopold meant the release year there.