Analogy Bank for AI Safety
By utilistrutil @ 2024-01-29T02:35 (+14)
This is a crosspost, probably from LessWrong. Try viewing it there.
nullAgustín Covarrubias @ 2024-01-29T21:29 (+3)
This is great! I've publicly spoken about AI Safety a couple of times, and I've found some analogies to be tremendously useful. There's one (which I've just submitted), that I particularly like:
I find myself thinking back to the early days of Covid. There were weeks when it was clear that lockdowns were coming, that the world was tilting into crisis, and yet normalcy reigned, and you sounded like a loon telling your family to stock up on toilet paper. There was the difficulty of living in exponential time, the impossible task of speeding policy and social change to match the rate of viral replication. I suspect that some of the political and social damage we still carry from the pandemic reflects that impossible acceleration. There is a natural pace to human deliberation. A lot breaks when we are denied the luxury of time.
But that is the kind of moment I believe we are in now. We do not have the luxury of moving this slowly in response, at least not if the technology is going to move this fast.
From this op-ed by Ezra Klein.
Rocket @ 2024-01-30T05:41 (+1)
Thanks, Agustín! This is great.
dschwarz @ 2024-01-29T03:02 (+2)
Nice post. I also have been exploring reasoning by analogy. I like some of the ones in your sheet, like "An international AI regulatory agency" --> "The International Atomic Energy Agency".
I think this effort could be a lot more concrete. The "AGI" section could have a lot more about specific AI capabilities (math, coding, writing) and compare them to recent technological capabilities (e.g. Google Maps, Microsoft Office) or human professions (accountant, analyst).
The more concrete it is, the more inferential power. I think the super abstract ones like "AGI" --> "Harnessing fire" don't give much more than a poetic flair to the nature of AGI.
Rocket @ 2024-01-29T19:00 (+1)
Please submit more concrete ones! I added "poetic" and "super abstract" as an advantage and disadvantage for fire.
SummaryBot @ 2024-01-30T13:35 (+1)
Executive summary: Analogies can be useful for explaining AI safety concepts, but they should be chosen carefully to avoid misrepresenting key ideas or distracting from the main arguments.
Key points:
- Analogies highlight certain aspects of a concept while obscuring others, so analogies for AI safety should minimize lost nuance.
- Simple comparisons stating two things share some property often add little, but arguments drawing unintuitive conclusions via analogy can be insightful.
- Analogies serve purposes like providing evidence, aiding intuition, saving time, and consistency checking.
- Common attacks on analogies question if the base conceptimplies the conclusion, if the analogues are sufficiently similar, or the conclusion itself.
- Entrenched analogies can mislead policymakers, so new analogies should be scrutinized, especially if expressing AI safety skepticism.
This comment was auto-generated by the EA Forum Team. Feel free to point out issues with this summary by replying to the comment, and contact us if you have feedback.