[linkpost] "What Are Reasonable AI Fears?" by Robin Hanson, 2023-04-23

By Arjun Panickssery @ 2023-04-14T23:26 (+41)

This is a linkpost to https://quillette.com/2023/04/14/what-are-reasonable-ai-fears/

Selected quotes (all emphasis mine):

Why are we so willing to “other” AIs? Part of it is probably prejudice: some recoil from the very idea of a metal mind. We have, after all, long speculated about possible future conflicts with robots. But part of it is simply fear of change, inflamed by our ignorance of what future AIs might be like. Our fears expand to fill the vacuum left by our lack of knowledge and understanding.

The result is that AI doomers entertain many different fears, and addressing them requires discussing a great many different scenarios. Many of these fears, however, are either unfounded or overblown. I will start with the fears I take to be the most reasonable, and end with the most overwrought horror stories, wherein AI threatens to destroy humanity.

 

As an economics professor, I naturally build my analyses on economics, treating AIs as comparable to both laborers and machines, depending on context. You might think this is mistaken since AIs are unprecedentedly different, but economics is rather robust. Even though it offers great insights into familiar human behaviors, most economic theory is actually based on the abstract agents of game theory, who always make exactly the best possible move. Most AI fears seem understandable in economic terms; we fear losing to them at familiar games of economic and political power.

He separates a few concerns:

Finally he gets to the part where he dunks on foom:

When I polled my 77K Twitter followers recently, most respondents’ main AI fear was not any of the above. Instead, they fear an eventuality about which I’ve long expressed great skepticism:

The AI “foom” fear, however, postulates an AI system that tries to improve itself, and finds a new way to do so that is far faster than any prior methods. Furthermore, this new method works across a very wide range of tasks, and over a great many orders of magnitude of gain. In addition, this AI somehow becomes an agent, who acts generally to achieve its goals, instead of being a mere tool controlled by others. Furthermore, the goals of this agent AI change radically over this growth period.

. . .

From humans’ point of view, this would admittedly be a suboptimal outcome. But to my mind, such a scenario is implausible (much less than one percent probability overall) because it stacks up too many unlikely assumptions in terms of our prior experiences with related systems. Very lumpy tech advances, techs that broadly improve abilities, and powerful techs that are long kept secret within one project are each quite rare. Making techs that meet all three criteria even more rare. In addition, it isn’t at all obvious that capable AIs naturally turn into agents, or that their values typically change radically as they grow. Finally, it seems quite unlikely that owners who heavily test and monitor their very profitable but powerful AIs would not even notice such radical changes.


Geoffrey Miller @ 2023-04-15T22:38 (+15)

My fundamental disagreement with Robin Hanson here is that he tends to view AIs either as 'passive, predictably controllable tools of humans' or as 'sentient agents with their own rights & interests'. This dichotomy traces back to our basic human tendency to classify things as either 'objects' or 'people/animals', or 'inanimate' versus 'animate'.

My worry is that the most dangerous near-term AIs will fall into the grey area between these two categories -- they'll have enough agency and autonomy to take powerful, tactically savvy actions on behalf of human individuals and groups telling them what to do (but whose instructions may not be followed accurately), but not quite enough agency and autonomy and wisdom to qualify as sentient agents in their own right, which could be granted rights and responsibilities as 'digital citizens'.

In other words, the most dangerous near-term AIs will be kind of like the henchmen who are given semi-autonomous tasks by evil masterminds in criminal thriller movies. Those henchmen tend to be extremely strong, formidable, and scary, but they don't always follow instructions very well, they tend to be rather over-literal and unimaginative, they often act with impulsive violence that's unaligned with their boss's long-term interests, and they often create more problems than they solve. 'Good help is hard to find', as they say. 

For example, nation-states might use AI henchmen as cyberwarfare agents that attack foreign infrastructure. Now suppose a geopolitical crisis happens. Assume nation A believes there is a clear and present danger form enemy nation-state B. In nation A, there is urgent political and military pressure to 'do something'. The 'fog of war' envelops the crisis situation, limiting the reliability of information, and making rational decision-making difficult. Nation A instructs its newest AI cyberwarfare Agent X to 'degrade nation B's ability to inflict damage on nation A', subject to whatever constraints nation A's leaders happen to think of in the heat of the crisis. Agent X is let lose upon the world.  Now, suppose agent X is not an ASI, or even an AGI; there's been no 'foom'; it's just very very good at cyberwarfare applied to enemy infrastructure, and it can think and act (digitally) a million times faster than its human 'masters' in nation A.  

So Agent X, being the good henchman that it is, sets about wreaking havoc in Nation B. It follows the constraints it's been given, at the literal level, but it doesn't follow their spirit. It hacks whatever control systems it can for cars, trucks, airplanes, ships, subways, rail stations, buildings, installations, ports, traffic control systems, airports, bridges, dams, power plants, military bases, etc. Within a few hours, Nation B is paralyzed by mass chaos, with millions dead. As instructed, Agent X has degraded nation B's ability to inflict damage on Nation A. 

But then Nation B figures out what happened, and they unleash their own henchman, Agent Y, upon Nation A.... leading to a cycle of vengeance with escalating cyber-attacks between major nation-states, and colossal loss of life.

These kinds of scenarios, in my opinion, are much more likely in the next decade or so than a foom-based ASI takeover of the sort often envisioned by AI alignment thinkers. The problem, in short, isn't that an AI becomes an evil genius in its own right. The problem is that AI henchmen, with (metaphorically) the strength of Superman and the speed of the Flash, get unleashed by political or corporate leaders who can't fully anticipate or control what their AI minions end up doing.

more better @ 2023-04-15T02:53 (+3)

Interesting, thanks for sharing. I'm curious about how the distribution of people that would see and vote on this Robin Hanson twitter poll compares with other populations.

Vasco Grilo @ 2023-04-22T15:06 (+2)

Thanks for sharing!

Very lumpy tech advances, techs that broadly improve abilities, and powerful techs that are long kept secret within one project are each quite rare.

It looks the 1st 2 are quite correlated. Self-improvement ability would lead to fast (lumpy?) progress.