The Unilateralist's Curse, An Explanation

By Aaron_Scher @ 2022-06-09T18:25 (+27)

Motivation: This post seeks to explain the Unilateralist’s Curse more thoroughly than is done on the tag and more simply than is done in the original paper. This post might be helpful to anybody trying to do good, and the examples used apply most to those doing AI Safety or community building work.

Epistemic Effort: ~10 hours of reading + some hours writing/revising

 

In short: The Unilateralist’s Curse happens when multiple people could take an action with far reaching consequences, they could do so without the permission of others, and the people are not all in agreement about whether doing the action is good – in this situation it is more likely that the action happens than is ideal. Understanding how to identify and respond to Unilateralist’s Curse situations is important for those trying to do good effectively because it will help them avoid causing accidental harm. 
 

The Unilateralist's Curse occurs when 

The curse: given the above situation, the action is more likely to be taken than is ideal, and the chances of it being taken increase as the number of actors capable of taking the action increases.

Examples: 

Dealing with the Unilateralist's Curse, practical advice: 

  1. Learn more about what’s going on. If you find yourself in the minority opinion of a unilateral scenario as described above, you should seek to find out why the people who aren’t doing the action aren’t doing it. Do they disagree on the empirical likelihood of various outcomes? Do they disagree about how good/bad various outcomes would be? Maybe they haven’t thought about it very much – if so, why not? Maybe they think it’s a good idea but that their comparative advantage lies elsewhere? Probably they have some information that you don’t have which is altering their decision – knowing this information would be useful! Maybe there are information hazards associated with the action and they can’t tell you about said hazards – if so maybe you should probably just trust their judgment and read below. 
  2. Take actions with the agreement of others. The heuristic of “comply in action, defy in thought,” or “think independently and act collectively” applies here. In unilateral situations, you can often coordinate with others; you should probably do this. There are some cases where this does not apply, but if you are going to do a thing unilaterally, you should definitely know why others haven’t done it yet.
  3. Relatedly, Use group decision making processes. For example, make an agreement ahead of time that you and the other actors in the unilateralist situation will deliberate on the action, take a vote, and you all abide by the majority’s decision to act or not to act. You could also create a single decision maker responsible for centrally planning everything – for instance consolidating all the grant-making to one person – but this has significant tradeoffs. 
  4. Simply defer to what the group is doing. If nobody has published the genome for SuperVirus35, don’t publish the genome for SuperVirus35. Again, it’s better if you can understand why others haven’t taken the action, but it’s often the safer option to defer to others.
  5. Recognize that you’re in a Unilateralist's Curse scenario and respond accordingly. This means reflecting on the fact that your views vary from others’ views and are subject to error. You should decrease the probability that you do an action accordingly. 
  6. When you can’t talk talk about it: If you can’t talk about the fact that you are in this position or can’t get information about others in this position (maybe you have the blueprint for an AGI and you’re considering deploying it; if you were in this position you probably wouldn’t want to tell many people because it would lead to people trying to steal it from you or accelerating timelines as everybody else realizes AGI is easily within reach, also other adversarial dynamics): ask yourself how likely it is that others are in a similar position. If others have the blueprint for AGI and haven’t deployed it, there’s probably a good reason. Generally, the more people in the unilateralist position, the higher the chance that somebody does the action. Therefore, if lots of people are in the position but nobody has done it, you should update your belief away from doing the action yourself because others have deemed it not worth doing. 

A little bit of nuance:

More examples from AI safety:

 

The EA community sometimes promotes a culture of “if you see a promising thing that could be done, go do it.” But this has a tradeoff with the Unilateralist's Curse. Probably somebody else has already thought of this – why didn’t they do it? If you find yourself in a unilateralist situation, first think, see if anybody else has done the thing, then consider talking to others about it (do so unless there’s huge info hazards), and try to learn why nobody else has done the thing. Please don’t kill us all. 

Thank you for reading! Please feel free to send me any feedback, positive or negative. Thank you to Lizka for feedback on an earlier draft.


 


Jakub Kraus @ 2023-03-03T04:11 (+1)

You could think about a random normal distribution of estimated-value clustered around the true value of the action. The more actors (estimated-values you draw from the normal distribution) the more likely you are to get an outlier who thinks the value is positive when it is actually negative. 

Additionally, the people willing to act unilaterally are more likely to have positively biased estimates of the value of the action:

As you noted, some curse arguments are symmetric, in the sense that they also provide reason to expect unilateralists to do more good. Notably, the bias above is asymmetric; it provides a reason to expect unilateralists to do less good, with no corresponding reason to expect unilateralists to do more good.

Jáchym Fibír @ 2023-10-22T12:34 (+1)

@JakubK I think that your interpretation of OP's quote is somewhat less useful, in the sense that it only retroactively explains a behavior of a "unilateralist" - i.e. an actor that had already made a decision. I find that for the purposes of a generic actor, it has less utility to ask themselves "Am I (acting as) a unilateralist?" instead of "How many other actors are there capable of, yet not acting?"

Beside the abstracntess of it, I see that there is actually quite some overlap with what you would call "unilateralists" and simply "courageous actors". This is because there usually is what I would refer to as an "activation energy" for actions to be made, which is basically a bias where the majority of actors are more likely not to act when the true value of an action is net neutral or slightly positive. And precisely in the scenario where the true value of an action is only slightly (but significantly) positive, you would then need a courageous actor to "overcome the activation energy" and do the right thing. 
 

I'm not sure if activation energy is the right term, but it is an observed phenomenon in ethics, see the comparison of the Fat Man with the classical Trolley Switch scenario. To sum it up, I think that looking at the unilateralist's curse simply as a statistical phenomenon is fundamentally wrong, and one must include moral dimensions in deciding whether to take the action no one has taken despite having the option to. That said, the general advice in the post like information sharing etc. are indeed helpful and applicable.