[Link and commentary] Beyond Near- and Long-Term: Towards a Clearer Account of Research Priorities in AI Ethics and Society
By MichaelAđ¸ @ 2020-03-14T09:04 (+18)
Earlier this year, Carina Prunkl (FHI) and Jess Whittlestone (Leverhulme Centre for the Future of Intelligence) put out a paper called Beyond Near- and Long-Term: Towards a Clearer Account of Research Priorities in AI Ethics and Society. The paper itself is 6 pages long and easy to read, and I thought it made several interesting and probably valuable points, so you may wish to just read the whole thing. But here Iâll summarise what I saw as the key points, and then provide my own commentary on some specific parts of the paper.
The abstract
One way of carving up the broad âAI ethics and societyâ research space that has emerged in recent years is to distinguish between ânear-termâ and âlong-termâ research. While such ways of breaking down the research space can be useful, we put forward several concerns about the near/long-term distinction gaining too much prominence in how research questions and priorities are framed. We highlight some ambiguities and inconsistencies in how the distinction is used, and argue that while there are differing priorities within this broad research community, these differences are not well-captured by the near/long-term distinction. We unpack the near/long-term distinction into four different dimensions, and propose some ways that researchers can communicate more clearly about their work and priorities using these dimensions. We suggest that moving towards a more nuanced conversation about research priorities can help establish new opportunities for collaboration, aid the development of more consistent and coherent research agendas, and enable identification of previously neglected research areas.
The four dimensions
To me, a lot of the paperâs value came from the following subsection about what those âfour different dimensionsâ actually are:
As commonly used, the terms ânear-termâ and âlong-termâ in fact appear to capture four different dimensions of differing priorities within the AI [ethics and society] research community:
-Capabilities: whether to focus on the impacts and challenges of current AI systems, or those relating to much more advanced AI systems
-Impacts: whether to focus mostly on the immediate impacts of AI for society, or whether to consider possible impacts much further into the future.
-Certainty: whether to focus on impacts and issues that are relatively certain and well-understood, or those that are more uncertain and speculative.
-Extremity: whether to focus on impacts at all scales, or to prioritise focusing on those that may be particularly large in scale.
None of these four dimensions are binary: one can choose research questions that focus on AI systems that are more or less advanced, exploring impacts on varying time horizons, with varying degrees of certainty and extremity. [...]
Of course, these dimensions are not entirely independent of one another: for example, if one wants to focus on particularly large-scale impacts of AI, one may have to be willing to consider more speculative questions. However, it is still useful to unpack these different dimensions, particularly because there are many possible views and research questions which are not easily captured by the near/long-term distinction as commonly used.
Table 1 and Figure 1 also interestingly represent how these dimensions can be unpacked and disentangled, so Iâd suggest checking those out.
My commentary
Long-term impacts from near-term capabilities?
Prunkl and Whittlestone write:
Of course the timescale of technological advances and their impacts on society will be related. However, conflating long-term capabilities and impacts may mean the research community neglect important questions about the potential long-term impacts of current AI systems and their applications in society. For example, in what ways could injustices perpetuated by increased use of current algorithmic systems have very long-lasting and irreversible consequences for inequality in society? What are the possible longer-term implications of how data-intensive AI applications are beginning to change norms around privacy and personal data? [23]
I think itâs interesting and valuable to note that timescales âof technological advances and their impacts on societyâ can be disaggregated. I also appreciate that the authors note that these timescales will be related.
But I think I suspect the relationship between these timescales to be stronger than the authors do. Personally, I donât see the questions they raise as compelling examples of potential long-term impacts from near-term capabilities. Itâs not apparent to me how âinjustices perpetuated by increased use of current algorithmic systemsâ could âhave very long-lasting and irreversible consequences for inequality in societyâ. Nor is it apparent to me how similarly long-lasting effects could result from âhow data-intensive AI applications are beginning to change norms around privacy and personal dataâ.
One potential explanation for this: In the context of AI and longtermism, I take âvery long-lastingâ to mean something like âlasting at least centuries and perhaps millions of yearsâ. The authors may instead mean the phrase as something like âat least decades, and upwards of a century or twoâ. But if thatâs the case, then I donât think this is merely a linguistic confusion, because then what they mean by âlong-term impacts of AI systemsâ is not really the key thing Iâm concerned about, so I wonât necessarily be very concerned if those questions get limited attention.
Another potential explanation for this: Perhaps there actually is some mechanism by which those current AI systems and applications could have impacts that last at substantial strength for at least centuries and perhaps millions of years. Or perhaps thereâs at least a high enough chance that thereâs such a mechanism for it to indeed be worth giving those questions more attention, even from longtermists.
As noted, this seems quite unlikely to me. But the authorsâ point is precisely that the research community may have neglected these sorts of questions, and itâs true that I havenât spent much time learning or thinking about them. So Iâm open to being convinced on this point, though I donât see that as likely.
Each dimension is continuous, not binary
I think that this point is true and important. Iâd like to highlight in particular one important implication the authors draw from this point:
it is important to recognise that both capabilities and impacts lie on a spectrum between near- and long-term. [...] there are many ways in which AI systems could become more advanced over the coming years, before reaching anything close to âsuperintelligenceâ or âAGIâ. We need to consider what issues these intermediate advances might raise, and what kinds of intermediate advances in capabilities are mostly likely to be of harm or benefit for society.
âAI ethics and societyâ
There is a growing community of researchers focused on ensuring that advances in AI are safe and beneficial for humanity (which we call the AI ethics and society community, or AI E&S for short.) The AI E&S research space is very broad, encompassing many different questions: from how AI is likely to impact society in practice to what ethical issues that might raise, and what technical and governance solutions might be needed.
It makes sense to me to label a conference as a âConference on AI, Ethics, and Societyâ. But it doesnât seem to me that âthe AI ethics and society communityâ is a good label for the âgrowing community of researchers focused on ensuring that advances in AI are safe and beneficial for humanityâ. (Note that Iâm unsure whether the authors are coining or just adopting that usage of this term.)
My core objection is that that label doesnât seem to neatly capture technical AI safety work.
But I also feel like itâs a bit strange to explicitly include âsocietyâ in the label. It does seem true that AIâs impacts will be partly a function of how AI interacts with and is used by society. But that seems like it can be taken as implied, as itâs also true of a very wide range of things, and we donât use labels like âbioengineering and societyâ or âclimate change and societyâ. (Though one could reply that in fact we should use labels like that for those cases as well, to highlight how those issues interact with society.)
So I think Iâd prefer to stick with something like âAI safety and governanceâ, or âAI alignmentâ (see also MacAskillâs three-part breakdown of âthe alignment problemâ).
It might be that the authors are intentionally coining/adopting a separate term in order to point at a somewhat distinct community, which has less emphasis on the technical safety work and/or less emphasis on uncertain, extreme impacts from long-term capabilities. But it seems to me that people focused on the more technical and longtermist work are clearly part of the âgrowing community of researchers focused on ensuring that advances in AI are safe and beneficial for humanityâ.
Prioritarianism/egalitarianism and longtermism
(Tangential philosophical point I may be wrong about.)
Further dimensions of disagreement may help explain why the AI E&S community has divided in certain ways. Disagreement on normative issues may be relevant here, such as around whether we have a special moral obligation to help those who are alive today over those who will live in future [26], or to prioritise helping those worst off in society [25]. Someone who holds the more fundamental philosophical belief that we should prioritise helping the worst off in society, for example, is likely to choose to work on the implications of AI for global fairness or social justice, regardless of their position on the certainty/extremity tension as outlined above.
I think that all of the above statements are true. But I disagree with one claim that could be inferred from that last sentence (and which may or may not also have been what the authors meant): The claim that a belief in the intrinsic value of promoting equality, or of improving the wellbeing of worse-off rather than better-off people, would itself or necessarily lead people to a focus on global fairness or social justice (as opposed to a focus on things like existential risk reduction).
I havenât really thought about this before, but it seems to me that a focus on existential risk reduction could well be compatible with, or possibly even follow from, such an egalitarian or prioritarian ethic, as long as you also hold some combination of the following views:
- rejection of person-affecting views (though see also this post)
- rejection of pure time discounting for moral matters
- belief that death counts as âworse for a personâ than the issues that could be alleviated by work towards global fairness or social justice
- belief that some existential risk reduction efforts have a non-negligible chance of preventing a âsuffering riskâ, long-lasting totalitarianism, or something like that, in which there would be many people whoâd suffer with similar or greater intensity as the people who are currently worst-off
Making beliefs and disagreements more clear and explicit
Based on this analysis, we have a few concrete recommendations for how the AI E&S research community can encourage more nuanced and productive discussion about different priorities and assumptions:
[...]
-Make an effort to understand the underlying motivations and assumptions of others with different research priorities, again using some of the questions outlined in the previous section as a starting point. Conferences, workshops and journals could potentially help support this by providing fora for researchers to debate such fundamental questions and disagreements, improving mutual understanding. (emphasis in original)
I strongly agree that this would be valuable; in fact, a decent portion of the work Iâve recently been involved with has had something similar as one of its primary goals (e.g., this and this). I also think that this paper by Prunkl and Whittlestone should indeed help with that goal.
See also Rohin Shahâs summary of and commentary on the paper here, which I found after writing this but before posting it.