Katana VentraIP

Friendly artificial intelligence

Friendly artificial intelligence (also friendly AI or FAI) is hypothetical artificial general intelligence (AGI) that would have a positive (benign) effect on humanity or at least align with human interests or contribute to fostering the improvement of the human species. It is a part of the ethics of artificial intelligence and is closely related to machine ethics. While machine ethics is concerned with how an artificially intelligent agent should behave, friendly artificial intelligence research is focused on how to practically bring about this behavior and ensuring it is adequately constrained.

Coherent extrapolated volition[edit]

Yudkowsky advances the Coherent Extrapolated Volition (CEV) model. According to him, our coherent extrapolated volition is "our wish if we knew more, thought faster, were more the people we wished we were, had grown up farther together; where the extrapolation converges rather than diverges, where our wishes cohere rather than interfere; extrapolated as we wish that extrapolated, interpreted as we wish that interpreted".[16]


Rather than a Friendly AI being designed directly by human programmers, it is to be designed by a "seed AI" programmed to first study human nature and then produce the AI which humanity would want, given sufficient time and insight, to arrive at a satisfactory answer.[16] The appeal to an objective through contingent human nature (perhaps expressed, for mathematical purposes, in the form of a utility function or other decision-theoretic formalism), as providing the ultimate criterion of "Friendliness", is an answer to the meta-ethical problem of defining an objective morality; extrapolated volition is intended to be what humanity objectively would want, all things considered, but it can only be defined relative to the psychological and cognitive qualities of present-day, unextrapolated humanity.

Public policy[edit]

James Barrat, author of Our Final Invention, suggested that "a public-private partnership has to be created to bring A.I.-makers together to share ideas about security—something like the International Atomic Energy Agency, but in partnership with corporations." He urges AI researchers to convene a meeting similar to the Asilomar Conference on Recombinant DNA, which discussed risks of biotechnology.[17]


John McGinnis encourages governments to accelerate friendly AI research. Because the goalposts of friendly AI are not necessarily eminent, he suggests a model similar to the National Institutes of Health, where "Peer review panels of computer and cognitive scientists would sift through projects and choose those that are designed both to advance AI and assure that such advances would be accompanied by appropriate safeguards." McGinnis feels that peer review is better "than regulation to address technical issues that are not possible to capture through bureaucratic mandates". McGinnis notes that his proposal stands in contrast to that of the Machine Intelligence Research Institute, which generally aims to avoid government involvement in friendly AI.[20]

Yudkowsky, E. . In Global Catastrophic Risks, Oxford University Press, 2008.
Discusses Artificial Intelligence from the perspective of Existential risk. In particular, Sections 1-4 give background to the definition of Friendly AI in Section 5. Section 6 gives two classes of mistakes (technical and philosophical) which would both lead to the accidental creation of non-Friendly AIs. Sections 7-13 discuss further related issues.

Artificial Intelligence as a Positive and Negative Factor in Global Risk

Omohundro, S. 2008 The Basic AI Drives Appeared in AGI-08 - Proceedings of the First Conference on Artificial General Intelligence

Mason, C. 2008 Archived 2022-01-09 at the Wayback Machine Appears in AAAI 2008 Workshop on Meta-Reasoning:Thinking About Thinking

Human-Level AI Requires Compassionate Intelligence

Froding, B. and Peterson, M 2021 Ethics and Information Technology volume 23, pp 207–214.

Friendly AI

by Nick Bostrom

Ethical Issues in Advanced Artificial Intelligence

— A brief description of Friendly AI by the Machine Intelligence Research Institute.

What is Friendly AI?

— A near book-length description from the MIRI

Creating Friendly AI 1.0: The Analysis and Design of Benevolent Goal Architectures

— by Bill Hibbard

Critique of the MIRI Guidelines on Friendly AI

— by Peter Voss.

Commentary on MIRI's Guidelines on Friendly AI

— On the motives for and impossibility of FAI; by Adam Keiper and Ari N. Schulman.

The Problem with ‘Friendly’ Artificial Intelligence