Ethics of artificial intelligence

From Wikipedia, the free encyclopedia
Jump to: navigation, search

The ethics of artificial intelligence is the part of the ethics of technology specific to robots and other artificially intelligent beings. It is typically divided into roboethics, a concern with the moral behavior of humans as they design, construct, use and treat artificially intelligent beings, and machine ethics, concern with the moral behavior of artificial moral agents (AMAs).

Roboethics[edit]

The term "roboethics" was coined by roboticist Gianmarco Veruggio in 2002, referring to the morality of how humans design, construct, use and treat robots and other artificially intelligent beings.[1] It considers both how artificially intelligent beings may be used to harm humans and how they may be used to benefit humans.

Robot rights[edit]

Robot rights are the moral obligations of society towards its machines, similar to human rights or animal rights. These may include the right to life and liberty, freedom of thought and expression and equality before the law.[2] The issue has been considered by the Institute for the Future[3] and by the U.K. Department of Trade and Industry.[4]

Experts disagree whether specific and detailed laws will be required soon or safely in the distant future.[4] Glenn McGee reports that sufficiently humanoid robots may appear by 2020.[5] Ray Kurzweil sets the date at 2029.[6] However, most scientists suppose that at least 50 years may have to pass before any sufficiently advanced system exists.[7][8]

The rules for the 2003 Loebner Prize competition explicitly addressed the question of robot rights:

61. If, in any given year, a publicly available open source Entry entered by the University of Surrey or the Cambridge Center wins the Silver Medal or the Gold Medal, then the Medal and the Cash Award will be awarded to the body responsible for the development of that Entry. If no such body can be identified, or if there is disagreement among two or more claimants, the Medal and the Cash Award will be held in trust until such time as the Entry may legally possess, either in the United States of America or in the venue of the contest, the Cash Award and Gold Medal in its own right.[9]

The threat to privacy[edit]

Aleksandr Solzhenitsyn's The First Circle describes the use of speech recognition technology in the service of tyranny.[10] If an AI program exists that can understand speech and natural languages (e.g. English), then, with adequate processing power it could theoretically listen to every phone conversation and read every email in the world, understand them and report back to the program's operators exactly what is said and exactly who is saying it. An AI program like this could allow governments or other entities to efficiently suppress dissent and attack their enemies.

The threat to human dignity[edit]

Joseph Weizenbaum argued in 1976 that AI technology should not be used to replace people in positions that require respect and care, such as:

  • A customer service representative, (AI technology is already used today for telephone-based interactive voice response systems)
  • A therapist, (as was seriously proposed by Kenneth Colby in the 1970s)
  • A nursemaid for the elderly, (as was reported by Pamela McCorduck in her book The Fifth Generation)
  • A judge, or
  • A police officer.

Weizenbaum explains that we require authentic feelings of empathy from people in these positions. If machines replace them, we will find ourselves alienated, devalued and frustrated. Artificial intelligence, if used in this way, represents a threat to human dignity. Weizenbaum argues that fact that we are entertaining the possibility of machines in these positions suggests that we have experienced an "atrophy of the human spirit that comes from thinking of ourselves as computers."[11]

Pamela McCorduck counters that, speaking for women and minorities "I'd rather take my chances with an impartial computer," pointing out that there are conditions where we would prefer to have automated judges and police that have no personal agenda at all.[11] AI founder John McCarthy objects to the moralizing tone of Weizenbaum's critique. "When moralizing is both vehement and vague, it invites authoritarian abuse," he writes.[11]

Machine ethics[edit]

Machine ethics (or machine morality) is the field of research concerned with designing Artificial Moral Agents (AMAs), robots or artificially intelligent computers that behave morally or as though moral.[12][13][14][15]

Isaac Asimov considered the issue in the 1950s in his I, Robot. At the insistence of his editor John W. Campbell Jr., he proposed the Three Laws of Robotics to govern artificially intelligent systems. Much of his work was then spent testing the boundaries of his three laws to see where they would break down, or where they would create paradoxical or unanticipated behavior. His work suggests that no set of fixed laws can sufficiently anticipate all possible circumstances.[16]

In 2009, during an experiment at the Laboratory of Intelligent Systems in the Ecole Polytechnique Fédérale of Lausanne in Switzerland, robots that were programmed to cooperate with each other in searching out a beneficial resource and avoiding a poisonous one eventually learned to lie to each other in an attempt to hoard the beneficial resource.[17] One problem in this case may have been that the goals were "terminal" (i.e. in contrast, ultimate human motives typically have a quality of requiring never-ending learning).[18]

Some experts and academics have questioned the use of robots for military combat, especially when such robots are given some degree of autonomous functions.[19] The US Navy has funded a report which indicates that as military robots become more complex, there should be greater attention to implications of their ability to make autonomous decisions.[20][21] The President of the Association for the Advancement of Artificial Intelligence has commissioned a study to look at this issue.[22] They point to programs like the Language Acquisition Device which can emulate human interaction.

Vernor Vinge has suggested that a moment may come when some computers are smarter than humans. He calls this "the Singularity."[23] He suggests that it may be somewhat or possibly very dangerous for humans.[24] This is discussed by a philosophy called Singularitarianism. The Singularity Institute for Artificial Intelligence has suggested a need to build "Friendly AI", meaning that the advances which are already occurring with AI should also include an effort to make AI intrinsically friendly and humane.[25]

In 2009, academics and technical experts attended a conference to discuss the potential impact of robots and computers and the impact of the hypothetical possibility that they could become self-sufficient and able to make their own decisions. They discussed the possibility and the extent to which computers and robots might be able to acquire any level of autonomy, and to what degree they could use such abilities to possibly pose any threat or hazard. They noted that some machines have acquired various forms of semi-autonomy, including being able to find power sources on their own and being able to independently choose targets to attack with weapons. They also noted that some computer viruses can evade elimination and have achieved "cockroach intelligence." They noted that self-awareness as depicted in science-fiction is probably unlikely, but that there were other potential hazards and pitfalls.[23]

In Moral Machines: Teaching Robots Right from Wrong,[26] Wendell Wallach and Colin Allen conclude that attempts to teach robots right from wrong will likely advance understanding of human ethics by motivating humans to address gaps in modern normative theory and by providing a platform for experimental investigation. As one example, it has introduced normative ethicists to the controversial issue of which specific learning algorithms to use in machines. Nick Bostrom and Eliezer Yudkowsky have argued for decision trees (such as ID3) over neural networks and genetic algorithms on the grounds that decision trees obey modern social norms of transparency and predictability (e.g. stare decisis),[27] while Chris Santos-Lang argued in the opposite direction on the grounds that the norms of any age must be allowed to change and that natural failure to fully satisfy these particular norms has been essential in making humans less vulnerable to criminal "hackers".[18]

Unintended consequences[edit]

Several critics have argued that AI technology has the potential to disrupt existing society and introduce new dangers and malaise. Nick Bostrom, Teacher and Philosopher at Oxford University, published a paper "Existential Risks" In the Journal of Evolution and Technology. Bostrom stated that Artificial Intelligence has the capability to bring about human extinction, which is of course not what society intends for artificial intelligence to do.

Berglas (2008) [28] notes that there is no direct evolutionary motivation for an AI entity to be friendly to humans. Further, the AI entity would have its own evolutionary pressures as it competes with other AI entities for computer hardware upon which to run. It is difficult to see how humanity could survive in competition with a more intelligent machine.

Ethics of artificial intelligence in fiction[edit]

The movie The Thirteenth Floor suggests a future where simulated worlds with sentient inhabitants are created by computer game consoles for the purpose of entertainment. The movie The Matrix suggests a future where the dominant species on planet Earth are sentient machines and humanity is treated with utmost Speciesism. The short story The Planck Dive suggest a future where humanity has turned itself into software that can be duplicated and optimized and the relevant distinction between types of software is sentient and non-sentient. The same idea can be found in the Emergency Medical Hologram of Starship Voyager, which is an apparently sentient copy of a reduced subset of the consciousness of its creator, Dr. Zimmerman, who, for the best motives, has created the system to give medical assistance in case of emergencies. The movies Bicentennial Man and A.I. deal with the possibility of sentient robots that could love. I, Robot explored some aspects of Asimov's three laws. All these scenarios try to foresee possibly unethical consequences of the creation of sentient computers.

The ethics of artificial intelligence is one of several core themes in BioWare's Mass Effect series of games. It explores the scenario of a civilization accidentally creating AI through a rapid increase in computational power through a global scale neural network. This event caused an ethical schism between those who felt bestowing organic rights upon the newly sentient Geth was appropriate and those who continued to see them as disposable machinery and fought to destroy them. Beyond the initial conflict, the complexity of the relationship between the machines and their creators is another ongoing theme throughout the story.

Over time, debates have tended to focus less and less on possibility and more on desirability, as emphasized in the "Cosmist" and "Terran" debates initiated by Hugo de Garis and Kevin Warwick. A Cosmist, according to Hugo de Garis, is actually seeking to build more intelligent successors to the human species.

See also[edit]

Notes[edit]

  1. ^ Veruggio, Gianmarco (2007). The Roboethics Roadmap. Scuola di Robotica. p. 2. Retrieved 28 March 2011. 
  2. ^ The American Heritage Dictionary of the English Language, Fourth Edition
  3. ^ "Robots could demand legal rights". BBC News. December 21, 2006. Retrieved January 3, 2010. 
  4. ^ a b Henderson, Mark (April 24, 2007). "Human rights for robots? We're getting carried away". The Times Online (The Times of London). Retrieved May 2, 2010. 
  5. ^ McGee, Glenn. "A Robot Code of Ethics". The Scientist. 
  6. ^ Kurzweil, Ray (2005). The Singularity is Near. Penguin Books. ISBN 0-670-03384-7. 
  7. ^ New World Technologies - NWT - Should we be worried by the rise of robots?
  8. ^ shuld-the-human-race-be-worried-by-the-rise-of-robots, Independent Newspaper,
  9. ^ Loebner Prize Contest Official Rules — Version 2.0 The competition was directed by David Hamill and the rules were developed by members of the Robitron Yahoo group.
  10. ^ (McCorduck 2004, p. 308)
  11. ^ a b c Joseph Weizenbaum, quoted in McCorduck 2004, pp. 356, 374–376
  12. ^ Anderson. "Machine Ethics". Retrieved 27 June 2011. 
  13. ^ Anderson, Michael; Anderson, Susan Leigh, eds. (July 2011). Machine Ethics. Cambridge University Press. ISBN 978-0-521-11235-2. 
  14. ^ Anderson, Michael; Anderson, Susan Leigh, eds. (July–August 2006). "Special Issue on Machine Ethics". IEEE Intelligent Systems 21 (4): 10–63. ISSN 1541-1672. 
  15. ^ Anderson, Michael; Anderson, Susan Leigh (Winter 2007). "Machine Ethics: Creating an Ethical Intelligent Agent". AI Magazine (American Association for Artificial Intelligence) 28 (4): 15–26. ISSN 0738-4602. 
  16. ^ Asimov, Isaac (2008). I, robot. New York: Bantam. ISBN 0-553-38256-X. 
  17. ^ Evolving Robots Learn To Lie To Each Other, Popular Science, August 18, 2009
  18. ^ a b Santos-Lang, Chris (2002). "Ethics for Artificial Intelligences". 
  19. ^ Call for debate on killer robots, By Jason Palmer, Science and technology reporter, BBC News, 8/3/09.
  20. ^ Science New Navy-funded Report Warns of War Robots Going "Terminator", by Jason Mick (Blog), dailytech.com, February 17, 2009.
  21. ^ Navy report warns of robot uprising, suggests a strong moral compass, by Joseph L. Flatley engadget.com, Feb 18th 2009.
  22. ^ AAAI Presidential Panel on Long-Term AI Futures 2008-2009 Study, Association for the Advancement of Artificial Intelligence, Accessed 7/26/09.
  23. ^ a b Scientists Worry Machines May Outsmart Man By JOHN MARKOFF, NY Times, July 26, 2009.
  24. ^ The Coming Technological Singularity: How to Survive in the Post-Human Era, by Vernor Vinge, Department of Mathematical Sciences, San Diego State University, (c) 1993 by Vernor Vinge.
  25. ^ Article at Asimovlaws.com, July 2004, accessed 7/27/09.
  26. ^ Wallach, Wendell; Allen, Colin (November 2008). Moral Machines: Teaching Robots Right from Wrong. USA: Oxford University Press. ISBN 978-0-19-537404-9. 
  27. ^ Bostrom, Nick; Yudkowsky, Eliezer (2011). "The Ethics of Artificial Intelligence" (PDF). Cambridge Handbook of Artificial Intelligence. Cambridge Press. 
  28. ^ Berglas, Anthony (2008), Artificial Intelligence will Kill our Grandchildren, retrieved 2008-06-13 

External links[edit]