Eliezer Yudkowsky

From Wikipedia, the free encyclopedia
Jump to: navigation, search
Eliezer Yudkowsky
Eliezer Yudkowsky, Stanford 2006 (square crop).jpg
Eliezer Yudkowsky at the 2006 Stanford Singularity Summit
Born (1979-09-11) September 11, 1979 (age 34)
Nationality American
Fields Machine ethics
Institutions Machine Intelligence Research Institute
Known for Friendly AI,[citation needed] Harry Potter fan fiction
Influences Judea Pearl, Vernor Vinge, E.T. Jaynes, I. J. Good[citation needed]

Eliezer Shlomo Yudkowsky (born September 11, 1979[citation needed]) is an American blogger, writer, and advocate for Friendly artificial intelligence.

Biography[edit]

Yudkowsky is a resident of Berkeley, California. Largely self-educated [1]:38, he co-founded the nonprofit Machine Intelligence Research Institute (formerly the Singularity Institute for Artificial Intelligence) in 2000 and continues to be employed there as a full-time Research Fellow.[2]:599

Work[edit]

Yudkowsky's interests focus on Artificial Intelligence theory for self-understanding, self-modification, and recursive self-improvement, and on artificial-intelligence architectures and decision theories for stable motivational structures (Friendly AI and Coherent Extrapolated Volition in particular).[2]:420 Yudkowsky's most recent work is on decision theory for problems of self-modification and Newcomblike problems.[clarification needed]

Yudkowsky was, along with Robin Hanson, one of the principal contributors to the blog Overcoming Bias[3][non-primary source needed] sponsored by the Future of Humanity Institute of Oxford University. In early 2009[citation needed], he helped to found LessWrong, a "community blog devoted to refining the art of human rationality".[1]:37

Yudkowsky contributed two chapters to Oxford philosopher Nick Bostrom's and Milan Ćirković's edited volume Global Catastrophic Risks.[4]

Fan fiction[edit]

Yudkowsky has also written several works[citation needed] of science fiction and other fiction. His lengthy Harry Potter fan fiction story Harry Potter and the Methods of Rationality illustrates topics in cognitive science and rationality[1]:37 (The New Yorker described it as "recast[ing] the original story in an attempt to explain Harry's wizardry through the scientific method"[5]), and has been reviewed by authors David Brin[6] and Rachel Aaron,[7][8] Robin Hanson,[9] Aaron Swartz,[10] and by programmer Eric S. Raymond.[11]

References[edit]

  1. ^ a b c Singularity Rising, by James Miller
  2. ^ a b Kurzweil, Ray (2005). The Singularity Is Near. New York, US: Viking Penguin. ISBN 0-670-03384-7. 
  3. ^ "Overcoming Bias: About". Robin Hanson. Retrieved 2012-02-01. 
  4. ^ Bostrom, Nick; Ćirković, Milan M., eds. (2008). Global Catastrophic Risks. Oxford, UK: Oxford University Press. pp. 91–119, 308–345. ISBN 978-0-19-857050-9. 
  5. ^ pg 54, "No Death, No Taxes: The libertarian futurism of a Silicon Valley billionaire"
  6. ^ David Brin (2010-06-21). "CONTRARY BRIN: A secret of college life... plus controversies and science!". Davidbrin.blogspot.com. Retrieved 2012-08-31. 
  7. ^ Authors (2012-04-02). "Rachel Aaron interview (April 2012)". Fantasybookreview.co.uk. Retrieved 2012-08-31. 
  8. ^ "Civilian Reader: An Interview with Rachel Aaron". Civilian-reader.blogspot.com. 2011-05-04. Retrieved 2012-08-31. 
  9. ^ Hanson, Robin (2010-10-31). "Hyper-Rational Harry". Overcoming Bias. Retrieved 2012-08-31. 
  10. ^ Swartz, Aaron. "The 2011 Review of Books (Aaron Swartz's Raw Thought)". archive.org. Retrieved 2013-04-10. 
  11. ^ "Harry Potter and the Methods of Rationality". Esr.ibiblio.org. 2010-07-06. Retrieved 2012-08-31. 

Publications[edit]

Further reading[edit]

  • Our Molecular Future: How Nanotechnology, Robotics, Genetics and Artificial Intelligence Will Transform Our World by Douglas Mulhall, 2002, p. 321.
  • The Spike: How Our Lives Are Being Transformed By Rapidly Advancing Technologies by Damien Broderick, 2001, pp. 236, 265-272, 289, 321, 324, 326, 337-339, 345, 353, 370.

External links[edit]