Jump to content

Eliezer Yudkowsky: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
Reverted to revision 590806797 by 59.180.19.218 (talk). (TW)
see Talk
Line 1: Line 1:
{{Notability|Academics|date=December 2013}}{{Infobox scientist
|name = Eliezer Yudkowsky
|image = Eliezer_Yudkowsky,_Stanford_2006_(square_crop).jpg
|image_size = 250px
|caption = Eliezer Yudkowsky at the 2006 Stanford [[Singularity Summit]]
|birth_date = {{Birth date and age|1979|9|11|mf=y}}
|birth_place =
|death_date =
|death_place =
|residence =
|citizenship = American
|nationality = American
|field = [[Machine ethics]]
|work_institutions = [[Machine Intelligence Research Institute]]
|known_for = [[Seed AI]], [[Friendly AI]], timeless decision theory,<ref name="intelligence.org">{{cite web|url=http://intelligence.org/files/Comparison.pdf|title=A Comparison of Decision Algorithms on Newcomblike Problems|accessdate=2013-08-26|publisher=Machine Intelligence Research Institute}}</ref> Harry Potter and the Methods of Rationality<ref>[http://www.fanfiction.net/s/5782108/1/Harry-Potter-and-the-Methods-of-Rationality Harry Potter and the Methods of Rationality]</ref>
|author_abbrev_bot =
|author_abbrev_zoo =
|influences = [[Judea Pearl]], [[Vernor Vinge]], [[E.T. Jaynes]], [[I. J. Good]]
|influenced =
|prizes =
|religion = [[Atheist]]<ref>{{cite web|url=http://www.goodreads.com/quotes/show/533670|title=Quote by Eliezer Yudkowsky|publisher=goodreads.com|accessdate=17 July 2012|author=Eliezer Yudkowsky|quote=[...] intelligent people only have a certain amount of time (measured in subjective time spent thinking about religion) to become atheists. After a certain point, if you're smart, have spent time thinking about and defending your religion, and still haven't escaped the grip of Dark Side Epistemology, the inside of your mind ends up as an Escher painting.}}</ref>
|footnotes =
|signature =
}}

'''Eliezer Shlomo Yudkowsky''' (born September 11, 1979<ref>[http://www.goodreads.com/author/show/4533716.Eliezer_Yudkowsky Goodreads author page]</ref>) is an American blogger, writer, and advocate for [[Friendly artificial intelligence]].<ref name=SingRising>{{cite book|last=Miller|first=James|title=Singularity Rising|year=2012|publisher=BenBella Books|location=Texas|isbn=1936661659|pages=35–44|url=http://www.singularityrising.com/}}</ref><ref name="singinst">{{cite web|url=http://www.singinst.org/aboutus/team|title=Singularity Institute for Artificial Intelligence: Team|publisher=Singularity Institute for Artificial Intelligence|accessdate = 2009-07-16}}</ref>
'''Eliezer Shlomo Yudkowsky''' (born September 11, 1979<ref>[http://www.goodreads.com/author/show/4533716.Eliezer_Yudkowsky Goodreads author page]</ref>) is an American blogger, writer, and advocate for [[Friendly artificial intelligence]].<ref name=SingRising>{{cite book|last=Miller|first=James|title=Singularity Rising|year=2012|publisher=BenBella Books|location=Texas|isbn=1936661659|pages=35–44|url=http://www.singularityrising.com/}}</ref><ref name="singinst">{{cite web|url=http://www.singinst.org/aboutus/team|title=Singularity Institute for Artificial Intelligence: Team|publisher=Singularity Institute for Artificial Intelligence|accessdate = 2009-07-16}}</ref>


Line 45: Line 20:
Yudkowsky has also written several works<ref>{{cite web|url=http://yudkowsky.net/other/fiction|title=Yudkowsky- Fiction|publisher=Eliezer Yudkowsky}}</ref> of science fiction and other fiction. His [[Harry Potter]] [[fan fiction]] story ''Harry Potter and the Methods of Rationality'' illustrates topics in [[cognitive science]] and [[rationality]] (''[[The New Yorker]]'' described it as "a thousand-page online 'fanfic' text called 'Harry Potter and the Methods of Rationality', which recasts the original story in an attempt to explain Harry's wizardry through the scientific method"<ref>pg 54, [http://www.newyorker.com/reporting/2011/11/28/111128fa_fact_packer "No Death, No Taxes: The libertarian futurism of a Silicon Valley billionaire"]</ref>), and has been reviewed by authors [[David Brin]]<ref>{{cite web|author=David Brin |url=http://davidbrin.blogspot.com/2010/06/secret-of-college-life-plus.html |title=CONTRARY BRIN: A secret of college life... plus controversies and science! |publisher=Davidbrin.blogspot.com |date=2010-06-21 |accessdate=2012-08-31}}</ref><ref>[http://www.theatlantic.com/entertainment/archive/2011/07/harry-potter-and-the-key-to-immortality/241972/ "'Harry Potter' and the Key to Immortality"], Daniel Snyder, ''[[The Atlantic]]''</ref><ref>{{cite web|author=David Brin |url=http://davidbrin.blogspot.com/2012/01/david-brins-list-of-greatest-science.html |title=CONTRARY BRIN: David Brin's List of "Greatest Science Fiction and Fantasy Tales" |publisher=Davidbrin.blogspot.com |date=2012-01-20 |accessdate=2012-08-31}}</ref><ref>http://davidbrin.blogspot.com/2013/02/science-fiction-and-our-duty-to-past.html</ref> and Rachel Aaron,<ref>{{cite web|author=Authors |url=http://www.fantasybookreview.co.uk/blog/2012/04/02/rachel-aaron-interview-april-2012/ |title=Rachel Aaron interview (April 2012) |publisher=Fantasybookreview.co.uk |date=2012-04-02 |accessdate=2012-08-31}}</ref><ref>{{cite web|url=http://civilian-reader.blogspot.com/2011/05/interview-with-rachel-aaron.html |title=Civilian Reader: An Interview with Rachel Aaron |publisher=Civilian-reader.blogspot.com |date=2011-05-04 |accessdate=2012-08-31}}</ref> [[Robin Hanson]],<ref>{{cite web|last=Hanson |first=Robin |url=http://www.overcomingbias.com/2010/10/hyper-rational-harry.html |title=Hyper-Rational Harry |publisher=Overcoming Bias |date=2010-10-31 |accessdate=2012-08-31}}</ref> [[Aaron Swartz]],<ref>{{cite web|last=Swartz |first=Aaron |url=http://web.archive.org/web/20130316081659/http://www.aaronsw.com/weblog/books2011 |title=The 2011 Review of Books (Aaron Swartz's Raw Thought) |publisher=archive.org |date= |accessdate=2013-04-10}}</ref> and by programmer [[Eric S. Raymond]].<ref>{{cite web|url=http://esr.ibiblio.org/?p=2100 |title=Harry Potter and the Methods of Rationality |publisher=Esr.ibiblio.org |date=2010-07-06 |accessdate=2012-08-31}}</ref>
Yudkowsky has also written several works<ref>{{cite web|url=http://yudkowsky.net/other/fiction|title=Yudkowsky- Fiction|publisher=Eliezer Yudkowsky}}</ref> of science fiction and other fiction. His [[Harry Potter]] [[fan fiction]] story ''Harry Potter and the Methods of Rationality'' illustrates topics in [[cognitive science]] and [[rationality]] (''[[The New Yorker]]'' described it as "a thousand-page online 'fanfic' text called 'Harry Potter and the Methods of Rationality', which recasts the original story in an attempt to explain Harry's wizardry through the scientific method"<ref>pg 54, [http://www.newyorker.com/reporting/2011/11/28/111128fa_fact_packer "No Death, No Taxes: The libertarian futurism of a Silicon Valley billionaire"]</ref>), and has been reviewed by authors [[David Brin]]<ref>{{cite web|author=David Brin |url=http://davidbrin.blogspot.com/2010/06/secret-of-college-life-plus.html |title=CONTRARY BRIN: A secret of college life... plus controversies and science! |publisher=Davidbrin.blogspot.com |date=2010-06-21 |accessdate=2012-08-31}}</ref><ref>[http://www.theatlantic.com/entertainment/archive/2011/07/harry-potter-and-the-key-to-immortality/241972/ "'Harry Potter' and the Key to Immortality"], Daniel Snyder, ''[[The Atlantic]]''</ref><ref>{{cite web|author=David Brin |url=http://davidbrin.blogspot.com/2012/01/david-brins-list-of-greatest-science.html |title=CONTRARY BRIN: David Brin's List of "Greatest Science Fiction and Fantasy Tales" |publisher=Davidbrin.blogspot.com |date=2012-01-20 |accessdate=2012-08-31}}</ref><ref>http://davidbrin.blogspot.com/2013/02/science-fiction-and-our-duty-to-past.html</ref> and Rachel Aaron,<ref>{{cite web|author=Authors |url=http://www.fantasybookreview.co.uk/blog/2012/04/02/rachel-aaron-interview-april-2012/ |title=Rachel Aaron interview (April 2012) |publisher=Fantasybookreview.co.uk |date=2012-04-02 |accessdate=2012-08-31}}</ref><ref>{{cite web|url=http://civilian-reader.blogspot.com/2011/05/interview-with-rachel-aaron.html |title=Civilian Reader: An Interview with Rachel Aaron |publisher=Civilian-reader.blogspot.com |date=2011-05-04 |accessdate=2012-08-31}}</ref> [[Robin Hanson]],<ref>{{cite web|last=Hanson |first=Robin |url=http://www.overcomingbias.com/2010/10/hyper-rational-harry.html |title=Hyper-Rational Harry |publisher=Overcoming Bias |date=2010-10-31 |accessdate=2012-08-31}}</ref> [[Aaron Swartz]],<ref>{{cite web|last=Swartz |first=Aaron |url=http://web.archive.org/web/20130316081659/http://www.aaronsw.com/weblog/books2011 |title=The 2011 Review of Books (Aaron Swartz's Raw Thought) |publisher=archive.org |date= |accessdate=2013-04-10}}</ref> and by programmer [[Eric S. Raymond]].<ref>{{cite web|url=http://esr.ibiblio.org/?p=2100 |title=Harry Potter and the Methods of Rationality |publisher=Esr.ibiblio.org |date=2010-07-06 |accessdate=2012-08-31}}</ref>


None of this has been accepted for publication by any legitimate [[academic journal]].
None of this has been submitted for publication in a peer-reviwed [[academic journal]].


==Criticism==
==Criticism==

Revision as of 11:04, 22 January 2014

Eliezer Shlomo Yudkowsky (born September 11, 1979[1]) is an American blogger, writer, and advocate for Friendly artificial intelligence.[2][3]

Biography

Yudkowsky, a resident of Berkeley, California has no formal education in computer science or artificial intelligence.[4] He co-founded the nonprofit Machine Intelligence Research Institute (formerly the Singularity Institute for Artificial Intelligence) in 2000 and continues to be employed there as a full-time Research Fellow.[5] He scored a 1410 on the SAT at age eleven[6] and a perfect 1600 four years later.[7]

Work

Yudkowsky's interests focus on Artificial Intelligence theory for self-understanding, self-modification, and recursive self-improvement (seed AI), and on artificial-intelligence architectures and decision theories for stable motivational structures (Friendly AI and Coherent Extrapolated Volition in particular).[8] Apart from his research work, Yudkowsky has written explanations of mathematical and philosophical topics in non-academic language, particularly on rationality, such as "An Intuitive Explanation of Bayes' Theorem".[9]

Publications

Yudkowsky was, along with Robin Hanson, one of the principal contributors to the blog Overcoming Bias[10] sponsored by the Future of Humanity Institute of Oxford University. In early 2009, he helped to found Less Wrong, a "community blog devoted to refining the art of human rationality".[11] The Sequences[12] on Less Wrong comprise over two years of blog posts on epistemology, Artificial Intelligence, and metaethics.

Yudkowsky's most recent work is on decision theory for problems of self-modification and Newcomblike problems, including "Tiling Agents for Self-Modifying AI, and the Löbian Obstacle"[13] and "Robust Cooperation in the Prisoner's Dilemma: Program Equilibrium via Provability Logic".[14] "A Comparison of Decision Algorithms on Newcomblike Problems" summarizes some of Yudkowsky's work on timeless decision theory.[15]

Yudkowsky contributed two chapters to Oxford philosopher Nick Bostrom's and Milan Ćirković's edited volume Global Catastrophic Risks,[16] and "Complex Value Systems are Required to Realize Valuable Futures"[17] to the conference AGI-11.

Yudkowsky is the author of the Singularity Institute publications "Creating Friendly AI"[18] (2001), "Levels of Organization in General Intelligence"[19] (2002), "Coherent Extrapolated Volition"[20] (2004), and "Timeless Decision Theory"[21] (2010).

Yudkowsky played the role of the AI in the first AI box experiments and wrote a page describing the rules he had used for the game.[22]

Yudkowsky has also written several works[23] of science fiction and other fiction. His Harry Potter fan fiction story Harry Potter and the Methods of Rationality illustrates topics in cognitive science and rationality (The New Yorker described it as "a thousand-page online 'fanfic' text called 'Harry Potter and the Methods of Rationality', which recasts the original story in an attempt to explain Harry's wizardry through the scientific method"[24]), and has been reviewed by authors David Brin[25][26][27][28] and Rachel Aaron,[29][30] Robin Hanson,[31] Aaron Swartz,[32] and by programmer Eric S. Raymond.[33]

None of this has been submitted for publication in a peer-reviwed academic journal.

Criticism

Yudkowsky's ideas have been criticized as self-serving and misleading by author Stephen Bond.[34]

References

  1. ^ Goodreads author page
  2. ^ Miller, James (2012). Singularity Rising. Texas: BenBella Books. pp. 35–44. ISBN 1936661659.
  3. ^ "Singularity Institute for Artificial Intelligence: Team". Singularity Institute for Artificial Intelligence. Retrieved 2009-07-16.
  4. ^ Singularity Rising, by James Miller, page 35
  5. ^ Kurzweil, Ray (2005). The Singularity Is Near. New York, US: Viking Penguin. p. 599. ISBN 0-670-03384-7.
  6. ^ Singularity Rising, by James Miller, page 38
  7. ^ The Spike, by Damien Broderick, page 265
  8. ^ Kurzweil, Ray (2005). The Singularity Is Near. New York, US: Viking Penguin. p. 420. ISBN 0-670-03384-7.
  9. ^ An Intuitive Explanation of Bayes' Theorem
  10. ^ "Overcoming Bias: About". Robin Hanson. Retrieved 2012-02-01.
  11. ^ "Welcome to Less Wrong". Less Wrong. Retrieved 2012-02-01.
  12. ^ "Sequences-Lesswrongwiki". Retrieved 2012-02-01.
  13. ^ "Tiling Agents for Self-Modifying AI, and the Löbian Obstacle" (PDF). Machine Intelligence Research Institute. Retrieved 2013-08-26.
  14. ^ "Robust Cooperation in the Prisoner's Dilemma: Program Equilibrium via Provability Logic" (PDF). Machine Intelligence Research Institute. Retrieved 2013-08-26.
  15. ^ Cite error: The named reference intelligence.org was invoked but never defined (see the help page).
  16. ^ Bostrom, Nick; Ćirković, Milan M., eds. (2008). Global Catastrophic Risks. Oxford, UK: Oxford University Press. pp. 91–119, 308–345. ISBN 978-0-19-857050-9.
  17. ^ Yudkowsky, Eliezer (2011). "Complex Value Systems are Required to Realize Valuable Futures" (PDF). AGI-11. {{cite conference}}: Unknown parameter |booktitle= ignored (|book-title= suggested) (help)
  18. ^ Yudkowsky, Eliezer. "Creating Friendly AI". Singularity Institute for Artificial Intelligence. Retrieved 2012-02-01.
  19. ^ Yudkowsky, Eliezer. "Levels of Organization in General Intelligence" (PDF). Singularity Institute for Artificial Intelligence. Retrieved 2012-02-01.
  20. ^ Yudkowsky, Eliezer. "Coherent Extrapolated Volition". Singularity Institute for Artificial Intelligence. Retrieved 2012-02-01.
  21. ^ Yudkowsky, Eliezer. "Timeless Decision Theory" (PDF). Singularity Institute for Artificial Intelligence. Retrieved 2012-02-01.
  22. ^ "The AI-Box Experiment". Retrieved 2013-08-26.
  23. ^ "Yudkowsky- Fiction". Eliezer Yudkowsky.
  24. ^ pg 54, "No Death, No Taxes: The libertarian futurism of a Silicon Valley billionaire"
  25. ^ David Brin (2010-06-21). "CONTRARY BRIN: A secret of college life... plus controversies and science!". Davidbrin.blogspot.com. Retrieved 2012-08-31.
  26. ^ "'Harry Potter' and the Key to Immortality", Daniel Snyder, The Atlantic
  27. ^ David Brin (2012-01-20). "CONTRARY BRIN: David Brin's List of "Greatest Science Fiction and Fantasy Tales"". Davidbrin.blogspot.com. Retrieved 2012-08-31.
  28. ^ http://davidbrin.blogspot.com/2013/02/science-fiction-and-our-duty-to-past.html
  29. ^ Authors (2012-04-02). "Rachel Aaron interview (April 2012)". Fantasybookreview.co.uk. Retrieved 2012-08-31.
  30. ^ "Civilian Reader: An Interview with Rachel Aaron". Civilian-reader.blogspot.com. 2011-05-04. Retrieved 2012-08-31.
  31. ^ Hanson, Robin (2010-10-31). "Hyper-Rational Harry". Overcoming Bias. Retrieved 2012-08-31.
  32. ^ Swartz, Aaron. "The 2011 Review of Books (Aaron Swartz's Raw Thought)". archive.org. Retrieved 2013-04-10.
  33. ^ "Harry Potter and the Methods of Rationality". Esr.ibiblio.org. 2010-07-06. Retrieved 2012-08-31.
  34. ^ Bond, Stephen. "THE CULT OF BAYES' THEOREM". Retrieved 27 November 2013.

Further reading

  • Our Molecular Future: How Nanotechnology, Robotics, Genetics and Artificial Intelligence Will Transform Our World by Douglas Mulhall, 2002, p. 321.
  • The Spike: How Our Lives Are Being Transformed By Rapidly Advancing Technologies by Damien Broderick, 2001, pp. 236, 265-272, 289, 321, 324, 326, 337-339, 345, 353, 370.

Template:Persondata