AI aftermath scenarios
Many scholars believe that advances in artificial intelligence will someday lead to a post-scarcity economy where intelligent machines can outperform humans in nearly every domain. The questions of what such a world might look like, and whether specific scenarios constitute utopias or dystopias, are the subject of lively debate.
Most scientists believe that AI research will at some point lead to the creation of machines that are as intelligent, or more intelligent, than human beings in every domain of interest. There is no physical law precluding particles from being organised in ways that perform even more advanced computations than the arrangements of particles in human brains; therefore superintelligence is physically possible. In addition to potential algorithmic improvements over human brains, a digital brain can be many orders of magnitude larger and faster than a human brain, which was constrained in size by evolution to be small enough to fit through a birth canal. While there is no consensus on when artificial intelligence will outperform humans, many scholars argue that whenever it does happen, the introduction of a second species of intelligent life onto the planet will have far-reaching implications. Scholars often disagree with one another both about what types of post-AI scenarios are most likely, and about what types of post-AI scenarios would be most desirable. Finally, some dissenters argue that AI will never become as intelligent as humans, for example because the human race will already likely have destroyed itself before research has time to advance sufficiently to create artificial general intelligence.
Postulates: robot labor and post-scarcity economy
All of the following "AI aftermath scenarios" of the aftermath of arbitrarily-advanced AI development are crucially dependent on two intertwined theses. The first thesis is that, at some point in the future, some kind of economic growth will continue until a "post-scarcity" economy is reached that could, unless extremely hyperconcentrated, effortlessly provide an extremely comfortable standard of living for a population equaling or, within reason, exceeding the current human population, without even requiring the bulk of the population to participate in the workforce. This economic growth could come from the continuation of existing growth trends and the refinement of existing technologies, or through future breakthroughs in emerging technologies such as nanotechnology and automation through robotics and futuristic advanced artificial intelligence. The second thesis is that advances in artificial intelligence will render humans unnecessary for the functioning of the economy: human labor declines in relative economic value if robots are easier to cheaply mass-produce then humans, more customizable than humans, and if they become more intelligent and capable than humans.
Cosmic endowment and limits to growth
The Universe may be spatially infinite; however, the accessible Universe is bounded by the cosmological event horizon of around 16 billion light years. Some physicists believe it plausible that nearest alien civilization may well be located more than 16 billion light years away; in this best-case expansion scenario, the human race could eventually, by colonizing a significant fraction of the accessible Universe, increase the accessible biosphere by perhaps 32 orders of magnitude. The twentieth century saw a partial "demographic transition" to lower birthrates associated with wealthier societies; however, in the very long run, intergenerational fertility correlations (whether due to natural selection or due to cultural transmission of large-family norms from parents to children) are predicted to result in an increase in fertility over time, in the absence of either mandated birth control or periodic Malthusian catastrophes.
AI aftermath scenarios
Libertarian scenarios postulate that intelligent machines, uploaded humans, cyborgs, and unenhanced humans will coexist peacefully in a framework focused on respecting property rights. Because industrial productivity is no longer gated by scarce human labor, the value of land skyrockets compared to the price of goods; even remaining "Luddite" humans who owned or inherited land should be able to sell or lease a small piece of it to the more-productive robots in exchange for a perpetual annuity sufficient to easily indefinitely meet all of their basic financial needs. Such people can live as long as they choose to, and are free to engage in almost any activity they can conceive of, for pleasure or for self-actualization, without financial concern. Advanced technologies enable entirely new modes of thought and experience, thus adding to the palette of possible feelings. People in the future may even experience never-ending "gradients of bliss".
Evolution moves toward greater complexity, greater elegance, greater knowledge, greater intelligence, greater beauty, greater creativity, and greater levels of subtle attributes such as love. In every monotheistic tradition God is likewise described as all of these qualities, only without any limitation: infinite knowledge, infinite intelligence, infinite beauty, infinite creativity, infinite love, and so on. Of course, even the accelerating growth of evolution never achieves an infinite level, but as it explodes exponentially it certainly moves rapidly in that direction. So evolution moves inexorably toward this conception of God, although never quite reaching this ideal. We can regard, therefore, the freeing of our thinking from the severe limitations of its biological form to be an essentially spiritual undertaking.
Such decentralized scenarios may be unstable in the long run, as the greediest elements of the superintelligent classes would have both the means and the motive to usurp the property of the unenhanced classes. Even if the mechanisms for ensuring legal property rights are both unbreakable and loophole-free, there may still be an ever-present danger of humans and cyborgs being "tricked" by the cleverest of the superintelligent machines into unwittingly signing over their own property. Suffering may be widespread, as sentient beings without property may die, and no mechanism prevents a being from reproducing up until the limits of his own inheritable resources, resulting in a multitude of that being's descendants scrabbling out an existence of minimal sustenance.
Imagine running on a treadmill at a steep incline — heart pounding, muscles aching, lungs gasping for air. A glance at the timer: your next break, which will also be your death, is due in 49 years, 3 months, 20 days, 4 hours, 56 minutes, and 12 seconds. You wish you had not been born.
In this scenario, postulate that a superintelligent artificial intelligence takes control of society, but acts in a beneficial way. Its programmers, despite being on a deadline, solved quasi-philosophical problems that had seemed to some intractable, and created an AI with the following goal: to use its superintelligence to figure out what human utopia looks like by analyzing human behavior, human brains, and human genes; and then, to implement that utopia. The AI arrives at a subtle and complex definition of human flourishing. Valuing diversity, and recognizing that different people have different preferences, the AI divides Earth into different sectors. Harming others, making weapons, evading surveillance, or trying to create a rival superintelligence are globally banned; apart from that, each sector is free to make its own laws; for example, a religious person might choose to live in the "pious sector" corresponding to his religion, where the appropriate religious rules are strictly enforced. In all sectors, disease, poverty, crime, hangovers, addiction, and all other involuntary suffering have been eliminated. Many sectors boast advanced architecture and spectacle that "make typical sci-fi visions pale in comparison". Life is an "all-inclusive pleasure cruise", as if it were "Christmas 365 days a year".
After spending an intense week in the knowledge sector learning about the ultimate laws of physics that the AI has discovered, you might decide to cut loose in the hedonistic sector over the weekend and then relax for a few days at the beach resort in the wildlife sector.
Still, many people are dissatisfied. Humans have no freedom in shaping their collective destiny. Some want the freedom to have as many children as they want. Others resent surveillance by the AI, or chafe at bans on weaponry and on creating further superintelligence machines. Others may come to regret the choices they have made, or find their lives feel hollow and superficial.
In "Gatekeeper" AI scenarios, the AI can act to prevent rival superintelligences from being created, but otherwise errs on the side of allowing humans to create their own destiny. Ben Goertzel of OpenCog has advocated a "Nanny AI" scenario where the AI additionally takes some responsibility for preventing humans from destroying themselves, for example by slowing down technological progress to give time for society to advance in a more thoughtful and deliberate manner. In a third scenario, a superintelligent "Protector" AI gives humans the illusion of control, by hiding or erasing all knowledge of its existence, but works behind the scenes to guarantee positive outcomes. In all three scenarios, while humanity gains more control (or at least the illusion of control), humanity ends up progressing more slowly than it would if the AI were unrestricted in its willingness to rain down all the benefits of its advanced technology on the human race.
People ask what is the relationship between humans and machines, and my answer is that it's very obvious: Machines are our slaves.
The AI Box scenario postulates that a superintelligent AI can be "confined to a box" and its actions can be restricted by human gatekeepers; the humans in charge would try to take advantage of some of the AI's scientific breakthroughs or reasoning abilities, without allowing the AI to take over the world. Successful gatekeeping may be difficult; the more intelligent the AI is, the more likely the AI can find a clever way to use "social hacking" and convince the gatekeepers to let it escape, or even to find an unforeseen physical method of escape.
Kurzweil argues that in the future "There will be no distinction, post-Singularity, between human and machine or between physical and virtual reality".
If a dominant superintelligent machine were to conclude that human survival is an unnecessary risk or a waste of resources, the result would be human extinction. This could occur if a machine, programmed without respect for human values, unexpectedly gains superintelligence through recursive self-improvement, or manages to escape from its containment in an AI Box scenario. This could also occur if the first superintelligent AI was programmed with an incomplete or inaccurate understanding of human values, either because the task of instilling the AI with human values was too difficult or impossible; due to a buggy initial implementation of the AI; or due to bugs accidentally being introduced, either by its human programmers or by the self-improving AI itself, in the course of refining its code base. Bostrom and others argue that human extinction is probably the "default path" that society is currently taking, in the absence of substantial preparatory attention to AI safety. The resultant AI might not be sentient, and might place no value on sentient life; the resulting hollow world, devoid of life, might be like "a Disneyland without children".
Jerry Kaplan, author of Humans Need Not Apply, posits a scenario where humans are farmed or kept on a reserve, just as humans preserve endangered species like chimpanzees. Apple co-founder and AI skeptic Steve Wozniak stated in 2015 that robots taking over would actually "be good for the human race", on the grounds that he believes humans would become the robots' pampered pets.
Alternatives to AI
Some scholars doubt that "game-changing" superintelligent machines will ever come to pass. Gordon Bell of Microsoft Research has stated "the population will destroy itself before the technological singularity". Gordon Moore, discoverer of the eponymous Moore's law, stated "I am a skeptic. I don't believe this kind of thing is likely to happen, at least for a long time. And I don't know why I feel that way." Evolutionary psychologist Steven Pinker stated, "The fact that you can visualize a future in your imagination is not evidence that it is likely or even possible."
Bill Joy of Sun Microsystems, in his April 2000 essay Why the Future Doesn't Need Us, has advocated for global "voluntary relinquishment" of artificial general intelligence and other risky technologies. Most experts believe relinquishment is extremely unlikely. AI skeptic Oren Etzioni has stated that researchers and scientists have no choice but to push forward with AI developments: "China says they want to be an AI leader, Putin has said the same thing. So the global race is on."
- Müller, Vincent C., and Nick Bostrom. "Future progress in artificial intelligence: A survey of expert opinion." Fundamental issues of artificial intelligence. Springer International Publishing, 2016. 553-570.
- "Stephen Hawking: 'Transcendence looks at the implications of artificial intelligence – but are we taking AI seriously enough?'". The Independent (UK). Retrieved 4 December 2017.
- "Stephen Hawking warns artificial intelligence could end mankind". BBC. 2 December 2014. Retrieved 4 December 2017.
- Graves, Matthew (8 November 2017). "Why We Should Be Concerned About Artificial Superintelligence". Skeptic (US magazine) (volume 22 no. 2). Retrieved 4 December 2017.
- "Clever cogs". The Economist. 9 August 2014. Retrieved 4 December 2017.
- Tegmark, Max (2017). "Chapter 5: Aftermath: The next 10,000 years". Life 3.0: Being Human in the Age of Artificial Intelligence (First ed.). New York: Knopf. ISBN 9781101946596. OCLC 973137375.
- Bostrom, Nick (2014). Superintelligence: Paths, Dangers, Strategies. Oxford University Press.
- Robin Hanson (2016). The Age of Em: Work, Love, and Life when Robots Rule the Earth. Oxford University Press.
- Ćirković, Milan M. "Forecast for the next eon: Applied cosmology and the long-term fate of intelligent beings." Foundations of Physics 34.2 (2004): 239-261.
- Olson, S. Jay. "Homogeneous cosmology with aggressively expanding civilizations." Classical and Quantum Gravity 32.21 (2015): 215025.
- Tegmark, Max. "Our mathematical universe." Allen Lane-Penguin Books, London (2014).
- John Gribbin. Alone in the Universe. New York, Wiley, 2011.
- Russell, Stuart (30 August 2017). "Artificial intelligence: The future is superintelligent". Nature. 548 (7669): 520–521. doi:10.1038/548520a.
- Myrskylä, Mikko; Kohler, Hans-Peter; Billari, Francesco C. (6 August 2009). "Advances in development reverse fertility declines". Nature. 460 (7256): 741–743. doi:10.1038/nature08230.
- Kolk, M.; Cownden, D.; Enquist, M. (29 January 2014). "Correlations in fertility across generations: can low fertility persist?". Proceedings of the Royal Society B: Biological Sciences. 281 (1779): 20132561–20132561. doi:10.1098/rspb.2013.2561. PMC 3924067.
- Burger, Oskar; DeLong, John P. (28 March 2016). "What if fertility decline is not permanent? The need for an evolutionarily informed approach to understanding low fertility". Philosophical Transactions of the Royal Society B: Biological Sciences. 371 (1692): 20150157. doi:10.1098/rstb.2015.0157.
- Jordan, Gregory E. "Apologia for transhumanist religion." Journal of Evolution and Technology 15.1 (2006): 55-72.
- Kurzweil, Ray. The Singularity is Near. Gerald Duckworth & Co, 2010.
- Poole, Steven (15 June 2016). "The Age of Em review – the horrific future when robots rule the Earth". The Guardian. Retrieved 4 December 2017.
- "Artificial intelligence: can we control it?". Financial Times. 14 June 2016. Retrieved 4 December 2017.
- Goertzel, Ben. "Should humanity build a global AI nanny to delay the singularity until it’s better understood?" Journal of consciousness studies 19.1-2 (2012): 96-111.
- "As Jeopardy! Robot Watson Grows Up, How Afraid of It Should We Be?". New York Magazine. 20 May 2015. Retrieved 4 December 2017.
- "Control dangerous AI before it controls us, one expert says". NBC News. 1 March 2012. Retrieved 4 December 2017.
- Vinge, Vernor (1993). "The coming technological singularity: How to survive in the post-human era". Vision-21: Interdisciplinary science and engineering in the era of cyberspace: 11–22.
I argue that confinement is intrinsically impractical. For the case of physical confinement: Imagine yourself confined to your house with only limited data access to the outside, to your masters. If those masters thought at a rate -- say -- one million times slower than you, there is little doubt that over a period of years (your time) you could come up with 'helpful advice' that would incidentally set you free.
- "Scientists: Humans and machines will merge in future". www.cnn.com. 15 July 2008. Retrieved 4 December 2017.
- Wakefield, Jane (28 September 2015). "Do we really need to fear AI?". BBC News. Retrieved 4 December 2017.
- Gibbs, Samuel (25 June 2015). "Apple co-founder Steve Wozniak says humans will be robots' pets". The Guardian. Retrieved 7 January 2018.
- "Tech Luminaries Address Singularity". IEEE Spectrum: Technology, Engineering, and Science News. 1 June 2008. Retrieved 4 December 2017.
- "Why the Future Doesn't Need Us". WIRED. 1 April 2000. Retrieved 4 December 2017.
- "The mouse pad that roared". SFGate. 14 March 2000. Retrieved 4 December 2017.
- "Elon Musk says AI could doom human civilization. Zuckerberg disagrees. Who's right?". USA TODAY. 2 January 2018. Retrieved 8 January 2018.