Global catastrophic risk

From Wikipedia, the free encyclopedia
  (Redirected from End of civilization)
Jump to: navigation, search
Artist's impression of a major asteroid impact. An asteroid with an impact strength of a billion atomic bombs may have caused the extinction of the dinosaurs.[1]

A global catastrophic risk is a hypothetical future event that has the potential to damage human well-being on a global scale.[2] Some events could cripple or destroy modern civilization.[3] Any event that could cause human extinction or permanently and drastically curtail humanity's potential is known as an existential risk.[4]

Potential global catastrophic risks include anthropogenic risks (technology risks, governance risks) and natural or external risks.[3] Examples of technology risks are hostile artificial intelligence, biotechnology risks, or nanotechnology weapons. Insufficient global governance creates risks in the social and political domain (potentially leading to a global war with or without a nuclear holocaust, bioterrorism using genetically modified organisms, cyberterrorism destroying critical infrastructures like the electrical grid, or the failure to manage a natural pandemic) as well as problems and risks in the domain of earth system governance (with risks resulting from global warming, environmental degradation, including extinction of species, or famine as a result of non-equitable resource distribution, human overpopulation, crop failures and non-sustainable agriculture). Examples for non-anthropogenic risks are an asteroid impact event, a supervolcanic eruption, a lethal gamma-ray burst, a geomagnetic storm destroying all electronic equipment, natural long-term climate change, or extraterrestrial life impacting life on Earth.


Scope/intensity grid from Bostrom's paper "Existential Risk Prevention as Global Priority"[5]

Global catastrophic vs. existential[edit]

Philosopher Nick Bostrom classifies risks according to their scope and intensity.[5] A "global catastrophic risk" is any risk that is at least "global" in scope, and is not subjectively "imperceptible" in intensity. Those that are at least "trans-generational" (affecting all future generations) in scope and "terminal"[clarification needed] in intensity are classified as existential risks. While a global catastrophic risk may kill the vast majority of life on earth, humanity could still potentially recover. An existential risk, on the other hand, is one that either destroys humanity (and, presumably, all but the most rudimentary species of non-human lifeforms and/or plant life) entirely or at least prevents any chance of civilization recovering. Bostrom considers existential risks to be far more significant.[6]

Similarly, in Catastrophe: Risk and Response, Richard Posner singles out and groups together events that bring about "utter overthrow or ruin" on a global, rather than a "local or regional" scale. Posner singles out such events as worthy of special attention on cost-benefit grounds because they could directly or indirectly jeopardize the survival of the human race as a whole.[7] Posner's events include meteor impacts, runaway global warming, grey goo, bioterrorism, and particle accelerator accidents.

Researchers experience difficulty in studying near human extinction directly, since humanity has never been destroyed before.[8] While this does not mean that it will not be in the future, it does make modelling existential risks difficult, due in part to survivorship bias.

Other classifications[edit]

Bostrom identifies four types of existential risk. "Bangs" are sudden catastrophes, which may be accidental or deliberate. He thinks the most likely sources of bangs are malicious use of nanotechnology, nuclear war, and the possibility that the universe is a simulation that will end. "Crunches" are scenarios in which humanity survives but civilization is irreversibly destroyed. The most likely causes of this, he believes, are exhaustion of natural resources, a stable global government that prevents technological progress, or dysgenic pressures that lower average intelligence. "Shrieks" are undesirable futures. For example, if a single mind enhances its powers by merging with a computer, it could dominate human civilization. Bostrom believes that this scenario is most likely, followed by flawed superintelligence and a repressive totalitarian regime. "Whimpers" are the gradual decline of human civilization or current values. He thinks the most likely cause would be evolution changing moral preference, followed by extraterrestrial invasion.[4]


Some risks, such as that from asteroid impact, with a one-in-a-million chance of causing humanity's extinction in the next century,[9] have had their probabilities predicted with considerable precision (although some scholars claim the actual rate of large impacts could be much higher than originally calculated).[10] Similarly, the frequency of volcanic eruptions of sufficient magnitude to cause catastrophic climate change, similar to the Toba Eruption, which may have almost caused the extinction of the human race,[11] has been estimated at about 1 in every 50,000 years.[12] The 2016 annual report by the Global Challenges Foundation estimates that an average American is more than five times more likely to die during a human-extinction event than in a car crash.[13][14]

The relative danger posed by other threats is much more difficult to calculate. In 2008, an informal survey of small but illustrious group of experts on different global catastrophic risks at the Global Catastrophic Risk Conference at the University of Oxford suggested a 19% chance of human extinction by the year 2100. The conference report cautions that the results should be taken "with a grain of salt".[15] In November 2017, a statement by 15,364 scientists from 184 countries indicated that increasing levels of greenhouse gases from use of fossil fuels, human population growth, deforestation, and overuse of land for agricultural production, particularly by farming ruminants for meat consumption, are trending in ways that forecast an increase in human misery over coming decades.[3]

Risk Estimated probability
for human extinction
before 2100
Overall probability
Molecular nanotechnology weapons
Superintelligent AI
Non-nuclear wars
Engineered pandemic
Nuclear wars
Nanotechnology accident
Natural pandemic
Nuclear terrorism
Table source: Future of Humanity Institute, 2008.[15]

There are significant methodological challenges in estimating these risks with precision. Most attention has been given to risks to human civilization over the next 100 years, but forecasting for this length of time is difficult. The types of threats posed by nature may prove relatively constant, though new risks could be discovered. Anthropogenic threats, however, are likely to change dramatically with the development of new technology; while volcanoes have been a threat throughout history, nuclear weapons have only been an issue since the 20th century. Historically, the ability of experts to predict the future over these timescales has proved very limited. Man-made threats such as nuclear war or nanotechnology are harder to predict than natural threats, due to the inherent methodological difficulties in the social sciences. In general, it is hard to estimate the magnitude of the risk from this or other dangers, especially as both international relations and technology can change rapidly.

Existential risks pose unique challenges to prediction, even more than other long-term events, because of observation selection effects. Unlike with most events, the failure of a complete extinction event to occur in the past is not evidence against their likelihood in the future, because every world that has experienced such an extinction event has no observers, so regardless of their frequency, no civilization observes existential risks in its history.[8] These anthropic issues can be avoided by looking at evidence that does not have such selection effects, such as asteroid impact craters on the Moon, or directly evaluating the likely impact of new technology.[5]

Moral importance of existential risk[edit]

Some scholars have strongly favored reducing existential risk on the grounds that it greatly benefits future generations. Derek Parfit argues that extinction would be a great loss because our descendants could potentially survive for four billion years before the expansion of the Sun makes the Earth uninhabitable.[16][17] Nick Bostrom argues that there is even greater potential in colonizing space. If future humans colonize space, they may be able to support a very large number of people on other planets, potentially lasting for trillions of years.[6] Therefore, reducing existential risk by even a small amount would have a very significant impact on the expected number of people who will exist in the future.

Exponential discounting might make these future benefits much less significant. However, Gaverick Matheny has argued that such discounting is inappropriate when assessing the value of existential risk reduction.[9]

Some economists have discussed the importance of global catastrophic risks, though not existential risks. Martin Weitzman argues that most of the expected economic damage from climate change may come from the small chance that warming greatly exceeds the mid-range expectations, resulting in catastrophic damage.[18] Richard Posner has argued that we are doing far too little, in general, about small, hard-to-estimate risks of large-scale catastrophes.[19]

Numerous cognitive biases can influence people's judgment of the importance of existential risks, including scope insensitivity, hyperbolic discounting, availability heuristic, the conjunction fallacy, the affect heuristic, and the overconfidence effect.[20]

Scope insensitivity influences how bad people consider the extinction of the human race to be. For example, when people are motivated to donate money to altruistic causes, the quantity they are willing to give does not increase linearly with the magnitude of the issue: people are roughly as concerned about 200,000 birds getting stuck in oil as they are about 2,000.[21] Similarly, people are often more concerned about threats to individuals than to larger groups.[20]

There are economic reasons that can explain why so little effort is going into existential risk reduction. It is a global good, so even if a large nation decreases it, that nation will only enjoy a small fraction of the benefit of doing so. Furthermore, the vast majority of the benefits may be enjoyed by far future generations, and though these quadrillions of future people would in theory perhaps be willing to pay massive sums for existential risk reduction, no mechanism for such a transaction exists.[5]

Potential sources of risk[edit]

Some sources of catastrophic risk are natural, such as meteor impacts or supervolcanos. Some of these have caused mass extinctions in the past.

On the other hand, some risks are man-made, such as global warming,[22] environmental degradation, engineered pandemics and nuclear war. According to the Future of Humanity Institute, human extinction is more likely to result from anthropogenic causes than natural causes.[5][23]


In 2012, Cambridge University created The Cambridge Project for Existential Risk which examines threats to humankind caused by developing technologies.[24] The stated aim is to establish within the University a multidisciplinary research centre, Centre for the Study of Existential Risk, dedicated to the scientific study and mitigation of existential risks of this kind.[24]

The Cambridge Project states that the "greatest threats" to the human species are man-made; they are artificial intelligence, global warming, nuclear war, and rogue biotechnology.[25]

Artificial intelligence[edit]

It has been suggested that learning computers that rapidly become superintelligent may take unforeseen actions or that robots would out-compete humanity (one technological singularity scenario).[26] Because of its exceptional scheduling and organizational capability and the range of novel technologies it could develop, it is possible that the first Earth superintelligence to emerge could rapidly become matchless and unrivaled: conceivably it would be able to bring about almost any possible outcome, and be able to foil virtually any attempt that threatened to prevent it achieving its objectives.[27] It could eliminate, wiping out if it chose, any other challenging rival intellects; alternatively it might manipulate or persuade them to change their behavior towards its own interests, or it may merely obstruct their attempts at interference.[27] In Bostrom's book, Superintelligence: Paths, Dangers, Strategies, he defines this as the control problem.[28]

Vernor Vinge has suggested that a moment may come when computers and robots are smarter than humans. He calls this "the Singularity."[29] He suggests that it may be somewhat or possibly very dangerous for humans.[30] This is discussed by a philosophy called Singularitarianism.

Physicist Stephen Hawking, Microsoft founder Bill Gates and SpaceX founder Elon Musk have expressed concerns about the possibility that AI could evolve to the point that humans could not control it, with Hawking theorizing that this could "spell the end of the human race".[31] In 2009, experts attended a conference hosted by the Association for the Advancement of Artificial Intelligence (AAAI) to discuss whether computers and robots might be able to acquire any sort of autonomy, and how much these abilities might pose a threat or hazard. They noted that some robots have acquired various forms of semi-autonomy, including being able to find power sources on their own and being able to independently choose targets to attack with weapons. They also noted that some computer viruses can evade elimination and have achieved "cockroach intelligence." They noted that self-awareness as depicted in science-fiction is probably unlikely, but that there were other potential hazards and pitfalls.[29] Various media sources and scientific groups have noted separate trends in differing areas which might together result in greater robotic functionalities and autonomy, and which pose some inherent concerns.[32][33] Eliezer Yudkowsky believes that risks from artificial intelligence are harder to predict than any other known risks. He also argues that research into artificial intelligence is biased by anthropomorphism. Since people base their judgments of artificial intelligence on their own experience, he claims that they underestimate the potential power of AI. He distinguishes between risks due to technical failure of AI, which means that flawed algorithms prevent the AI from carrying out its intended goals, and philosophical failure, which means that the AI is programmed to realize a flawed ideology.[34]


Biotechnology can pose a global catastrophic risk in the form of bioengineered organisms (viruses, bacteria, fungi, plants or animals). In many cases the organism will be a pathogen of humans, livestock, crops or other organisms we depend upon (e.g. pollinators or gut bacteria). However, any organism able to catastrophically disrupt ecosystem functions, e.g. highly competitive weeds, outcompeting essential crops, poses a biotechnology risk.

A biotechnology catastrophe may be caused by accidentally releasing a genetically engineered organism escaping from controlled environments, by the planned release of such an organism which then turns out to have unforeseen and catastrophic interactions with essential natural or agro-ecosystems, or by intentional usage of biological agents in biological warfare, bioterrorism attacks.[35] Terrorist applications of biotechnology have historically been infrequent.[35] To what extent this is due to a lack of capabilities or motivation is not resolved.[35]

Exponential growth has been observed in the biotechnology sector and Noun and Chyba predict that this will lead to major increases in biotechnological capabilities in the coming decades.[35] They argue that risks from biological warfare and bioterrorism are distinct from nuclear and chemical threats because biological pathogens are easier to mass-produce and their production is hard to control (especially as the technological capabilities are becoming available even to individual users).[35]

Given current development, more risk from novel, engineered pathogens is to be expected in the future.[35] Pathogens may be intentionally or unintentionally genetically modified to change virulence and other characteristics.[35] For example, a group of Australian researchers unintentionally changed characteristics of the mousepox virus while trying to develop a virus to sterilize rodents.[35] The modified virus became highly lethal even in vaccinated and naturally resistant mice.[36][37] The technological means to genetically modify virus characteristics are likely to become more widely available in the future if not properly regulated.[35]

Noun and Chyba propose three categories of measures to reduce risks from biotechnology and natural pandemics: Regulation or prevention of potentially dangerous research, improved recognition of outbreaks and developing facilities to mitigate disease outbreaks (e.g. better and/or more widely distributed vaccines).[35]

(See also Natural pathogens below.)


Cyberattacks have the potential to destroy everything from personal data to electric grids. Christine Peterson, co-founder and past president of the Foresight Institute, believes a cyberattack on electric grids has the potential to be a catastrophic risk.[38] Peterson also identifies attacks on Internet of Things devices as potentially catastrophic.

Global warming[edit]

Global warming refers to the warming caused by human technology since the 19th century or earlier. Global warming reflects abnormal variations to the expected climate within the Earth's atmosphere and subsequent effects on other parts of the Earth. Projections of future climate change suggest further global warming, sea level rise, and an increase in the frequency and severity of some extreme weather events and weather-related disasters. Effects of global warming include loss of biodiversity, stresses to existing food-producing systems, increased spread of known infectious diseases such as malaria, and rapid mutation of microorganisms.

It has been suggested that runaway global warming (runaway climate change) might cause Earth to become searingly hot like Venus. In less extreme scenarios, it could cause the end of civilization as we know it.[39]

Environmental disaster[edit]

An environmental or ecological disaster, such as world crop failure and collapse of ecosystem services, could be induced by the present trends of overpopulation, economic development,[40] and non-sustainable agriculture. An October 2017 report published in The Lancet stated that toxic air, water, soils, and workplaces were collectively responsible for 9 million deaths worldwide in 2015, particularly from air pollution which was linked to deaths by increasing susceptibility to non-infectious diseases, such as heart disease, stroke, and lung cancer.[41] The report warned that the pollution crisis was exceeding "the envelope on the amount of pollution the Earth can carry" and “threatens the continuing survival of human societies”.[41]

Most environmental scenarios involve one or more of the following: Holocene extinction event,[42] scarcity of water that could lead to approximately one half of the Earth's population being without safe drinking water, pollinator decline, overfishing, massive deforestation, desertification, climate change, or massive water pollution episodes. Detected in the early 21st century, a threat in this direction is colony collapse disorder,[43] a phenomenon that might foreshadow the imminent extinction[44] of the Western honeybee. As the bee plays a vital role in pollination, its extinction would severely disrupt the food chain.

Mineral resource exhaustion[edit]

Romanian American economist Nicholas Georgescu-Roegen, a progenitor in economics and the paradigm founder of ecological economics, has argued that the carrying capacity of Earth — that is, Earth's capacity to sustain human populations and consumption levels — is bound to decrease sometime in the future as Earth's finite stock of mineral resources is presently being extracted and put to use; and consequently, that the world economy as a whole is heading towards an inevitable future collapse, leading to the demise of human civilization itself.[45]:303f Ecological economist and steady-state theorist Herman Daly, a student of Georgescu-Roegen, has propounded the same argument by asserting that "... all we can do is to avoid wasting the limited capacity of creation to support present and future life [on Earth]."[46]:370

Ever since Georgescu-Roegen and Daly published these views, various scholars in the field have been discussing the existential impossibility of distributing Earth's finite stock of mineral resources evenly among an unknown number of present and future generations. This number of generations is likely to remain unknown to us, as there is little way of knowing in advance if or when mankind will eventually face extinction. In effect, any conceivable intertemporal distribution of the stock will inevitably end up with universal economic decline at some future point.[47]:253–256 [48]:165 [49]:168–171 [50]:150–153 [51]:106–109 [52]:546–549 [53]:142–145

Experimental technology accident[edit]

Nick Bostrom suggested that in the pursuit of knowledge, humanity might inadvertently create a device that could destroy Earth and the Solar System.[54] Investigations in nuclear and high-energy physics could create unusual conditions with catastrophic consequences. For example, scientists worried that the first nuclear test might ignite the atmosphere.[55][56] More recently, others worried that the RHIC[57] or the Large Hadron Collider might start a chain-reaction global disaster involving black holes, strangelets, or false vacuum states. These particular concerns have been refuted,[58][59][60][61] but the general concern remains.

Biotechnology could lead to the creation of a pandemic, chemical warfare could be taken to an extreme, nanotechnology could lead to grey goo in which out-of-control self-replicating robots consume all living matter on earth while building more of themselves—in both cases, either deliberately or by accident.[62]


Many nanoscale technologies are in development or currently in use.[63] The only one that appears to pose a significant global catastrophic risk is molecular manufacturing, a technique that would make it possible to build complex structures at atomic precision.[64] Molecular manufacturing requires significant advances in nanotechnology, but once achieved could produce highly advanced products at low costs and in large quantities in nanofactories of desktop proportions.[63][64] When nanofactories gain the ability to produce other nanofactories, production may only be limited by relatively abundant factors such as input materials, energy and software.[63]

Molecular manufacturing could be used to cheaply produce, among many other products, highly advanced, durable weapons.[63] Being equipped with compact computers and motors these could be increasingly autonomous and have a large range of capabilities.[63]

Phoenix and Treder classify catastrophic risks posed by nanotechnology into three categories:

  1. From augmenting the development of other technologies such as AI and biotechnology.
  2. By enabling mass-production of potentially dangerous products that cause risk dynamics (such as arms races) depending on how they are used.
  3. From uncontrolled self-perpetuating processes with destructive effects.

At the same time, nanotechnology may be used to alleviate several other global catastrophic risks.[63]

Several researchers state that the bulk of risk from nanotechnology comes from the potential to lead to war, arms races and destructive global government.[36][63][65] Several reasons have been suggested why the availability of nanotech weaponry may with significant likelihood lead to unstable arms races (compared to e.g. nuclear arms races):

  1. A large number of players may be tempted to enter the race since the threshold for doing so is low;[63]
  2. The ability to make weapons with molecular manufacturing will be cheap and easy to hide;[63]
  3. Therefore, lack of insight into the other parties' capabilities can tempt players to arm out of caution or to launch preemptive strikes;[63][66]
  4. Molecular manufacturing may reduce dependency on international trade,[63] a potential peace-promoting factor;
  5. Wars of aggression may pose a smaller economic threat to the aggressor since manufacturing is cheap and humans may not be needed on the battlefield.[63]

Since self-regulation by all state and non-state actors seems hard to achieve,[67] measures to mitigate war-related risks have mainly been proposed in the area of international cooperation.[63][68] International infrastructure may be expanded giving more sovereignty to the international level. This could help coordinate efforts for arms control. International institutions dedicated specifically to nanotechnology (perhaps analogously to the International Atomic Energy Agency IAEA) or general arms control may also be designed.[68] One may also jointly make differential technological progress on defensive technologies, a policy that players should usually favour.[63] The Center for Responsible Nanotechnology also suggests some technical restrictions.[69] Improved transparency regarding technological capabilities may be another important facilitator for arms-control.

A grey goo is another catastrophic scenario, which was proposed by Eric Drexler in his 1986 book Engines of Creation[70] and has been a theme in mainstream media and fiction.[71][72] This scenario involves tiny self-replicating robots that consume the entire biosphere using it as a source of energy and building blocks. Nowadays, however, nanotech experts - including Drexler - discredit the scenario. According to Chris Phoenix a "so-called grey goo could only be the product of a deliberate and difficult engineering process, not an accident".[73]

Warfare and mass destruction[edit]

The scenarios that have been explored most frequently are nuclear warfare and doomsday devices. Although the probability of a nuclear war per year is slim, Professor Martin Hellman has described it as inevitable in the long run; unless the probability approaches zero, inevitably there will come a day when civilization's luck runs out.[74] During the Cuban missile crisis, U.S. president John F. Kennedy estimated the odds of nuclear war at "somewhere between one out of three and even".[75] The United States and Russia have a combined arsenal of 14,700 nuclear weapons,[76] and there is an estimated total of 15,700 nuclear weapons in existence worldwide.[76]

While popular perception sometimes takes nuclear war as "the end of the world", experts assign low probability to human extinction from nuclear war.[77][78] In 1982, Brian Martin estimated that a US–Soviet nuclear exchange might kill 400–450 million directly, mostly in the United States, Europe and Russia, and maybe several hundred million more through follow-up consequences in those same areas.[77]

Nuclear war could yield unprecedented human death tolls and habitat destruction. Detonating such large numbers of nuclear weapons would have a long-term effect on the climate, causing cold weather and reduced sunlight[79] that may generate significant upheaval in advanced civilizations.[80]

Beyond nuclear, other threats to humanity include biological warfare (BW) and bioterrorism. By contrast, chemical warfare, while able to create multiple local catastrophes, is unlikely to create a global one.

World population and agricultural crisis[edit]

The 20th century saw a rapid increase in human population due to medical developments and massive increases in agricultural productivity[81] such as the Green Revolution.[82] Between 1950 and 1984, as the Green Revolution transformed agriculture around the globe, world grain production increased by 250%. The Green Revolution in agriculture helped food production to keep pace with worldwide population growth or actually enabled population growth. The energy for the Green Revolution was provided by fossil fuels in the form of fertilizers (natural gas), pesticides (oil), and hydrocarbon fueled irrigation.[83] David Pimentel, professor of ecology and agriculture at Cornell University, and Mario Giampietro, senior researcher at the National Research Institute on Food and Nutrition (INRAN), place in their 1994 study Food, Land, Population and the U.S. Economy the maximum U.S. population for a sustainable economy at 200 million. To achieve a sustainable economy and avert disaster, the United States must reduce its population by at least one-third, and world population will have to be reduced by two-thirds, says the study.[84]

The authors of this study believe that the mentioned agricultural crisis will begin to impact us after 2020, and will become critical after 2050. Geologist Dale Allen Pfeiffer claims that coming decades could see spiraling food prices without relief and massive starvation on a global level such as never experienced before.[85][86]

Wheat is humanity's 3rd most produced cereal. Extant fungal infections such as Ug99[87] (a kind of stem rust) can cause 100% crop losses in most modern varieties. Little or no treatment is possible and infection spreads on the wind. Should the world's large grain producing areas become infected then there would be a crisis in wheat availability leading to price spikes and shortages in other food products.[88]


Asteroid impact[edit]

Several asteroids have collided with earth in recent geological history. The Chicxulub asteroid, for example, is theorized to have caused the extinction of the non-avian dinosaurs 66 million years ago at the end of the Cretaceous. No sufficiently large asteroid currently exists in an Earth-crossing orbit; however, a comet of sufficient size to cause human extinction could impact the Earth, though the annual probability may be less than 10-8.[89] Geoscientist Brian Toon estimates that a 60-mile meteorite would be large enough to "incinerate everybody".[90] Asteroids with around a 1 km diameter have impacted the Earth on average once every 500,000 years; these are probably too small to pose an extinction risk, but might kill billions of people.[89][91] Larger asteroids are less common. Small near-Earth asteroids are regularly observed. As of 2013, Spaceguard estimates it has identified 95% of all NEOs over 1 km in size.[92]

In 1.4 million years, the star Gliese 710 is expected to start causing an increase in the number of meteoroids in the vicinity of Earth when it passes within 1.1 light years of the Sun, perturbing the Oort cloud. Dynamic models by García-Sánchez predict a 5% increase in the rate of impact.[93] Objects perturbed from the Oort cloud take millions of years to reach the inner Solar System.

Extraterrestrial invasion[edit]

Extraterrestrial life could invade Earth[94] either to exterminate and supplant human life, enslave it under a colonial system, steal the planet's resources, or destroy the planet altogether.

Although evidence of alien life has never been documented, scientists such as Carl Sagan have postulated that the existence of extraterrestrial life is very likely. In 1969, the "Extra-Terrestrial Exposure Law" was added to the United States Code of Federal Regulations (Title 14, Section 1211) in response to the possibility of biological contamination resulting from the U.S. Apollo Space Program. It was removed in 1991.[95] Scientists consider such a scenario technically possible, but unlikely.[96]

An article in The New York Times discussed the possible threats for humanity of intentionally sending messages aimed at extraterrestrial life into the cosmos in the context of the SETI efforts. Several renowned public figures such as Stephen Hawking and Elon Musk have argued against sending such messages on the grounds that extraterrestrial civilizations are probably technologically far more advanced than humanity and could pose an existential threat to humanity.[97]

Natural climate change[edit]

Climate change refers to a lasting change in the Earth's climate. The climate has ranged from ice ages to warmer periods when palm trees grew in Antarctica. It has been hypothesized that there was also a period called "snowball Earth" when all the oceans were covered in a layer of ice. These global climatic changes occurred slowly, prior to the rise of human civilization about 10 thousand years ago near the end of the last Major Ice Age when the climate became more stable. However, abrupt climate change on the decade time scale has occurred regionally. Since civilization originated during a period of stable climate, a natural variation into a new climate regime (colder or hotter) could pose a threat to civilization.

In the history of the Earth, many ice ages are known to have occurred. More ice ages will be possible at an interval of 40,000–100,000 years. An ice age would have a serious impact on civilization because vast areas of land (mainly in North America, Europe, and Asia) could become uninhabitable. It would still be possible to live in the tropical regions, but with possible loss of humidity and water. Currently, the world is existing in an interglacial period within a much older glacial event. The last glacial expansion ended about 10,000 years ago, and all civilizations evolved later than this. Scientists do not predict that a natural ice age will occur anytime soon.

Cosmic threats[edit]

A number of astronomical threats have been identified. Massive objects, e.g. a star, large planet or black hole, could be catastrophic if a close encounter occurred in the Solar System. In April 2008, it was announced that two simulations of long-term planetary movement, one at Paris Observatory and the other at University of California, Santa Cruz indicate a 1% chance that Mercury's orbit could be made unstable by Jupiter's gravitational pull sometime during the lifespan of the Sun. Were this to happen, the simulations suggest a collision with Earth could be one of four possible outcomes (the others being Mercury colliding with the Sun, colliding with Venus, or being ejected from the Solar System altogether). If Mercury were to collide with Earth, all life on Earth could be obliterated entirely: an asteroid 15 km wide is believed to have caused the extinction of the non-avian dinosaurs, whereas Mercury is 4,879 km in diameter.[98]

Another cosmic threat is a gamma-ray burst, typically produced by a supernova when a star collapses inward on itself and then "bounces" outward in a massive explosion. Under certain circumstances, these events are thought to produce massive bursts of gamma radiation emanating outward from the axis of rotation of the star. If such an event were to occur oriented towards the Earth, the massive amounts of gamma radiation could significantly affect the Earth's atmosphere and pose an existential threat to all life. Such a gamma ray burst may have been the cause of the Ordovician–Silurian extinction events. Neither this scenario nor Mercury's orbit destabilizing threats are likely in the foreseeable future.[99]

If the Solar System were to pass through a dark nebula, a cloud of cosmic dust, severe global climate change would occur.[100]

A powerful solar flare or solar superstorm, which is a drastic and unusual decrease or increase in the Sun's power output, could have severe consequences for life on Earth.

If our universe lies within a false vacuum, a bubble of lower-energy vacuum could come to exist by chance or otherwise in our universe, and catalyze the conversion of our universe to a lower energy state in a volume expanding at nearly the speed of light, destroying all that we know without forewarning.[101][further explanation needed] Such an occurrence is called a vacuum metastability event.

Geomagnetic reversal[edit]

The magnetic poles of the Earth shifted many times in geologic history. The duration of such a shift is still debated. Theories exist that during such times, the Earth's magnetic field would be substantially weakened, threatening civilization by allowing radiation from the Sun, especially solar wind, solar flares or cosmic radiation, to reach the surface. These theories have been somewhat discredited, as statistical analysis shows no evidence for a correlation between past reversals and past extinctions.[102][103]

Global pandemic[edit]

Numerous historical examples of pandemics[104] had a devastating effect on a large number of people. The present, unprecedented scale and speed of human movement make it more difficult than ever to contain an epidemic through local quarantines. A global pandemic has become a realistic threat to human civilization.

Naturally evolving pathogens will ultimately develop an upper limit to their virulence.[105] Pathogen with the highest virulence, quickly killing their hosts reduce their chances of spread the infection to new hosts or carriers.[106] This simple model predicts that - if virulence and transmission are not genetically linked - pathogens will evolve towards low virulence and rapid transmission. However, this is not necessarily a safeguard against a global catastrophe, for the following reasons:

1. The fitness advantage of limited virulence is primarily a function of a limited number of hosts. Any pathogene with a high virulence, high transmission rate and long incubation time may have already caused a catastrophic pandemic before ultimately virulence is limited through natural selection. 2. In models where virulence level and rate of transmission are related, high levels of virulence can evolve.[107] Virulence is instead limited by the existence of complex populations of hosts with different susceptibilities to infection, or by some hosts being geographically isolated.[105] The size of the host population and competition between different strains of pathogens can also alter virulence.[108] 3. A pathogen that infects humans as a secondary host and primarily infects another species (a zoonosis) has no constraints on its virulence in people, since the accidental secondary infections do not affect its evolution.[109]

Naturally arising pathogens and Neobiota[edit]

In a similar scenario to biotechnology risks, naturally evolving organisms can disrupt essential ecosystem functions.

An example of a pathogen able to threaten global food security is the wheat rust Ug99.

Other examples are neobiota (invasive species), i.e. organisms that become disruptive to ecosystems once transported – often as a result of human activity – to a new geographical region. Normally the risk is a local disruption. If it becomes coupled with serious crop failures and a global famine it may, however, pose a global catastrophic risk.


A remote possibility is a megatsunami. It has been suggested that a megatsunami caused by the collapse of a volcanic island could, for example, destroy the entire East Coast of the United States, but such predictions are based on incorrect assumptions and the likelihood of this happening has been greatly exaggerated in the media.[110] While none of these scenarios are likely to destroy humanity completely, they could regionally threaten civilization. There have been two recent high-fatality tsunamis—after the 2011 Tōhoku earthquake and the 2004 Indian Ocean earthquake. A megatsunami could have astronomical origins as well, such as an asteroid impact in an ocean.[111]


A geological event such as massive flood basalt, volcanism, or the eruption of a supervolcano[112] could lead to a so-called volcanic winter, similar to a nuclear winter. One such event, the Toba eruption,[113] occurred in Indonesia about 71,500 years ago. According to the Toba catastrophe theory,[114] the event may have reduced human populations to only a few tens of thousands of individuals. Yellowstone Caldera is another such supervolcano, having undergone 142 or more caldera-forming eruptions in the past 17 million years.[115] A massive volcano eruption would eject extraordinary volumes of volcanic dust, toxic and greenhouse gases into the atmosphere with serious effects on global climate (towards extreme global cooling; volcanic winter if short term, and ice age if long term) or global warming (if greenhouse gases were to prevail).

When the supervolcano at Yellowstone last erupted 640,000 years ago, the thinnest layers of the ash ejected from the caldera spread over most of the United States west of the Mississippi river and part of northeastern Mexico. The magma covered much of Yellowstone park and extended beyond, covering much of the ground from Yellowstone river in the East to the Idaho falls in the West with some of the flows extending north beyond Mammoth springs.[116]

According to a recent study, if it were to erupt as a supervolcano again, you’d get 1–3 millimeters thickness of ash right out to New York, which is enough to “reduce traction on roads and runways, short out electrical transformers and cause respiratory problems”. There would be centimeters of thickness over much of the mid west, enough to disrupt crops and livestock, especially if it happened at critical time in the growing season. and a meter of thickness out to quite a distance. The worst affected in their list of cites is Billings, population 109,000, which their model predicted would get an estimated 1.03 to 1.8 meters thickness of ash.[117]

The main long term effect is through global climate change, which reduces the temperature globally by about 5 -15 degrees C for a decade, together with the direct effects of the deposits of ash on their crops. A large supervolcano like Toba would deposit one or two meters thickness of ash over an area of several million square kilometers.(1000 cubic kilometers is equivalent to a one meter thickness of ash spread over a million square kilometers). If that happened in some densely populated agricultural area, such as India, it could destroy one or two seasons of crops for two billion people.[118]

Yellowstone however shows no signs of erupting as a supervolcano at present and there is no certainty that it will erupt like that again.[119][120]

Research published in 2011 finds evidence that massive volcanic eruptions caused massive coal combustion, supporting models for significant generation of greenhouse gases. Researchers have suggested that massive volcanic eruptions through coal beds in Siberia would generate significant greenhouse gases and cause a runaway greenhouse effect.[121] Massive eruptions can also throw enough pyroclastic debris and other material into the atmosphere to partially block out the sun and cause a volcanic winter, as happened on a smaller scale in 1816 following the eruption of Mount Tambora, the so-called Year Without a Summer. Such an eruption might cause the immediate deaths of millions of people several hundred miles from the eruption, and perhaps billions of deaths[122] worldwide, due to the failure of the monsoon[citation needed], resulting in major crop failures causing starvation on a massive scale.[122]

A much more speculative concept is the Verneshot: a hypothetical volcanic eruption caused by the buildup of gas deep underneath a craton. Such an event may be forceful enough to launch an extreme amount of material from the crust and mantle into a sub-orbital trajectory.

Precautions and prevention[edit]

Planetary management and respecting planetary boundaries have been proposed as approaches to preventing ecological catastrophes. Within the scope of these approaches, the field of geoengineering encompasses the deliberate large-scale engineering and manipulation of the planetary environment to combat or counteract anthropogenic changes in atmospheric chemistry. Space colonization is a proposed alternative to improve the odds of surviving an extinction scenario.[123] Solutions of this scope may require megascale engineering. Food storage has been proposed globally, but the monetary cost would be high. Furthermore, it would likely contribute to the current millions of deaths per year due to malnutrition.

Precautions being taken include:

  • Some survivalists stocking survival retreats with multiple-year food supplies.
  • The Svalbard Global Seed Vault is buried 400 feet (120 m) inside a mountain on an island in the Arctic. It is designed to hold 2.5 billion seeds from more than 100 countries as a precaution to preserve the world's crops. The surrounding rock is −6 °C (21 °F) (as of 2015) but the vault is kept at −18 °C (0 °F) by refrigerators powered by locally sourced coal.[124][125]

Global catastrophic risks and global governance[edit]

Insufficient global governance creates risks in the social and political domain, but the governance mechanisms develop more slowly than technological and social change. There are concerns from governments, the private sector, as well as the general public about the lack of governance mechanisms to efficiently deal with risks, negotiate and adjudicate between diverse and conflicting interests. This is further underlined by an understanding of the interconnectedness of global systemic risks.[126]


The Bulletin of the Atomic Scientists (est. 1945) is one of the oldest global risk organizations, founded after the public became alarmed by the potential of atomic warfare in the aftermath of WWII. It studies risks associated with nuclear war and energy and famously maintains the Doomsday Clock established in 1947. The Foresight Institute (est. 1986) examines the risks of nanotechnology and its benefits. It was one of the earliest organizations to study the unintended consequences of otherwise harmless technology gone haywire at a global scale. It was founded by K. Eric Drexler who postulated "grey goo".[127][128]

Beginning after 2000, a growing number of scientists, philosophers and tech billionaires created organizations devoted to studying global risks both inside and outside of academia.[129]

Independent non-governmental organizations (NGOs) include the Machine Intelligence Research Institute (est. 2000) which aims to reduce the risk of a catastrophe caused by artificial intelligence and the Singularity.[130] The top donors include Peter Thiel and Jed McCaleb.[131] The Lifeboat Foundation (est. 2009) funds research into preventing a technological catastrophe.[132] Most of the research money funds projects at universities.[133] The Global Catastrophic Risk Institute (est. 2011) is a think tank for all things catastrophic risk. It is funded by the NGO Social and Environmental Entrepreneurs. The Global Challenges Foundation (est. 2012), based in Stockholm and founded by Laszlo Szombatfalvy, releases a yearly report on the state of global risks.[13][14] The Future of Life Institute (est. 2014) aims to support research and initiatives for safeguarding life considering new technologies and challenges facing humanity.[134] Elon Musk is one of its biggest donors.[135] The Nuclear Threat Initiative seeks to reduce global threats from nuclear, biological and chemical threats, and containment of damage after an event.[136] It maintains a nuclear material security index.[137]

University-based organizations include the Future of Humanity Institute (est. 2005) which researches the questions of humanity's long-term future, particularly existential risk. It was founded by Nick Bostrom and is based at Oxford University. The Centre for the Study of Existential Risk (est. 2012) is a Cambridge-based organization which studies four major technological risks: artificial intelligence, biotechnology, global warming and warfare. All are man-made risks, as Huw Price explained to the AFP news agency, "It seems a reasonable prediction that some time in this or the next century intelligence will escape from the constraints of biology". He added that when this happens "we're no longer the smartest things around," and will risk being at the mercy of "machines that are not malicious, but machines whose interests don't include us."[138] Stephen Hawking is an acting adviser. The Millennium Alliance for Humanity and the Biosphere is a Stanford University-based organization focusing on many issues related to global catastrophe by bringing together members of academic in the humanities.[139][140] It was founded by Paul Ehrlich among others.[141] Stanford University also has the Center for International Security and Cooperation focusing on political cooperation to reduce global catastrophic risk.[142]

Other risk assessment groups are based in or are part of governmental organizations. The World Health Organization (WHO) includes a division called the Global Alert and Response (GAR) which monitors and responds to global epidemic crisis.[143] GAR helps member states with training and coordination of response to epidemics.[144] The United States Agency for International Development (USAID) has its Emerging Pandemic Threats Program which aims to prevent and contain naturally generated pandemics at their source.[145] The Lawrence Livermore National Laboratory has a division called the Global Security Principal Directorate which researches on behalf of the government issues such as bio-security, counter-terrorism, etc.[146]

See also[edit]


  1. ^ Schulte, P.; et al. (5 March 2010). "The Chicxulub Asteroid Impact and Mass Extinction at the Cretaceous-Paleogene Boundary". Science. 327 (5970): 1214–1218. Bibcode:2010Sci...327.1214S. doi:10.1126/science.1177265. PMID 20203042. 
  2. ^ Bostrom, Nick (2008). Global Catastrophic Risks (PDF). Oxford University Press. p. 1. 
  3. ^ a b c Ripple WJ, Wolf C, Newsome TM, Galetti M, Alamgir M, Crist E, Mahmoud MI, Laurance WF (13 November 2017). "World Scientists' Warning to Humanity: A Second Notice". BioScience. doi:10.1093/biosci/bix125. 
  4. ^ a b Bostrom, Nick (March 2002). "Existential Risks: Analyzing Human Extinction Scenarios and Related Hazards". Journal of Evolution and Technology. 9. 
  5. ^ a b c d e Bostrom, Nick (2013). "Existential Risk Prevention as Global Priority" (PDF). Global Policy. Future of Humanity Institute. 4 (1): 15–3. doi:10.1111/1758-5899.12002 – via Existential Risk. 
  6. ^ a b Bostrom, Nick (2009). "Astronomical Waste: The opportunity cost of delayed technological development". Utilitas. 15 (3): 308–314. doi:10.1017/s0953820800004076. 
  7. ^ Posner, Richard A. (2006). Catastrophe : risk and response. Oxford: Oxford University Press. ISBN 978-0195306477. , Introduction, "What is Catastrophe?"
  8. ^ a b "Observation Selection Effects and Global Catastrophic Risks", Milan Cirkovic, 2008
  9. ^ a b Matheny, Jason Gaverick (2007). "Reducing the Risk of Human Extinction" (PDF). Risk Analysis. 27 (5): 1335–1344. doi:10.1111/j.1539-6924.2007.00960.x. PMID 18076500. 
  10. ^ Asher, D.J.; Bailey, M.E.; Emel'yanenko, V.; Napier, W.M. (2005). "Earth in the cosmic shooting gallery" (PDF). The Observatory. 125: 319–322. Bibcode:2005Obs...125..319A. 
  11. ^ Ambrose 1998; Rampino & Ambrose 2000, pp. 71, 80.
  12. ^ Rampino, M.R.; Ambrose, S.H. (2002). "Super eruptions as a threat to civilizations on Earth-like planets" (PDF). Icarus. 156 (2): 562–569. Bibcode:2002Icar..156..562R. doi:10.1006/icar.2001.6808. 
  13. ^ a b Robinson Meyer (April 29, 2016). "Human Extinction Isn't That Unlikely". The Atlantic. Retrieved April 30, 2016. 
  14. ^ a b "Global Challenges Foundation website". Retrieved April 30, 2016. 
  15. ^ a b Global Catastrophic Risks Survey, Technical Report, 2008, Future of Humanity Institute
  16. ^ Parfit, Derek (1984). Reasons and Persons. Oxford University Press. pp. 453–454. 
  17. ^ Carrington, Damian (21 February 2000). "Date set for desert Earth". BBC News Online. 
  18. ^ Weitzman, Martin (2009). "On modeling and interpreting the economics of catastrophic climate change" (PDF). The Review of Economics and Statistics. 91 (1): 1–19. doi:10.1162/rest.91.1.1. 
  19. ^ Posner, Richard (2004). Catastrophe: Risk and Response. Oxford University Press. 
  20. ^ a b Eliezer Yudkowsky, 2008, "Cognitive Biases potentially affecting judgments of global risks"
  21. ^ Desvousges, W.H., Johnson, F.R., Dunford, R.W., Boyle, K.J., Hudson, S.P., and Wilson, N. 1993, Measuring natural resource damages with contingent valuation: tests of validity and reliability. In Hausman, J.A. (ed), Contingent Valuation:A Critical Assessment, pp. 91−159 (Amsterdam: North Holland).
  22. ^ IPCC (11 November 2013): D. "Understanding the Climate System and its Recent Changes", in: Summary for Policymakers (finalized version), in: IPCC AR5 WG1 2013, p. 13
  23. ^ "Frequently Asked Questions". Existential Risk. Future of Humanity Institute. Retrieved 26 July 2013. 
  24. ^ a b "The Cambridge Project for Existential Risk". Cambridge University. 
  25. ^ "'Terminator center' to open at Cambridge University". Fox News. 2012-11-26. 
  26. ^ Bill Joy, Why the future doesn't need us. Wired magazine.
  27. ^ a b Nick Bostrom 2002 "Ethical Issues in Advanced Artificial Intelligence"
  28. ^ Bostrom, Nick. Superintelligence: Paths, Dangers, Strategies. 
  29. ^ a b Scientists Worry Machines May Outsmart Man By JOHN MARKOFF, NY Times, July 26, 2009.
  30. ^ The Coming Technological Singularity: How to Survive in the Post-Human Era, by Vernor Vinge, Department of Mathematical Sciences, San Diego State University, (c) 1993 by Vernor Vinge.
  31. ^ Rawlinson, Kevin. "Microsoft's Bill Gates insists AI is a threat". BBC News. Retrieved 30 January 2015. 
  32. ^ Gaming the Robot Revolution: A military technology expert weighs in on Terminator: Salvation., By P. W. Singer, Thursday, May 21, 2009.
  33. ^ robot page,
  34. ^ Yudkowsky, Eliezer. "Artificial Intelligence as a Positive and Negative Factor in Global Risk". Retrieved 26 July 2013. 
  35. ^ a b c d e f g h i j Ali Noun; Christopher F. Chyba (2008). "Chapter 20: Biotechnology and biosecurity". In Bostrom, Nick; Cirkovic, Milan M. Global Catastrophic Risks. Oxford University Press. 
  36. ^ a b Sandberg, Anders. "The five biggest threats to human existence". Retrieved 13 July 2014. 
  37. ^ Jackson, Ronald J.; Ramsay, Alistair J.; Christensen, Carina D.; Beaton, Sandra; Hall, Diana F.; Ramshaw, Ian A. (2001). "Expression of Mouse Interleukin-4 by a Recombinant Ectromelia Virus Suppresses Cytolytic Lymphocyte Responses and Overcomes Genetic Resistance to Mousepox". Journal of Virology. 75 (3): 1205–1210. doi:10.1128/jvi.75.3.1205-1210.2001. PMC 114026Freely accessible. PMID 11152493. Retrieved 13 July 2014. 
  38. ^ UCLA Engineering (June 28, 2017). "Scholars assess threats to civilization, life on Earth". UCLA. Retrieved June 30, 2017. 
  39. ^ Isaac M. Held, Brian J. Soden, "Water Vapor Feedback and Global Warming", In: Annu. Rev. Energy Environ 2000. Page 449.
  40. ^ Chiarelli, B. (1998). "Overpopulation and the Threat of Ecological Disaster: the Need for Global Bioethics". Mankind Quarterly. 39 (2): 225–230. 
  41. ^ a b Carrington, Damian (20 October 2017). "Global pollution kills 9m a year and threatens 'survival of human societies'". London, UK: The Guardian. Retrieved 20 October 2017. 
  42. ^ Graham, Chris (July 11, 2017). "Earth undergoing sixth 'mass extinction' as humans spur 'biological annihilation' of wildlife". The Telegraph. Retrieved October 20, 2017. 
  43. ^ Evans-Pritchard, Ambrose (6 February 2011). "Einstein was right - honey bee collapse threatens global food security". The Daily Telegraph. London. 
  44. ^ Lovgren, Stefan. "Mystery Bee Disappearances Sweeping U.S." National Geographic News. URL accessed March 10, 2007.
  45. ^ Georgescu-Roegen, Nicholas (1971). The Entropy Law and the Economic Process (Full book accessible in three parts at SlideShare). Cambridge, Massachusetts: Harvard University Press. ISBN 0674257804. 
  46. ^ Daly, Herman E., ed. (1980). Economics, Ecology, Ethics. Essays Towards a Steady-State Economy (PDF contains only the introductory chapter of the book) (2nd ed.). San Francisco: W.H. Freeman and Company. ISBN 0716711788. 
  47. ^ Rifkin, Jeremy (1980). Entropy: A New World View (PDF). New York: The Viking Press. ISBN 0670297178. Archived from the original (PDF contains only the title and contents pages of the book) on 2016-10-18. 
  48. ^ Boulding, Kenneth E. (1981). Evolutionary Economics. Beverly Hills: Sage Publications. ISBN 0803916485. 
  49. ^ Martínez-Alier, Juan (1987). Ecological Economics: Energy, Environment and Society. Oxford: Basil Blackwell. ISBN 0631171460. 
  50. ^ Gowdy, John M.; Mesner, Susan (1998). "The Evolution of Georgescu-Roegen's Bioeconomics" (PDF). Review of Social Economy. London: Routledge. 56 (2): 136–156. doi:10.1080/00346769800000016. 
  51. ^ Schmitz, John E.J. (2007). The Second Law of Life: Energy, Technology, and the Future of Earth As We Know It (Author's science blog, based on his textbook). Norwich: William Andrew Publishing. ISBN 0815515375. 
  52. ^ Kerschner, Christian (2010). "Economic de-growth vs. steady-state economy" (PDF). Journal of Cleaner Production. Amsterdam: Elsevier. 18: 544–551. doi:10.1016/j.jclepro.2009.10.019. 
  53. ^ Perez-Carmona, Alexander (2013). "Growth: A Discussion of the Margins of Economic and Ecological Thought". In Meuleman, Louis, ed. Transgovernance. Advancing Sustainability Governance (Article accessible at SlideShare). Heidelberg: Springer. pp. 83–161. doi:10.1007/978-3-642-28009-2_3. ISBN 9783642280085. 
  54. ^ Bostrom 2002, section 4.8
  55. ^ Richard Hamming. "Mathematics on a Distant Planet". 
  56. ^ "Report LA-602, ''Ignition of the Atmosphere With Nuclear Bombs''" (PDF). Retrieved 2011-10-19. 
  57. ^ New Scientist, 28 August 1999: "A Black Hole Ate My Planet"
  58. ^ Konopinski, E. J; Marvin, C.; Teller, Edward (1946). "Ignition of the Atmosphere with Nuclear Bombs" (PDF) (Declassified February 1973) (LA–602). Los Alamos National Laboratory. Retrieved 23 November 2008. 
  59. ^ "Statement by the Executive Committee of the DPF on the Safety of Collisions at the Large Hadron Collider." Archived 2009-10-24 at the Wayback Machine.
  60. ^ "Safety at the LHC". 
  61. ^ J. Blaizot et al., "Study of Potentially Dangerous Events During Heavy-Ion Collisions at the LHC", CERN library record CERN Yellow Reports Server (PDF)
  62. ^ Eric Drexler, Engines of Creation, ISBN 0-385-19973-2, available online
  63. ^ a b c d e f g h i j k l m n Chris Phoenix; Mike Treder (2008). "Chapter 21: Nanotechnology as global catastrophic risk". In Bostrom, Nick; Cirkovic, Milan M. Global catastrophic risks. Oxford: Oxford University Press. ISBN 978-0-19-857050-9. 
  64. ^ a b "Frequently Asked Questions - Molecular Manufacturing". Retrieved 19 July 2014. 
  65. ^ Drexler, Eric. "A Dialog on Dangers". Retrieved 19 July 2014. 
  66. ^ Drexler, Eric. "ENGINES OF DESTRUCTION (Chapter 11)". Retrieved 19 July 2014. 
  67. ^ "Dangers of Molecular Manufacturing". Retrieved 19 July 2014. 
  68. ^ a b "The Need for International Control". Retrieved 19 July 2014. 
  69. ^ "Technical Restrictions May Make Nanotechnology Safer". Retrieved 19 July 2014. 
  70. ^ Joseph, Lawrence E. (2007). Apocalypse 2012. New York: Broadway. p. 6. ISBN 978-0-7679-2448-1. 
  71. ^ Rincon, Paul (2004-06-09). "Nanotech guru turns back on 'goo'". BBC News. Retrieved 2012-03-30. 
  72. ^ Hapgood, Fred (November 1986). "Nanotechnology: Molecular Machines that Mimic Life" (PDF). Omni. Retrieved 19 July 2014. 
  73. ^ "Leading nanotech experts put 'grey goo' in perspective". Retrieved 19 July 2014. 
  74. ^ "On the Probability of Nuclear War" by Martin E. Hellman
  75. ^ Nuclear Weapons and the Future of Humanity: The Fundamental Questions by Avner Cohen, Steven Lee, p. 237, at Google Books
  76. ^ a b Federation of American Scientists (28 April 2015). "Status of World Nuclear Forces". Federation of American Scientists. Archived from the original on 18 June 2015. Retrieved 4 June 2015. 
  77. ^ a b Martin, Brian (1982). "Critique of nuclear extinction". Journal of Peace Research. 19 (4): 287–300. doi:10.1177/002234338201900401. Retrieved 25 October 2014. 
  78. ^ Shulman, Carl (5 Nov 2012). "Nuclear winter and human extinction: Q&A with Luke Oman". Overcoming Bias. Retrieved 25 October 2014. 
  79. ^ "Atmospheric effects and societal consequences of regional-scale nuclear conflicts and acts of individual nuclear terrorism", Atmospheric Chemistry and Physics
  80. ^ Bostrom 2002, section 4.2.
  81. ^ "The end of India's green revolution?". BBC News. 2006-05-29. Retrieved 2012-01-31. 
  82. ^ April 8th, 2000 by admin (2000-04-08). "Food First/Institute for Food and Development Policy". Archived from the original on July 14, 2009. Retrieved 2012-01-31. 
  83. ^ "How peak oil could lead to starvation". 2009-05-27. Archived from the original on May 27, 2009. Retrieved 2012-01-31. 
  84. ^ "Eating Fossil Fuels". 2003-10-02. Retrieved 2012-01-31. 
  85. ^ The Oil Drum: Europe. "Agriculture Meets Peak Oil". Retrieved 2012-01-31. 
  86. ^ "Drawing Momentum from the Crash" by Dale Allen Pfeiffer
  87. ^ "Cereal Disease Laboratory : Ug99 an emerging virulent stem rust race". Retrieved 2012-01-31. 
  88. ^ "Durable Rust Resistance in Wheat". Retrieved 2012-01-31. 
  89. ^ a b Gehrels, Tom; Matthews, Mildred Shapley; Schumann, A. M. (1994). Hazards Due to Comets and Asteroids. University of Arizona Press. p. 71. ISBN 9780816515059. 
  90. ^ "How Big Would A Meteorite Have To Be To Wipe Out All Human Life?". Popular Science. 26 February 2015. Retrieved 13 February 2018. 
  91. ^ Bostrom 2002, section 4.10
  92. ^ "Committee on Science, Space and Technology" (PDF). NASA. 19 March 2013. Retrieved 13 February 2018. 
  93. ^ García-Sánchez, Joan; et al. (February 1999). "Stellar Encounters with the Oort Cloud Based on HIPPARCOS Data". The Astronomical Journal. 117 (2): 1042–1055. Bibcode:1999AJ....117.1042G. doi:10.1086/300723. 
  94. ^ Twenty ways the world could end suddenly Archived 2004-09-24 at the Wayback Machine., Discover Magazine
  95. ^ Urban Legends Reference Pages: Legal Affairs (E.T. Make Bail)
  96. ^ Bostrom 2002, section 7.2
  97. ^ Johnson, Steven (2017-06-28). "Greetings, E.T. (Please Don't Murder Us.)". The New York Times. ISSN 0362-4331. Retrieved 2017-06-29. 
  98. ^ Ken Croswell, Will Mercury Hit Earth Someday?, April 24, 2008, accessed April 26, 2008
  99. ^ Bostrom 2002, section 4.7
  100. ^ Fraser Cain (2003-08-04). "Local Galactic Dust is on the Rise". Universe Today. 
  101. ^ Coleman, Sidney; De Luccia, Frank (1980-06-15). "Gravitational effects on and of vacuum decay" (PDF). Physical Review D. D21 (12): 3305–3315. Bibcode:1980PhRvD..21.3305C. doi:10.1103/PhysRevD.21.3305. 
  102. ^ Plotnick, Roy E. (1 January 1980). "Relationship between biological extinctions and geomagnetic reversals". Geology. 8 (12): 578. Bibcode:1980Geo.....8..578P. doi:10.1130/0091-7613(1980)8<578:RBBEAG>2.0.CO;2. 
  103. ^ Glassmeier, Karl-Heinz; Vogt, Joachim (29 May 2010). "Magnetic Polarity Transitions and Biospheric Effects". Space Science Reviews. 155 (1–4): 387–410. Bibcode:2010SSRv..155..387G. doi:10.1007/s11214-010-9659-6. 
  104. ^ "Near Apocalypse Causing Diseases, a Historical Look:". Retrieved 2012-05-05. 
  105. ^ a b Frank SA (March 1996). "Models of parasite virulence" (PDF). Q Rev Biol. 71 (1): 37–78. doi:10.1086/419267. PMID 8919665. 
  106. ^ Brown NF, Wickham ME, Coombes BK, Finlay BB (May 2006). "Crossing the Line: Selection and Evolution of Virulence Traits". PLoS Pathogens. 2 (5): e42. doi:10.1371/journal.ppat.0020042. PMC 1464392Freely accessible. PMID 16733541. 
  107. ^ Ebert D, Bull JJ (January 2003). "Challenging the trade-off model for the evolution of virulence: is virulence management feasible?". Trends Microbiol. 11 (1): 15–20. doi:10.1016/S0966-842X(02)00003-3. PMID 12526850. 
  108. ^ André JB, Hochberg ME (July 2005). "Virulence evolution in emerging infectious diseases". Evolution. 59 (7): 1406–12. doi:10.1554/05-111. PMID 16153027. 
  109. ^ Gandon S (March 2004). "Evolution of multihost parasites". Evolution. 58 (3): 455–69. doi:10.1111/j.0014-3820.2004.tb01669.x. PMID 15119430. 
  110. ^ Pararas-Carayannis, George (2002). "Evaluation of the threat of mega tsunami generation from postulated massive slope failures of island volcanoes on La Palma, Canary Islands, and on the island of Hawaii". Retrieved 2008-12-20. 
  111. ^ Prehistoric Asteroid "Killed Everything"
  112. ^ Kate Ravilious (2005-04-14). "What a way to go". The Guardian. 
  113. ^ 2012 Admin (2008-02-04). "Toba Supervolcano". 2012 Final Fantasy. Archived from the original on 2010-08-22. 
  114. ^ Science Reference. "Toba Catastrophe Theory". Science Daily. 
  115. ^ Greg Breining (10 November 2007). "The Next Big Blast". Super Volcano: The Ticking Time Bomb Beneath Yellowstone National Park. MBI Publishing Company. ISBN 978-1-61673-898-3. 
  116. ^ Greg Breining (10 November 2007). "Distant Death". Super Volcano: The Ticking Time Bomb Beneath Yellowstone National Park. MBI Publishing Company. ISBN 978-1-61673-898-3. 
  117. ^ "Modeling the Ash Distribution of a Yellowstone Supereruption". USGS Volcanic Observatory. 
  118. ^ "Extreme Geohazards: Reducing the Disaster Risk and Increasing Resilience" (PDF). European Space Foundation. 
  119. ^ "Questions About Future Volcanic Activity at Yellowstone". USGA Volcanic Observatory FAQ. 
  120. ^ "Steam Explosions, Earthquakes, and Volcanic Eruptions—What's in Yellowstone's Future?". USGS Yellowstone Volcanic Observatory. The USGS puts it like this: “If another large caldera-forming eruption were to occur at Yellowstone, its effects would be worldwide. Thick ash deposits would bury vast areas of the United States, and injection of huge volumes of volcanic gases into the atmosphere could drastically affect global climate. Fortunately, the Yellowstone volcanic system shows no signs that it is headed toward such an eruption. The probability of a large caldera-forming eruption within the next few thousand years is exceedingly low.” 
  121. ^ "World's biggest extinction event: Massive volcanic eruption, burning coal and accelerated greenhouse gas choked out life -- ScienceDaily". Https:. Retrieved 28 September 2016. 
  122. ^ a b Breining, Greg (2007). "The Next Big Blast". Super Volcano: The Ticking Time Bomb Beneath Yellowstone National Park. St. Paul, MN.: Voyageur Press. p. 256 pg. ISBN 978-0-7603-2925-2. 
  123. ^ "Mankind must abandon earth or face extinction: Hawking",, August 9, 2010, retrieved 2012-01-23 
  124. ^ Lewis Smith (2008-02-27). "Doomsday vault for world's seeds is opened under Arctic mountain". London: The Times Online. Archived from the original on 2008-05-12. 
  125. ^ Suzanne Goldenberg (May 20, 2015). "The doomsday vault: the seeds that could save a post-apocalyptic world". The Guardian. Retrieved June 30, 2017. 
  126. ^ "Global Challenges Foundation | Understanding Global Systemic Risk". Retrieved 2017-08-15. 
  127. ^ Fred Hapgood (November 1986). "Nanotechnology: Molecular Machines that Mimic Life" (PDF). Omni. Retrieved June 5, 2015. 
  128. ^ Giles, Jim (2004). "Nanotech takes small step towards burying 'grey goo'". Nature. 429 (6992): 591. Bibcode:2004Natur.429..591G. doi:10.1038/429591b. PMID 15190320. 
  129. ^ Sophie McBain (September 25, 2014). "Apocalypse soon: the scientists preparing for the end times". New Statesman. Retrieved June 5, 2015. 
  130. ^ "Reducing Long-Term Catastrophic Risks from Artificial Intelligence". Machine Intelligence Research Institute. Retrieved June 5, 2015. The Machine Intelligence Research Institute aims to reduce the risk of a catastrophe, should such an event eventually occur. 
  131. ^ Angela Chen (September 11, 2014). "Is Artificial Intelligence a Threat?". The Chronicle of Higher Education. Retrieved June 5, 2015. 
  132. ^ "About the Lifeboat Foundation". The Lifeboat Foundation. Retrieved 26 April 2013. 
  133. ^ Ashlee Vance (July 20, 2010). "The Lifeboat Foundation: Battling Asteroids, Nanobots and A.I." New York Times. Retrieved June 5, 2015. 
  134. ^ "The Future of Life Institute". Retrieved May 5, 2014. 
  135. ^ Nick Bilton (May 28, 2015). "Ava of 'Ex Machina' Is Just Sci-Fi (for Now)". New York Times. Retrieved June 5, 2015. 
  136. ^ "Nuclear Threat Initiative". Nuclear Threat Initiative. Retrieved June 5, 2015. 
  137. ^ Alexander Sehmar (May 31, 2015). "Isis could obtain nuclear weapon from Pakistan, warns India". The Independent. Retrieved June 5, 2015. 
  138. ^ Hui, Sylvia (25 November 2012). "Cambridge to study technology's risks to humans". Associated Press. Archived from the original on 1 December 2012. Retrieved 30 January 2012. 
  139. ^ Scott Barrett (2014). Environment and Development Economics: Essays in Honour of Sir Partha Dasgupta. Oxford University Press. p. 112. Retrieved June 5, 2015. 
  140. ^ "Millennium Alliance for Humanity & The Biosphere". Millennium Alliance for Humanity & The Biosphere. Retrieved June 5, 2015. 
  141. ^ Guruprasad Madhavan (2012). Practicing Sustainability. Springer Science & Business Media. p. 43. Retrieved June 5, 2015. 
  142. ^ "Center for International Security and Cooperation". Center for International Security and Cooperation. Retrieved June 5, 2015. 
  143. ^ "Global Alert and Response (GAR)". World Health Organization. Retrieved June 5, 2015. 
  144. ^ Kelley Lee (2013). Historical Dictionary of the World Health Organization. Rowman & Littlefield. p. 92. Retrieved June 5, 2015. 
  145. ^ "USAID Emerging Pandemic Threats Program". USAID. Retrieved June 5, 2015. 
  146. ^ "Global Security". Lawrence Livermore National Laboratory. Retrieved June 5, 2015. 


Further reading[edit]

External links[edit]