Existential Risks


What are the risks that may destroy our civilization and Humanity? In most general terms, these risks could be grouped as global catastrophic or existential (terminal), whenclassified according to their scope and severity. A “global catastrophic risk” is any risk that is “global” in scope, lasting for some time (endurable) and hence may kill the vast majority of life on earth but humanity could still potentially recover. An “existential” risk on the other hand, is such that its intensity is terminal and its effect is transgenerational. It can destroy all human, non-human and even plant life.

An example of a global catastrophic risk that could destroy the world in a material sense and at the same time potentially eliminate human race is a global nuclear war. It could immediately wipe out most of the current material substance of civilization, i.e. towns, infrastructure, food crops, etc. and in the longer term, through radiation, lack of food and the emergence of post-nuclear winter, causing the death of all remaining people.

A global pandemic, is an example of an existential risk. It could be caused by an accidental or intentional release from laboratories of a deadly virus that would wipe out the human race in weeks but leave the infrastructure undamaged, at least for a few years.

Furthermore, both global catastrophic and existential risks could be divided into two broad groups: anthropogenic, which Humanity could to some extent control and mitigate (e.g. global warming), and non-anthropogenic over which we have no control (e.g. asteroid impact).

Some catastrophic risks are natural, such as super-volcanos or an impact of a large asteroid that can cause extinction of many species. Earth has experienced such mass extinctions in the past. This also includes global warming, which in the past was natural and today is anthropogenic i.e. man-made. Other such anthropogenic risks are pandemics caused by artificially created biologically active viruses, or nuclear wars. Perhaps that’s why Nick Bostrom of Future of Humanity Institute believes that human extinction is more likely to result from anthropogenic causes than natural causes.

If our civilization is to survive, we need to apply some powerful risk mitigation strategies. We have hardly any control over the non-anthropogenic (natural) risks. But we do have control over political, social, economic and technological risks. The anthropogenic (man-made) existential risks can be split into three categories, which may require different approaches:

Risks that are immediate and may become existential within days or even in hours:

  1. Global nuclear war
  2. Weaponized AI or cyber wars
  3. Engineered pandemics and synthetic biology
  4. Nanotechnology and experimental technology accident
  5. Unknown risks, mainly technology-orientated

Risks that may become existential progressively:

  1. Climate Change over a long time (at least over a century)
  2. Superintelligence in a short time (over a decade)

Leave a Reply

Your email address will not be published.

five × four =