Our Delicate Future: Handle with Care Our Delicate Future: Handle with Care
Our Delicate Future: Handle with Care
Why we need to get serious about the threats that could wipe out humanity.
Part nine in a GOOD miniseries on the singularity by Michael Anissimov and Roko Mijic. New posts every Monday from November 16 to January 23.
In his book Reasons and Persons, the philosopher Derek Parfit asks us to compare two scenarios: A nuclear war that kills 99 percent of the world's existing population, or a nuclear war that kills 100 percent. The first outcome would be a horrible disaster, but the second would be an existential disaster: one that destroys the human race or irreversibly curtails our whole future.
Call the chance of such an event an "existential risk." The future promises a whole host of new ways the human race could be snuffed out: from AI gone wrong (see this previous article) to nanotechnology to synthetic biology and engineered viruses.
What makes existential disasters worse than even widespread personal disasters like cancer, which has killed billions of people? The future potential of the human race is the key difference: If a disaster kills all the people on the planet, then there will be no one to continue the human race. Our species has come a long way since the dawn of history, and if we work to preserve our humanity, our civilization, and our values, we may go a long way yet. There is a whole universe out there and it is huge beyond our wildest dreams-our own galaxy contains more solar systems than there are people on the planet, and our galaxy itself is merely a tiny mote of dust in the great sea of galaxies. If the human race can spread out and colonize a few dozen systems, it seems likely that it will also be able to colonize the entire reachable universe, which contains a whopping 1,000,000,000,000,000,000,000 stars, and will last for perhaps 100,000,000,000 years.
The number of good human lives that could be lived in this time is simply too large to comfortably contemplate, and all those lives currently hang in the balance between existence and nonexistence, like an innumerable audience of ghostly figures looking down anxiously at the early twenty-first century. If any of the existential disasters actually occur, this future will be wiped out. In effect, the indirect death toll from an existential disaster is so big that all other disasters or humanitarian causes pale into complete insignificance in comparison.
Cosmologists, physicists who study the origin and evolution of the universe, know this better than anyone else. One world-renowned cosmologist after another has spoken out about existential risks. Stephen Hawking sits on the board of the Bulletin of the Atomic Scientists, and has warned that mistakes in constructing artificial intelligence risk human extinction as well. Carl Sagan was active in efforts to avert nuclear war. Martin Rees, in his book Our Final Hour, has warned of risks of bioterrorism and biowarfare.
It is a tall order to preserve not just human civilization, but also our human values, through the disruptive forces of time, technology, and Darwinian selection. The "fragility of human values" thesis claims that many desirable properties that a possible future could have are delicate properties of human brains and culture that could easily be disrupted by technological, economic, and geopolitical forces. We value laughter, friendship, children, sexuality, music, art, family, humor, and nature, amongst many other things. All of these things exist only because the human brain happens to create them, and the human brain is also currently the most intelligent thinking machine in the world. Each day, women give birth to children who also have brains which contain that same magical combination of human value and useful intelligence. But if we create entities that are more competitive and intelligent than humans-entities that did not also value laughter, friendship, children, sexuality, music, art, family, humor, and nature-then there would be a serious risk that humans would lose control of the future, and that these new, inhuman minds would create the world as they saw fit, or as emergent political and economic trends dictated; a grim universe of shuffling electrons and economic transactions with no people and no joy. These dark futures where human civilization gradually shifts away from human values without a "bang" also count as existential risks, and are perhaps the most insidious and horrific to think about.
Existential risks are, from one point of view, the most important and pressing problem in the world, for they threaten humanity as a whole. Unfortunately, they are also grossly under-addressed in terms of research, funding, and effort. Gaverick Matheny writes, in his paper on reducing human extinction risks (pdf), that "A search of EconLit and the Social Sciences Citation Index suggests that virtually nothing has been written about the cost effectiveness of reducing human extinction risks," and Nick Bostrom and Anders Sandberg noted, in a personal communication, that there are orders of magnitude more papers on coleoptera-the study of beetles-than "human extinction." Anyone can confirm this for themselves with a Google Scholar search: coleoptera gets 245,000 hits, and "human extinction" gets fewer than 1,200.
This means that small groups or even individuals can make a difference to the outcome, by reading and understanding the subject or by supporting research into understanding and avoiding these risks at places like Oxford University's Future of Humanity Institute.
The potential of the human race is virtually limitless, but first we have to survive into the next century.
Roko Mijic is a Cambridge University mathematics graduate, and has worked in ultra low-temperature engineering, pure mathematics, digital evolution and artificial intelligence. In his spare time he blogs about the future of the human race and the philosophical foundations of ethics and human values.
When Humans Fight, but Animals Win Penguins have resorted to using landmines to keep pesky humans away.
So You Think You’re a Foodie? Pop culture was onto these trends way before you were. A sampling of the screwball comedies, sob stories, and sci-fis that anticipated our culinary moment
Dear Nine-Year-Old Me The transition is going to be difficult for you, but whenever you feel a little lonely and left out, take comfort in the knowledge that you are honing one of your greatest superpowers.
What to Do When Your Country is Drowning The wild and desperate ways island nations are fighting the effects of climate change
The Rise of Drone Pizza Delivery Why the skies will soon be filled with flying, snack-bearing robots
How Helsinki Became a Public Transporation Paradise One European city plans to make car ownership obsolete within a decade.
Follow the Crowd NanoCrafter and the rise of group intelligence Why online gaming may just be the future of science
The Empathy Mirror Neurofeedback enables us to better see ourselves in the other. Recent discoveries in neurofeedback can teach you to be less of a dick.
Robots On Ice Probe the Arctic Why a team of research robots is investigating disappearing sea ice, and why you should care
Don’t Turn Away Colin Finlay photographs the consequences of climate change. You will never see more beautiful photos of the deteriorating state of our planet than the ones in this photo feature.
Puppy Love How dogecoin spawned an improbable community of giving What a canine-emblazoned cryptocurrency can teach about philanthropy
Positive In, Positive Out: How a USC Alumna is Coping with Lymphoma Coast Guard Reserves member Cassie Sulfridge, 28, had just graduated from MSW@USC, the Southern California university’s web-based Master of Social Work program, and was working two jobs when her life was turned upside down.