15/05/2020

Existential Risk and the Future of Humanity









The world is in the early stages of what may be the most deadly pandemic of the past 100 years. In China, thousands of people have already died; large outbreaks have begun in South Korea, Iran and Italy; and the rest of the world is bracing for impact. We do not yet know whether the final toll will be measured in thousands or hundreds of thousands. For all our advances in medicine, humanity remains much more vulnerable to pandemics than we would like to believe.

To understand our vulnerability, and to determine what steps must be taken to end it, it is useful to ask about the very worst-case scenarios. Just how bad could a pandemic be? In science fiction, we sometimes encounter the idea of a pandemic so severe that it could cause the end of civilisation, or even of humanity itself. Such a risk to humanity’s entire future is known as an existential risk. We can say with certainty that the novel coronavirus, named Covid-19, does not pose such a risk. But could the next pandemic? To find out, and to put the current outbreak into greater context, let us turn to the past.

In 1347, death came to Europe. It entered through the Crimean town of Caffa, brought by the besieging Mongol army. Fleeing merchants unwittingly carried it back to Italy. From there, it spread to France, Spain and England. Then up as far as Norway and across the rest of Europe – all the way to Moscow. Within six years, the Black Death had taken the continent.

Tens of millions fell gravely ill, their bodies succumbing to the disease in different ways. Some bore swollen buboes on their necks, armpits and thighs; some had their flesh turn black from haemorrhaging beneath the skin; some coughed blood from the necrotic inflammation of their throats and lungs. All forms involved fever, exhaustion and an intolerable stench from the material that exuded from the body.

There were so many dead that mass graves needed to be dug and, even then, cemeteries ran out of room for the bodies. The Black Death devastated Europe. In those six years, between a quarter and half of all Europeans were killed. The Middle East was ravaged, too, with the plague killing about one in three Egyptians and Syrians. And it may have also laid waste to parts of central Asia, India and China. Due to the scant records of the 14th century, we will never know the true toll, but our best estimates are that somewhere between 5% and 14% of all the world’s people were killed, in what may have been the greatest catastrophe humanity has seen.

The Black Death was not the only biological disaster to scar human history. It was not even the only great bubonic plague. In AD541 the plague of Justinian struck the Byzantine empire. Over three years, it took the lives of roughly 3% of the world’s people.

When Europeans reached the Americas in 1492, the two populations exposed each other to completely novel diseases. Over thousands of years, each population had built up resistance to their own set of diseases, but were extremely susceptible to the others. The American peoples got by far the worse end of the exchange, through diseases such as measles, influenza and, especially, smallpox.

During the next 100 years, a combination of invasion and disease took an immense toll – one whose scale may never be known, due to great uncertainty about the size of the pre-existing population. We can’t rule out the loss of more than 90% of the population of the Americas during that century, though the number could also be much lower. And it is very difficult to tease out how much of this should be attributed to war and occupation, rather than disease. At a rough estimate, as many as 10% of the world’s people may have been killed.





Centuries later, the world had become so interconnected that a truly global pandemic was possible. Towards the end of the first world war, a devastating strain of influenza, known as the 1918 flu or Spanish flu, spread to six continents, and even remote Pacific islands. About a third of the world’s population were infected and between 3% and 6% were killed. This death toll outstripped that of the first world war.

Yet even events like these fall short of being a threat to humanity’s long-term potential. In the great bubonic plagues we saw civilisation in the affected areas falter, but recover. The regional 25%-50% death rate was not enough to precipitate a continent-wide collapse. It changed the relative fortunes of empires, and may have substantially altered the course of history, but if anything, it gives us reason to believe that human civilisation is likely to make it through future events with similar death rates, even if they were global in scale.

The Spanish flu pandemic was remarkable in having very little apparent effect on the world’s development, despite its global reach. It looks as if it was lost in the wake of the first world war, which, despite a smaller death toll, seems to have had a much larger effect on the course of history.

The full history of humanity covers at least 200,000 years. While we have scarce records for most of these 2,000 centuries, there is a key lesson we can draw from the sheer length of our past. The chance of human extinction from natural catastrophes of any kind must have been very low for most of this time – or we would not have made it so far. But could these risks have changed? Might the past provide false comfort?

Our population now is a thousand times greater than it was for most of human history, so there are vastly more opportunities for new human diseases to originate. And our farming practices have created vast numbers of animals living in unhealthy conditions within close proximity to humans. This increases the risk, as many major diseases originate in animals before crossing over to humans. Examples include HIV (chimpanzees), Ebola (bats), Sars (probably civets or bats) and influenza (usually pigs or birds). We do not yet know where Covid-19 came from, though it is very similar to coronaviruses found in bats and pangolins. Evidence suggests that diseases are crossing over into human populations from animals at an increasing rate.




Modern civilisation may also make it much easier for a pandemic to spread. The higher density of people living together in cities increases the number of people each of us may infect. Rapid long-distance transport greatly increases the distance pathogens can spread, reducing the degrees of separation between any two people. Moreover, we are no longer divided into isolated populations as we were for most of the past 10,000 years.

Together these effects suggest that we might expect more new pandemics, for them to spread more quickly, and to reach a higher percentage of the world’s people.

But we have also changed the world in ways that offer protection. We have a healthier population; improved sanitation and hygiene; preventative and curative medicine; and a scientific understanding of disease. Perhaps most importantly, we have public health bodies to facilitate global communication and coordination in the face of new outbreaks. We have seen the benefits of this protection through the dramatic decline of endemic infectious disease over the past century (though we can’t be sure pandemics will obey the same trend). Finally, we have spread to a range of locations and environments unprecedented for any mammalian species. This offers special protection from extinction events, because it requires the pathogen to be able to flourish in a vast range of environments and to reach exceptionally isolated populations such as uncontacted tribes, Antarctic researchers and nuclear submarine crews.

It is hard to know whether these combined effects have increased or decreased the existential risk from pandemics. This uncertainty is ultimately bad news: we were previously sitting on a powerful argument that the risk was tiny; now we are not.

We have seen the indirect ways that our actions aid and abet the origination and spread of pandemics. But what about cases where we have a much more direct hand in the process – where we deliberately use, improve or create the pathogens?

Our understanding and control of pathogens is very recent. Just 200 years ago, we didn’t even understand the basic cause of pandemics – a leading theory in the west claimed that disease was produced by a kind of gas. In just two centuries, we discovered it was caused by a diverse variety of microscopic agents and we worked out how to grow them in the lab, to breed them for different traits, to sequence their genomes, to implant new genes and to create entire functional viruses from their written code.

This progress is continuing at a rapid pace. The past 10 years have seen major qualitative breakthroughs, such as the use of the gene editing tool Crispr to efficiently insert new genetic sequences into a genome, and the use of gene drives to efficiently replace populations of natural organisms in the wild with genetically modified versions.

This progress in biotechnology seems unlikely to fizzle out anytime soon: there are no insurmountable challenges looming; no fundamental laws blocking further developments. But it would be optimistic to assume that this uncharted new terrain holds only familiar dangers.

To start with, let’s set aside the risks from malicious intent, and consider only the risks that can arise from well-intentioned research. Most scientific and medical research poses a negligible risk of harms at the scale we are considering. But there is a small fraction that uses live pathogens of kinds that are known to threaten global harm. These include the agents that cause the Spanish flu, smallpox, Sars and H5N1 or avian flu. And a small part of this research involves making strains of these pathogens that pose even more danger than the natural types, increasing their transmissibility, lethality or resistance to vaccination or treatment.

In 2012, a Dutch virologist, Ron Fouchier, published details of an experiment on the recent H5N1 strain of bird flu. This strain was extremely deadly, killing an estimated 60% of humans it infected – far beyond even the Spanish flu. Yet its inability to pass from human to human had so far prevented a pandemic. Fouchier wanted to find out whether (and how) H5N1 could naturally develop this ability. He passed the disease through a series of 10 ferrets, which are commonly used as a model for how influenza affects humans. By the time it passed to the final ferret, his strain of H5N1 had become directly transmissible between mammals.

The work caused fierce controversy. Much of this was focused on the information contained in his work. The US National Science Advisory Board for Biosecurity ruled that his paper had to be stripped of some of its technical details before publication, to limit the ability of bad actors to cause a pandemic. And the Dutch government claimed that the research broke EU law on exporting information useful for bioweapons. But it is not the possibility of misuse that concerns me here. Fouchier’s research provides a clear example of well-intentioned scientists enhancing the destructive capabilities of pathogens known to threaten global catastrophe.

Of course, such experiments are done in secure labs, with stringent safety standards. It is highly unlikely that in any particular case the enhanced pathogens would escape into the wild. But just how unlikely? Unfortunately, we don’t have good data, due to a lack of transparency about incident and escape rates. This prevents society from making well-informed decisions balancing the risks and benefits of this research, and it limits the ability of labs to learn from each other’s incidents.

Security for highly dangerous pathogens has been deeply flawed, and remains insufficient. In 2001, Britain was struck by a devastating outbreak of foot-and-mouth disease in livestock. Six million animals were killed in an attempt to halt its spread, and the economic damages totalled £8bn. Then, in 2007, there was another outbreak, which was traced to a lab working on the disease. Foot-and-mouth was considered a highest-category pathogen, and required the highest level of biosecurity. Yet the virus escaped from a badly maintained pipe, leaking into the groundwater at the facility. After an investigation, the lab’s licence was renewed – only for another leak to occur two weeks later.

In my view, this track record of escapes shows that even the highest biosafety level (BSL-4) is insufficient for working on pathogens that pose a risk of global pandemics on the scale of the Spanish flu or worse. Thirteen years since the last publicly acknowledged outbreak from a BSL-4 facility is not good enough. It doesn’t matter whether this is from insufficient standards, inspections, operations or penalties. What matters is the poor track record in the field, made worse by a lack of transparency and accountability. With current BSL-4 labs, an escape of a pandemic pathogen is only a matter of time.



One of the most exciting trends in biotechnology is its rapid democratisation – the speed at which cutting-edge techniques can be adopted by students and amateurs. When a new breakthrough is achieved, the pool of people with the talent, training, resources and patience to reproduce it rapidly expands: from a handful of the world’s top biologists, to people with PhDs in the field, to millions of people with undergraduate-level biology.

The Human Genome Project was the largest ever scientific collaboration in biology. It took 13 years and $500m to produce the full DNA sequence of the human genome. Just 15 years later, a genome can be sequenced for under $1,000, and within a single hour. The reverse process has become much easier, too: online DNA synthesis services allow anyone to upload a DNA sequence of their choice then have it constructed and shipped to their address. While still expensive, the price of synthesis has fallen by a factor of 1,000 in the past two decades, and continues to drop. The first ever uses of Crispr and gene drives were the biotechnology achievements of the decade. But within just two years, each of these technologies were used successfully by bright students participating in science competitions.

Such democratisation promises to fuel a boom of entrepreneurial biotechnology. But since biotechnology can be misused to lethal effect, democratisation also means proliferation. As the pool of people with access to a technique grows, so does the chance it contains someone with malign intent.

People with the motivation to wreak global destruction are mercifully rare. But they exist. Perhaps the best example is the Aum Shinrikyo cult in Japan, active between 1984 and 1995, which sought to bring about the destruction of humanity. It attracted several thousand members, including people with advanced skills in chemistry and biology. And it demonstrated that it was not mere misanthropic ideation. It launched multiple lethal attacks using VX gas and sarin gas, killing more than 20 people and injuring thousands. It attempted to weaponise anthrax, but did not succeed. What happens when the circle of people able to create a global pandemic becomes wide enough to include members of such a group? Or members of a terrorist organisation or rogue state that could try to build an omnicidal weapon for the purposes of extortion or deterrence?




The main candidate for biological existential risk in the coming decades thus stems from technology – particularly the risk of misuse by states or small groups. But this is not a case in which the world is blissfully unaware of the risks. Bertrand Russell wrote of the danger of extinction from biowarfare to Einstein in 1955. And, in 1969, the possibility was raised by the American Nobel laureate for medicine, Joshua Lederberg: “As a scientist I am profoundly concerned about the continued involvement of the United States and other nations in the development of biological warfare. This process puts the very future of human life on earth in serious peril.”

In response to such warnings, we have already begun national and international efforts to protect humanity. There is action through public health and international conventions, and self-regulation by biotechnology companies and the scientific community. Are they adequate?

National and international work in public health offers some protection from engineered pandemics, and its existing infrastructure could be adapted to better address them. Yet even for existing dangers this protection is uneven and under-provided.

Despite its importance, public health is underfunded worldwide, and poorer countries remain vulnerable to being overwhelmed by outbreaks. Biotechnology companies are working to limit the dark side of the democratisation of their field. For example, unrestricted DNA synthesis would help bad actors overcome a major hurdle in creating extremely deadly pathogens. It would allow them to get access to the DNA of controlled pathogens such as smallpox (whose genome is readily available online) and to create DNA with modifications to make the pathogen more dangerous. Therefore, many synthesis companies make voluntary efforts to manage this risk, screening their orders for dangerous sequences. But the screening methods are imperfect, and they only cover about 80% of orders. There is significant room for improving this process, and a strong case for making screening mandatory.

We might also look to the scientific community for careful management of biological risks. Many of the dangerous advances usable by states and small groups have come from open science. And we’ve seen that science produces substantial accident risk. The scientific community has tried to regulate its dangerous research, but with limited success. There are a variety of reasons why this is extremely hard, including difficulty in knowing where to draw the line, lack of central authorities to unify practice, a culture of openness and freedom to pursue whatever is of interest, and the rapid pace of science outpacing that of governance. It may be possible for the scientific community to overcome these challenges and provide strong management of global risks, but it would require a willingness to accept serious changes to its culture and governance – such as treating the security around biotechnology more like that around nuclear power. And the scientific community would need to find this willingness before catastrophe strikes.

Threats to humanity, and how we address them, define our time. The advent of nuclear weapons posed a real risk of human extinction in the 20th century. There is strong reason to believe the risk will be higher this century, and increasing with each century that technological progress continues. Because these anthropogenic risks outstrip all natural risks combined, they set the clock on how long humanity has left to pull back from the brink.

I am not claiming that extinction is the inevitable conclusion of scientific progress, or even the most likely outcome. What I am claiming is that there has been a robust trend towards increases in the power of humanity, which has reached a point where we pose a serious risk to our own existence. How we react to this risk is up to us. Nor am I arguing against technology. Technology has proved itself immensely valuable in improving the human condition.

The problem is not so much an excess of technology as a lack of wisdom. Carl Sagan put this especially well: “Many of the dangers we face indeed arise from science and technology – but, more fundamentally, because we have become powerful without becoming commensurately wise. The world-altering powers that technology has delivered into our hands now require a degree of consideration and foresight that has never before been asked of us.”

Because we cannot come back from extinction, we cannot wait until a threat strikes before acting – we must be proactive. And because gaining wisdom takes time, we need to start now.

I think that we are likely to make it through this period. Not because the challenges are small, but because we will rise to them. The very fact that these risks stem from human action shows us that human action can address them. Defeatism would be both unwarranted and counterproductive – a self-fulfilling prophecy. Instead, we must address these challenges head-on with clear and rigorous thinking, guided by a positive vision of the longterm future we are trying to protect.

This is an edited extract from The Precipice: Existential Risk and the Future of Humanity by Toby Ord, published by Bloomsbury.

Why we need worst-case thinking to prevent pandemics. By Toby Ord. The Guardian, March 6, 2020.





We live during the most important era of human history. In the twentieth century, we developed the means to destroy ourselves – without developing the moral framework to ensure we won't. This is the Precipice, and how we respond to it will be the most crucial decision of our time.

Oxford moral philosopher Toby Ord explores the risks to humanity's future, from the familiar man-made threats of climate change and nuclear war, to the potentially greater, more unfamiliar threats from engineered pandemics and advanced artificial intelligence. With clear and rigorous thinking, Ord calculates the various risk levels, and shows how our own time fits within the larger story of human history. We can say with certainty that the novel coronavirus does not pose such a risk. But could the next pandemic? And what can we do, in our present moment, to face the risks head on?


A major work that brings together the disciplines of physics, biology, earth and computer science, history, anthropology, statistics, international relations, political science and moral philosophy, The Precipice is a call for a new understanding of our age: a major reorientation in the way we see the world, our history, and the role we play in it.

Bloomsbury







Toby Ord on the precipice and humanity’s potential futures.

By Robert Wiblin, Arden Koehler and Keiran Harris  

This week Oxford academic and 80,000 Hours trustee Toby Ord released his new book The Precipice: Existential Risk and the Future of Humanity. It’s about how our long-term future could be better than almost anyone believes, but also how humanity’s recklessness is putting that future at grave risk, in Toby’s reckoning a 1 in 6 chance of being extinguished this century.
I loved the book and learned a great deal from it.


While preparing for this interview I copied out 87 facts that were surprising to me or seemed important. Here’s a sample of 16:

1.       The probability of a supervolcano causing a civilisation-threatening catastrophe in the next century is estimated to be 100x that of asteroids and comets combined.
2.       The Biological Weapons Convention — a global agreement to protect humanity — has just four employees, and a smaller budget than an average McDonald’s.
3.       In 2008 a ‘gamma ray burst’ reached Earth from another galaxy, 10 billion light years away. It was still bright enough to be visible to the naked eye. We aren’t sure what generates gamma ray bursts but one cause may be two neutron stars colliding.
4.       Before detonating the first nuclear weapon, scientists in the Manhattan Project feared that the high temperatures in the core, unprecedented for Earth, might be able to ignite the hydrogen in water. This would set off a self-sustaining reaction that would burn off the Earth’s oceans, killing all life above ground. They thought this was unlikely, but many atomic scientists feared their calculations could be missing something. As far as we know, the US President was never informed of this possibility, but similar risks were one reason Hitler stopped pursuing the Bomb.
5.       If we eventually burn all the fossil fuels we’re confident we can access, the leading Earth-system models suggest we’d experience 9–13°C of warming by 2300, an absolutely catastrophic increase.
6.       In 1939, the renowned nuclear scientist Enrico Fermi told colleagues that a nuclear chain reaction was but a ‘remote possibility’. Four years later Fermi himself was personally overseeing the world’s first nuclear reactor. Wilbur Wright predicted heavier-than-air flight was at least fifty years away — just two years before he himself invented it.
7.       The Japanese bioweapons programme in the Second World War — which included using bubonic plague against China — was directly inspired by an anti-bioweapons treaty. The reasoning ran that if Western powers felt the need to outlaw their use, these weapons must especially good to have.
8.       In the early 20th century the Spanish Flu killed 3-6% of the world’s population. In the 14th century the Black Death killed 25-50% of Europeans. But that’s not the worst pandemic to date: that’s the passage of European diseases to the Americans, which may have killed as much as 90% of the local population.
9.       A recent paper estimated that even if honeybees were completely lost — and all other pollinators too — this would only create a 3 to 8 percent reduction in global crop production.
10.   In 2007, foot-and-mouth disease, a high-risk pathogen that can only be studied in labs following the top level of biosecurity, escaped from a research facility leading to an outbreak in the UK. An investigation found that the virus had escaped from a badly-maintained pipe. After repairs, the lab’s licence was renewed — only for another leak to occur two weeks later.
11.   Toby estimates that ‘great power wars effectively pose more than a percentage point of existential risk over the next century. This makes it a much larger contributor to total existential risk than all the natural risks like asteroids and volcanos combined.
12.   During the Cuban Missile Crisis, Kennedy and Khrushchev found it so hard to communicate, and the long delays so dangerous, that they established the ‘red telephone’ system so they could write to one another directly, and better avoid future crises coming so close to the brink.
13.   A US Airman claims that during a nuclear false alarm in 1962 that he himself witnessed, two airmen from one launch site were ordered to run through the underground tunnel to the launch site of another missile, with orders to shoot a lieutenant if he continued to refuse to abort the launch of his missile.
14.   In 2014 GlaxoSmithKline accidentally released 45 litres of concentrated polio virus into a river in Belgium. In 2004, SARS escaped from the National Institute of Virology in Beijing. In 2005 at the University of Medicine and Dentistry in New Jersey, three mice infected with bubonic plague went missing from the lab and were never found.
15.   The Soviet Union covered 22 million square kilometres, 16% of the world’s land area. At its height, during the reign of Genghis Khan’s grandson, Kublai Khan, the Mongol Empire had a population of 100 million, around 25% of world’s population at the time.
16.   All the methods we’ve come up with for deflecting asteroids wouldn’t work on one big enough to cause human extinction.


The 80,000 Hours Podcast.  Published March 7th, 2020




Many people have recently found that they want to read books offering the grandest perspectives possible on human existence, such as Sapiens by Yuval Noah Harari, for, however speculative or unreliable they may be, the long views they offer make all other studies seem piddling.

Toby Ord’s new book is a startling and rigorous contribution to this genre (he pointedly never mentions Noah Harari) that deserves to be just as widely read. Australian-born, Ord is a moral philosopher at the Future of Humanity Institute at Oxford University who has advised organisations such as the World Health Organisation, the World Bank and the World Economic Forum. He is a rationalist and meliorist who believes in looking at the biggest picture possible. Studying global health and global poverty, he concluded that his own money could do hundreds of times as much good for those in poverty as it could for him, so he made a lifelong pledge to donate at least a 10th of his earnings and founded a society, Giving What We Can, that has raised £100 million for charities. So he walks the walk, as well as talks the talk.

 The Precipice, boldly dedicated to the 100 billion people before us, the seven billion alive and “the trillions to come, whose existence lies in the balance”, addresses nothing less than the fundamental threats to humanity itself. It is clearly written, approachable and concise for a work tackling such an immense subject, since Ord has confined all subsidiary questions to extensive footnotes and appendices.

He begins by setting out “the stakes”, saying that “we live at a time uniquely important to humanity’s future”, ­dubbing our era the Precipice, as the time of high risk of destroying ­ourselves. In the last century, we faced a one in a hundred risk of human extinction, he reckons, while in this century it is one in six. The book that follows ­presents his reasoning for this estimate, “the risks”.

He begins with natural risks, like asteroids and comets, supervolcanic eruptions and stellar explosions and finds them, taken altogether, to be low-risk, one in a million per century. But, significantly, he considers the risk represented by pandemics to be anthropogenic, not natural.

The most obvious risk to human survival is nuclear war, and Ord provides an alarming litany of the times we have come near it, mainly by accident, but, unexpectedly, he doubts that even nuclear winter would lead to total human extinction or the global unrecoverable collapse of civilisation. Likewise, he contends that while climate change has the capacity to be “a global calamity of unprecedented scale”, it would not necessarily have that result either. Environmental damage doesn’t show “a direct mechanism for existential risk” either.

Nonetheless, Ord concludes that “each of these three risks has a higher probability than that of all natural risks put together” — and then turns to future risks. These include pandemics, “un­aligned artificial intelligence”, ­“dystopian scenarios” (his term for “a world with civilisation intact, but locked into a terrible form, with little or no value”, in which category he includes a single fundamentalist religion) and, more briefly, nanotechnology and extraterrestrial life.

Ultimately he charts the risks as he sees them, very surprisingly rating the chances of existential catastrophe through nuclear war or climate change in the next hundred years as just one in 1,000, while the risk from engineered pandemics is one in 30 and unaligned artificial intelligence is one in 10, adding up together to that one-in-six chance.

Then he delivers a tremendous pep talk, The Path Forward, about what we can and should do to safeguard humanity, to reach existential security, have a “long reflection” and achieve our potential in deep time and space. “We need to take responsibility for our future,” he urges, saying he finds it “useful to consider our predicament from humanity’s point of view: casting humanity as a coherent agent, and considering the strategic choices were it sufficiently rational and wise.” He admits: “This frame is highly idealised.” Not half.

Among his main recommendations (“International coordination”, “technological progress”), is that we need much more research on existential risk itself, a study in its infancy. That seems inarguable. Let us hope this book is not quite so timely as it feels.

Book review: The Precipice by Toby Ord​ :  What are the chances that the end is truly nigh?
By David Sexton.  Evening Standard,  March 5, 2020. 








According to Toby Ord’s  The Precipice: Existential Risk and the Future of Humanity, the probability of the annihilation of billions of people and/or the unrecoverable collapse of civilization in the coming century is one in six: “[T]he roll of a die, or Russian roulette.”

The Precipice was released to British audiences on March 5 and reached American shelves by the end of the month. It’s unclear whether the timing of Ord’s intervention, in the midst of the current global pandemic, will damn or consecrate the book — if The Precipice is precisely what the world needs right now or the last book anyone is going to place on their quarantine reading list. Ord includes a section on how we’ve exacerbated the risk of natural pandemics.

In an alternative universe, untouched by COVID-19, whether one would find the title of The Precipice and its conclusion alarming or irritating would depend, at least in part, on what a reader thinks of where Ord works. The Precipice is the latest public missive delivered by an affiliate of Oxford’s Future for Humanity Institute (FHI), founded and directed by Nick Bostrom and dedicated to studying “big picture questions.” FHI shares an address with the Centre for Effective Altruism. The two back up against Westgate, the city’s titanic and improbably open-aired, Golden State–style shopping complex.

Pioneered by Ord and fellow Oxford philosopher William MacAskill, effective altruism is a movement that defines itself as “the use of reason and evidence to help you help others as much as possible.” This sounds especially intriguing to young people hoping to do meaningful work. To those already involved in the endeavor, it can come across as a provocation — as if no one was using reason and evidence before the late 2000s. EA’s advocates often poke at provincial moral intuitions. Famously, their impartial “p-values” suggest that it’s more effective to become a banker and donate well than to work for change in your own neighborhood — happy news for financiers hoping to have their zero-guilt cake and eat it too. The group has been energetically critiqued.

Ord is by no means the first of the effective altruists to focus on existential risk — i.e., dangers that threaten “existential catastrophe,” a class of disasters that includes but is not identical to human extinction. The topic follows in a straightforward way from EA principles. If people matter equally regardless of their position in time, if there will be many more people in the future, and if our current actions can influence how those people live, then in order to maximize the good you can do, your actions should be good for those future people. The greatest good you can do for them is ensuring that they will exist, and that they won’t live in a dystopia. Note that the community once called existential risks “x-risks,” a term with the advantage of combining what many like least about economics, TED Talks, and tech, while also making the shared spirit of global capitalism and EA explicit. Ord calls them “existential risks.”

This is the short chain of logic that leads from the Centre for Effective Altruism to FHI. At its end are some curious alliances: researchers bringing marginal utility to bear on international poverty and futurists chasing after digital minds and cosmic imperium. Ord, who publicly pledged to live on £18,000 a year (adjusted for inflation) and has given away to date over a quarter of his total earnings, works for a center that received a donation from billionaire Elon Musk. A shared interest in artificial intelligence and deep learning is what brings Silicon Valley to FHI’s doorstep. Bostrom and Ord both believe that an AI unaligned with human values is the greatest threat to human potential. And while it is, of course, a good thing for those sounding the AI alarm bells to be within earshot of the technology’s engineers and indecently rich patrons, these alliances haven’t helped to clear the air of suspicion. Their critics are particularly wary of math-intensive products glazed with the user-friendly minimalism favored by the well off.



The Precipice is both an artifact of this institutional context and something altogether different. EA tools are still there: priorities are set with reference to the steepness of curves; there’s an appendix titled “Surprising Effects when Combining Risks.” And EA principles are still there: Ord confirms the overwhelming moral significance of securing humanity’s future, or, in Bostromian terms, “reducing an existential threat by a billionth of a billionth of one per cent” in a way that “would be worth a hundred billion times the value of a billion present-day lives.”

The book’s difference begins with its approach and attitude. Ord suggests that “helping humanity over the long run could be one of the best ways to do good in the world.” Chapter Seven includes a brief section titled “What not to do.” “Don’t be fanatical,” he writes (to his colleagues). Safeguarding humanity isn’t the only priority. “Boring others with endless talk about this cause is counterproductive. Cajoling them about why it is more important than a cause they hold dear is even worse.” In a footnote citing the work of Stuart Russell, Professor of Computer Science at the University of California, Berkeley, Ord suggests that one promising way to address the AI alignment issue is to build machines with a reward function that encourages them to act as if in a state of moral uncertainty. There’s something very touching about the idea of an algorithm being better because it’s not sure of how to do the most good. Ord’s book is certainly more persuasive. In fact, the book persuades its reader that its title comes from a place of deep conviction, not self-importance. The term “effective altruism” appears once in the main body of text. The word “uncertainty” over 60 times.

But Ord is certain that risk levels are unsustainably high, and that there will be resistance to his book’s conclusion — that this is largely the stuff of Marvel comics or science fiction and that many readers, even at a time of global crisis, will stumble over the word “omnicide.” He does his best to anticipate objections. Those looking to challenge Ord’s calculations will have to work through the book’s footnotes and appendices, which, taken together, are as long as the main text, and which synthesize the research of the experts listed in his vast acknowledgments section. “One in six” takes our estimated response into account, so the fatalist should note that, if we do nothing, the odds are closer to two in six. Ord argues that there is insufficient evidence for pessimism. We can’t be sure that the future will be so bad, or whether billions of lives would be better off unlived. Nihilism sounds impressive but isn’t a principle of governance. How can we determine the budget for the Biological Weapons Convention according to the principle that nothing matters? You might agree that the book is alarming but fail to see how it concerns you. Ord compares the issue to climate change and includes a section titled “What you can do.” Concrete policy suggestions are discussed throughout the main text and reiterated in an appendix.

Ord is not simply explaining a probability. He wants to encourage his readers to see ethics from the perspective of humanity. That is, he wants them to imagine the species as a coherent actor and to identify with it. That’s why Ord speaks of “our newfound knowledge” and “our potential,” a decision that seems tone-deaf or wilful, and occasionally completely insane.

Ord writes: “There will be great challenges in getting people to look far enough ahead and see beyond the parochial conflicts of the day.” We need to “accept the fresh responsibilities that come with our unprecedented power.” “Every day we are the beneficiaries of uncountable innovations made by people over hundreds of thousands of years. […] This is a stunning inheritance.” Tonto’s famous riposte to the Lone Ranger comes to mind: “What do you mean we, pale-face?” Those un-ensnared in today’s parochial conflicts have commendable vision. The view from Oxford’s tower is long indeed. From a different perspective, or for those who’ve read Ta-Nehisi Coates, our inheritance may look a little grim. I suspect that the greatest resistance to The Precipice will be to its use of the first-person plural. It isn’t a stylistic objection; it takes us, rather, to the heart of the project. Getting you to occupy this “we,” where Ord’s argument is clarified and its urgency most deeply felt, is the book’s most incredible work. A reader’s resistance won’t be overcome with cost-benefit analyses.



To put a face to this resistance and clarify what’s at issue, let’s imagine a reader. Call her Ann. Ann’s ethical attention is directed to the inequalities between persons and groups. Some people are impermissibly disadvantaged by these fissures. Some are concerned and trying to help. Others resent that they’re finally being roused from their moral stupor. Even if you manage to convince Ann that these inequalities are decreasing, she’ll maintain that our awareness of them has never been greater, and that it needs to continue to grow. Ann is a bit suspicious of universal claims. She looks at “we” and her mind shouts “illusion of consensus!” Ann has read Sontag and Chomsky. Weapons of mass destruction have race, class, and gender aspects. Ann’s background is, moreover, in postcolonial theory. Any call-to-arms on behalf of humanity is going to raise an ugly ghost — a ghost that could forget his principles in order to purportedly spread those principles to the great advantage of his economic agenda. Ann lives in the United Kingdom, or in the United States, or in Europe.

How is Ord going to get Ann to join his “we”? Ord might observe that our moral vision is improving and that Ann herself is evidence of this fact. Our concern is stretching into the past. Why shouldn’t it extend into the future, he might ask, a future that, by all indications, will be more humane? Ann could refuse to listen to the importance of buying us more time. Inequality isn’t a matter of time, she might say, but of moral courage. But there’s a one in six chance that there will be no more time for moral courage or, in fact, anything else, Ord replies. Moral courage is needed to make our future look more democratic. We all need to care to survive. Ann describes racialized capitalism. Ord describes three kinds of unrecoverable dystopia. Ann accuses Ord’s “we” of ignoring systemic injustice. Ord asks Ann, with a nod to Richard Rorty, if she also accuses atheists of blasphemy. Ann thinks this “we” is a dangerous fiction. Ord declares that it’s the most dangerous time in history to throw it away. “Don’t be tribal,” he warns. Ann says that the warning signals his tribal affiliation. “Framing it as a political issue on one side of a contentious divide would be a disaster,” Ord says. Ann insists that there are no politically neutral claims. Ord tells her that she insists that at our collective peril.

The author of The Precipice walks out of the room and takes a few deep breaths.

It’s easy to stay within one’s own moral framework and still prioritize existential risk. We may have obligations to preserve the past or address past evils, or honor past actions by paying it forward. The future could seem particularly valuable to those for whom the past is like a “massive depository that sticks to the present,” to quote philosopher of history Berber Bevernage. Even those who drew humanity’s most bitter stick could be made to care. Perhaps a few readers secretly think that apocalypse wouldn’t be so bad, for at least extinction would be equally distributed. You might find that perspective a little repulsive. You might ask them to think of the enormous waste, or how painful the end would be, or to keep in mind that billionaires are currently building bunkers to weather such a storm. They might be among the few who succeed. Imagine all of humanity’s descendants to be of Bezos.

People from an astonishingly wide range of moral perspectives will agree that existential risk is a great concern. But Ord really wants you to see it in a certain way. “In optimism lies urgency.”

Ord walks back into the room. He invites Ann to take a seat. He turns on the overhead projector.

Philosopher and novelist Iris Murdoch thought that there was something to morality prior to the question of “what should we do.” In her 1964 essay “The Idea of Perfection,” Murdoch suggests that ethical thinking is not simply a matter of “opening one’s eyes” to the relevant facts. Ethical thinking is, rather, the result of “just discernment and exploration.” What characterizes this exploration are “little peering efforts of imagination.” This is a slow and cumulative and often unremarkable process.




Consider two people looking at a piece of art. Certain facts about the object, like its size and material, are listed on the plaque. The first person assesses the sculpture and decides that it’s a little crude. Her companion begins to talk. He points to features of the object and uses words like “bold” and “unexpected.” She listens to him because he’s an authority on art and because she likes his tone, which is soft and probing. Her experience is improved as a result of his descriptions. She wants to go off around the museum and try to see things as he has seen them. There are moments when the face of a thing changes, but not because the face has objectively changed. We see not a different problem, but a different aspect of it.



Both Ann and Ord stand before the following picture: there’s a one in six chance that we will forfeit our potential this century, and that the reality of this “we” is, to quote Ord, “messy and fraught.” He’s not trying to paper over the fissures that preoccupy Ann. But he wants us to ask ourselves what humanity ought to do. Can we imagine ourselves as a collective that responds to an urgent situation? Do we see the person making the call as a man of privilege or of principle? The facts have already been reported. But it matters how they’re being described.

Try to think of humanity, suggests Ord, in terms of the lifespan of a single human. In that life the 20th century is but three days. We went from horse and buggy to Apollo 11 in countable hours. We are, in fact, 16 years old: “[J]ust old enough to get ourselves in serious trouble.” Ord asks us to measure our concern over 20th-century bioweaponry. And yet, 10 decades ago “we had only just discovered viruses and were yet to discover DNA.” How do we feel about “the next hundred years of improvements”? “Would we expect to get through 2,000 centuries like this one?” Did you know that we hold nuclear weapons on hair-trigger alert, the global equivalent of keeping an arsenal of loaded guns within reach of toddlers, just in case of an intruder? Can you imagine existential catastrophe? Climate change will bring “reduced agricultural yields, sea level rises, water scarcity, increased tropical diseases, ocean acidification, and the collapse of the Gulf Stream.” This is tragedy on a scale we’ve never seen. “None of these,” however, “threaten extinction or irrevocable collapse.” The one in six chance we’re facing is something far worse.

Ord includes the following sketch at the beginning of his third chapter, in a section on “Asteroids & Comets”:

An asteroid, ten kilometers across, speeds toward the Earth. The chance of a direct collision is tiny — for millions of years it has swung through the solar system, missing the Earth on every single pass. But given such deep time the chances compound, and this is the day.

It slams into the Earth’s surface off the coast of Mexico at more than 60,000 kilometers an hour. A trillion tons of rock moving so fast strikes with the energy of a hundred times its own weight in TNT. In just seconds, it releases the energy of 10 billion Hiroshima blasts: 10,000 times the entire Cold War nuclear arsenal. It smashes a hole thirty kilometers deep into the Earth’s crust — over sixty times the height of the Empire State Building; three times taller than Everest. Everything living within 1,000 kilometers is killed by heat from the impact fireball. A tsunami devastates the Caribbean. Trillions of tons of rock and dust are thrown far up into the sky. Some of this superheated rock rains down over millions of square kilometers, burning the animals to death and igniting fires that spread the devastation still further. But much more deadly is the dust that stays aloft. A billowing cloud of dust and ash rises all the way to the upper atmosphere, blocking out the Sun’s light.




It’s in the long, cold darkness that “regional catastrophe” turns to “mass extinction.”

What is the effect of narrating the extinction of the dinosaurs in the present tense? It produces a small mental hiccup in the mind of the reader. She recalibrates, adjusting from what she thought was the description of a hypothetical event to that of an event that has already happened. She doesn’t understand 30 kilometers deep, but she can try to imagine 60 Empire State Buildings bored down into her world. She doesn’t understand a trillion tons, or a hundred times that in TNT. But she can try to recall images of the Hiroshima blast and multiply that by 10 billion. The point is not that there’s a real risk of an asteroid striking Earth this century (it’s much, much smaller than the risks we’re engineering ourselves). The point is that, like seeing footage from World War I in color, what was tucked away in history is now brought to our senses. We realize what’s possible. There have been five large mass extinction events in the last 540 million years. The End-Permian extinction, in which 96 percent of species disappeared, was 250 million years ago.

According to the fossil record, mammals last on average about one million years, and species generally last between one and 10 million years. Humans have been around for about 200,000 years. What if we, or the species that follows us “on the evolutionary family tree,” had another 800,000? To put that number into perspective, consider that Stonehenge was built about 5,000 years ago and we’re not even sure why.

Ten million years could be a lower bound for us. We might be like coelacanths and sponges, species that have lasted for hundreds of millions of years essentially unchanged. In 10 million years, Africa will be torn in two along the Rift Valley. In 50 million years, the Mediterranean Sea will be crushed to mountain. In 250 million years, there will be a new Pangaea, a new congress of the continents. Days will stretch to 25 hours. Five hundred million years from now, the land will break apart and the world will see a new configuration. New constellations. Maybe we could live long enough to see it. “If this feels unimaginable, consider that the horseshoe crab has already witnessed such change.”

When we think of humanity through “the slow clock of cosmological expansion,” our future becomes “a canvas vast in time and space.” Perhaps we could live long enough for humans to “heal our society and our planet of the wounds we have caused in our immaturity.” We could prolong life on Earth by saving the biosphere from a brightening sun. When the sun grows too vast, perhaps we could carry seeds and cells to other planets to “make green the barren places of the galaxy,” with a space-fleet called Noah’s Ark. If you think space travel is impossible, try to imagine the “perseverance of the Polynesian sailors who, a thousand years ago, sailed vast stretches of the Pacific.” In the life of our species, we went from discovering that the Earth revolved around the sun to building computers better than us at chess in a matter of weeks. Wait until we’re in our late 30s. Just see what we can do.

Imagine that our descendants will look on the planets and moons of our solar system as I look on the national parks of my home country: “[A]s monuments, jewels. To be explored and treasured. To fill us with wonder and inspire us to journey further.” Shine a beam of light into space and trace it to the edge of influence. Imagine a future human, whose lineage we trace back to you, who can hear “[m]usic that we lack the ears to hear,” journeying beyond halfway to the edge of the affectable universe in a “final diaspora.” Some stars last for trillions of years. Maybe our time is “astonishingly close to the very start of the universe.” Maybe we could last long enough to reach “some external insurmountable limit — perhaps the death of the last stars, the decay of matter into energy.”

The epigraph to Ord’s final chapter is a quote from H. G. Wells: “It is possible to believe that all the past is but the beginning of a beginning. […] It is possible to believe that all that the human mind has ever accomplished is but the dream before the awakening.” That we are all paddling out to some new horizon, with something unwritten before. How much farther can we go? Ord believes “we have barely begun the ascent.” “A vast and extraordinary adulthood awaits.”



Ann blinks and can’t believe that she just had a moment of cosmic optimism. Perhaps humanity needs a common enemy in order to come together. In which case, mankind is engineering its own great problem as well as the conditions for its solution. Ord repeats “we” until she starts to see how it could be true. His nosism may be the most provocative feature of The Precipice. It also signals the book’s real project and serves as a shorthand for its greatest achievement.

Ord may simply deploy these images and this language in order to combat what’s known in behavioral psychology as availability heuristics and scope neglect. An availability heuristic is a mental shortcut that allows us to predict things based on examples that come to mind quickly. Scope neglect is a phenomenon in which our compassion doesn’t scale well with numbers. Together these two phenomena encourage us to undervalue the importance of existential risk. This imaginative work may be an example of a rational man momentarily relying on emotional appeal in order to answer to a meta logic of cold calculus. Perhaps.

And yet, by the end of The Precipice, a reader sees that what Ord is arguing to preserve is itself the stuff of his book. Humans, he suggests, may be cosmically significant. Without us, the universe would lose the buttery sound of a brassy instrument and the noise of New York City on a May afternoon. It would also lose the only things in all its vastness that are capable of “an upwards force,” of imagining horizons of expectation unshaped by our spheres of experience, of imagining justice. The use of this capacity is how an EA doomsday book becomes, against all odds, a vision. The Precipice may be the Silent Spring that the futurists have been waiting for.

In Optimism Lies Urgency: Toby Ord and the Future of Humanity.  By Alexa Hazel.
























No comments:

Post a Comment