Skip to main content

The context you need, when you need it

When news breaks, you need to understand what actually matters — and what to do about it. At Vox, our mission to help you make sense of the world has never been more vital. But we can’t do it on our own.

We rely on readers like you to fund our journalism. Will you support our work and become a Vox Member today?

Join now

10 years ago, “AI risk” was a joke. Now it’s a serious concern. Thank Jaan Tallinn.

The philanthropist explains why he decided to fund ideas no other donor would touch.

Cars driving on a highway through high-rise buildings with superimposed squares over many of the cars.
Cars driving on a highway through high-rise buildings with superimposed squares over many of the cars.
Honestly the closest thing I could find to a good photograph representing “deep learning.” Sorry, folks.
Dong Wenjie / Getty Images

Most charity is focused on the near term — it goes to universities educating people now, or arts organizations putting on shows and exhibits now, or food pantries helping the hungry now.

So what happens when you try to only give to charities that will help humans a long time from now — not just in 100 years, but in a million years?

That’s exactly what Jaan Tallinn, a founding engineer of Skype, has done with his fortune. He was one of the first donors to take seriously arguments that advanced artificial intelligence poses a threat to human existence — not now, maybe not in 50 years, but certainly somewhere in the future. He has come to believe we might be entering the first era in human history where we are not the dominant force on the planet, and that as we hand off our future to advanced AI, we should be damned sure its morality is aligned with our own.

He has donated more than $600,000 to the Machine Intelligence Research Institute, a prominent organization working on “AI alignment” (that is, aligning the interests of artificial intelligence with the interests of human society) and more than $310,000 to the Future of Humanity Institute at Oxford, which works on similar subjects. He’s also co-founded two new organizations studying AI and other extinction threats: the Centre for the Study of Existential Risk at Cambridge and the Future of Life Institute.

Tallinn came on the latest episode of Vox’s Future Perfect podcast to talk about the rationale behind his philanthropy and how he was persuaded to care so much about AI:

To people unfamiliar with the argument that AI poses an existential risk, Tallinn’s actions might seem bizarre. So we had Kelsey Piper, a Vox reporter who’s written extensively on AI risk, walk through the argument. And Robert Reich, a Stanford philosopher who has been highly critical of big philanthropy, explains why this kind of big experiment in giving might be the best the charitable sector has to offer us.


Read more


Sign up for the Future Perfect newsletter. Twice a week, you’ll get a roundup of ideas and solutions for tackling our biggest challenges: improving public health, decreasing human and animal suffering, easing catastrophic risks, and — to put it simply — getting better at doing good.

Future Perfect
We’re asking the wrong question about the hantavirus outbreakWe’re asking the wrong question about the hantavirus outbreak
Future Perfect

The problem with hantavirus coverage isn’t the alarmism.

By Bryan Walsh
Future Perfect
“I’m disgusted to be a human”: What to do when you hate your own species“I’m disgusted to be a human”: What to do when you hate your own species
Future Perfect

Yes, it hurts to be human right now. That’s actually the assignment.

By Sigal Samuel
Future Perfect
The surprisingly strong case for feeling great about your coffee habitThe surprisingly strong case for feeling great about your coffee habit
Future Perfect

Your morning coffee is one of modern life’s underrated miracles.

By Bryan Walsh
Future Perfect
The old tech that could help stop the next airborne pandemicThe old tech that could help stop the next airborne pandemic
Future Perfect

Glycol vapors, explained.

By Shayna Korol
Future Perfect
Elon Musk could lose his case against OpenAI — and still get what he wantsElon Musk could lose his case against OpenAI — and still get what he wants
Future Perfect

It’s not about who wins. It’s about the dirty laundry you air along the way.

By Sara Herschander
Future Perfect
The backlash to Billie Eilish’s vegan comments explains a lot about the American left (and everyone else)The backlash to Billie Eilish’s vegan comments explains a lot about the American left (and everyone else)
Future Perfect

Why are American leftists so reluctant to confront the meat industry?

By Kenny Torrella