Skip to main content

The context you need, when you need it

When news breaks, you need to understand what actually matters — and what to do about it. At Vox, our mission to help you make sense of the world has never been more vital. But we can’t do it on our own.

We rely on readers like you to fund our journalism. Will you support our work and become a Vox Member today?

Join now

Google wants to improve artificial intelligence to prevent robot screw-ups

Five ways your future robot might go awry.

More and more artificial intelligence will soon enter our lives. And Google would very much like its AI systems to be front and center.

That’s why the company is putting resources into making sure AI systems don’t go off the rails.

Last month, Google researchers teamed with scientists from two universities and OpenAI, the consortium backed by Elon Musk, to investigate potential pitfalls for AI that may arrive in the not-too-distant future.

The paper walks through five different ways a household cleaning robot could encounter safety problems, then details some math to avoid these problems. (Google hasn’t released plans for this kind of robot, but one of its robotics projects — no, not the frightening humanoid one — fits the bill.)

Here are the five issues that Google flags:

  • Whoops: If you train a robot to clean your house as quickly as possible, how do you ensure that it doesn’t knock over the lamp?
  • “Reward hacking”: What if the robots cheat? Google wonders aloud if a robot could, say, sweep a mess under a rug or disable its vision (nothing to see here!).
  • Reasonable human feedback: Can robots figure out how to improvise on the job — like toss a candy wrapper, but not a cellphone — without constant human oversight?
  • Physical safety: We probably don’t want our robots shoving wet mops in sockets.
  • Different environments: Once the robot learns to clean a room in a house, can the same robot clean an office?

It’s important research — before these types of robots arrive en masse, they will surely need to pass certain safety requirements. And Google needs to prove that AI is a force for good. The Google paper also nods to other important research topics like ensuring that AI doesn’t fall into “malicious” hands or “attack or harm people.”

This article originally appeared on Recode.net.

More in Technology

Podcasts
Are humanoid robots all hype?Are humanoid robots all hype?
Podcast
Podcasts

AI is making them better — but they’re not going to be doing your chores anytime soon.

By Avishay Artsy and Sean Rameswaram
Future Perfect
The old tech that could help stop the next airborne pandemicThe old tech that could help stop the next airborne pandemic
Future Perfect

Glycol vapors, explained.

By Shayna Korol
Future Perfect
Elon Musk could lose his case against OpenAI — and still get what he wantsElon Musk could lose his case against OpenAI — and still get what he wants
Future Perfect

It’s not about who wins. It’s about the dirty laundry you air along the way.

By Sara Herschander
Life
Why banning kids from AI isn’t the answerWhy banning kids from AI isn’t the answer
Life

What kids really need in the age of artificial intelligence.

By Anna North
Culture
Anthropic owes authors $1.5B for pirating work — but the claims process is a Kafkaesque messAnthropic owes authors $1.5B for pirating work — but the claims process is a Kafkaesque mess
Culture

“Your AI monster ate all our work. Now you’re trying to pay us off with this piece of garbage that doesn’t work.”

By Constance Grady
Future Perfect
Some deaf children are hearing again because of a new gene therapySome deaf children are hearing again because of a new gene therapy
Future Perfect

A medical field that almost died is quietly fixing one disease at a time.

By Bryan Walsh