Skip to main content

The context you need, when you need it

When news breaks, you need to understand what actually matters — and what to do about it. At Vox, our mission to help you make sense of the world has never been more vital. But we can’t do it on our own.

We rely on readers like you to fund our journalism. Will you support our work and become a Vox Member today?

Join now

These surveillance robots will work together to chase down suspects

Networks of drones, humanoids, wheeled robots and stationary smart cameras may one day keep watch.

Cornell University

When security cameras capture someone leaving a suspicious package at a train station, a person monitoring the camera feeds may, if they act fast enough, be able to coordinate with an agent on the ground to follow the perpetrator before he dodges out of the security camera’s field of vision.

Not super efficient. But in the future, the job may be better be done by robots.

Imagine if the camera that saw the crime was a wheeled robot equipped with facial recognition technology that can share information instantly with other nearby robotic cameras — all programed to surveil a scene and pursue suspects to keep them in sight.

Researchers from Cornell University are building a system for networked coordination between camera robots, drones and mounted smart cameras that can swap information instantly and move around a scene to chase a suspect, change their perspective and even reason about their environment as the machines look for for questionable activity.

Funded by a $1.7 million grant from the U.S. Office of Naval Research, the researchers will be using Segue robots with automatic cameras that can be programmed to pan, tilt and zoom for their experiments. The research team is led by Professor Silvia Ferrari, director of the Laboratory for Intelligent Systems and Controls at Cornell.

“We are trying to teach robots to follow things of interest, like people, cars and animals, and to reason about what they are seeing, what kind of activity is happening and what the agent might be doing next,” Ferrari told Recode in an interview.

One day, the software might be able to manage and coordinate hundreds of robotic cameras, but for the initial experiment, the team at Cornell plans to trial up to 12 camera systems operating simultaneously.

In an area under surveillance that is quite large, a network of mobile robots could be a big help, since one robot — or even an array of mounted cameras — can’t capture everything.

The idea here, says Ferrari, is to make the robotic cameras as autonomous as possible. The researchers are programming their surveillance robots to fuse together all available video data to reason about a scene, and the robots will be connected to the web so they can access more data for when they detect holes in their understanding.

Typically, surveillance systems send data back to a human operator, who interprets the scene to make tactical decisions about what other information is needed and how to collect it.

“Our intention is to automate that side of the network so that the robots are actually in charge of perception,” said Ferrari.

The surveillance robots will be communicating to each other in a computer language, Ferrari says, but will also be able able to translate what they’re thinking into “some syntax” that a human can understand.

“This is basically the only time they’ll be interfacing with a human being,” says Ferrari.

To get robots to reason and make decisions about what to pursue and where to go, the team at Cornell is building artificial intelligence navigation algorithms that are coupled with the ability to perceive and understand the information they are collecting.

That means these robots won’t be programed to simply know how to avoid obstacles or get from point A to point B, like most navigation algorithms, but they also will be able to deduce what needs to be focused on and what agent in its video feed is the right one to pursue.

For now, though, this roving robot surveillance fleet technology is still being built and there’s a lot of work to do before it’s ready to be deployed in the field.

Ferrari says her team should have a working demonstration in the next three years.


This article originally appeared on Recode.net.

More in Technology

Podcasts
Are humanoid robots all hype?Are humanoid robots all hype?
Podcast
Podcasts

AI is making them better — but they’re not going to be doing your chores anytime soon.

By Avishay Artsy and Sean Rameswaram
Future Perfect
The old tech that could help stop the next airborne pandemicThe old tech that could help stop the next airborne pandemic
Future Perfect

Glycol vapors, explained.

By Shayna Korol
Future Perfect
Elon Musk could lose his case against OpenAI — and still get what he wantsElon Musk could lose his case against OpenAI — and still get what he wants
Future Perfect

It’s not about who wins. It’s about the dirty laundry you air along the way.

By Sara Herschander
Life
Why banning kids from AI isn’t the answerWhy banning kids from AI isn’t the answer
Life

What kids really need in the age of artificial intelligence.

By Anna North
Culture
Anthropic owes authors $1.5B for pirating work — but the claims process is a Kafkaesque messAnthropic owes authors $1.5B for pirating work — but the claims process is a Kafkaesque mess
Culture

“Your AI monster ate all our work. Now you’re trying to pay us off with this piece of garbage that doesn’t work.”

By Constance Grady
Future Perfect
Some deaf children are hearing again because of a new gene therapySome deaf children are hearing again because of a new gene therapy
Future Perfect

A medical field that almost died is quietly fixing one disease at a time.

By Bryan Walsh