A New York Lawmaker Wants to Ban Police Use of Armed Robots

New York City councilmember Ben Kallos says he “watched in horror” last month when city police responded to a hostage situation in the Bronx using Boston Dynamics’ Digidog, a remotely operated robotic dog equipped with surveillance cameras. Pictures of the Digidog went viral on Twitter, in part due to their uncanny resemblance with world-ending machines in the Netflix sci-fi series Black Mirror.

Now Kallos is proposing what may be the nation’s first law banning police from owning or operating robots armed with weapons.

“I don’t think anyone was anticipating that they’d actually be used by the NYPD right now,” Kallos says. ”I have no problem with using a robot to defuse a bomb, but it has to be the right use of a tool and the right type of circumstance.”

Kallos’ bill would not ban unarmed utility robots like the Digidog, only weaponized robots. But robotics experts and ethicists say he has tapped into concerns about the increasing militarization of police: their increasing access to sophisticated robots through private vendors and a controversial military equipment pipeline. Police in Massachusetts and Hawaii are testing the Digidog as well.

“Nonlethal robots could very well morph into lethal ones,” says Patrick Lin, director of the Ethics and Emerging Sciences Group at California Polytechnic University, San Luis Obispo. Lin briefed CIA employees on autonomous weapons during the Obama administration and supports a ban on armed robots. He worries their increased availability poses a serious concern.

“Robots can save police lives, and that’s a good thing,” he says. “But we also need to be careful it doesn’t make a police force more violent.”

In the Bronx incident last month, police used the Digidog to gather intel on the house where two men were holding two others hostage, scoping out hiding places and tight corners. Police ultimately apprehended the suspects, but privacy advocates raised concerns about the technical capabilities of the robot and policies governing its use.

The ACLU questioned why the Digidog was not listed on the police department’s disclosure of surveillance devices under a city law passed last year. The robot was only mentioned in passing in a section on “situational awareness cameras.” The ACLU called that disclosure “highly inadequate,” criticizing the “weak data protection and training sections” regarding Digidog.

In a statement, the NYPD said it “has been using robots since the 1970s to save lives in hostage situations and hazmat incidents. This model of robot is being tested to evaluate its capabilities against other models in use by our Emergency Service Unit and Bomb Squad.” Boston Dynamics did not respond to a request for comment.

Local response to the use of the Digidog was mixed, says councilmember Kevin Riley, who represents the Bronx neighborhood where the incident ocurred. Some residents opposed police use of the robot and others wanted more human police presence. A third group thought the robots might help prevent police misconduct by creating distance between officers and suspects.

Riley says he’s continuing to speak with residents, who want to feel safe in the neighborhood. “It’s our job as elected officials to educate residents and make sure they have a seat at the table” in discussions, he told WIRED.

The diversity of concerns mirror those in Dallas in 2016. During a standoff with a sniper, local law enforcement used a robot to remotely deliver and detonate an explosive device, killing him. The sniper had shot and killed five police officers.

The incident raised questions about how police acquire robots. Dallas police had at least three bomb robots in 2016. Two were acquired from the defense contractor Northrop Grumman, according to Reuters. The third came through the federal government’s 1033 program, which permits the transfer of surplus military equipment to local police departments. Since 1997, over 8,000 police departments have received over $7 billion in equipment.

www.wired.com

Leave a Reply