Death robots: Where next after Dallas?

The use of a robot to deliver an explosive device and kill the Dallas shooting suspect has intensified the debate over a future of "killer robots".

While robots and unmanned systems have been used by the military before, this is the first time the police within the US have used such a technique with lethal intent

"Other options would have exposed our officers to greater danger," the Dallas police chief said.

Robots are spreading fast. What might that mean?


Killer drones

Remote killing is not new in warfare. Technology has always been driven by military application, including allowing killing to be carried out at distance - prior examples might be the introduction of the longbow by the English at Crecy in 1346, then later the Nazi V1 and V2 rockets.

More recently, unmanned aerial vehicles (UAVs) or drones such as the Predator and the Reaper have been used by the US outside of traditional military battlefields.

Since 2009, the official US estimate is that about 2,500 "combatants" have been killed in 473 strikes, along with perhaps more than 100 non-combatants. Critics dispute those figures as being too low.

Back in 2008, I visited the Creech Air Force Base in the Nevada desert, where drones are flown from.

During our visit, the British pilots from the RAF deployed their weapons for the first time.

One of the pilots visibly bristled when I asked him if it ever felt like playing a video game - a question that many ask.

Supporters of drones argue that they are more effective than manned planes because they can usually loiter longer and ensure they strike the right target.

And, of course, there is the understandable desire to reduce risks to pilots, just as in Dallas the police officers could stay protected.

But critics argue that the lack of risk fundamentally changes the nature of operations since it lowers the threshold for lethal force to be used.


Gun bots

Robots have also been deployed on the ground militarily.

South Korea pioneered using robots to guard the demilitarised zone with North Korea. These are equipped with heat and motion detectors as well as weapons.

The advantage, proponents say, is that the robots do not get tired or fall asleep, unlike human sentries.

When the Korean robot senses a potential threat, it notifies a command centre

Crucially though, it still requires a decision by a human to fire.

And this gets back to the crucial point about the Dallas robot. It was still under human control.

The real challenge for the future is not so much the remote-controlled nature of weapons but automation - two concepts often wrongly conflated.

Truly autonomous robotic systems would involve no person taking the decision to shoot a weapon or detonate an explosive.

The next step for the Korean robots may be to teach them to tell friend from foe and then fire themselves.

Futurologists imagine swarms of target-seeking nano-bots being unleashed pre-programmed with laws of warfare and rules of engagement.

There are still questions both about how such machines could be programmed to deal with complex situations and the ethical dilemmas involved when you have to choose whether or not to shoot or make calculations over potential civilian casualties.

There's a parallel here with the challenge about what self-driving cars should do when faced with crashing into a group of children or harming their passengers.

The fears over automation are not new.

One of the earliest use of computers was during the Cold War to automate as far as possible the response to a Soviet nuclear attack.


Dawn of cybersecurity

A system called Semi-Automatic Ground Environment (Sage) was designed using networked computers to help spot incoming Soviet planes.

Soon, missiles were also connected up to the systems to shoot the planes down.

One air force captain queried the fact that computers controlled the launch of such missiles and asked if that was dangerous.

Could someone get inside such a computer system and subvert it to send the missiles back into US cities rather than at Soviet bombers?

That question, over whether automated and remote systems could be subverted, led to some of the earliest work on what we now call cybersecurity.