The threat of 'killer robots' is real and closer than you might think

The threat of 'killer robots' is real and closer than you might think
Credit: cottonbro studio from Pexels

From self-driving cars, to digital assistants, artificial intelligence (AI) is fast becoming an integral technology in our lives today. But this same technology that can help to make our day-to-day life easier is also being incorporated into weapons for use in combat situations.

Weaponised AI features heavily in the security strategies of the US, China and Russia. And some existing weapons systems already include autonomous capabilities based on AI, developing weaponised AI further means machines could potentially make decisions to harm and kill people based on their programming, without .

Countries that back the use of AI weapons claim it allows them to respond to emerging threats at greater than human speed. They also say it reduces the risk to military personnel and increases the ability to hit targets with greater precision. But outsourcing use-of-force decisions to machines violates human dignity. And it's also incompatible with international law which requires human judgment in context.

Indeed, the role that humans should play in use of force decisions has been an increased area of focus in many United Nations (UN) meetings. And at a recent UN meeting, states agreed that it's unacceptable on ethical and legal grounds to delegate use-of-force decisions to machines—"without any human control whatsoever".

But while this may sound like good news, there continues to be major differences in how states define "human control".

The problem

A closer look at different governmental statements shows that many states, including key developers of weaponised AI such as the US and UK, favor what's known as a distributed perspective of human control.

This is where human control is present across the entire life-cycle of the weapons—from development, to use and at various stages of military -making. But while this may sound sensible, it actually leaves a lot of room for human control to become more nebulous.

Taken at face value, recognizing human control as a process rather than a single decision is correct and important. And it reflects operational reality, in that there are multiple stages to how modern militaries plan attacks involving a human chain of command. But there are drawbacks to relying upon this understanding.

It can, for example, uphold the illusion of human control when in reality it has been relegated to situations where it does not matter as much. This risks making the overall quality of human control in warfare dubious. In that it is exerted everywhere generally and nowhere specifically.

This could allow states to focus more on early stages of research and development and less so on specific decisions around the use of force on the battlefield, such as distinguishing between civilians and combatants or assessing a proportional military response—which are crucial to comply with international law.

And while it may sound reassuring to have human control from the research and development stage, this also glosses over significant technological difficulties. Namely, that current algorithms are not predictable and understandable to human operators. So even if human operators supervise systems applying such algorithms when using force, they are not able to understand how these systems have calculated targets.

Life and death with data

Unlike machines, human decisions to use force cannot be pre-programmed. Indeed, the brunt of international humanitarian law obligations apply to actual, specific battlefield decisions to use force, rather than to earlier stages of a weapons system's lifecycle. This was highlighted by a member of the Brazilian delegation at the recent UN meetings.

Adhering to international humanitarian law in the fast-changing context of warfare also requires constant human assessment. This cannot simply be done with an algorithm. This is especially the case in urban warfare, where civilians and combatants are in the same space.

Ultimately, to have machines that are able to make the decision to end people's lives violates human dignity by reducing people to objects. As Peter Asaro, a philosopher of science and technology, argues: "Distinguishing a 'target' in a field of data is not recognizing a human person as someone with rights." Indeed, a machine cannot be programmed to appreciate the value of human life.

Many states have argued for new legal rules to ensure human control over autonomous weapons systems. But a few others, including the US, hold that existing international law is sufficient. Though the uncertainty surrounding what meaningful human control actually is shows that more clarity in the form of new international law is needed.

This must focus on the essential qualities that make human control meaningful, while retaining human judgment in the context of specific use-of-force decisions. Without it, there's a risk of undercutting the value of new aimed at curbing weaponised AI.

This is important because without specific regulations, current practices in military decision-making will continue to shape what's considered "appropriate"—without being critically discussed.

Provided by The Conversation

This article is republished from The Conversation under a Creative Commons license. Read the original article.The Conversation

Citation: The threat of 'killer robots' is real and closer than you might think (2020, October 15) retrieved 28 March 2024 from https://techxplore.com/news/2020-10-threat-killer-robots-real-closer.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

AI has already been weaponised – and it shows why we should ban 'killer robots'

13 shares

Feedback to editors