The Law That Applies to Autonomous Weapon Systems
Even though most robotics experts predict fully autonomous weapon systems will not be available for use on the battlefield for a number of years, the debate over the lawfulness of such systems has nonetheless begun. In November 2012, Human Rights Watch released a report entitled âLosing Humanity: The Case against Killer Robots.â In the report, which was produced in association with Harvard Law Schoolâs International Human Rights Clinic, Human Rights Watch asserts that fully autonomous weapons will be unable to comply with key law of armed conflict principles. Specifically, it claims that âsuch revolutionary weapons . . . would increase the risk of death or injury to civilians during armed conflict.â Accordingly, Human Rights Watch advocates for an international treaty prohibiting the development, production, and future use of fully autonomous systems.
Only a few days after the Human Rights Watch report was released, the U.S. Department of Defense (âDoDâ) seemingly reached a different conclusion about the lawfulness of such weapons when it released DoD Directive 3000.09. Entitled âAutonomy in Weapon Systems,â the directive established policies and guidelines for the development of autonomous functions in weapons. Although the directive does not specifically call for the development of fully autonomous weapons, it provides the required approval chain and details the necessary legal reviews to be conducted before such systems may be designed. It further mandates the enactment of a series of precautionary measures, such as safety measures and anti-tampering mechanisms, designed to keep autonomous systems from striking unintended targets.
These two documents have brought to the forefront the underlying question of whether autonomous weapon systems will ever be able to fully comply with the tenets of the law of armed conflict. This Insight explores these broader legal concerns over autonomous weapons by articulating the relevant precepts of law. It is instructive, however, to first begin with a brief introduction of the specific technology involved with these systems.
The Context: Autonomous Technology and What It Means to Be Fully Autonomous
The emergence of autonomous technology represents a potential sea change for modern warfare. DoD defines autonomous systems as those that âonce activated, can select and engage targets without further human intervention by a human operator.â Both Human Rights Watch and DoD acknowledge that such systems do not currently exist. However, in the past few years, developments in artificial intelligence have shown remarkable advances. Researchers are beginning to develop ânovel algorithms,â which allow unmanned systems to operate without human intervention. Using so called âmachine learningâ and other innovative approaches, computer systems are now coming closer to replicating human thought processes. These systems are better able to decipher and determine answers to complex problems by taking approaches similar to the way humans learn by example.
As these technological advances have emerged, militaries have begun embedding many autonomous features into weapon systems. Human Rights Watch fears that this trajectory will ultimately lead to fully autonomous weapon systems designed to make lethal targeting decisions. It asserts that such autonomous weapon systems will be unable to comply with key law of armed conflict principles. Given that charge, it is essential to examine more closely the relevant law.
The Law: How to Determine the Lawfulness of a Weapon System
It is incontrovertible that the law of armed conflict applies to autonomous weapon systems. When determining the overall lawfulness of a weapon system, there are two distinct aspects of the law that need to be analyzed: weapons law and targeting law. The former verifies that the weapon itself is lawful. The latter determines whether the use of the weapon system during hostilities might be prohibited in some manner under the law of armed conflict. A weapon must satisfy both aspects before it may be lawfully used on a battlefield.
When analyzing whether the weapon system itself is lawful, there are two distinct rules that apply. The first rule is that the weapon system must not be indiscriminate by its very nature. A weapon is deemed indiscriminate by nature if it cannot be aimed at a specific target and would be as likely to strike civilians as combatants. Found in Article 51(4)(b) of Additional Protocol I to the Geneva Conventions, the rule is considered to be reflective of customary international law. Accordingly, all states, even those not a party to the Protocol (such as the United States), are bound to comply with this customary law rule against indiscriminate attack. The mere fact that an autonomous weapon system rather than a human might be making the final targeting decision would not render the weapon indiscriminate by nature. Instead, as long as it is possible to supply the autonomous system with sufficiently reliable and accurate data to ensure it can be aimed at a military objective, then the system would not be deemed indiscriminate by nature. In the end, any proposed autonomous weapon system must comply with this provision to be lawful.
The second rule, codified in Article 35(2) of Additional Protocol I, is that a weapon system cannot cause unnecessary suffering or superfluous injury. This rule, which is also reflective of customary international law, seeks to prevent needless or inhumane injuries to combatants. A classic example of an unlawful weapon under this rule is a warhead that is filled with glass. Such a warhead would unnecessarily complicate medical treatment and would consequently be unlawful. This rule only presents a problem for an autonomous system if the specific warheads or weapons installed on the system would violate the rule. The fact that the system autonomously decides to engage a target does not itself affect or violate the prohibition on unnecessary suffering or superfluous injury. To potentially be deemed lawful, a fully autonomous weapon system must only be armed with weapons and ammunition that comply with this rule.
To verify compliance with the two rules listed above, a state intent on fielding a new weapon must conduct a thorough legal review. This requirement for a legal review, which appears in Article 36 of Additional Protocol I, ensures that the weapon is not indiscriminate and that it would not cause unnecessary suffering or superfluous injury. The review also determines whether there is any other particular provision under the law of armed conflict which would prohibit the use of the weapon. Customary law requires this legal review of weapons and weapon systems, also referred to as the means of warfare, and these reviews are thereby required by all states, including those not a party to the Protocol. The DoD Directive sets forth the U.S. policy of requiring such a review at the early stages of development and again just prior to actually fielding the system. Furthermore, if a weapon system is significantly modified after its initial fielding, then an additional review would be necessary. The development of any fully autonomous weapon system would clearly require such legal reviews.
Assuming the particular weapon satisfies the above weapons law rules, the weapon must still be examined under targeting law to determine whether the actual use of the weapon might be prohibited in some manner. To conduct this analysis, three core law of armed conflict requirements are particularly salient: distinction, proportionality, and precautions in the attack. A weapon system, even one deemed lawful under the above tenets of weapons law, may not be lawfully used if, based on the circumstances, its use would violate any one of these three requirements.
The first requirement is distinction. Distinction is the most fundamental principle of the law of armed conflict. A customary law principle, distinction obliges a combatant to distinguish between combatants and civilians, as well as between military and civilian objects. The rule is also codified in Article 48 of Additional Protocol I, with companion rules in Articles 51 and 52. The principle is intended to protect the civilian population by directing military attacks against only military targets. The context and environment in which the weapon system operates play a significant role in this analysis. There may be situations in which an autonomous weapon system could satisfy this rule with a considerably low level ability to distinguish between civilian and military targets. Examples would include during high intensity conflicts against declared hostile forces or in battles that occur in remote regions, such as underwater, deserts, or areas like the Demilitarized Zone in Korea. At other times, such as on a complex counterinsurgency-type battlefield or in an urban environment, the demands on the systems to distinguish would be drastically higher. In the latter situations, even autonomous systems equipped with the most robust sensor packages may have difficulty fulfilling this requirement. Ultimately, for the use of an autonomous weapon system to be lawful, the system would be expected to reasonably distinguish between combatants and civilians (and between military objectives and civilian objects) given the particular environment and circumstances of the battlefield ruling at the time.
The second requirement, proportionality, requires combatants to examine whether the expected collateral damage from an attack would be excessive in relation to the anticipated military gain. This complex principle is reflective of customary international law and has traditionally involved a human judgment call evaluated on the basis of reasonableness. This rule is also codified in both Article 51(5)(b) and Article 57(2)(iii) of Additional Protocol I. Many critics of autonomous weapon systems, including Human Rights Watch, have questioned whether the systems can fulfill this requirement. To comply with the principle, autonomous weapon systems would, at a minimum, need to be able to estimate the expected amount of collateral harm that might come to civilians from an attack. Additionally, if civilian casualties were likely to occur, the autonomous systems would need to be able to compare the amount of collateral harm against some predetermined military advantage value of the target. This step may present a significant challenge for autonomous weapon systems. The military advantage of a particular target is extremely contextual, and its value can change rapidly based upon developments on the battlefield. Human operators may be able to develop sliding scale-type mechanisms which regularly update and provide the autonomous weapon system with the relative military advantage value of a given target. Operators might also help fulfill this principle by detailing strict rules of engagement for these systems and establishing other controls, such as geographic or time limits on use. Nevertheless, these complicated issues would need to be resolved, if the future use of autonomous weapon systems is to comport with the principle of proportionality.
The third and final core requirement is the obligation to take feasible precautions in the attack. These precautions, which are customary in nature and codified in Article 57 of Additional Protocol I, present certain challenges to autonomous weapon systems. One such challenge will be the requirement to do everything feasible to verify that a target is a military one. Feasible, in this context, generally means âthat which is practicable or practically possible, taking into account all circumstances prevailing at the time, including humanitarian and military considerations.â There may be times where the robust recognition capabilities of an autonomous system would be more precise (and, thus, more reliable) than a human in fulfilling this requirement. In other cases, depending on the circumstances (and what is practically possible), a force may have to augment the autonomous system with other sensors to help verify the target. Another significant challenge is the requirement to do everything feasible to choose a means of attack âwith a view to avoiding, and in any event minimizing,â collateral damage. This precaution may, under certain circumstances, preclude the use of an autonomous system if a different type of a system would better protect civilians. With all of the required precautions in attack, there is inherently a value judgment about whether all feasible steps have been taken. How autonomous systems will reasonably make this value judgment may prove to be one of the biggest challenges in terms of compliance. Ultimately if a country intends to use an autonomous weapon system on a battlefield, it must ensure that the system can adequately take these feasible precautions.
In this debate, which is often infused with policy, morality, or ethical arguments, it is important for individuals in the international law field to distinguish those arguments from purely legal ones. The above rules outline the basic legal standards a weapon must meet to be deemed lawful. It remains to be seen whether in the future autonomous weapon systems will be banned or will flourish on the battlefield. Both the United States and groups like Human Rights Watch recognize the significance and the unique nature of these legal concerns, but, at least initially, they have staked vastly different approaches and positions towards handling autonomous weapon systems. The only thing clear at this point is that this legal debate will likely continue.
Jeffrey S. Thurnher is a Judge Advocate in the U.S. Army and a faculty member in the International Law Department at the U.S. Naval War College. Major Thurnher was the winner of the 2009 ASIL Lieber Society Military Prize. The opinions expressed herein are solely those of the author, and are not intended in any way to reflect an official position of the U.S. Naval War College, the U.S. Department of Defense, or any branch of the U.S. Government.