New Military Technologies Require That We Rewrite the “Rules of War”

What's to come?
Nov. 12 2012 9:29 AM

Why We Need New “Rules of War”

Individually addressing new technologies like drones isn’t enough. We need to rewrite the rule book.

U.S. Marines in Helmand Province, Afghanistan, on March 3, 2011

Photo by Adek Berry/Getty Images.

Also in Slate: Fritz Allhoff describes the paradox of nonlethal weapons: International law sometimes restricts military technologies that don't kill more than those that do.

Some people worry about the use of drones—unmanned aerial vehicles—in warfare. Others are concerned about potential deployment of lethal autonomous robots that are programmed to identify, track, and destroy targets and people without a human decision-maker in the loop. Many fret about cyberconflict. Some fear the effects of cognitive enhancers—pharmaceuticals that will reduce the sense of risk that troops carry with them, or enable manipulation of memories, or change moral postures—on the battlefield. Others wonder about the implications of augmented cognition as the complexity and speed of combat environments grow beyond the ability of human perceptual and cognitive capabilities. And there are other war technologies worth discussing, some of which already exist, some of which may never come to fruition: self-guiding bullets, exoskeletons for burdened soldiers, “telepathic” brain-reading helmets, hummingbird- and insect-size cyborg platforms for surveillance or for attack, software to protect your networks and to attack those of your enemy.

It’s more than enough to feed technophobic dystopian fantasies or, alternatively, to generate techno-optimistic dreams of cultural conquest—especially, it seems, among those who have never served in military organizations. But the hyperventilation about specific technologies such as UAVs or lethal autonomous robots obscures two more subtle but fundamental challenges. The first is the need to understand, and manage, not just individual technologies but an accelerating rate of change across the entire technological frontier—indeed, each individual technology is only a small part of this much more complex reality. Second is the destabilizing impact these technologies have on the rules and norms that have evolved over centuries to try to reduce the awful impacts of war, especially on civilian populations.


Worse yet for those who would rather do their thinking in terms of bumper stickers, these challenges arise in an environment where unpredictable change in institutions, cultures, and practices of combat are all in fundamental flux. Currently, international behavior in cases of war is covered by the so-called the “laws of war,” which govern when a state may go to war, how to carry out such a conflict, and, increasingly, how to end it ethically and legally. For example, a state may protect itself or, under some circumstances, clearly signal a potential strike, but it may not simply attack another state out of spite. In conducting a war, the force used may not be more than is necessary to achieve military objectives, and noncombatants may not be targeted or attacked, although collateral damage resulting from the otherwise lawful and necessary pursuit of military objectives is permitted. These laws of war have been developed over many centuries with contributions by many civilizations and are embodied in treaties and agreements, such as the Geneva Conventions and the U.N. Charter, as well as customary international law.

But consider these rules and their underlying assumptions in light of current conditions. For example, today’s practices and international laws assume a state-based regime—the traditional wars among countries—but a very different pattern emerges when nonstate global actors, such as radical jihadist Islam, become a major source of conflict. Worse yet, since nations are losing their traditional monopoly on military technologies, nonstate actors and even individuals are gaining the ability to impose damage on a far larger scale. The laws of war tell you when a country can respond to an attack from another country—but they say nothing about how to respond to attacks by terrorist networks woven into the fabric of nations around the world. The geographical assumption that ties combatant status to a particular physical battlefield, core to the existing framework of the laws of war, is questionable in a world of global terrorism and cyberspace confrontation. Similarly, does building software bombs or software backdoors into a potential opponent’s Internet systems, civilian and military, constitute an attack that may be responded to by force even if they remain unactivated? And what happens when nonstate actors use the laws of war against those that consider themselves bound by them (a practice called “lawfare”)?

We are also seeing an eroding of the clear differences between a state of peace and a state of war, creating substantial institutional confusion. Drones operated by a single government such as the United States may be operated by the military or by intelligence entities, with private contractors involved to some extent. The U.S. armed forces operate under codified laws of war and strict rules of engagement, while intelligence organizations abide by very different standards (a complication that also exists in the cyberconflict domain). And it is not at all clear what formal rules may govern other players (private military contractors, for example, or nongovernmental organizations intervening in conflict zones for their own purposes). Sophisticated cyberattacks are not only difficult to attribute to any particular actor, but they often involve quasi-governmental organizations that may or may not be completely under the control, or even the sponsorship, of a particular government (like the pro-Putin youth organization Nashi). Contrary to popular belief, privateering is not just an 18th-century phenomenon.

There is an ongoing revolution in the nature of conflict as the battlefield, freed from the constraint of traditional weapons, extends globally through cyberconnections. The terrorist on leave from the Middle East may be tracked down and terminated in Peru by a drone operated from Nevada. Combat mixes with counterinsurgency mixes with policing mixes with nation-building, each a physically, culturally, and psychologically very different space subject to varying norms and rules. There is a revolution in civilian systems as well, as technologies developed for combat ooze into the civilian sector and societies re-evaluate the balance between security and personal freedom in spaces ranging from airports to social networking sites. Cartoons or films, protected free speech in Western societies, cause explosions of violence in Islamic societies on the other side of the world. The military itself is not immune. How, for example, does a military culture of obedience and discipline fit itself to gamers and cybergeeks?

The assumptions built into existing rules and norms about the clearly bounded geographies of battlefields, or the simple status of “combatant” vs. “noncombatant,” or the dominant role of states in international affairs, or even about what conflict actually is, are, at the least, unstable. At the worst, they are overthrown, with no obvious substitutes. Perhaps this is inevitable given the accelerating rate of disruptive technological change and the human desire to cling to institutions that may be growing obsolete but at least worked in the past. But given the importance of trying to reduce the impacts of warfare, this area requires far more serious attention than it has received to date. It isn’t that the established laws of war, many centuries in development, have suddenly become completely inapplicable—at the least, some future conflicts will consist of traditional state-to-state combat. But, clearly, technological evolution and concomitant changes in military, cultural, and social domains have rendered virtually all of the fundamental assumptions underlying the laws of war at least potentially contingent. It is unlikely that such fundamental and pervasive change will not affect in some meaningful way doctrines and principles that were formulated and tested under very different and less complex conditions.

We need to develop a sophisticated and adaptive institutional capability to recognize critical change as it happens, understand the implications across multiple domains, and respond in ways that are rational, ethical, and responsible. Calls for new treaties regarding specific technologies—cyberconflict, lethal autonomous robots, or drones—to some extent reflect this inchoate need. But useful as they may be as learning experience, they are mere attempts to update an already obsolete international regime. They neither appreciate, nor respond to, the enormity of the challenge before us: to create new and viable laws of conflict that represent a modern, sentient, and moral response to the human condition known as war.

This article was inspired by the 2012 Chautauqua Council on Emerging Technologies and 21st Century Conflict, sponsored by Arizona State University’s Consortium for Emerging Technologies, Military Operations, and National Security and Lincoln Center for Applied Ethics and held at the Chautauqua Institution in New York. Future Tense is a partnership of Arizona State, the New America Foundation, and Slate magazine.

Brad Allenby is President’s Professor of Sustainable Engineering, and Lincoln Professor of Engineering and Ethics, at Arizona State University.

Carolyn Mattick is a Ph.D. candidate in environmental engineering and a graduate fellow in emerging technologies at the Lincoln Center for Applied Ethics at Arizona State University.


Justice Ginsburg’s Crucial Dissent in the Texas Voter ID Case

Even When They Go to College, the Poor Sometimes Stay Poor

Here’s Just How Far a Southern Woman May Have to Drive to Get an Abortion

The Most Ingenious Teaching Device Ever Invented

Marvel’s Civil War Is a Far-Right Paranoid Fantasy

It’s also a mess. Can the movies do better?


Sprawl, Decadence, and Environmental Ruin in Nevada

Space: The Next Generation

An All-Female Mission to Mars

As a NASA guinea pig, I verified that women would be cheaper to launch than men.

Watching Netflix in Bed. Hanging Bananas. Is There Anything These Hooks Can’t Solve?

The Procedural Rule That Could Prevent Gay Marriage From Reaching SCOTUS Again

  News & Politics
Oct. 20 2014 3:53 PM Smash and Grab Will competitive Senate contests in Kansas and South Dakota lead to more late-breaking races in future elections?
Oct. 20 2014 5:39 PM Whole Foods Desperately Wants Customers to Feel Warm and Fuzzy Again
Oct. 20 2014 3:16 PM The Catholic Church Is Changing, and Celibate Gays Are Leading the Way
  Double X
The XX Factor
Oct. 20 2014 6:17 PM I Am 25. I Don't Work at Facebook. My Doctors Want Me to Freeze My Eggs.
  Slate Plus
Tv Club
Oct. 20 2014 7:15 AM The Slate Doctor Who Podcast: Episode 9 A spoiler-filled discussion of "Flatline."
Brow Beat
Oct. 20 2014 5:03 PM Marcel the Shell Is Back and as Endearing as Ever
Future Tense
Oct. 20 2014 4:59 PM Canadian Town Cancels Outdoor Halloween Because Polar Bears
  Health & Science
Medical Examiner
Oct. 20 2014 11:46 AM Is Anybody Watching My Do-Gooding? The difference between being a hero and being an altruist.
Sports Nut
Oct. 20 2014 5:09 PM Keepaway, on Three. Ready—Break! On his record-breaking touchdown pass, Peyton Manning couldn’t even leave the celebration to chance.