Why military 鈥榙rone swarms鈥 raise ethical concerns in future wars
Intelligent drones equipped with AI offer military advantages while raising ethical concerns about autonomous computer warfare.聽
Intelligent drones equipped with AI offer military advantages while raising ethical concerns about autonomous computer warfare.聽
The proliferation of cheap drones in conflicts in Ukraine and the Middle East has sparked a scramble to perfect uncrewed vehicles that can plan and work together on the battlefield.聽
These next-generation, intelligent 鈥渟warms鈥 would represent a breakthrough in warfare. Rather than soldiers piloting individual uncrewed vehicles, they could deploy air and seaborne swarms on missions 鈥渨ith limited need for human attention and control,鈥 according to a recent U.S. government report. It鈥檚 the 鈥渉oly grail鈥 for the military, says Samuel Bendett, an adviser to the Center for Naval Analysis, a federally funded research and development center.聽
It鈥檚 also an ethical minefield. As researchers apply artificial intelligence and autonomy to lethal machines, their systems raise the specter of drone armies and pose new questions about the role human control should play in modern combat. And while Pentagon officials have long promised that humans will always be 鈥渋n the loop鈥 when it comes to decisions to kill, the Defense Department last year updated its guidance to address AI autonomy in weapons.聽
鈥淚t鈥檚 a very high level of approval to even proceed with testing of a fully autonomous weapons system,鈥 says Duane T. Davis, a senior lecturer in the computer science department at the Naval Postgraduate School in Monterey, California. But it does 鈥減rovide for the possibility of completely autonomous weapons systems.鈥澛犅
That鈥檚 largely because much U.S. military research is driven by fears of how adversaries may exploit their own swarm technology in a future conflict with the United States or its allies. The question going forward is whether the Pentagon can overcome the myriad technological challenges of drone warfare while also maintaining the ethics of a democratic state.
The concern is that China 鈥渋s not going to wrestle with these same ethical decisions in the way that we will,鈥 says Dr. Davis.聽
What makes a swarm
Current instances of uncrewed military group attacks over battlefields 鈥 as well as the drone light shows now popping up as entertainment in night skies over the U.S. 鈥 are not intelligent swarms. The former are essentially salvos of slow-moving aerial 鈥渕issiles,鈥 each one operated by a human, with no machine-to-machine coordination or communication. The latter 鈥撀燼 high-tech alternative to fireworks 鈥撀燼re preprogrammed displays in near-ideal conditions, which aren鈥檛 particularly useful in a military setting, since an adversary can figure out how to counter them.
鈥淔or an enemy, that just means I鈥檝e got a pattern of things I can shoot at, or they鈥檙e operating similarly, so it鈥檚 easier to predict what they鈥檙e going to do,鈥 notes Bryan Clark, senior fellow at Hudson Institute.聽
Swarms instead use an array of sensors to communicate drone to drone 鈥 and then switch to AI to plan and collaborate for attacks on the fly. They鈥檙e programmed to create a siege of overwhelming force from 鈥渁 bunch of different angles 鈥 the way ants crawl all over a beetle, or whatever, to eat it,鈥 says Zachary Kallenborn, a fellow at George Mason University鈥檚 Schar School of Policy and Government.
A big challenge for current drone operators on Ukraine鈥檚 battlefields is Russian jamming technology, which can prevent operator-drone and, thus, drone-to-drone communication. To address this challenge, some researchers are working on ways for drones to observe and infer what other drones are doing.聽
The fog of war complicates visual observation. That鈥檚 why Theodore Pavlic of Arizona State University recently began studying weaver ants in Australia at the behest of U.S. Special Operations Command. As the ants swarm and transport their prey up trees, they sense each other鈥檚 presence without constantly looking around.聽
They also cooperate and make decisions as a team. 鈥淚f we can replicate that [with drones], you can basically hit go, and they will plan their own way,鈥 says Dr. Pavlic, who also studies stingless bees and other types of ants. 鈥淚f new challenges occur, then they can [set] temporary short-term goals to get around those challenges.鈥澛
Bang for the buck
Building smart drones, with more onboard intelligence and computing power, means bigger and more expensive machines, and that has a downside. 鈥淐omputers can only be so small, and you can only put so much power and payload onto a drone,鈥 says Nisar Ahmed, director of the Research and Engineering Center for Unmanned Vehicles at the University of Colorado Boulder.聽
Just for a drone to take off, for starters, requires roughly 10 times the energy that a world-class sprinter expends to run a 100-meter race, says Vijay Kumar, dean of the University of Pennsylvania鈥檚 engineering school. The result: Missions with aerial drones are currently limited in terms of distance and time. Since longer-range drones are expensive, cheaper drones that can stay aloft for an hour 鈥 or even 30 minutes 鈥 offer more bang for the buck.
Despite the challenges, researchers are making progress. Red Cat Holdings, a drone technology company in Puerto Rico, announced last year a system in which one person could operate four of its Teal drones, as opposed to today鈥檚 1-1 ratio. The company aims to increase that ratio by pushing even more autonomy onto the machines themselves.
Embedding such autonomy in lethal machines, however, also poses ethical challenges about maintaining human oversight 鈥 particularly as the speed and complexity of drone decision-making increases. Humans, after all, don鈥檛 process information as quickly as machines, which may increase pressure to take humans out of the loop if, say, China or another adversary deploys AI-equipped drones capable of full autonomy.
The Pentagon hired an ethics officer in 2020 to grapple with precisely such challenges. Still, 鈥淲e need more people thinking about them in the context of the military, in the context of international law, in the context of ethics,鈥 says Margaret E. Kosal, a professor at the Sam Nunn School of International Affairs at the Georgia Institute of Technology and former science and technology adviser at the Defense Department.聽
A machine gun analogy
What is clear is that the technology will continue to develop at breakneck speed, even as researchers wrestle with challenges specific to the battlefield of the day. Drones will change war the way the machine gun did more than a century ago, says George Matus, chief technology officer of Red Cat and founder of its Teal subsidiary.聽
鈥淏ack then, a handful of gunners could defeat large numbers of even the mightiest cavalry. [Sometimes, even] today, a handful of drones can defeat a battalion of the mightiest armored vehicles before they even reach the front line.鈥 In the future, intelligent swarms will prove even more effective, he adds.
While many researchers worry the technology is one more step toward all-out swarm warfare, Mr. Matus embraces the vision.
鈥淭he front line is going to become majority automated, if not fully automated,鈥 he says. 鈥淭here鈥檚 no doubt in my mind at least for the next couple of decades, this is going to be a very large part of the future of war.鈥
Others see it as an evolutionary step with more limited battlefield applications. 鈥淚t is not fundamentally going to be a revolution in military affairs,鈥 says Dr. Kosal. 鈥淭hat doesn鈥檛 mean we shouldn鈥檛 be worried.鈥