Machines are starting to slowly replace humans on the battlefield. It’s believed fully autonomous weapons may be ready in 20-30 years. And Switzerland is moving ahead with its own research into mobile robot technology.This content was published on June 9, 2014 - 11:00
- Deutsch Bald auf einem Schlachtfeld in Ihrer Nähe?
- Español ¿Nos acercamos a un campo de batalla?
- Português Em breve num campo de batalhas perto de você?
- 中文 引入全自动杀人武器？且慢！
- عربي نحو استبدال البشر في ساحات المعارك بالأسلحة المُستقلة
- Français Les robots tueurs bientôt en action près de chez vous?
- Pусский Роботы – братья по оружию?
- 日本語 殺人ロボット兵器 規制を巡り初の専門家会議
- Italiano Il futuro dei campi di battaglia
At a recent meeting in Geneva on the threat of killer robot technology (lethal autonomous weapons systems or LAWS), a Swiss foreign ministry disarmament expert said the international community was only starting to grasp the possible repercussions of the robotic revolution on the future battlefield.
Laurent Masmejean added that it was crucial to further examine which applications of autonomous capabilities would be desirable, legal and acceptable, and which raised concerns.
Swiss representatives at the conference stressed that their ongoing defence research programmes did not include work on fully autonomous weapons.
“But we’ll see how different research fields, if considered together could help make progress in technology towards more autonomy in systems - lethal or not,” said Quentin Ladetto, director of defence future technologies at armasuisse, which oversees technology procurement at the Swiss defence ministry.
Switzerland currently pursues small research programmes for unmanned air and ground vehicles, conducted with numerous academic and industrial partners, including the Federal Institute of Technology in Zurich (ETHZ) and US firms Black-I Robotics and iRobot.
In its 2012-2016 research strategy the Swiss defence ministry underlines the importance of unmanned vehicles, especially for reconnaissance, surveillance and communication, which it says is likely to increase.
These unmanned air and ground vehicle programmes aim at “the autonomy of unmanned platforms, and creating the conditions for the approval and widespread use of mobile robots for security policy tasks in Switzerland”.
“If we have a closer look at research in robotics and artificial intelligence, the trend and focus is clearly on autonomy. Even if the ultimate research goals for academic institutions and universities respectively are not of military nature, the different building blocks that are being developed today and their integration tomorrow could lead to advanced autonomous capabilities,” said Ladetto.
But some experts have their doubts. Stuart Casey-Maslen head of research at the Geneva Academy of International Humanitarian Law and Human Rights, said he would be “astonished” if Switzerland wasn’t researching LAWS.
“More autonomy is clearly the future for the military and security forces. The question is whether, and if so how and where, the technology is restrained for legal or ethical reasons. I'm particularly concerned about use of lethal autonomous systems for law enforcement, which will not be considered within the CCW [Convention on Conventional Weapons] context,” he said.
Joe Farha, a military weapons specialist with the Manchester-based Omega Research Foundation, agreed that any civilian research into autonomy was likely to “spill over” into military uses or law enforcement.
“It is possible that more and more will be asked of unmanned systems so that a gradual advance in capability may lead to greater degrees of autonomy by default,” he declared, adding that unmanned ground or air systems could have dual usage as a weapons platform, “if the right technology is applied”.
The military of the future is likely to be increasingly unmanned. In recent years the US has spent $6 billion annually on unmanned systems for war. The US Defense Advanced Research Projects Agency (DARPA) has been developing military robotics and funding projects like the LS3 rough terrain robot created by the US firm Boston Dynamics.
And US policy documents reflect plans to increase the autonomy of weapons systems. In its Unmanned Systems Integrated Roadmap FY2011-2036 the US Department of Defense wrote that it “envisions unmanned systems seamlessly operating with manned systems while gradually reducing the degree of human control and decision-making require for the unmanned portion of the force structure.”
Ever more autonomous defence technology already exists. The US Navy’s MK 15 Phalanx Close-In Weapons System, and its land-based version, the Counter Rocket, Artillery and Mortar System (C-RAM), have been operational for years.
Elsewhere, Israel has deployed its Iron Dome automatic weapons defence system, and Germany has designed the NBS Mantis system to be used to protect forward operating bases in Afghanistan.
Talking about C-RAM, robotic warfare expert Peter W Singer said humans are part of the decision-making but mainly in the initial programming of the robot.
“During the actual operation of the machine, the operator only exercises veto power, and a decision to override a robot’s decision must be made in only half a second,” he said. The danger is that this can lead to “automation bias” – the tendency to trust an automated system.
Other precursors to fully autonomous weapons include South Korea and Israel’s sentry robots that can detect people and, if a human grants the command, fire its weapons.
“As nations develop the technological capability, many may choose to go down the path toward full autonomy, because of the benefits these weapons could provide: rapid response time, reduced risk to their own soldiers, fewer costs, and insulation from the effects of human emotions involved in decisions to use force,” said Mary Wareham, an arms expert at Human Rights Watch.
Unmanned aircraft are also moving beyond existing drones to have greater autonomy. The US Navy’s X-47B plane can take off from and land on an aircraft carrier and refuel on its own power, while Britain has unveiled the “autonomous and stealthy unmanned” Taranis aircraft. Both reportedly have weapons bays that could be adapted.
"Meaningful human control"
For NGOs there is no way of checking what countries mean by “meaningful human control” of more autonomous systems. Efforts to curtail such weapons have so far come to nothing and a pre-emptive ban is the only solution, they say.
“I have tried to engage with weapons manufacturers but they are frankly not interested in international humanitarian law (IHL),” said Juergen Altmann, co-founder of the International Committee for Robot Arms Control.
But Masmejean said the current legal basis was “quite solid” to handle future changes.
“Whether autonomous weapons are involved or not, all use of force or use of weapon systems must be in accordance with the entire framework of international law, and in particular IHL, most notably the principles of distinction, proportionality and precaution,” he said.
But he added that the international community should carefully examine the legal reviews of new weapons, the means and methods of warfare, and possible responsibility gaps.
Altmann is not convinced, however.
“They may be interested in the idea of limiting the effects of war if it’s for better precision which might lead to better products. But they rather want to innovate and the trend is towards unmanned vehicles which can react faster, which is exactly the direction we are trying to prevent.”
“There needs to be a political decision by society and states to block this and not let military and technological trends evolve freely.”
Some 87 countries out of the 117 that have signed up to the Convention on Conventional Weapons (CCW) attended the meeting of experts on lethal autonomous weapons systems (LAWS) at the United Nations in Geneva from 13-16 May, 2014.
The aim was to start to define the limits and responsibilities of LAWS. At their next annual meeting on 14 November CCW members will decide whether to continue the process.
Campaign groups are calling for a pre-emptive ban on such future weapons. They put forward the example of blinding lasers, which the international community banned by adopting an international protocol in 1995 before they could be used.
So far only five states, including Cuba and Pakistan, have joined activists calling for a ban on LAWS. Many others, including France and Britain, highlighted in Geneva the need to keep meaningful human control over targeting and attack decisions. The United States said there should be “appropriate” human control over autonomy in weapons systems, while Israel talked about the desirability of autonomous systems.End of insertion
This article was automatically imported from our old content management system. If you see any display errors, please let us know: email@example.com