{"id":1193,"date":"2013-03-18T13:41:14","date_gmt":"2013-03-18T19:41:14","guid":{"rendered":"http:\/\/peaceeconomyproject.org\/wordpress\/?p=1193"},"modified":"2013-03-18T13:41:14","modified_gmt":"2013-03-18T19:41:14","slug":"smart-drones","status":"publish","type":"post","link":"https:\/\/peaceeconomyproject.org\/wordpress\/smart-drones\/","title":{"rendered":"Smart Drones"},"content":{"rendered":"<p>by Bill Keller, New York Times<br \/>\n<a href=\"http:\/\/www.nytimes.com\/2013\/03\/17\/opinion\/sunday\/keller-smart-drones.html?pagewanted=all&amp;_r=0\">click here for original article<\/a><\/p>\n<p>IF you find the use of remotely piloted warrior drones troubling, imagine that the decision to kill a suspected enemy is not made by an operator in a distant control room, but by the machine itself. Imagine that an aerial robot studies the landscape below, recognizes hostile activity, calculates that there is minimal risk of collateral damage, and then, with no human in the loop, pulls the trigger.<\/p>\n<div id=\"readerscomment\">\n<h3><span style=\"font-size: 16px;\">Welcome to the future of warfare. While Americans are debating the president\u2019s power to order assassination by drone, powerful momentum \u2014 scientific, military and commercial \u2014 is propelling us toward the day when we cede the same lethal authority to software.<\/span><\/h3>\n<\/div>\n<div>\n<p>Next month, several human rights and arms control organizations are meeting in London to introduce a campaign to ban killer robots before they leap from the drawing boards. Proponents of a ban include many of the same people who succeeded in building a civilized-world consensus against the use of crippling and indiscriminate land mines. This time they are taking on what may be the trickiest problem arms control has ever faced.<\/p>\n<p>The arguments against developing fully autonomous weapons, as they are called, range from moral (\u201cthey are evil\u201d) to technical (\u201cthey will never be that smart\u201d) to visceral (\u201cthey are creepy\u201d).<\/p>\n<p>\u201cThis is something people seem to feel at a very gut level is wrong,\u201d says Stephen Goose, director of the arms division of Human Rights Watch, which has assumed a leading role in\u00a0<a href=\"http:\/\/www.hrw.org\/reports\/2012\/11\/19\/losing-humanity-0\">challenging the dehumanizing of warfare<\/a>. \u201cThe ugh factor comes through really strong.\u201d<\/p>\n<p>Some robotics experts doubt that a computer will ever be able to reliably distinguish between an enemy and an innocent, let alone judge whether a load of explosives is the right, or proportional, response. What if the potential target is already wounded, or trying to surrender? And even if artificial intelligence achieves or surpasses a human level of competence, the critics point out, it will never be able to summon compassion.<\/p>\n<p>Noel Sharkey, a computer scientist at the University of Sheffield and chairman of the International Committee for Robot Arms Control, tells the story of an American patrol in Iraq that came upon a group of insurgents, leveled their rifles, then realized the men were carrying a coffin off to a funeral. Killing mourners could turn a whole village against the United States. The Americans lowered their weapons. Could a robot ever make that kind of situational judgment?<\/p>\n<p>Then there is the matter of accountability. If a robot bombs a school, who gets the blame: the soldier who sent the machine into the field? His commander? The manufacturer? The inventor?<\/p>\n<p>At senior levels of the military there are misgivings about weapons with minds of their own. Last November the Defense Department\u00a0<a href=\"http:\/\/www.dtic.mil\/whs\/directives\/corres\/pdf\/300009p.pdf\">issued<\/a>\u00a0what amounts to a 10-year moratorium on developing them while it discusses the ethical implications and possible safeguards. It\u2019s a squishy directive, likely to be cast aside in a minute if we learn that China has sold autonomous weapons to Iran, but it is reassuring that the military is not roaring down this road without giving it some serious thought.<\/p>\n<p>Compared with earlier heroic efforts to outlaw land mines and curb nuclear proliferation, the campaign against licensed-to-kill robots faces some altogether new obstacles.<\/p>\n<p>For one thing, it\u2019s not at all clear where to draw the line. While the Terminator scenario of cyborg soldiers is decades in the future, if not a complete fantasy, the militaries of the world are already moving along a spectrum of autonomy, increasing, bit by bit, the authority of machines in combat.<\/p>\n<p>The military already lets machines make critical decisions when things are moving too fast for deliberate human intervention. The United States has long had Aegis-class warships with automated antimissile defenses that can identify, track and shoot down incoming threats in seconds. And the role of machinery is expanding toward the point where that final human decision to kill will be largely predetermined by machine-generated intelligence.<\/p>\n<p>\u201cIs it the finger on the trigger that\u2019s the problem?\u201d asks Peter W. Singer, a specialist in the future of war at the Brookings Institution. \u201cOr is it the part that tells me \u2018that\u2019s a bad guy\u2019?\u201d<\/p>\n<p>Israel is the first country to make and deploy (and sell, to China, India, South Korea and others) a weapon that can attack pre-emptively without a human in charge. The hovering drone called the Harpy is programmed to recognize and automatically divebomb any radar signal that is not in its database of \u201cfriendlies.\u201d No reported misfires so far, but suppose an adversary installs its antiaircraft radar on the roof of a hospital?<\/p>\n<p>Professor Sharkey points to the Harpy as a weapon that has already crossed a worrisome threshold and probably can\u2019t be called back. Other systems are close, like the Navy\u2019s X-47B, a pilotless, semi-independent, carrier-based combat plane that is in the testing stage. For now, it is unarmed but it is built with two weapons bays. We are already ankle-deep in the future.<\/p>\n<p>For military commanders the appeal of autonomous weapons is almost irresistible and not quite like any previous technological advance. Robots are cheaper than piloted systems, or even drones, which require scores of technicians backing up the remote pilot. These systems do not put troops at risk of death, injury or mental trauma. They don\u2019t get tired or frightened. A weapon that is not tethered to commands from home base can continue to fight after an enemy jams your communications, which is increasingly likely in the age of electromagnetic pulse and cyberattacks.<\/p>\n<p>And no military strategist wants to cede an advantage to a potential adversary. More than 70 countries currently have drones, and some of them are hard at work on the technology to let those drones off their virtual leashes.<\/p>\n<p>\u201cEven if you had a ban, how would you enforce it?\u201d asks Ronald Arkin, a computer scientist and director of the Mobile Robot Laboratory at Georgia Tech. \u201cIt\u2019s just software.\u201d<\/p>\n<p>THE military \u2014 and the merchants of war \u2014 are not the only ones invested in this technology. Robotics is a hyperactive scientific frontier that runs from the most sophisticated artificial intelligence labs down to middle-school computer science programs. Worldwide, organized robotics competitions engage a quarter of a million school kids. (My 10-year-old daughter is one of them.) And the science of building killer robots is not so easily separated from the science of making self-driving cars or computers that excel at \u201cJeopardy.\u201d<\/p>\n<p>Professor Arkin\u00a0<a href=\"http:\/\/www.cc.gatech.edu\/ai\/robot-lab\/online-publications\/Arkin_ethical_autonomous_systems_final.pdf\">argues<\/a>\u00a0that automation can also make war more humane. Robots may lack compassion, but they also lack the emotions that lead to calamitous mistakes, atrocities and genocides: vengefulness, panic, tribal animosity.<\/p>\n<p>\u201cMy friends who served in Vietnam told me that they fired \u2014 when they were in a free-fire zone \u2014 at anything that moved,\u201d he said. \u201cI think we can design intelligent, lethal, autonomous systems that can potentially do better than that.\u201d<\/p>\n<p>Arkin argues that autonomous weapons need to be constrained, but not by abruptly curtailing research. He advocates a moratorium on deployment and a full-blown discussion of ways to keep humans in charge.<\/p>\n<p>Peter Singer of Brookings is also wary of a weapons ban: \u201cI\u2019m supportive of the intent, to draw attention to the slippery slope we\u2019re going down. But we have a history that doesn\u2019t make me all that optimistic.\u201d<\/p>\n<p>Like Singer, I don\u2019t hold out a lot of hope for an enforceable ban on death-dealing robots, but I\u2019d love to be proved wrong. If war is made to seem impersonal and safe, about as morally consequential as a video game, I worry that autonomous weapons deplete our humanity. As unsettling as the idea of robots\u2019 becoming more like humans is the prospect that, in the process, we become more like robots.<\/p>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>by Bill Keller, New York Times click here for original article IF you find the use of remotely piloted warrior drones troubling, imagine that the decision to kill a suspected enemy is not made by an operator in a distant control room, but by the machine itself. Imagine that an aerial robot studies the landscape [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":1195,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"nf_dc_page":"","om_disable_all_campaigns":false,"_monsterinsights_skip_tracking":false,"_monsterinsights_sitenote_active":false,"_monsterinsights_sitenote_note":"","_monsterinsights_sitenote_category":0,"_jetpack_memberships_contains_paid_content":false,"footnotes":"","jetpack_publicize_message":"","jetpack_publicize_feature_enabled":true,"jetpack_social_post_already_shared":false,"jetpack_social_options":{"image_generator_settings":{"template":"highway","default_image_id":0,"font":"","enabled":false},"version":2}},"categories":[8],"tags":[],"class_list":["post-1193","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-news"],"acf":[],"aioseo_notices":[],"jetpack_publicize_connections":[],"jetpack_featured_media_url":"https:\/\/i0.wp.com\/peaceeconomyproject.org\/wordpress\/wp-content\/uploads\/2013\/03\/17KELLER-articleLarge1.jpg?fit=600%2C356&ssl=1","jetpack_sharing_enabled":true,"jetpack_likes_enabled":true,"_links":{"self":[{"href":"https:\/\/peaceeconomyproject.org\/wordpress\/wp-json\/wp\/v2\/posts\/1193","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/peaceeconomyproject.org\/wordpress\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/peaceeconomyproject.org\/wordpress\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/peaceeconomyproject.org\/wordpress\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/peaceeconomyproject.org\/wordpress\/wp-json\/wp\/v2\/comments?post=1193"}],"version-history":[{"count":1,"href":"https:\/\/peaceeconomyproject.org\/wordpress\/wp-json\/wp\/v2\/posts\/1193\/revisions"}],"predecessor-version":[{"id":1196,"href":"https:\/\/peaceeconomyproject.org\/wordpress\/wp-json\/wp\/v2\/posts\/1193\/revisions\/1196"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/peaceeconomyproject.org\/wordpress\/wp-json\/wp\/v2\/media\/1195"}],"wp:attachment":[{"href":"https:\/\/peaceeconomyproject.org\/wordpress\/wp-json\/wp\/v2\/media?parent=1193"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/peaceeconomyproject.org\/wordpress\/wp-json\/wp\/v2\/categories?post=1193"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/peaceeconomyproject.org\/wordpress\/wp-json\/wp\/v2\/tags?post=1193"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}