why not have the robot war where you live, do you not know that the Sahara الصحراء الكبرى - is not a dustbin but part of many populated sovereign counties.
Photograph by Luke MacGregor, Reuters
Published June 4, 2013
In a world of robotic cars and robotic assembly lines, could warriors be next? Some countries, including the U.S., Israel, and the U.K., are developing lethal autonomous robots that would mimic human soldiers and could be sent into battle.
Several robot models have been designed to carry machine guns and advanced rifles. The company iRobot, which makes the Roomba vacuum cleaner, has developed a robot to taser an enemy. The U.S. Navy has tested an unmanned drone that is capable of taking off and landing on an aircraft carrier and can deploy two tons of artillery.
Proponents see benefits, including the ability to wage combat with an indefatigable military, as well as the advantage of fewer casualties. But international organizations are hesitant about such dangerous technology and the risks it may pose. National Geographic asked Christof Heyns—a human rights lawyer and the United Nations' Special Rapporteur on extrajudicial, summary, or arbitrary executions—about lethal autonomous robots (LARs), as these warrior robots are known, and whether developers should tread lightly.
You've called for a moratorium on lethal autonomous robots. What concerns you about this technology?
Can robots make the kind of decisions that are required in war, [such as] distinguish between combatants and civilians? They could make it easier for states to go to war, [and] thus could be used by despots to repress their own people. Who would be responsible if a robot goes wild? Is it acceptable that machines kill people?
Many countries have long had lethal automatic instruments of warfare like missiles and bombs, not to mention the increasing use of drones. What would make these robots so different?
Bombs are not autonomous, and neither are drones. This is a crucial distinction and the very basis of the concern about LARs. Drones have a human in the loop. LARs don't, so with LARs, the machine makes the decision whether and who to kill.
Despite the legal, human, and foreign policy risks, is there an upside to automated warriors?
Yes, those who promote them say they can be more targeted and as a result reduce civilian casualties. They can also obtain [tactical] information [about a particular enemy] that we may be able to get or process. But then again, they can make it easier to go to war.
It's all hands (and paws) on deck when it comes to the poaching crisis in Africa.
In this new series, writers and photographers from around the world reflect on places that hold special meaning for them.
For Sebastián García Iglesias, the ghosts of his ancestors are stitched to the tapestry of the land they pioneered.
The Future of Food
Food. It's driven nearly everything we've ever done as a species, and yet it's one of the most overlooked aspects of human history.
We've made our magazine's best stories about the future of food available in a free iPad app.