Design World

  • Home
  • Technologies
    • ELECTRONICS • ELECTRICAL
    • Fastening • joining
    • FLUID POWER
    • LINEAR MOTION
    • MOTION CONTROL
    • SENSORS
    • TEST & MEASUREMENT
    • Factory automation
    • Warehouse automation
    • DIGITAL TRANSFORMATION
  • Learn
    • Tech Toolboxes
    • Learning center
    • eBooks • Tech Tips
    • Podcasts
    • Videos
    • Webinars • general engineering
    • Webinars • Automated warehousing
    • Voices
  • LEAP Awards
  • 2025 Leadership
    • 2024 Winners
    • 2023 Winners
    • 2022 Winners
    • 2021 Winners
  • Design Guides
  • Resources
    • 3D Cad Models
      • PARTsolutions
      • TraceParts
    • Digital Issues
      • Design World
      • EE World
    • Educational Assets
    • Engineering diversity
    • Reports
    • Trends
  • Supplier Listings
  • Advertise
  • SUBSCRIBE
    • MAGAZINE
    • NEWSLETTER

How I learned to stop worrying and love “killer robots”

By atesmeh | June 20, 2013

Stars and Stripes recently ran an article from the Pittsburgh Tribune-Review with the provocative title “Militaries’ growing use of ground robots raises ethics concerns.” It rehashes old concerns about “killer robots” — lack of accountability, ethical responsibility — but its thesis is a fear of modern technology.

This irrational fear assigns higher moral standards to robots vs. flesh-and-blood troops, retarding technological progress and putting lives at risk. We should never downplay anything with the power to kill, but fear of the unknown shouldn’t paralyze us.

In the case of unmanned weapons systems (aka, “killer robots”), which have the power to reduce collateral damage and save lives, we should support and encourage their development, not preemptively ban them and set disproportionately high ethical standards as a function of their deployment.

And these tricky ethical concerns — responsibility, accountability — won’t get away anytime soon. The war on terror — Afghanistan and Waziristan, in particular — has dramatically raised the profile of unmanned weapons systems. Since 2004, we’ve launched 354 UAV strikes in the vicinity of northwest Pakistan, and the frequency has shot up under the present administration.

A wide swath of human rights groups, political factions, and interested parties have charged these artificial warriors with collateral rates up to 35%. Human Rights Watch has called for a preemptive ban. But drones are almost certainly more humane than 20th century tools of warfare, which killed up to three civilians for every enemy soldier.

Let’s be clear — every civilian death is tragic; collateral damage always has the same result, regardless of the source. But shouldn’t we emphasize weaponry that causes fewer civilian casualties?

In the minds of critics, it’s preferable to die a noble death at the behest of a human than suffer the indignity of losing your life to a cold, unfeeling robot. Here’s a hint: The poor sap is dead either way.

This quote from Steve Goose, Arms Division director at Human Rights Watch, says it all: “Giving machines the power to decide who lives and dies on the battlefield would take technology too far.” “Human control of robotic warfare is essential to minimizing civilian deaths and injuries,” he said.

How does Mr. Goose know that humans are more adept at preventing collateral damage? Why would unmanned weapons systems — operating with robotic precision — cause more civilian casualties?

If Human Rights Watch and other NGOs were truly concerned with reducing collateral damage, they would support the development of military robots.

Robots aren’t prone to mental distractions and have far more information at their disposal than we do. Their identification friend or foe (IFF) programs are infinity more sophisticated than the fallible judgment of a human.

The IEEE Spectrum nails it: “A robot can use high-resolution cameras, infrared imaging, ultraviolet imaging, radar, LIDAR, data feeds from other robots, and anything else you can think of all at once to determine very quickly how tall a person is, how much they weigh, and whether they’re holding an ice cream made of ice cream or a gun made of metal.”

Robots don’t get fatigued. They don’t experience stress. They won’t succumb to post-traumatic stress disorder (PTSD) and the associated dangers. And humans define their technical and moral parameters.

Forget Hollywood’s killer-robot hyperbole. Real autonomous military robots — like the Navy’s PHALANX Close-in Weapon System (CIWS) — operate in a highly restrictive environment and can only make reactive “decisions”. The CIWS, for example, detects incoming projectiles and autonomously defeats them.

Their “autonomy” is based entirely on hypothetical scenarios. They can’t formulate their own strategies, and they won’t plot the enslavement of mankind. Even autonomous robots — which act without human input — cannot help but follow their programming. They have no choice in the matter.

As The Spectrum points out, calling a robot “killer” ascribes a sinister motive to it, but robots have no will of their own.

So why the irrational fear of “killer” robots? Fear of the unknown — of advanced technology that critics don’t fully understand — is the biggest obstacle. And we should never let fear stand in the way of progress.

You might also like


Filed Under: Aerospace + defense

 

LEARNING CENTER

Design World Learning Center
“dw
EXPAND YOUR KNOWLEDGE AND STAY CONNECTED
Get the latest info on technologies, tools and strategies for Design Engineering Professionals.
Motor University

Design World Digital Edition

cover

Browse the most current issue of Design World and back issues in an easy to use high quality format. Clip, share and download with the leading design engineering magazine today.

EDABoard the Forum for Electronics

Top global problem solving EE forum covering Microcontrollers, DSP, Networking, Analog and Digital Design, RF, Power Electronics, PCB Routing and much more

EDABoard: Forum for electronics

Sponsored Content

  • Robot Integration with Rotary Index Tables and Auxiliary Axes
  • How to Choose the Right Rotary Index Table for Your Application
  • Designing a Robust Rotary Index Table: Engineering Best Practices for Long-Term Performance
  • Custom Integration Options for your New and Existing Rotary Table Applications
  • Tech Tips: Crossed Roller Bearing Update
  • Five Uses for the Parvalux Modular Range
View More >>
Engineering Exchange

The Engineering Exchange is a global educational networking community for engineers.

Connect, share, and learn today »

Design World
  • About us
  • Contact
  • Manage your Design World Subscription
  • Subscribe
  • Design World Digital Network
  • Control Engineering
  • Consulting-Specifying Engineer
  • Plant Engineering
  • Engineering White Papers
  • Leap Awards

Copyright © 2026 WTWH Media LLC. All Rights Reserved. The material on this site may not be reproduced, distributed, transmitted, cached or otherwise used, except with the prior written permission of WTWH Media
Privacy Policy | Advertising | About Us

Search Design World

  • Home
  • Technologies
    • ELECTRONICS • ELECTRICAL
    • Fastening • joining
    • FLUID POWER
    • LINEAR MOTION
    • MOTION CONTROL
    • SENSORS
    • TEST & MEASUREMENT
    • Factory automation
    • Warehouse automation
    • DIGITAL TRANSFORMATION
  • Learn
    • Tech Toolboxes
    • Learning center
    • eBooks • Tech Tips
    • Podcasts
    • Videos
    • Webinars • general engineering
    • Webinars • Automated warehousing
    • Voices
  • LEAP Awards
  • 2025 Leadership
    • 2024 Winners
    • 2023 Winners
    • 2022 Winners
    • 2021 Winners
  • Design Guides
  • Resources
    • 3D Cad Models
      • PARTsolutions
      • TraceParts
    • Digital Issues
      • Design World
      • EE World
    • Educational Assets
    • Engineering diversity
    • Reports
    • Trends
  • Supplier Listings
  • Advertise
  • SUBSCRIBE
    • MAGAZINE
    • NEWSLETTER
We use cookies to personalize content and ads, to provide social media features, and to analyze our traffic. We share information about your use of our site with our social media, advertising, and analytics partners who may combine it with other information you’ve provided to them or that they’ve collected from your use of their services. You consent to our cookies if you continue to use this website.