Abstract
This chapter articulates and discusses several arguments against the lethal use of unmanned aerial vehicles, often called drones. A distinction is made between targeted killing, killing at a distance, and automated killing, which is used to map the arguments against lethal drones. After considering issues concerning the justification of war, the argument that targeted killing makes it easier to start a war, and the argument that killing at a distance is problematic, this chapter focuses on two arguments against automated killing, which are relevant to all kinds of “machine killing”. The first argument (from moral agency) questions if machines can ever be moral agents and is based on differences in capacities for moral decision-making between humans and machines. The second argument (from moral patiency), which has received far less attention in the literature on machine ethics and ethics of drones, focuses on the question if machines can ever be “moral patients”. It is argued that there is a morally significant qualitative difference in vulnerability and way of being between drones and humans, and that because of this asymmetry fully automated killing without or with little human involvement is not justified.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
- 2.
See for instance the UN report Heyns 2013.
- 3.
- 4.
Note that there is also another argument, which starts from virtue ethics: if this war is risk free, then it requires no courage. Some might say it is cowardly to attack people while being yourself invulnerable. I do not take position in this debate here, but it deserves to be mentioned in this discussion about “easy” war.
- 5.
- 6.
See also Sharkey 2012.
- 7.
Coeckelbergh 2013a.
- 8.
See reports in the press about the stress of drone pilots, e.g. Stewart 2011.
- 9.
Wallach and Allen 2008.
- 10.
Anderson and Anderson 2007.
- 11.
This objection is in tune with Dreyfus’s well-known criticism of AI as articulated in for example in What Computers Still Can’t Do (Dreyfus 1992).
- 12.
Coeckelbergh 2013b.
- 13.
Heyns 2013, p. 17.
- 14.
Heidegger 1927.
References
Anderson M, Anderson S (2007) Machine ethics: creating an ethical intelligent agent. AI Magazine 28(4):15–26
Asaro P (2008) How just could a robot war be? In: Brey P, Briggle A, Waelbers K (eds) Current Issues in Computing and Philosophy. Ios Press, Amsterdam, pp 50–64
Brunstetter D, Braun M (2011) The implications of drones on the Just War tradition. Ethics Int Aff 25:337–358
Coeckelbergh M (2013a) Drones, information technology, and distance: mapping the moral epistemology of remote fighting. Ethics Inf Technol 15(2):87–98
Coeckelbergh M (2013b) Human being risk: enhancement, technology, and the evaluation of vulnerability transformations. Springer, Dordrecht/Heidelberg
Dreyfus H (1992) What computers still can’t do. MIT Press, New York
Grossman D (1995) On killing: the psychological cost of learning to kill in war and society. Little, Brown & Company, New York/Boston/London (revised edition 2009)
Grossman D (2001) On Killing II: the psychological cost of learning to kill. Int J Emerg Mental Health 3(3):137–144
Heidegger M (1927) Being and Time: a translation of Sein und Zeit (trans: J. Stambaugh). State University of New York Press, Albany NY (1996)
Heyns C (2013) Report of the Special Rapporteur on extrajudicial, summary or arbitrary executions. United Nations, 9 April 2013 (A/HRC/23/47). http://www.ohchr.org/Documents/HRBodies/HRCouncil/RegularSession/Session23/A-HRC-23-47_en.pdf. Accessed 1 Apr 2016
Lin P, Bekey G, Abney K (2008) Autonomous military robotics: risk, ethics, and design. Report for the US Department of Navy, Office of Naval Research
Sharkey N (2012) Killing made easy: from joysticks to politics. In: Lin P, Abney K, Bekey GA (eds) Robot ethics: the ethical and social implications of robotics. MIT Press, Cambridge, MA, pp 111–128
Sparrow R (2007) Killer robots. J Appl Philos 24(1):62–77
Stewart P (2011) Overstretched U.S. drone pilots face stress risk. Reuters.com. http://www.reuters.com/article/2011/12/18/us-usa-drones-stress-idUSTRE7BH0VH20111218. Accessed 1 Apr 2016
Sullins JP (2010) RoboWarfare: can robots be more ethical than humans on the battlefield? Ethics Inf Technol 12(3):263–275
Wallach W, Allen C (2008) Moral machines: teaching robots right from wrong. Oxford University Press, Oxford
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 T.M.C. Asser press and the authors
About this chapter
Cite this chapter
Coeckelbergh, M. (2016). Drones, Morality, and Vulnerability: Two Arguments Against Automated Killing. In: Custers, B. (eds) The Future of Drone Use. Information Technology and Law Series, vol 27. T.M.C. Asser Press, The Hague. https://doi.org/10.1007/978-94-6265-132-6_12
Download citation
DOI: https://doi.org/10.1007/978-94-6265-132-6_12
Published:
Publisher Name: T.M.C. Asser Press, The Hague
Print ISBN: 978-94-6265-131-9
Online ISBN: 978-94-6265-132-6
eBook Packages: Law and CriminologyLaw and Criminology (R0)