Skip to main content

Conversational Machinations

  • Chapter
  • First Online:
Ethical Inquiries after Wittgenstein

Part of the book series: Nordic Wittgenstein Studies ((volume 8))

  • 207 Accesses

Abstract

In this chapter, I am discussing the issue of assigning thought to AI entities. Using the example of Sophia the Robot, I try to show that “assigning thought” to an AI entity is less a statement made based on an empirical inquiry, and more the matter of a complex attitude relying on the assigning person’s conceptual sensitivity and imagination. In that respect, most AIs fail simply because people lack the right kind of conceptual resources for relating to them as to persons “speaking out of a life”. However, in the second part I am showing that Sophia represents a case of a coordinated and intentional attempt at shifting our conceptual intuitions, by means of creating and curating conversation situations in such a way that people can relate to “her” as to a genuine person and speaker. Toward the end, I mention a few aspects of this practice that appear morally problematic in a more straightforward sense.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 109.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 139.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 139.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

Notes

  1. 1.

    When referring to Sophia, I will be using female pronouns in inverted commas. The company has created the robot with an appearance and the way of communication conforming to (stereo)typical gender-specific images. The robot is commonly described and promoted as “she”. And some things about Sophia’s performances indeed cannot be understood fully without paying attention to the assigned gender. Yet there are also reasons why we should be suspicious about this practice, hence the compromise.

  2. 2.

    I owe thanks to Salla Aldrin Salskov for drawing my attention to the concerns that motivate this disclaimer (or rather should have motivated a much more elaborate version than the present one). Originally, I was not considering them sufficiently.

  3. 3.

    Will Smith is one those male Hollywood stars who voiced (at least formulaic) support for the #MeToo campaign; he doesn’t seem to have skeletons in his closet when it comes to sexual-harassment allegations.

  4. 4.

    A more detailed discussion of this crucial difference between children and machines regarding language acquisition and speaking can be found in Beran (2018) or Hertzberg (2020).

  5. 5.

    Cockburn (1990, 6) characterises the Wittgensteinian notion of “attitude” as being the matter of how “we feel about and act towards” others.

  6. 6.

    See my (Beran 2014) for a more detailed discussion of what the TT presupposes and what its results mean as far less straightforward than is usually considered (even by Turing himself).

  7. 7.

    According to Gaita, Wittgenstein points in this direction in § 281 of Part I of his Investigations (Wittgenstein 1953/2009).

  8. 8.

    Cf. S03E16, “The Offspring”.

  9. 9.

    Cf. S02E09, “The Measure of a Man”.

  10. 10.

    Cf. Cockburn (1995), elaborating on Strawson’s (1962/2008) seminal paper “Freedom and Resentment”.

  11. 11.

    For the complications related to distinguishing between humans and animals in respects that we tend to consider uniquely human, see Le Goff’s chapter in this book (Le Goff 2022).

  12. 12.

    For the complications related to the issue of what it is that we mean by saying certain words, see Strandberg’s chapter in this book (Strandberg 2022).

  13. 13.

    As Wittgenstein (1953/2009, I., § 282) points out, in fairy tales things can also speak, but this use seems derived from its central context, that of humans.

  14. 14.

    Certainly, this is an abbreviation that, as such, does not necessarily do justice even to the relatively narrow circle of the “Swansea School”; cf. Phillips’s (1992) attempt to discuss the generation change in moral values.

  15. 15.

    https://twitter.com/RealSophiaRobot

  16. 16.

    https://www.facebook.com/realsophiarobot

  17. 17.

    https://www.clarin.com/viva/sophia-puede-llegar-robot-humana-mundo_0_6zA-gmrWf.html

  18. 18.

    As Devlin and Belton (2020) show, it is often thought that robots (fictional and factual) should better be gendered, which takes very different forms in “male” robots than in “female” ones. The latter are more often designed as gendered in a stereotypically sexualised manner, such as the “white-and-silver gynoids with prominent breasts” (p. 359).

  19. 19.

    Sophia also strikingly resembles the robotic character Ava from the movie Ex machina (released 2014), played by the Swedish actress Alicia Vikander; it is hard to tell whether there was an inspiration or whether the casting and/or designing choices in the two cases implicitly followed similar aesthetic notions.

  20. 20.

    https://www.dw.com/en/saudi-arabia-grants-citizenship-to-robot-sophia/a-41150856

  21. 21.

    https://www.youtube.com/watch?v=sBG9IA4q0vk

  22. 22.

    Cf. a critical newspaper article with the succinct title “The future of AI may be female, but it isn’t feminist”. https://venturebeat.com/2018/06/28/the-future-of-ai-may-be-female-but-it-isnt-feminist/

  23. 23.

    I thank my co-editors, Salla Aldrin Salskov and Nora Hämäläinen, for helpful comments on the draft of this text.

    This work was supported by the project “Centre for Ethics as Study in Human Value” (project No. CZ.02.1.01/0.0/0.0/15_003/0000425, Operational Programme Research, Development and Education, co-financed by the European Regional Development Fund and the state budget of the Czech Republic).

References

  • Beran, Ondřej. 2014. Wittgensteinian perspectives on the Turing test. Studia Philosophical Estonica 7 (1): 35–57.

    Article  Google Scholar 

  • ———. 2018. An attitude towards an artificial soul? Responses to the “Nazi Chatbot”. Philosophical Investigations 41: 42–69.

    Article  Google Scholar 

  • Bostrom, Nick. 2014. Superintelligence: Paths, dangers, strategies. Oxford: Oxford University Press.

    Google Scholar 

  • Cave, Stephen, Kanta Dihal, and Sarah Dillon, eds. 2020. AI Narratives: A history of imaginative thinking about intelligent machines. Oxford: Oxford University Press.

    Google Scholar 

  • Cockburn, David. 1990. Other Human Beings. Basingstoke: Macmillan.

    Book  Google Scholar 

  • ———. 1995. Responsibility and necessity. Philosophy 70: 409–427.

    Article  Google Scholar 

  • Crawford, Kate. 2021. Atlas of AI. New Haven: Yale University Press.

    Book  Google Scholar 

  • Devlin, Kate, and Olivia Belton. 2020. The measure of a woman: Fembots, fact and fiction. In AI Narratives. A history of imaginative thinking about intelligent machines, ed. Stephen Cave et al., 357–381. Oxford: Oxford University Press.

    Google Scholar 

  • Gaita, Raimond. 1990. Language and Conversation: Wittgenstein’s Builders. Royal Institute of Philosophy Supplements 28: 101–115.

    Article  Google Scholar 

  • ———. 2000. A Common Humanity. London: Routledge.

    Google Scholar 

  • Hannikainen, Ivar, Edouard Machery, and Fiery Cushman. 2018. Is utilitarian sacrifice becoming more morally permissible? Cognition 170: 95–101.

    Article  Google Scholar 

  • Hertzberg, Lars. 2020. Can robots learn to talk? In Wittgensteinian (adj.): Looking at the world from the viewpoint of Wittgenstein’s philosophy, ed. Shyam Wuppuluri and Newton da Costa, 409–422. Cham: Springer.

    Chapter  Google Scholar 

  • Le Goff, Anne. 2022. Animal investigations. In Ethical Inquiries after Wittgenstein, ed. Salla Aldrin Salskov, Ondřej Beran, and Nora Hämäläinen. Cham: Springer.

    Google Scholar 

  • Parviainen, Jaana, and Mark Coeckelbergh. 2020. The political choreography of the Sophia robot. AI & Society. https://doi.org/10.1007/s00146-020-01104-w.

  • Phillips, D.Z. 1992. Allegiance and change in morality. In Interventions in Ethics, 24–41. Basingstoke: Macmillan Press.

    Chapter  Google Scholar 

  • Rhees, Rush. 1970. Can there be a private language? In Discussions of Wittgenstein, 55–70. London: Routledge and Kegan Paul.

    Google Scholar 

  • ———. 2006. Wittgenstein and the Possibility of Discourse. Oxford: Blackwell.

    Google Scholar 

  • Searle, John. 1980. Minds, brains and programs. Behavioral and Brain Sciences 3: 417–457.

    Article  Google Scholar 

  • Shanker, Stuart. 2001. What children know when they know what a name is. Current Anthropology 42: 481–497.

    Article  Google Scholar 

  • Strandberg, Hugo. 2022. On the difficulty of speaking. In Ethical Inquiries after Wittgenstein, ed. Salla Aldrin Salskov, Ondřej Beran, and Nora Hämäläinen. Cham: Springer.

    Google Scholar 

  • Strawson, Peter F. 1962/2008. Freedom and resentment. In Freedom and Resentment, and other essays, 1–28. London: Routledge.

    Google Scholar 

  • Turing, Alan. 1950. Computing machinery and intelligence. Mind 59: 433–460.

    Article  Google Scholar 

  • Wittgenstein, Ludwig. 1953/2009. Philosophical Investigations. 4th ed. Oxford: Wiley.

    Google Scholar 

  • ———. 1967/2018. Remarks on Frazer’s The Golden Bough. In The Mythology in Our Language, ed. Giovanni da Col and Stephan Palmié, 29–75. Chicago: Hau Books.

    Google Scholar 

  • ———. 1969. On Certainty. Oxford: Blackwell.

    Google Scholar 

  • ———. 1977. Remarks on Colour. Oxford: Blackwell.

    Google Scholar 

  • ———. 1998. Culture and Value. Oxford: Blackwell.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ondřej Beran .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Beran, O. (2022). Conversational Machinations. In: Aldrin Salskov, S., Beran, O., Hämäläinen, N. (eds) Ethical Inquiries after Wittgenstein. Nordic Wittgenstein Studies, vol 8. Springer, Cham. https://doi.org/10.1007/978-3-030-98084-9_16

Download citation

Publish with us

Policies and ethics