# References

Agre, Philip E. 1997. “Toward a Critical Technical Practice: Lessons Learned in Trying to Reform AI.” In *Social Science, Technical Systems, and Cooperative Work: Beyond the Great Divide*, 131–57. New York, London: Psychology Press.

———. 1998a. “Beyond the Mirror World: Privacy and the Representational Practices of Computing.” In *Technology and Privacy: The New Landscape*, edited by Philip E. Agre and Marc Rotenberg, 29–61. Cambridge, Mass: MIT Press.

———. 1998b. “Introduction.” In *Technology and Privacy: The New Landscape*, edited by Philip E. Agre and Marc Rotenberg, 1–28. Cambridge, Mass: MIT Press.

Agre, Philip E., and Marc Rotenberg, eds. 1998. *Technology and Privacy: The New Landscape*. Cambridge, Mass: MIT Press.

Bender, Emily M., Timnit Gebru, Angelina McMillan-Major, and Shmargaret Shmitchell. 2021. “On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜.” In *Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency*, 610–23. Virtual Event Canada: ACM. <https://doi.org/10.1145/3442188.3445922>.

Bijker, Wiebe E. 1995. *Of Bicycles, Bakelites, and Bulbs: Toward a Theory of Sociotechnical Change*. Inside Technology. Cambridge, Mass: MIT Press.

Bijker, Wiebe E., Thomas Parke Hughes, and Trevor J. Pinch, eds. 1987. *The Social Construction of Technological Systems: New Directions in the Sociology and History of Technology*. Cambridge, MA; London: MIT Press.

Born, Rainer, ed. 1987. *Artificial Intelligence: The Case Against*. New York: St. Martin’s Press.

———, ed. 2018. *Artificial Intelligence: The Case Against*. New York: Routledge.

Born, Rainer, and Ilse Born-Lechleitner. 1987. “Introduction.” In *Artificial Intelligence: The Case Against*, edited by Rainer Born, i–xxxv. New York: St. Martin’s Press.

Brown, Tom B., Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, et al. 2020. “Language Models Are Few-Shot Learners.” *ArXiv:2005.14165 \[Cs]*, July. <http://arxiv.org/abs/2005.14165>.

Bryson, Joanna J. 2010. “Robots Should Be Slaves.” *Close Engagements with Artificial Companions: Key Social, Psychological, Ethical and Design Issues*, 63–74. <https://doi.org/10.1075/nlp.8.11bry>.

Bryson, Joanna J., and Philip P. Kime. 2011. “Just an Artifact: Why Machines Are Perceived as Moral Agents.” In *Twenty-Second International Joint Conference on Artificial Intelligence*.

“Buolamwini, Gebru and Raji Win AI Innovation Award for Research into Biometric Bias | Biometric Update.” n.d. Accessed January 25, 2023. <https://www.biometricupdate.com/201907/buolamwini-gebru-and-raji-win-ai-innovation-award-for-research-into-biometric-bias>.

Bur, Tatiana Claudine Danielle. 2016. “Mechanical Miracles: Automata in Ancient Greek Religion.” Thesis. <https://ses.library.usyd.edu.au/handle/2123/15398>.

Cave, Stephen, Claire Craig, Kanta Sarasvati Dihal, Sarah Dillon, Jessica Montgomery, Beth Singler, and Lindsay Taylor. 2018. *Portrayals and Perceptions of AI and Why They Matter*. London: The Royal Society. <https://royalsociety.org/-/media/policy/projects/ai-narratives/AI-narratives-workshop-findings.pdf>.

Cave, Stephen, and Kanta Dihal. 2020. “The Whiteness of AI.” *Philosophy & Technology* 33 (4): 685–703. <https://doi.org/10.1007/s13347-020-00415-6>.

Cave, Stephen, Kanta Sarasvati Monique Dihal, and Sarah Dillon, eds. 2020. *AI Narratives: A History of Imaginative Thinking about Intelligent Machines*. First edition. Oxford: Oxford University Press.

Chandler, Daniel. 2017. *Semiotics: The Basics*. Taylor & Francis.

Chesterman, Simon. 2021. *We, the Robots? Regulating Artificial Intelligence and the Limits of the Law*. Cambridge, United Kingdom ; New York, NY, USA: Cambridge University Press.

Dale, Robert. 2021. “GPT-3: What’s It Good for?” *Natural Language Engineering* 27 (1): 113–18. <https://doi.org/10.1017/S1351324920000601>.

Davison, Joe. 2018. “No, Machine Learning Is Not Just Glorified Statistics.” Medium. June 27, 2018. <https://towardsdatascience.com/no-machine-learning-is-not-just-glorified-statistics-26d3952234e3>.

DoD. 2012. “Autonomy in Weapon Systems. Department of Defense Directive Number 3000.09.” United States. Department of Defense. <https://www.hsdl.org/?abstract&did=726163>.

Elkins, Katherine, and Jon Chun. 2020. “Can GPT-3 Pass a Writer’s Turing Test.” *Journal of Cultural Analytics* 2371: 1–16.

Fragaki, Hélène. 2012. “Automates et statues merveilleuses dans l’Alexandrie antique.” *Journal des Savants* 1 (1): 29–67. <https://doi.org/10.3406/jds.2012.6293>.

Gardner, Nikolas. 2021. “Clausewitzian Friction and Autonomous Weapon Systems.” *Comparative Strategy* 40 (1): 86–98. <https://doi.org/10.1080/01495933.2021.1853442>.

Gell, Alfred. 1994. “The Technology of Enchantment and the Enchantment of Technology.” In *Anthropology, Art, and Aesthetics*, edited by Jeremy Coote. Clarendon Press.

Geraci, Robert M. 2008. “Apocalyptic AI: Religion and the Promise of Artificial Intelligence.” *Journal of the American Academy of Religion* 76 (1): 138–66. <https://doi.org/10.1093/jaarel/lfm101>.

Glickman, Moshe, and Tali Sharot. 2022. “Biased AI Systems Produce Biased Humans.” OSF Preprints. <https://doi.org/10.31219/osf.io/c4e7r>.

Goddard, Kate, Abdul Roudsari, and Jeremy C Wyatt. 2012. “Automation Bias: A Systematic Review of Frequency, Effect Mediators, and Mitigators.” *Journal of the American Medical Informatics Association* 19 (1): 121–27. <https://doi.org/10.1136/amiajnl-2011-000089>.

Goodman, C. P. 2003. “The Tacit Dimension.” *Polanyiana* 2 (1): 133–57.

GPT-3. 2020. “A Robot Wrote This Entire Article. Are You Scared yet, Human?” *The Guardian*, September 8, 2020. <http://www.theguardian.com/commentisfree/2020/sep/08/robot-wrote-this-article-gpt-3>.

Gray, Chris Hables. 1997. “Artificial Intelligence at War: An Analysis of the Aegis System in Combat.” In *Reinventing Technology, Rediscovering Community: Critical Explorations of Computing as a Social Practice*, edited by Philip E. Agre and Douglas Schuler, 127–42. London: Ablex Publishing Corporation.

Greene, Tristan. 2022. “DeepMind’s New Gato AI Makes Me Fear Humans Will Never Achieve AGI.” TNW | Neural. May 13, 2022. <https://thenextweb.com/news/deepminds-astounding-new-gato-ai-makes-fear-humans-will-never-achieve-agi>.

Harbers, Maaike, Marieke M. M. Peeters, and Mark A. Neerincx. 2017. “Perceived Autonomy of Robots: Effects of Appearance and Context.” In *A World with Robots: International Conference on Robot Ethics: ICRE 2015*, edited by Maria Isabel Aldinhas Ferreira, Joao Silva Sequeira, Mohammad Osman Tokhi, Endre E. Kadar, and Gurvinder Singh Virk, 19–33. Intelligent Systems, Control and Automation: Science and Engineering. Cham: Springer International Publishing. <https://doi.org/10.1007/978-3-319-46667-5_2>.

Horowitz, Michael C., Lauren Kahn, Julia Macdonald, and Jacquelyn Schneider. 2022. “COVID-19 and Public Support for Autonomous Technologies—Did the Pandemic Catalyze a World of Robots?” *PLOS ONE* 17 (9): e0273941. <https://doi.org/10.1371/journal.pone.0273941>.

Irvine, Martin. 2022. “Semiotics in Computing and Information Systems.” In *Bloomsbury Semiotics*, edited by Jamin Pelkey, 2:203–37. Bloomsbury Publishing.

Johansen, Sigrid Redse. 2018. “So Man Created Robot in His Own Image: The Anthropomorphism of Autonomous Weapon Systems and the Law of Armed Conflict.” *Oslo Law Review* 5 (2): 89–102. <https://doi.org/10.18261/issn.2387-3299-2018-02-03>.

Johnson, Deborah G., and Mario Verdicchio. 2017. “Reframing AI Discourse.” *Minds and Machines* 27 (4): 575–90. <https://doi.org/10.1007/s11023-017-9417-6>.

Johnson, James. 2022. “Delegating Strategic Decision-Making to Machines: Dr. Strangelove Redux?” *Journal of Strategic Studies* 45 (3): 439–77. <https://doi.org/10.1080/01402390.2020.1759038>.

Kang, Minsoo. 2011. *Sublime Dreams of Living Machines: The Automaton in the European Imagination*. Cambridge, Mass: Harvard University Press.

Katz, Yarden. 2020. *Artificial Whiteness: Politics and Ideology in Artificial Intelligence*. New York: Columbia University Press.

Kennedy, J. S. 1992. *The New Anthropomorphism*. Cambridge \[England] ; New York: Cambridge University Press.

Kline, Ronald. 2010. “Cybernetics, Automata Studies, and the Dartmouth Conference on Artificial Intelligence.” *IEEE Annals of the History of Computing* 33 (4): 5–16. <https://doi.org/10.1109/MAHC.2010.44>.

Korngiebel, Diane M., and Sean D. Mooney. 2021. “Considering the Possibilities and Pitfalls of Generative Pre-Trained Transformer 3 (GPT-3) in Healthcare Delivery.” *Npj Digital Medicine* 4 (1): 1–3. <https://doi.org/10.1038/s41746-021-00464-x>.

Krishnan, Armin. 2016. *Killer Robots: Legality and Ethicality of Autonomous Weapons*. London, England ; New York, New York: Routledge.

Leveson, Nancy G. 2017. “The Therac-25: 30 Years Later.” *Computer* 50 (11): 8–11. <https://doi.org/10.1109/MC.2017.4041349>.

Leveson, Nancy G., and C.S. Turner. 1993. “An Investigation of the Therac-25 Accidents.” *Computer* 26 (7): 18–41. <https://doi.org/10.1109/MC.1993.274940>.

Lovelace, Douglas C., ed. 2016. *Autonomous and Semiautonomous Weapons Systems*. Vol. 144. Terrorism: Commentary on Security Documents. New York: Oxford University Press.

Martin, C. Dianne. 1993. “The Myth of the Awesome Thinking Machine.” *Communications of the ACM* 36 (4): 120–33. <https://doi.org/10.1145/255950.153587>.

Mayor, Adrienne. 2018. *Gods and Robots: Myths, Machines, and Ancient Dreams of Technology*. Princeton: Princeton University Press.

McCarthy, John, Marvin L. Minsky, Nathaniel Rochester, and Claude E. Shannon. 1955. “A Proposal for the Dartmouth Summer Research Project on Artificial Intelligence.” <http://raysolomonoff.com/dartmouth/boxa/dart564props.pdf>.

———. 2006. “A Proposal for the Dartmouth Summer Research Project on Artificial Intelligence, August 31, 1955.” *AI Magazine* 27 (4): 12–14.

McGuffie, Kris, and Alex Newhouse. 2020. “The Radicalization Risks of GPT-3 and Advanced Neural Language Models.” *ArXiv Preprint ArXiv:2009.06807*.

Meunier, Jean-Guy. 1989. “Artificial Intelligence and Sign Theory.” *Semiotica* 77 (1–3): 43–64. <https://doi.org/10.1515/semi.1989.77.1-3.43>.

Mosco, Vincent. 2004. *The Digital Sublime: Myth, Power, and Cyberspace*. Cambridge, Mass: MIT Press.

Natale, Simone, and Andrea Ballatore. 2020. “Imagining the Thinking Machine: Technological Myths and the Rise of Artificial Intelligence.” *Convergence* 26 (1): 3–18. <https://doi.org/10.1177/1354856517715164>.

Payne, Kenneth. 2021. *I, Warbot: The Dawn of Artificially Intelligent Conflict*. London: Hurst & Company.

Penn, Jonathan. 2021. “Inventing Intelligence: On the History of Complex Information Processing and Artificial Intelligence in the United States in the Mid-Twentieth Century.” Thesis, University of Cambridge. <https://doi.org/10.17863/CAM.63087>.

Pinch, Trevor J., and Wiebe E. Bijker. 1984. “The Social Construction of Facts and Artefacts: Or How the Sociology of Science and the Sociology of Technology Might Benefit Each Other.” *Social Studies of Science* 14 (3): 399–441. <https://doi.org/10.1177/030631284014003004>.

Polanyi, Michael. 1966. *The Tacit Dimension*. Garden City, N.Y: Doubleday.

Polanyi, Michael, and Amartya Sen. 2009. *The Tacit Dimension*. Chicago ; London: University of Chicago Press.

Raji, Inioluwa Deborah, I. Elizabeth Kumar, Aaron Horowitz, and Andrew Selbst. 2022. “The Fallacy of AI Functionality.” In *2022 ACM Conference on Fairness, Accountability, and Transparency*, 959–72. Seoul Republic of Korea: Association for Computing Machinery. <https://doi.org/10.1145/3531146.3533158>.

Roberts, John Michael, and Colin Cremin. 2017. “Contested Meanings, Myths and Hyperimages of the Apocalypse: The Bakhtin Circle and the Politicisation of Catastrophism.” *Social Semiotics* 27 (2): 178–94. <https://doi.org/10.1080/10350330.2016.1182303>.

Salvatore, Sergio, Jaan Valsiner, and Giuseppe Alessandro Veltri. 2019. “The Theoretical and Methodological Framework. Semiotic Cultural Psychology, Symbolic Universes and Lines of Semiotic Forces.” In *Symbolic Universes in Time of (Post) Crisis*, 25–49. Springer.

Sayler, Kelley M. 2021. “Defense Primer: U.S. Policy on Lethal Autonomous Weapon Systems.” CRS In Focus, IF11150. Library of Congress. Congressional Research Service. <https://www.hsdl.org/?abstract&did=861619>.

Scharre, Paul. 2018. *Army of None: Autonomous Weapons and the Future of War*. First edition. New York ; London: W. W. Norton & Company.

Scharre, Paul, and Michael Horowitz. 2015. “An Introduction to Autonomy in Weapon Systems.” Project on Ethical Autonomy. Center for a New American Security. <http://www.files.ethz.ch/isn/188865/Ethical%20Autonomy%20Working%20Paper_021015_v02.pdf>.

Shane, Janelle. 2019. *You Look like a Thing and I Love You: How Artificial Intelligence Works and Why It’s Making the World a Weirder Place*. First edition. New York: Voracious/Little, Brown and Company.

Sharkey, Amanda, and Noel Sharkey. 2020. “We Need to Talk about Deception in Social Robotics!” *Ethics and Information Technology*, 1–8.

———. 2021. “Sunlight Glinting on Clouds: Deception and Autonomous Weapons Systems.” In *Counter-Terrorism, Ethics and Technology*, 35–47. Springer, Cham.

Sharkey, Noel. 2014. “Killing Made Easy: From Joysticks to Politics.” In *Robot Ethics: The Ethical and Social Implications of Robotics*, 111–28. Cambridge, MA; London: MIT Press.

Sharkey, Noel E. 2012. “The Evitability of Autonomous Robot Warfare.” *International Review of the Red Cross* 94 (886): 787–800. <https://doi.org/10.1017/S1816383112000732>.

Singler, Beth. 2020a. “The AI Creation Meme: A Case Study of the New Visibility of Religion in Artificial Intelligence Discourse.” *Religions* 11 (5): 253. <https://doi.org/10.3390/rel11050253>.

———. 2020b. “‘Blessed by the Algorithm’: Theistic Conceptions of Artificial Intelligence in Online Discourse.” *AI & SOCIETY* 35 (4): 945–55. <https://doi.org/10.1007/s00146-020-00968-2>.

Sismondo, Sergio. 2010. *An Introduction to Science and Technology Studies*. 2nd ed. Chichester, West Sussex, U.K. ; Malden, MA: Wiley-Blackwell.

Smith, Gary. 2018. *The AI Delusion*. First edition. Oxford: Oxford University Press.

Stahl, Bernd Carsten, Tonii Leach, Oluyinka Oyeniji, and George Ogoh. 2023. “AI Policy as a Response to AI Ethics? Addressing Ethical Issues in the Development of AI Policies in North Africa.” In *Responsible AI in Africa: Challenges and Opportunities*, edited by Damian Okaibedi Eke, Kutoma Wakunuma, and Simisola Akintoye, 141–67. Social and Cultural Studies of Robots and AI. Cham: Springer International Publishing. <https://doi.org/10.1007/978-3-031-08215-3_7>.

Taylor, Trevor. 2019. “Artificial Intelligence in Defence.” *The RUSI Journal* 164 (5–6): 72–81. <https://doi.org/10.1080/03071847.2019.1694229>.

Tiku, Nitasha. 2021. “Google Fired Its Star AI Researcher One Year Ago. Now She’s Launching Her Own Institute.” *Washington Post*, December 2, 2021. <https://www.washingtonpost.com/technology/2021/12/02/timnit-gebru-dair/>.

Tresch, John. 2011. “The Machine Awakens: The Science and Politics of the Fantastic Automaton.” *French Historical Studies* 34 (1): 87–123. <https://doi.org/10.1215/00161071-2010-024>.

Truitt, Elly Rachel. 2004. “‘Trei Poete, Sages Dotors, Qui Mout Sorent Di Nigromance’: Knowledge and Automata in Twelfth-Century French Literature.” *Configurations* 12 (2): 167–93. <https://doi.org/10.1353/con.2006.0006>.

———. 2015. *Medieval Robots*. Philadelphia: University of Pennsylvania Press.

Uexküll, Jakob von. 2001. “The New Concept of Umwelt: A Link between Science and the Humanities.” *Semiotica* 2001 (134). <https://doi.org/10.1515/semi.2001.018>.

Vannuccini, Simone, and Ekaterina Prytkova. 2020. “Artificial Intelligence’s New Clothes? From General Purpose Technology to Large Technical System.” SSRN Scholarly Paper ID 3704011. Rochester, NY: Social Science Research Network. <https://doi.org/10.2139/ssrn.3704011>.

Viidalepp, Auli. 2020. “Representations of Robots in Science Fiction Film Narratives as Signifiers of Human Identity.” *Információs Társadalom* 20 (4): 19. <https://doi.org/10.22503/inftars.XX.2020.4.2>.

———. 2022. “Sociocommunicative Functions of a Generative Text: The Case of GPT-3.” *Lexia. Rivista Di Semiotica* 39–40: 177–92. <https://doi.org/10.53136/979122180426310>.

Visser, Ewart J. de, Samuel S. Monfort, Ryan McKendrick, Melissa A. B. Smith, Patrick E. McKnight, Frank Krueger, and Raja Parasuraman. 2016. “Almost Human: Anthropomorphism Increases Trust Resilience in Cognitive Agents.” *Journal of Experimental Psychology: Applied* 22 (3): 331–49. <https://doi.org/10.1037/xap0000092>.

Waal, Frans B. M. de. 1999. “Anthropomorphism and Anthropodenial: Consistency in Our Thinking about Humans and Other Animals.” *Philosophical Topics* 27 (1): 255–80. <https://doi.org/10.5840/philtopics199927122>.

Winner, Langdon. 1978. *Autonomous Technology: Technics-out-of-Control as a Theme in Political Thought*. 9. printing. Cambridge, Mass.: MIT Pr.

Zhai, Yujia, Jiaqi Yan, Hezhao Zhang, and Wei Lu. 2020. “Tracing the Evolution of AI: Conceptualization of Artificial Intelligence in Mass Media Discourse.” *Information Discovery and Delivery* 48 (3): 137–49. <https://doi.org/10.1108/IDD-01-2020-0007>.
