Journal Article


Modelling a socialised chatbot using trust development in children: Lessons learnt from Tay

Abstract

In 2016 Microsoft released Tay.ai to the Twittersphere, a conversational chatbot that was intended to act like a millennial girl. However, they ended up taking Tay's account down in less than 24 h because Tay had learnt to tweet racist and sexist statements from its online interactions. Taking inspiration from the theory of morality as cooperation, and the place of trust in the developmental psychology of socialisation, we offer a multidisciplinary and pragmatic approach to build on the lessons learnt from Tay's experiences, to create a chatbot that is more selective in its learning, and thus resistant to becoming immoral the way Tay did.

Attached files

Authors

Bridge, Oliver
Raper, Rebecca
Strong, Nicola
Nugent, Selin E.

Oxford Brookes departments

School of Education
Faculty of Technology, Design and Environment

Dates

Year of publication: 2021
Date of RADAR deposit: 2021-10-04


Creative Commons License This work is licensed under a Creative Commons Attribution 4.0 International License


Related resources

This RADAR resource is Identical to Modelling a socialised chatbot using trust development in children: Lessons learnt from Tay

Details

  • Owner: Joseph Ripp
  • Collection: Outputs
  • Version: 1 (show all)
  • Status: Live
  • Views (since Sept 2022): 319