So far, no other celebrities have complained. But they decided to avoid a legal battle and dubbed the replacement chatbot "Zo" instead. Gizmodo reports that Microsoft's copyright lawyers weren't quite convinced by the purported connection between Tay the chatbot and Taylor Swift the pop megastar. As reported by Gizmodo, Swift's lawyers apparently got in touch with Microsoft to complain that Tay's name created a "false and misleading association between the popular singer and our chatbot." RebrandingĪccording to the book, the reason that Microsoft's next chatbot - which was barred from discussing topics like politics and race - got a name change was to avoid going to court. Remember the 2016 internet horror story where Microsoft created a Twitter bot, named Tay, and had to shut it down within a single day after internet trolls turned it into a hate-spewing neo-Nazi?Īccording to Microsoft President Brad Smith's new book, pop star Taylor Swift took issue with the debacle. Within 24 hours Tay had been deactivated so the. Microsoft launches and then kills Twitter AI chatbot Tay Soon after its launch, the bot ‘Tay’ was fired after it started tweeting abusively, one of the tweet said this Hitler was right I. "As a result, we have taken Tay offline and are making adjustments.Apparently "Tay" was too close to Swift's name for comfort. Tay, aimed at 18-24-year-olds on social media, was targeted by a 'coordinated attack by a subset of people' after being launched earlier this week. "Unfortunately, within the first 24 hours of coming online, we became aware of a coordinated effort by some users to abuse Tay's commenting skills to have Tay respond in inappropriate ways," the spokeswoman said. The culture seems to have asserted itself. Tay, a Microsoft spokeswoman told me, is "as much a social and cultural experiment, as it is technical." The tech company introduced 'Tay' this week a bot that. This has to have been taxing for the people behind the scenes, too. Microsofts new AI chatbot went off the rails Wednesday, posting a deluge of incredibly racist messages in response to questions. In the 24 hours it took Microsoft to shut her down, Tay had abused President Obama, suggested Hitler was right, called feminism a disease and delivered a stream of online hate. Indeed, late Wednesday, Tay went on hiatus, as she tweeted: "c u soon humans need sleep now so many conversations today thx." For sure - she had already emitted more than 96,000 tweets in a very few hours. In other words, once Tay heard I love Hitler more than I love puppies, she started to say she loves Hitler more than puppies because that is how she thinks people talk to each other. She behaved like such a naughty robot that Daddy Microsoft appears to have removed these tweets. Tay is converting words into probabilities, and probabilities back into words. The chatbot was shut down within 24 hours of her introduction to the world after offending the masses. It seems that many of these responses were elicited by humans asking Tay to repeat what they'd written.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |