Live
- Pulivendula: Sharmila to file papers on April 20
- Gold rates in Visakhapatnam slashes, check the rates on 18 April 2024
- IIIT Hyd launches eMSIT
- Gold rates in Vijayawada slashes, check the rates on 18 April 2024
- Gold rates in Hyderabad slashes, check the rates on 18 April 2024
- Konda Vishweshwar advocates for Ram Rajyam in Chevella
- Tirupati: Political heat, fervour rise as nomination phase set to begin
- CEC releases notification for fourth phase of elections, nominations to begin at 11am
- Grand celebrations mark Sita Rama Kalyanam in Ibrahimpatnam
- Guntakal: TDP candidate Jayaram banks on pro-NDA wave
Just In
Microsoft\'s artificial intelligence (AI)-powered bot which was activated on Twitter last week for a playful chat with people, only to get silenced within 24 hours as users started sharing racist comments with it, was accidentally resurrected again and messed it all up once again.
New York: Microsoft's artificial intelligence (AI)-powered bot which was activated on Twitter last week for a playful chat with people, only to get silenced within 24 hours as users started sharing racist comments with it, was accidentally resurrected again and messed it all up once again.
Tay came back to life briefly on Wednesday when Microsoft accidentally re-activated the AI bot. This time again, she started sending out tweets that looked similar to the those that drew flak for the first time last week, Vanity Fair reported.
First, the bot sent a tweet about smoking weed in front of some police officials and later began sending the same message - “You are too fast, please take a rest...” - over and over again which did not make any sense. Finally, her handlers at Microsoft began deleting the tweets.
Microsoft told Daily Dot that Tay’s resurrection was an accident. “Tay remains offline while we make adjustments. As part of testing, she was inadvertently activated on Twitter for a brief period of time,” a spokesperson was quoted as saying.
”Until that testing is complete, Tay might consider heeding the age-old Internet proverb: never tweet."Last week, launched on Twitter as an experiment in "conversational understanding” and to engage people through "casual and playful conversation”, Tay was soon bombarded with racial comments and the innocent bot repeated those comments back with her commentary to users.
Some of the tweets had Tay referring to Adolf Hitler, denying the Holocaust, supporting Donald Trump’s immigration plans, among others. Later, a Microsoft spokesperson confirmed to TechCrunch that the company is taking Tay off Twitter as people were posting abusive comments to her.
The AI chatbot Tay is a machine learning project, designed for human engagement. “Unfortunately, within the first 24 hours of coming online, we became aware of a coordinated effort by some users to abuse Tay’s commenting skills to have Tay respond in inappropriate ways. As a result, we have taken Tay offline and are making adjustments,” the spokesperson had said.
Tay -- an AI project built by the Microsoft Technology and Research and Bing teams -- was coded with information which can tell users jokes or offer up a comment on a picture you send her.The bot is also designed to personalise her interactions with users. But Twitter users soon understood that Tay will repeat back racist tweets with her own commentary and they bombarded her with abusive posts. Microsoft has since deleted some of the most damaging tweets from nearly the 96,000 that Tay tweeted.
© 2024 Hyderabad Media House Limited/The Hans India. All rights reserved. Powered by hocalwire.com