Microsoft just can't seem to get AI chatbots right

Gandalf_The_Grey

Level 76
Thread author
Verified
Honorary Member
Top Poster
Content Creator
Well-known
Apr 24, 2016
6,596
These past few weeks, news items related to artificial intelligence (AI) have dominated the headlines. This was mostly spurred by Microsoft-backed OpenAI's chatbot as well as the integrated chatbot in the new Bing. Although many people have been impressed by the capabilities demonstrated by generative AI - with millions signing up for a limited preview of the new Bing -, the past few days have surfaced notable problems with the current implementation too.

Users of the new Bing have managed to make the integrated chatbot say some truly unhinged stuff, including claims that it spied on its developers through their PC's webcams and even fell in love with some of them. The AI expressed the desire to become human as well, which is something we have seen other chatbots do in the past too. The AI also displayed factual mistakes while answering objective questions. All of this became problematic to the extent that Microsoft had to enforce hard limits on the length and nature of conversations that you have with the AI, in an effort to reduce rampancy.

Of course, none of this means that Microsoft's Bing has become sentient. Weird responses by the chatbot are just the by-product of a large language model scraping information from all over the internet (including forums with user-generated content) to identify patterns in conversations and generate a response accordingly. However, Microsoft's latest experiment does show that well-behaved AI chatbots continue to be a challenge for the company, and maybe the tech's pioneers as a whole.

Way back in 2016 - when Cortana was still alive and well - Microsoft launched a chatbot called "Tay" on Twitter. It was similar to the new Bing AI in nature, in the sense that you could engage in free-flowing conversations with it, even via Direct Messages.

However, within 16 hours of launch, Microsoft took the bot offline due to the AI model making racist and sexist remarks. The company was forced to issue an apology, with Microsoft's Corporate Vice President at Microsoft Research, Peter Lee, claiming that the unwanted results were due to "a coordinated attack by a subset of people [who] exploited a vulnerability" in the chatbot. In hindsight, this is not surprising at all considering that the AI had been unleashed to practically everyone on the internet and was learning on-the-go.

A successor named "Zo" was launched across multiple social media platforms in late 2016, but it eventually suffered the same fate as Tay in 2019, following a tirade of controversial religious remarks.

Despite these failures, Microsoft has had some success in this area too. It has another older AI chatbot project called "Xiaoice" that's geared more towards Asian markets such as Japan and China. Although Microsoft later spun Xiaoice off into a separate company, the chatbot has had its share of controversies too. The bot has made comments critical of the Chinese government in the past, which led to it being taken offline temporarily. And given its target market and commercial use-cases, it is much more restrictive and attempts to dodge conversations related to potentially sensitive topics, just like the new Bing.

It's clear that while Microsoft is making major headway in terms of what AI chatbots can do for you, it is still grappling with major challenges related to the generation of inappropriate responses, accuracy, and biasness. The recent hard limits imposed on its Bing chatbot indicate that free-flowing conversations with AI may still be some way off and maybe it is better to tailor your chatbot to specific use-cases rather than giving them free reign over what they can scrape, in real-time.
 

About us

  • MalwareTips is a community-driven platform providing the latest information and resources on malware and cyber threats. Our team of experienced professionals and passionate volunteers work to keep the internet safe and secure. We provide accurate, up-to-date information and strive to build a strong and supportive community dedicated to cybersecurity.

User Menu

Follow us

Follow us on Facebook or Twitter to know first about the latest cybersecurity incidents and malware threats.

Top