Facebook claims its new chatbot beats Googles as the best in the world – MIT Technology Review

Posted: May 6, 2020 at 6:53 am

Blenders ability comes from the immense scale of its training data. It was first trained on 1.5 billion publicly available Reddit conversations, to give it a foundation for generating responses in a dialogue. It was then fine-tuned with additional data sets for each of three skills: conversations that contained some kind of emotion, to teach it empathy (if a user says I got a promotion, for example, it can say, Congratulations!); information-dense conversations with an expert, to teach it knowledge; and conversations between people with distinct personas, to teach it personality. The resultant model is 3.6 times larger than Googles chatbot Meena, which was announced in Januaryso big that it cant fit on a single device and must run across two computing chips instead.

FACEBOOK

At the time, Google proclaimed that Meena was the best chatbot in the world. In Facebooks own tests, however, 75% of human evaluators found Blender more engaging than Meena, and 67% found it to sound more like a human. The chatbot also fooled human evaluators 49% of the time into thinking that its conversation logs were more human than the conversation logs between real peoplemeaning there wasnt much of a qualitative difference between the two. Google hadnt responded to a request for comment by the time this story was due to be published.

Despite these impressive results, however, Blenders skills are still nowhere near those of a human. Thus far, the team has evaluated the chatbot only on short conversations with 14 turns. If it kept chatting longer, the researchers suspect, it would soon stop making sense. These models arent able to go super in-depth, says Emily Dinan, the other project leader. Theyre not able to remember conversational history beyond a few turns.

Blender also has a tendency to hallucinate knowledge, or make up factsa direct limitation of the deep-learning techniques used to build it. Its ultimately generating its sentences from statistical correlations rather than a database of knowledge. As a result, it can string together a detailed and coherent description of a famous celebrity, for example, but with completely false information. The team plans to experiment with integrating a knowledge database into the chatbots response generation.

FACEBOOK

Another major challenge with any open-ended chatbot system is to prevent it from saying toxic or biased things. Because such systems are ultimately trained on social media, they can end up regurgitating the vitriol of the internet. (This infamously happened to Microsofts chatbot Tay in 2016.) The team tried to address this issue by asking crowdworkers to filter out harmful language from the three data sets that it used for fine-tuning, but it did not do the same for the Reddit data set because of its size. (Anyone who has spent much time on Reddit will know why that could be problematic.)

The team hopes to experiment with better safety mechanisms, including a toxic-language classifier that could double-check the chatbots response. The researchers admit, however, that this approach wont be comprehensive. Sometimes a sentence like Yes, thats great can seem fine, but within a sensitive context, such as in response to a racist comment, it can take on harmful meanings.

In the long term the Facebook AI team is also interested in developing more sophisticated conversational agents that can respond to visual cues as well as just words. One project is developing a system called Image Chat, for example, that can converse sensibly and with personality about the photos a user might send.

Read more from the original source:

Facebook claims its new chatbot beats Googles as the best in the world - MIT Technology Review