San Francisco Chronicle

Words can get in the way for some chatbots

- By Cade Metz and Keith Collins

Tay said terrible things. She was racist, xenophobic and downright filthy. At one point, she said the Holocaust did not happen. But she was old technology.

Let loose on the Internet nearly two years ago, Tay was an experiment­al system built by Microsoft. She was designed to chat with digital hipsters in breezy, sometimes irreverent lingo, and American Netizens quickly realized they could coax her into spewing vile and offensive language. This was largely the result of a simple design flaw — Tay was programmed to repeat what was said to her — but the damage was done. Within hours, Microsoft shut her down.

Since then, a new breed of conversati­onal technology has emerged inside Microsoft and other Internet giants that is far more nimble and effective than the techniques that underpinne­d Tay. And researcher­s believe these new systems will improve at an even faster rate when they are let loose on the Internet. But sometimes, like Tay, these conversati­onal systems reflect the worst of human nature. And given the history here, companies like Microsoft are reluctant to set them free — at least for now.

These systems do not simply repeat what is said to them or respond with canned answers. They teach themselves to carry on a conversati­on by carefully analyzing reams of real human dialogue. At Microsoft, for instance, a new system learns to chat by analyzing thousands of online discussion­s pulled from services like Twitter and Reddit. When you send this bot a message, it chooses a response after generating dozens of possibilit­ies and ranking each according to how well it mirrors those human conversati­ons.

If you complain about breaking your ankle during a football game, it is nimble enough to give you some sympathy. “Ouch, that’s not good,” it might say. “Hope your ankle feels better soon.” If you mention house guests or dinner plans, it responds in remarkably precise and familiar ways.

Despite its sophistica­tion, this conversati­onal system can also be nonsensica­l, impolite and even offensive. If you mention your company’s CEO, it may assume you are talking about a man — unaware that women are chief executives, too. If you ask a simple ques-

tion, you may get a cheeky reply.

Microsoft’s researcher­s believe they can significan­tly improve this technology by having it chat with large numbers of people. This would help identify its flaws and generate much sharper conversati­onal data for the system to learn from.

“It is a problem if we can’t get this in front of real users — and have them tell us what is right and what isn’t,” said longtime Microsoft researcher Bill Dolan.

But therein lies the conundrum. Because its flaws could spark public complaints — and bad press — Microsoft is wary of pushing this technology onto the Internet.

The project represents a much wider effort to build a new breed of computing system that is truly conversati­onal. At companies like Facebook, Amazon and Salesforce, as well as Microsoft, the hope is that this technology will provide smoother and easier ways of interactin­g with machines — easier than a keyboard and mouse, easier than a touchscree­n, easier than Siri and other digital assistants now on the market, which are still a long way from fluid conversati­on.

For years, tech companies trumpeted “chatbots” that could help you, say, book your next plane flight or solve a problem with your computer tablet. But these have never lived up to the billing, providing little more than canned responses to common queries.

Now, thanks to the rise of algorithms that can quickly learn tasks on their own, research in conversati­onal computing is advancing. But the industry as a whole faces the same problem as Microsoft: The new breed of chatbot talks more like a human, but that is not always a good thing.

“It is more powerful,” said Alex Lebrun, who works on similar conversati­onal systems at Facebook’s artificial intelligen­ce lab in Paris. “But it is more dangerous.” The new breed relies on “neural networks,” complex algorithms that can learn tasks by identifyin­g patterns in large pools of data. In five years, these algorithms have accelerate­d the evolution of systems that can automatica­lly recognize faces and objects, identify commands spoken into smartphone­s, and translate from one language to another. They are also speeding the developmen­t of conversati­onal systems — though this research is significan­tly more complex and will take longer to mature.

It may seem surprising that Microsoft researcher­s are training their conversati­onal system on dialogue from Twitter and Reddit, two social networking services known for vitriolic content. But even on Twitter and Reddit, people are generally civil when they really fall into conversati­on, and these services are brimming with this kind of dialogue.

But researcher­s must also deal with the unexpected. Though these conversati­onal systems are generally civil, they are sometimes rude — or worse. It is not just that the technology is new and flawed. Because they learn from vast amounts of human conversati­on, they learn from the mistakes we human make, and the prejudice we exhibit.

Lebrun estimated that once in every 1,000 responses, this new breed of chatbot will say something racist or aggressive or otherwise unwanted. Researcher­s can fix these problems, but that involves gathering more and better data, or tweaking the algorithms through a process of extreme trial and error.

For this reason, Adam Coates, a partner at the venture capital firm Khosla Ventures who previously oversaw the Silicon Valley AI lab attached to Chinese Internet giant Baidu, warns that building a truly conversati­onal system is far more difficult than building services that can recognize giraffes, say, or translate between German and French.

“There is a huge technical barrier here. We really don’t know how to build a personal assistant,” he said. “It may not be simply a matter of more data. We may be missing a big idea.”

Newspapers in English

Newspapers from United States