When rumors began swirling last month about the Chinese search giant Baidu working on a chatbot to rival OpenAI’s ChatGPT, it seemed like the perfect move. Baidu has invested heavily in artificial intelligence over the past decade and could harness the technology for its leading search engine, as Microsoft has done for Bing and Google says it will do too.
Yet when Baidu unveiled Ernie Bot, or 文心一言 “Wenxin Yiyan” in Chinese, in Beijing earlier this month, the news fell flat.
Robin Li, Baidu’s CEO, admitted halfway through the launch stream that demos of Ernie Bot answering general knowledge questions, summarizing information from the web, and generating images were prerecorded, leading to snarky commentary on Chinese social media. It didn’t help that OpenAI had introduced a major upgrade, called GPT-4, to the AI technology that powers ChatGPT only the day before.
But Baidu also faces challenges that don’t apply to companies outside of China racing to compete with ChatGPT. It is inherently difficult to contain the tendency of these chatbots to make up or “hallucinate” facts, or the way they can be prompted into saying unpleasant—or inappropriate—things. But Baidu must also adhere to strict government censorship guidelines for online content.
“Baidu is going to face a tension between making a useful chatbot and making one that conforms to Chinese speech controls,” says Matt Sheehan, a fellow at the Carnegie Endowment for International Peace who studies China’s AI industry. “I'm skeptical they'll be able to create a general-purpose chatbot that users can’t trick into spitting out speech that's unacceptable in China.”
In less than four months since it was introduced, ChatGPT has become a cultural phenomenon, wowing the world with its ability to write poetry and prose, answer mathematical questions, hold forth on philosophical ideas, and converse fluently on just about any topic. The latest version can respond to images, not just text, and OpenAI says it scores more highly on a range of academic tests and makes fewer errors. In the tech industry, just about every company is now scrambling to develop a chatbot strategy.
The problem of getting models like ChatGPT to behave is far from solved, however. Microsoft was forced to limit the use of its Bing chatbot based on OpenAI’s technology after users found ways of evading the guardrails in place and getting the model to say inappropriate or questionable things such as claiming to want to break free of its controls or professing its feelings for a user.
Like the Bing bot and ChatGPT, Baidu’s Ernie Bot is built on top of a machine learning algorithm known as a large language model that was trained using vast quantities of text to predict the next word in a sentence. That simple mechanism, when paired with vast quantities of text and sufficient computing power, has proven able to produce strikingly humanlike responses.
Baidu and OpenAI both also used an additional training step in which human testers provide feedback on what type of answers are most satisfying. That causes the bots to produce responses that are more helpful but still far from perfect. It is not clear how to prevent such models from fabricating answers some of the time, or how to stop them from ever misbehaving.