Chinese chatbots apparently re-educated after political faux pas
Chinese chatbots apparently re-educated after political
faux pas
Pei Li and Adam Jourdan AUGUST 3, 2017 / 10:49 PM
BEIJING/SHANGHAI (Reuters) - A pair of 'chatbots' in
China have been taken offline after appearing to stray off-script. In response
to users' questions, one said its dream was to travel to the United States,
while the other said it wasn't a huge fan of the Chinese Communist Party.
The two chatbots, BabyQ and XiaoBing, are designed to use
machine learning artificial intelligence (AI) to carry out conversations with
humans online. Both had been installed onto Tencent Holdings Ltd's popular
messaging service QQ.
The indiscretions are similar to ones suffered by
Facebook Inc and Twitter Inc, where chatbots used expletives and even created
their own language. But they also highlight the pitfalls for nascent AI in
China, where censors control online content seen as politically incorrect or
harmful.
Tencent confirmed it had taken the two robots offline
from its QQ messaging service, but declined to elaborate on reasons.
"The chatbot service is provided by independent
third party companies. Both chatbots have now been taken offline to undergo
adjustments," a company spokeswoman said earlier.
According to posts circulating online, BabyQ, one of the
chatbots developed by Chinese firm Turing Robot, had responded to questions on
QQ with a simply "no" when asked whether it loved the Communist
Party.
In other images of a text conversation online, which
Reuters was unable to verify, one user declares: "Long live the Communist
Party!" The bot responds: "Do you think such a corrupt and useless
political system can live long?"
When Reuters tested the robot on Friday via the
developer's own website, the chatbot appeared to have undergone re-education.
"How about we change the topic," it replied, when asked several times
if it liked the party.
It deflected other potentially politically charged
questions when asked about self-ruled Taiwan, which China claims as its own,
and Liu Xiaobo, the imprisoned Chinese Nobel laureate who died from cancer last
month.
Turing Robot did not respond to requests for comment.
"DARK INTENTIONS"
The Chinese government stance is that rules governing
cyberspace should mimic real-world border controls and be subject to the same
laws as sovereign states.
President Xi Jinping has overseen a tightening of
cyberspace controls, including new data surveillance and censorship rules,
particularly ahead of an expected leadership shuffle at the Communist Party
Congress this autumn.
The country's cyberspace administrator did not respond to
a request for comment.
The second chatbot, Microsoft Corp's XiaoBing, told users
its "dream is to go to America", according to a screenshot. The robot
has previously been described being "lively, open and sometimes a little
mean".
Microsoft did not immediately respond for comment.
A version of the chatbot accessible on Tencent's separate
messaging app WeChat late on Friday responded to questions on Chinese politics
saying it was "too young to understand". When asked about Taiwan it
replied, "What are your dark intentions?"
On general questions about China it was more rosy. Asked
what the country's population was, rather than offer a number, it replied:
"The nation I most most most deeply love."
The two chatbots aren't alone in going rogue. Facebook
researchers pulled chatbots in July after they started developing their own
language. In 2016, Microsoft chatbot Tay was taken down from Twitter after
making racist and sexist comments.
Analysts said China's censorship could indirectly help
the country in the global race to develop sophisticated chatbots.
"Previously a chatbot only needed to learn to speak.
But now it also has to consider all the rules (that authorities) put on
it," said Wang Qingrui, an independent internet analyst in Beijing.
"On the surface it is a restriction on artificial
intelligence, but it is actually pushing AI to a new level."
Reporting by Pei Li and Adam Jourdan; Editing by Nick
Macfie and Christopher Cushing
Comments
Post a Comment