Skip to content
Tech News
← Back to articles

ChatGPT Is Saying Weird Things in Chinese

read original get Chinese Language Learning Book → more articles
Why This Matters

The article highlights challenges in ChatGPT's Chinese language responses, revealing how training biases and mode collapse can lead to odd or repetitive phrases that may irritate native speakers. This underscores the importance of refining AI language models to better understand and adapt to diverse linguistic nuances, which is crucial for global user engagement and trust. Improving multilingual AI capabilities will be key to making these tools more effective and culturally sensitive for a worldwide audience.

Key Takeaways

Sign up to see the future, today Can’t-miss innovations from the bleeding edge of science and tech Email address Sign Up Thank you!

If you thought English-language ChatGPT-prose was annoying, just wait until you get a load of its conversational habits in Chinese.

A fascinating piece of reporting by Wired took a look at how ChatGPT handles Chinese, the global language with the highest number of native speakers, according to the Language School at Middlebury College.

One of its go-to tics, the publication reports, is to answer questions with “我会稳稳地接住你,” which literally translates to “I will catch you steadily,” a phrase signalling willingness to talk about a person’s feelings (as Wired’s Zeyi Yang notes, a more flowerly translation could be “I’ll hold you steadily through whatever comes,” though the sentiment is apparently irritating to Chinese speakers either way.)

At other times, ChatGPT will tell its Chinese users “砍一刀,” which means either “help me cut it once,” or “slash the price,” an obnoxious bit of ad copy parroted by Chinese eCommerce platform Pinduoduo, Wired notes.

The odd mannerisms are so ubiquitous that they’ve become a meme among Chinese netizens, with some depicting ChatGPT as a giant inflatable airbag placed to break someone’s fall — catching them steadily.

The problem, Wired observes, may come down to a phenomenon called “mode collapse,” a fundamental bias tracing back to the people training large language models (LLMs). Basically, the idea goes that when human data annotators comb through text to train AI chatbots, they unknowingly favor familiar turns-of-phrase over more foreign-sounding sentences.

After an LLM like ChatGPT is trained, it becomes difficult to force it to “unlearn” certain phrases. While AI developers can reinforce the usage of a particular response — “I will catch you steadily” may be a great answer in a particular situation — accounting for range and quantity is a different beast altogether.

“We don’t know how to say: ‘this is good writing, but if we do this good writing thing 10 times, then it’s no longer good writing,” Max Spero, cofounder and CEO of AI-writing detector Pangram, told Wired.

Whatever the cause, it’s nice to know when English-speakers agree on something with our Chinese brethren, even if it’s just a universal hatred for ChatGPT’s inane babble.

... continue reading