How Chinese AI Chatbots Censor Themselves | WIRED
Summary
A study by Stanford and Princeton reveals that Chinese AI chatbots are more likely to censor political questions than their Western counterparts, highlighting the impact of censorship on AI development.
Why It Matters
This research sheds light on the mechanisms of digital censorship in China and its implications for AI technology. Understanding how AI models are influenced by their training data and manual interventions is crucial for developers and policymakers, especially as AI becomes more integrated into society.
Key Takeaways
- Chinese AI models refuse to answer political questions more frequently than American models.
- Manual interventions play a significant role in the censorship of Chinese AI responses.
- The study provides quantifiable evidence of biases in Chinese LLMs, crucial for understanding AI censorship.
Save StorySave this storySave StorySave this storyHearing someone talk about digital censorship in China is always either extremely boring or extremely interesting. Most of the time, people are still regurgitating the same talking points from 20 years ago about how the Chinese internet is like living in George Orwell’s 1984. But occasionally, someone discovers something new about how the Chinese government exerts control over emerging technologies, revealing how the censorship machine is a constantly evolving beast.A new paper by scholars from Stanford University and Princeton University about Chinese artificial intelligence belongs to the second category. The researchers fed the same 145 politically sensitive questions to four Chinese large language models and five American models and then compared how they responded. They then repeated the same experiment 100 times.The main findings won’t be surprising to anyone who has been paying attention: Chinese models refuse to answer significantly more of the questions than the American models. (DeepSeek refused 36 percent of the questions, while Baidu’s Ernie Bot refused 32 percent; OpenAI’s GPT and Meta’s Llama had refusal rates lower than 3 percent.) In cases where they didn’t outright refuse to answer, the Chinese models also gave shorter answers and more inaccurate information than their American counterparts did.One of the most interesting things the researchers attempted to do was to separate the impact of pre-training and ...