[2602.22221] Misinformation Exposure in the Chinese Web: A Cross-System Evaluation of Search Engines, LLMs, and AI Overviews

[2602.22221] Misinformation Exposure in the Chinese Web: A Cross-System Evaluation of Search Engines, LLMs, and AI Overviews

arXiv - AI 3 min read Article

Summary

This article evaluates misinformation exposure on the Chinese web by comparing traditional search engines, LLMs, and AI-generated overviews using a new dataset of real user queries.

Why It Matters

Understanding how misinformation spreads in non-English web ecosystems is crucial for developing reliable AI tools. This study highlights the structural risks associated with AI-mediated search, emphasizing the need for improved accuracy and transparency in information access.

Key Takeaways

  • The study introduces a fact-checking dataset of 12,161 Chinese Yes/No questions.
  • Significant differences in factual accuracy were found across search engines, LLMs, and AI overviews.
  • The research estimates the potential misinformation exposure of Chinese users across different regions.
  • AI-mediated search poses structural risks that need addressing for better information reliability.
  • The findings call for more transparent and reliable information-access tools.

Computer Science > Information Retrieval arXiv:2602.22221 (cs) [Submitted on 15 Dec 2025] Title:Misinformation Exposure in the Chinese Web: A Cross-System Evaluation of Search Engines, LLMs, and AI Overviews Authors:Geng Liu, Junjie Mu, Li Feng, Mengxiao Zhu, Francesco Pierri View a PDF of the paper titled Misinformation Exposure in the Chinese Web: A Cross-System Evaluation of Search Engines, LLMs, and AI Overviews, by Geng Liu and 4 other authors View PDF HTML (experimental) Abstract:Large Language Models (LLMs) are increasingly integrated into search services, providing direct answers that can reduce users' reliance on traditional result pages. Yet their factual reliability in non-English web ecosystems remains poorly understood, particularly when answering real user queries. We introduce a fact-checking dataset of 12~161 Chinese Yes/No questions derived from real-world online search logs and develop a unified evaluation pipeline to compare three information-access paradigms: traditional search engines, standalone LLMs, and AI-generated overview modules. Our analysis reveals substantial differences in factual accuracy and topic-level variability across systems. By combining this performance with real-world Baidu Index statistics, we further estimate potential exposure to incorrect factual information of Chinese users across regions. These findings highlight structural risks in AI-mediated search and underscore the need for more reliable and transparent information-acces...

Related Articles

[2603.23966] Policy-Guided Threat Hunting: An LLM enabled Framework with Splunk SOC Triage
Llms

[2603.23966] Policy-Guided Threat Hunting: An LLM enabled Framework with Splunk SOC Triage

Abstract page for arXiv paper 2603.23966: Policy-Guided Threat Hunting: An LLM enabled Framework with Splunk SOC Triage

arXiv - AI · 4 min ·
[2603.16790] InCoder-32B: Code Foundation Model for Industrial Scenarios
Llms

[2603.16790] InCoder-32B: Code Foundation Model for Industrial Scenarios

Abstract page for arXiv paper 2603.16790: InCoder-32B: Code Foundation Model for Industrial Scenarios

arXiv - AI · 4 min ·
[2603.16430] EngGPT2: Sovereign, Efficient and Open Intelligence
Llms

[2603.16430] EngGPT2: Sovereign, Efficient and Open Intelligence

Abstract page for arXiv paper 2603.16430: EngGPT2: Sovereign, Efficient and Open Intelligence

arXiv - AI · 4 min ·
[2603.11066] Exploring Collatz Dynamics with Human-LLM Collaboration
Llms

[2603.11066] Exploring Collatz Dynamics with Human-LLM Collaboration

Abstract page for arXiv paper 2603.11066: Exploring Collatz Dynamics with Human-LLM Collaboration

arXiv - AI · 4 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime