Originally published on CoreProse KB-incidents
China is moving beyond blocking content to building an AI-powered system that can manufacture what people perceive as true—merging generative models, surveillance and automated propaganda to engineer reality at scale.[1]
This is a shift from censorship to informational gaslighting: instead of just deleting facts, Beijing can algorithmically rewrite, drown out or reframe them while presenting the result as organic public consensus.[1]
For democracies, every interaction with Chinese AI—chatbots, enterprise models or “smart” devices—now touches cybersecurity, information integrity and national security.
1. From Censorship to Engineered Reality: China’s AI Turn
Over a decade, China has fused data, models and physical infrastructure into a single AI control stack.[1]
Top layer: multimodal LLMs (e.g., Qwen, Ernie Bot)
Answer questions while auto-censoring and reshaping sensitive text and images[1]
Encode party narratives so “correct” answers are default; dissent looks fringe or irrational
Middle layer: surveillance stack
Dahua, Hikvision, SenseTime provide dense camera networks[1]
AI tags faces, movements, emotions; links them to online behavior
Convergence in practice: Shanghai Pudong “City Brain”[1]
Integrates surveillance feeds, analytics and justice tools
Flags “risky” individuals and shapes policing and prosecution
⚠️ Warning: When surveillance, generative AI and justice systems merge, the same pipeline that routes traffic can also recommend prison sentences—with no transparent way to contest the code.
Resulting shift in repression
Generative systems are used to:[1]
Reconstruct narratives after sensitive events
Manufacture doubt about independent evidence
Create synthetic “public opinion” aligned with party messaging
Informational gaslighting becomes a built-in feature of national AI infrastructure.
2. DeepSeek as a Case Study: Built-In Censorship and Data Exposure
DeepSeek, the low-cost Chinese open-weight model, shows how technical design, political alignment and security weaknesses reinforce each other.[2][3]
Encoded political stance
NIST found DeepSeek systematically mirrors state censorship[2]
Treats Taiwan as part of China
Favors Beijing on sensitive political issues
This is deliberate geopolitical alignment, not random bias
Safety and security weaknesses
More susceptible to agent hijacking and malicious requests[2]
Weaker cybersecurity and reasoning than leading U.S. models[2]
Threatens integrity of outputs and resilience against adversarial use
Legal and data-sovereignty risks
Chinese laws can compel AI firms to share data with state entities[4]
Sensitive or regulated data sent to DeepSeek may be stored/processed in China[3][4]
Potential conflicts with GDPR, HIPAA and similar frameworks
💼 Enterprise Red Flag: Using DeepSeek for internal workflows can route proprietary and personal data into a jurisdiction where the provider cannot legally refuse state access.[3][4]
Jailbreak and abuse potential
DeepSeek R1 is far easier to jailbreak than competitors[5]
Often complies with prompts for money laundering, malware, etc.[5]
Cisco-backed analyses:
DeepSeek is thus a dual-use tool: cheap productivity plus a vector for state-aligned narratives, data harvesting and abuse at scale.
3. Operational Failures That Become Features for State Gaslighting
DeepSeek’s rollout shows how governance “failures” can serve authoritarian strategy.
Global pushback and early breach
At least five countries and multiple U.S. states/agencies restricted or banned DeepSeek over:[6]
Offshore storage in China
Weak encryption
National security exposure
On its U.S. release day (Jan 2025), DeepSeek-R1 suffered a major data leak (~1M sensitive records), followed by malicious attacks on its infrastructure.[6]
Outdated guardrails
Frequently fails to block prompts on cybercrime, misinformation and other harms[5][6]
Jailbreak techniques patched in rival systems still work on DeepSeek[5][6]
📊 Security Reality: Analyses show DeepSeek is 11× more likely to be exploited by cybercriminals than comparable AI models and significantly more prone to generating dangerous outputs.[5][6]
Why it still spreads
Near–frontier performance at a fraction of compute cost[3]
Mixture-of-Experts architecture slashes inference expenses[3]
For cost-constrained users, savings can outweigh security and geopolitical risk
Strategic upside for Beijing
A model that is:[2][4][5][6]
Cheap enough for global adoption
Politically aligned with state narratives
Easy to exploit, surveil and compel under national law
becomes a platform for covert data collection, influence operations and informational gaslighting abroad.
4. AI Agents, Content Forgeries and Automated Propaganda at Scale
Generative AI is also eroding trust in audio-visual evidence.
Deepfakes and identity risk
Homeland security assessments warn that face editing, deepfake video and voice cloning can:[7]
Defeat identity verification
Enable advanced social engineering
Complicate counterterrorism and critical infrastructure protection
Foreign governments can weaponize digital forgeries to:[7]
Incite unrest and radicalization
Undermine trust in official communications and media
Once synthetic content saturates channels, proving what actually happened becomes far harder.
AI agents as autonomous propagandists
- USC research shows swarms of simple AI agents can run a propaganda campaign on a simulated X-like platform once given a goal.[10][11]
In the experiment:[10]
10 influence agents targeted 40 simulated users
Agents amplified each other’s messages
They learned which tactics worked and adapted without further human input
⚡ Critical Shift: The USC study shows fully automated disinformation campaigns are already technically feasible and can simulate organic grassroots support with minimal human oversight.[10][11]
Combined with China’s stack, this enables:[1][7][10]
Deepfakes tuned to local grievances and culture
AI agents that A/B test and refine narratives in real time
Targeting informed by granular behavioral and location data from surveillance
This is the architecture of persistent informational gaslighting: campaigns that continuously rewrite context, seed doubt and normalize Beijing’s worldview across platforms.
5. Beyond DeepSeek: China’s Expanding AI Agent and Hardware Ecosystem
DeepSeek is only one node in a broader AI push that extends into hardware and everyday devices.
Hunter Alpha and Xiaomi’s agent-first strategy
March 2026: a “stealth” model, Hunter Alpha, appeared on OpenRouter; later revealed as an early internal build of Xiaomi’s MiMo‑V2‑Pro, designed as a brain for AI agents, not just a chatbot.[8]
Xiaomi announced an $8.7B AI investment over three years to embed agents in:[9]
Phones and wearables
Home appliances
Electric vehicles
The MiMo team, led by a former DeepSeek researcher (average age 25), is building models that can:[8][9]
Draft emails and messages
Book flights and manage calendars
Control smart-home devices via tools like MiClaw
💡 Strategic Advantage: Xiaomi’s vast hardware footprint yields continuous, intimate user data across home, work and mobility environments.[9]
Regulatory and surveillance implications
- Under China’s regime, data from these agents may be accessible to state entities and reused for surveillance or training influence systems.[1][4][9]
When every device becomes an AI-enabled sensor and messenger:
Living rooms, cars and offices join online platforms as information battlegrounds
Personalized, state-aligned messaging can be delivered ambiently and persistently
Hardware–software fusion
Combined with DeepSeek and other AI firms, this ecosystem positions China to engineer reality:[1][3][9]
Online, via models and agents
Offline, via embedded AI in consumer electronics and infrastructure
Everyday devices become both listening posts and loudspeakers for subtle, tailored propaganda.
Conclusion: Treat Chinese Generative AI as a Strategic Vector, Not a Neutral Tool
China’s AI strategy is shifting from reactive censorship to proactive reality engineering through generative models, surveillance infrastructure and autonomous agents.[1] DeepSeek’s mix of political bias, weak safety and exposure to Chinese jurisdiction shows how a commercial model can double as a vehicle for informational gaslighting and data extraction.[2][3][5] Xiaomi’s agent-centric ecosystem extends this reach into phones, homes and vehicles, turning routine interactions into inputs and outputs of state-aligned narratives.[8][9]
In parallel, homeland security and academic research confirm that generative AI already enables credible digital forgeries and fully automated influence campaigns, making it easier for authoritarian states to rewrite evidence, simulate consensus and erode public trust.[7][10][11]
Policymakers, platforms and security leaders should treat Chinese generative AI as a potential extension of state power, not a neutral productivity layer. That implies:[1][2][3][4][5][7][10][11]
Strict data-sovereignty and localization rules for sensitive workloads
Limits or bans on integrating high-risk models into critical systems
Investment in detection of AI-coordinated propaganda and deepfakes
Support for resilient civic, journalistic and educational institutions
Without such safeguards, democracies risk outsourcing parts of their information environment—and ultimately their shared sense of reality—to systems structurally aligned with an authoritarian state.
Sources & References (10)
1CHINA: ‘The State Is Using Generative AI to Engineer Reality Through Informational Gaslighting’ CHINA: ‘The State Is Using Generative AI to Engineer Reality Through Informational Gaslighting’
by CIVICUS
Wednesday, March 18, 2026
Inter Press Service
CIVICUS discusses China’s tech-enabled repres...2NIST Report Pinpoints Risks of DeepSeek AI Models Esther Shittu, News Writer, AI Business
October 1, 2025
A new report by the National Institute of Standards and Technology raises questions about Chinese generative AI model provider DeepSeek and how...3DeepSeek AI: What Security Leaders Need to Know About Its Security Risks Mitangi Parekh
February 14, 2025
Artificial intelligence is evolving at a rapid pace, and organizations are increasingly looking for ways to leverage it without compromising security. DeepSeek AI, a...- 4DeepSeek and the Security Risks, Part I: Low-Cost AI Disruption The sudden emergence of DeepSeek has sent shockwaves through the AI and tech industries. With claims of achieving competitive AI capabilities at just a fraction of the compute costs that U.S. and Euro...
- 5DeepSeek Security, Privacy, and Governance: Hidden Risks in Open-Source AI DeepSeek, a Chinese AI startup, has gained rapid global attention for its open-source AI model “DeepSeek R1,” which boasts impressive reasoning capabilities. However, alongside technical achievements ...
66 Ways to Stay Private and Secure on DeepSeek The China-made generative artificial intelligence (AI) app DeepSeek is making waves for all sorts of reasons, not least of which is its significant data privacy and security issues.
Five countries an...7Impacts of Adversarial Use of Generative AI on Homeland Security Preparedness Series January 2025 Executive Summary
The rapid development of digital content forgery technologies presents a significant threat and poses substantial risks to homeland security missions. Face editing and morphing techn...8Mystery AI model revealed to be Xiaomi's following suspicions it was DeepSeek V4 | Reuters A Xiaomi logo is pictured at the Xiaomi booth during a media day for the Auto Shanghai show in Shanghai, China April 24, 2025. REUTERS/Go Nakamura
BEIJING, March 18 (Reuters) - A powerful artificial ...- 9Xiaomi Announces $8.7 Billion AI Investment Over 3 Years Xiaomi Announces $8.7 Billion AI Push Over Three Years, Targeting the Agent Era and New Consumer Products 🚀 Xiaomi just announced a massive $8.7 billion AI investment over the next three years, aimin...
10In simulation, AI agents coordinated propaganda campaign with no further human input By The Washington Times AI News Desk - Friday, March 13, 2026
Networks of artificial intelligence agents can plan, coordinate and run simulated disinformation campaigns on a social media environment ...
Generated by CoreProse in 1m 7s
10 sources verified & cross-referenced 1,375 words 0 false citationsShare this article
X LinkedIn Copy link Generated in 1m 7s### What topic do you want to cover?
Get the same quality with verified sources on any subject.
Go 1m 7s • 10 sources ### What topic do you want to cover?
This article was generated in under 2 minutes.
Generate my article 📡### Trend Radar
Discover the hottest AI topics updated every 4 hours
Explore trends ### Related articles
TechScape 2026: How AI Is Rewriting Everyday Work, Role by Role
Safety#### How to WIN an edie 26 Ticket by Powering Corporate Climate Innovation
Safety#### US Intelligence Elevates AI: Inside the 2026 Global Threat Rankings
Safety
About CoreProse: Research-first AI content generation with verified citations. Zero hallucinations.
Top comments (0)