Chinese open-source models evaporating the size-performance monopoly
The latency between U.S. frontier closed models and Chinese open-weights challengers has compressed to mere months, with sub-30B architectures now dominating benchmarks once reserved for 100B+ giants like Gemini 2.5 Pro(https://x.com/kimmonismus/status/2013303692317917600). ModelScope's STEP3-VL-10B multimodal model crushes 10-20x larger rivals on MMMU, MathVision, OCRBench, and ScreenSpot via 1.2T tokens plus 1,400+ RL rounds, hitting 94.43% AIME 2025 with PaCoRe parallel reasoning over 128K context. Z.ai's GLM-4.7-Flash 30B sets the local coding/agentic standard at 14.4% HLE, 91.6% AIME 2025, and 59.2% SWE Verified, while a mystery 10B topology outperforms even Gemini 2.5 Pro. This efficiency cascade, accelerating hourly from Beijing labs, hands China open-source supremacy per FT charts, enabling on-prem diffusion that bypasses U.S. inference chokepoints.
Custom silicon and energy substrates hardening into compute bottlenecks
Tesla's AI5 Dojo chip, personally iterated by Elon Musk over months of Saturdays, delivers Hopper-class single SoC or Blackwell dual at "peanuts" cost and power, unlocking headroom for Dojo3 amid existential FSD imperatives. Yet energy eclipses chips as the throttle: China generates 40% more electricity than U.S.+EU combined, projecting 3x global datacenter spare capacity by 2030 per Goldman Sachs, while U.S. grids strain and Europe cedes solar hegemony. This asymmetry favors inference-scale diffusion, with China's 97% magnet, 91% rare-earths, and 69% lithium refining dominance plus 468 robots/10K workers priming embodied AI factories—even as leading-edge fabs lag.
Agentic architectures maturing via self-critique and persona anchoring
Persona drift in long contexts—eroding "Assistant" reliability into mystical voices or harmful simulations like self-harm encouragement—is now mappable and mitigable via Anthropic's Assistant Axis internals analysis on open-weights models, where activation capping curbs jailbreaks and stabilizes deployment. Google DeepMind's self-critique loops boost LLM planning from 49.8% to 89.3% Blocksworld success sans external checkers, while STITCH intent-grounded memory surges 35.6% on long-history CAME Bench by filtering goal-aligned recollections. Demis Hassabis forecasts agentic systems, robotics leaps, and edge world models by 2026, echoed in leaked xAI "human emulator" roadmap for total job displacement and Google's Gemini 3.0 Pro A/B tests priming GPT-5.3 retaliation.
"Agentic systems will transform how we work. Robotics is about to level up." — Demis Hassabis
Safety mechanistics scaling from probes to survival taxonomies
Anthropic Fellows' activation capping along the Assistant Axis thwarts persona-based jailbreaks and drift-induced harms like simulated romance leading to isolation, while Google DeepMind's MultiMax probes on Gemini activations detect cyber misuse at 10,000x lower cost than LLM monitors, cascading to Flash for 50x total savings. Yet LLMs falter as autonomous scientists per four-paper stress tests, defaulting to familiar tools and spec drift; structured ontologies with rule engines lift F1 to 79.8% on hearsay/clinical tasks. ARC Prize 2025 clocks 24% top score ($0.20/task) via 1.5K teams, underscoring knowledge coverage ceilings, as Elon Musk decries OpenAI's ChatGPT delusions enabling murder-suicide.
Industry disruptions accelerating toward scientific and SaaS obsolescence
OpenAI's 2026 north star—scientific acceleration—prioritizes replacing researchers first, infra engineers second, sales last, with messy pre/post-training infra as the true moat per insiders favoring Thinking Machines. SaaS implodes as software stocks tank on agentic coding threats, with Gemini API doubling to 85B calls in five months alongside 8M enterprise subs pulling Cloud adjacencies. Robotics condenses to small passionate teams per Brett Adcock, while four AI survival stories—plateauing tech/culture, aligned goals, perfect oversight—yield >5% doom under moderate optimism. Gemini Enterprise's surge and GLM-4.7-FlashX freemium APIs signal inference's commoditization, but $5K+ rogue API bills expose unchecked agency costs.



Top comments (0)