Skip to main content
Back to News
DeepSeek V4 Preview: 1.6 Trillion Parameters, Open-Weight Challenge to Frontier Labs
Technology
2 min read
CN

DeepSeek V4 Preview: 1.6 Trillion Parameters, Open-Weight Challenge to Frontier Labs

The AMW Read

Novelty 2: DeepSeek is a known player, but V4 represents a major technical iteration with full stack migration to domestic hardware. Significance 3: Cross-segment impact via cost collapse, sovereignty signaling, and open-weight pressure on frontier labs.
NoveltySignificance
Foundation Models · Player MapGeopoliticsScaling Laws
DeepSeek AI
DeepSeek AI

Foundation Models / LLMs

View Company Profile

DeepSeek V4 Preview: 1.6 Trillion Parameters, Open-Weight Challenge to Frontier Labs

DeepSeek has released the preview of its V4 model, a 1.6 trillion-parameter mixture-of-experts architecture trained on 32-33 trillion tokens, supporting a 1 million-token context window by default across both Pro and Flash tiers. The model marks the company's first flagship iteration since R1 in January 2025, after a 15-month development cycle that involved migrating training from Nvidia CUDA to Huawei Ascend chips. DeepSeek V4 is available as open-weight under a permissive license, with API pricing starting at ¥0.2 per million tokens for Flash cache hits. The model achieves inference FLOP reduction of 73% and KV cache memory reduction of 90% versus its predecessor, according to the technical report.

This launch updates the ongoing open-weight vs. closed-source debate within the foundation-model competitive landscape. DeepSeek's strategy of releasing frontier-capability models at dramatically lower cost — now paired with native Chinese silicon adaptation — directly challenges the capital-intensive scaling narrative of US frontier labs. The V4 preview arrives alongside reports that DeepSeek is seeking its first external funding round at a valuation exceeding $20 billion, after founder Liang Wenfeng's long-standing refusal of outside capital. The company faces headwinds including high hallucination rates (94-96%), absence of multimodal capabilities, and ongoing researcher attrition to competitors like ByteDance and Xiaomi.

The combination of open-weight licensing with aggressive API pricing exemplifies the 'price-squeeze' pattern that has compressed margins across the model layer. By demonstrating that a purely open-weight lab can achieve near-frontier performance on domestic hardware, DeepSeek provides concrete evidence for the China challenger frame in the foundation-model open debate — namely that sovereign-AI pressure and algorithmic efficiency can substitute for unrestricted compute access. Industry observers note that V4's long-context capability and agent-optimized architecture position it as infrastructure for enterprise and government deployments requiring data sovereignty.

#DeepSeek #FoundationModel #OpenSource #SovereignAI #AIPricing

#DeepSeek#V4#open-weight#1.6 trillion parameters#mixture-of-experts#Huawei Ascend#AI pricing#foundation model

How This Connects

Based on Foundation Models · Player Map

  1. 3h agoAnthropic in talks to buy breakthrough AI chip from Fractile, claims 100x speed, 90% cost cut over NvidiaAnthropic
  2. 3h agoAnthropic and Blackstone Launch Joint Venture to Accelerate Claude Adoption Among SMEsAnthropic
  3. 1d agoDeepSeek V4 Preview: 1.6 Trillion Parameters, Open-Weight Challenge to Frontier Labs · THIS ARTICLE
  4. 2d agoU.S. Department of Defense (DoD) selects 8 tech companies for classified AI agreement, excluding Anthropic.OpenAI
  5. 3w agoArcee has launched Trinity Large Thinking, a 400B-parameter sparse MoE LLM with only 13B active per...Arcee
  6. 1mo agoIndian AI startup Sarvam has open-sourced two reasoning models (30B and 105B parameters) trained ent...Sarvam AI

Related News

More news from DeepSeek AI

Stay updated with the latest news and announcements from DeepSeek AI.

View all DeepSeek AI news

Discover AI Startups

Explore 2,000+ AI companies with VC-grade analysis, funding data, and investment insights.

Explore Dashboard