A new AI architecture from Chinese startup DeepSeek promises to run million-token models with 73% fewer computing resources, directly threatening the cost structure that underpins the current AI hardware market.
Back
A new AI architecture from Chinese startup DeepSeek promises to run million-token models with 73% fewer computing resources, directly threatening the cost structure that underpins the current AI hardware market.

A new AI architecture from Chinese startup DeepSeek promises to run million-token models with 73% fewer computing resources, directly threatening the cost structure that underpins the current AI hardware market. The company claims its new V4 model can handle a one-million-token context using just 27% of the computing power and 10% of the memory of its predecessor, a structural shift that could significantly lower costs for developers and intensify competition for incumbents like Nvidia and Google.
"From now on, 1M (one million) context will be the standard configuration for all of DeepSeek's official services," the company said in its official announcement. This move is a direct challenge to the high costs associated with large-context AI, a problem Nvidia CEO Jensen Huang has highlighted as a critical barrier. While DeepSeek's benchmarks show it still trails Google's most advanced closed-source models in general knowledge, its efficiency gains represent a formidable new threat in the AI arms race.
The V4 model's efficiency stems from a new hybrid attention architecture. It reduces the computational load, measured in floating-point operations per second (FLOPs), to just 27% of the previous V3.2 model for a single token inference at a 1M token context. The required KV cache, a key bottleneck for memory, is reduced to just 10% of the prior version. The company released two versions: the V4-Pro, a 1.6 trillion parameter model, and a smaller V4-Flash model, both available under an open-source MIT license.
For investors, DeepSeek's breakthrough represents a potential disruption to the current market. By designing a model that is less reliant on brute-force computing power, the company creates an opening for alternative hardware, such as Huawei's Ascend chips. This aligns with warnings from Nvidia's own CEO about China building its own, independent AI stack. DeepSeek, reportedly seeking a valuation over $20 billion with backing from Alibaba and Tencent, could compress margins for cloud providers and chipmakers if its cost advantages prove scalable and drive widespread adoption.
The core innovation behind DeepSeek V4 is a two-pronged approach to redesigning the attention mechanism, the computational heart of a transformer model. Standard attention requires every token to calculate a relevance score with every other token in a sequence, leading to computational complexity that grows quadratically—a major barrier to commercializing million-token context windows.
DeepSeek's solution combines Compressed Sparse Attention (CSA) and Heavily Compressed Attention (HCA). CSA uses a trainable mechanism to learn which token connections are important enough for a full calculation, dynamically creating a sparse structure instead of calculating everything. HCA tackles the memory problem by compressing the Key-Value (KV) cache, the data that must be held in expensive GPU memory during inference. Together, these innovations allow DeepSeek to serve 3 to 4 times the number of concurrent users on the same hardware compared to traditional architectures.
While DeepSeek V4-Pro's efficiency is its main feature, its performance benchmarks paint a picture of a specialized competitor. The model excels in mathematics and coding, scoring 3206 on the Codeforces benchmark, outperforming reported scores for models from OpenAI and Google. However, in tests of general world knowledge and advanced reasoning, it lags. On the SimpleQA-Verified benchmark, V4 scored 57.9, well behind the 75.6 score of Google's Gemini 3.1 Pro.
This suggests DeepSeek is focusing its resources on specific, high-value capabilities where it can establish a clear lead, rather than trying to beat frontier models on all fronts. This strategy, combined with its open-source and low-cost approach, has already seen it top Apple's App Store download charts in its initial weeks, signaling a strong market appetite for alternatives to expensive, proprietary models from US tech giants. The rise of a potent, cost-efficient model optimized for non-US hardware is the exact scenario Nvidia's Jensen Huang described as a "horrible outcome for our nation," and it appears to be unfolding faster than many expected. The key question for investors is how quickly this architectural advantage translates into market share and revenue, and whether incumbents like Nvidia can adapt their own roadmaps to counter the threat of a more efficient, multi-polar AI hardware world.
This article is for informational purposes only and does not constitute investment advice.