DeepSeek V4 drops with 1M context, MIT license
DeepSeek has released DeepSeek V4, a frontier model suite featuring the flagship V4-Pro and high-efficiency V4-Flash built on a novel Deeply Sparse Attention (DSA) architecture. Supporting a 1-million-token context window under the MIT license, the models significantly reduce inference costs through architectural optimizations for long-range reasoning.
DeepSeek V4 is a masterclass in architectural efficiency that makes frontier-level intelligence affordable by outperforming competitors on cost-per-token rather than just raw scale. The 1.6T parameter MoE architecture of V4-Pro matches or beats Claude 4.6 Max in coding and agentic tasks. DSA and Manifold-Constrained Hyper-Connections solve the long-context cost bottleneck, enabling deep reasoning across massive repositories without exponential cost increases. At 97% lower cost than proprietary competitors, it effectively commoditizes high-end reasoning for production-scale agentic workflows. Hardware optimizations for Huawei's latest chips demonstrate a successful path for AI performance independent of Nvidia's dominance.
DISCOVERED
5h ago
2026-04-24
PUBLISHED
6h ago
2026-04-24
RELEVANCE
AUTHOR
impact_sy