Analyst memo

Models1 sourceDeveloping

DeepSeek-V4 Supports Million-Token Context

DeepSeek AI's release of DeepSeek-V4, leveraging compressed sparse attention techniques, allows handling one-million-token contexts efficiently, marking a significant advancement in large language model capabilities.

Published Apr 25, 2026, 2:54 AMUpdated Apr 25, 2026, 2:54 AM

What happened

DeepSeek AI introduced DeepSeek-V4, featuring two MoE models, DeepSeek-V4-Pro and DeepSeek-V4-Flash, designed to support one-million-token contexts through innovations in compressed sparse attention and more.

Why it matters

The capability to efficiently process one-million-token contexts marks a substantial breakthrough in AI, potentially transforming applications reliant on long context understanding.

Who is affected

This development impacts AI researchers, developers of large language models, and industries focusing on applications like language translation and content summarization.

Risks / uncertainty

It remains uncertain how these architectural changes will perform across varied real-world applications, and whether the computational efficiencies claimed can be consistently realized at scale.