AI
🚀DeepSeek V4 Launches With 1M-Token Context at 10% the Cost of ChatGPT
The Rundown: DeepSeek released V4 Pro with a 1-million-token context window and pricing at just $4/million output tokens — roughly 10% of what ChatGPT and Claude charge — while openly acknowledging it trails frontier models by 3-6 months.
The details:
- ●DeepSeek V4 Pro features 1.6 trillion parameters and uses novel Compressed Sparse Attention (CSA) and Heavily Compressed Attention (HCA) to achieve 1M-token context on a fraction of the VRAM of its predecessor
- ●Output pricing sits at $4/million tokens, compared to roughly $40/million for ChatGPT and Claude — a 90% cost reduction for developers building on top of frontier-class models
- ●DeepSeek self-disclosed that V4 trails GPT-5.4 and Gemini 3.1 Pro by 3-6 months on key benchmarks, a rare move of transparency from an AI lab
- ●Model is released under a permissive MIT license, making it freely deployable for commercial use without royalties
Why it matters: For founders building AI-native products, DeepSeek V4 is a direct shot across the bow at your cost structure. If you're paying OpenAI or Anthropic rates for inference, you now have a legitimate open-source alternative at a fraction of the price. The MIT license means you can self-host, fine-tune, and ship without vendor lock-in. The 1M-token context window also opens entirely new product categories — long-form document analysis, codebase-level reasoning, and persistent agent memory — that were economically unviable at previous price points.
📰 Source: The Neuron / AlphaSignal