DeepSeek Drops Two Powerful New AI Models With Record-Low Prices and 1 Trillion+ Parameter Architecture

May 04, 2026
Simon Willison’s Weblog
Article image for DeepSeek Drops Two Powerful New AI Models With Record-Low Prices and 1 Trillion+ Parameter Architecture

Summary

DeepSeek unleashes two groundbreaking AI models — DeepSeek-V4-Pro and V4-Flash — featuring over 1 trillion parameters and record-breaking low prices, with V4-Flash undercutting OpenAI's cheapest model at just $0.14 per million tokens, powered by dramatic efficiency gains that slash compute costs to a fraction of its predecessor.

Key Points

  • DeepSeek releases two new preview models, DeepSeek-V4-Pro and DeepSeek-V4-Flash, both featuring 1 million token context windows and Mixture of Experts architecture, with V4-Pro now standing as the largest open weights model at 1.6 trillion total parameters.
  • DeepSeek-V4-Flash is the cheapest small model on the market at $0.14/million input tokens, undercutting even OpenAI's GPT-5.4 Nano, while DeepSeek-V4-Pro at $1.74/million input tokens is the most affordable large frontier model available.
  • The extreme pricing is driven by major efficiency gains, with V4-Pro using only 27% of the compute FLOPs and 10% of the KV cache size of its predecessor DeepSeek-V3.2 in long-context scenarios, though benchmarks suggest V4-Pro trails the absolute state-of-the-art frontier models by approximately 3 to 6 months.

Tags

Read Original Article