DeepSeek V4 Flash
284B MoE / 13B active. Apache 2.0, 1M context. Among the cheapest frontier-adjacent models — $0.14 input, $0.28 output per 1M tokens.
1M
tokens
16K
tokens
$0.14
per 1M tokens
$0.28
per 1M tokens
218
tokens/sec
Apr 2026
2026-04-24
$0.21
per 1M tokens
371.4
quality per $
Capabilities
Benchmarks
Compare With
Open Source — Licensed under Apache 2.0
About DeepSeek V4 Flash
DeepSeek V4 Flash is a fast AI model by DeepSeek, released on April 24, 2026. It supports a context window of 1000K tokens and can generate up to 16K output tokens.
At $0.14 per million input tokens and $0.28 per million output tokens, its blended cost of $0.21/1M tokens makes it one of the most affordable models available. Its value score of 371.4 reflects the balance of quality and cost.
DeepSeek V4 Flash is available as an open-source model under the Apache 2.0 license, meaning you can self-host it for predictable costs or use it through API providers like Swfte Connect.
Using DeepSeek V4 Flash with Swfte
Access DeepSeek V4 Flash through Swfte Connect, our unified LLM gateway. Connect gives you a single API for 50+ models, with automatic routing, cost optimization, and fallback handling. You can also try DeepSeek V4 Flash in our AI Playground before integrating.