Deepseek V4 Flash
Deepseek V4 Flash is available through Ollama for local agent workflows, with support for text input. DeepSeek-V4-Flash is a preview of the DeepSeek-V4 series, a Mixture-of-Experts model with 284B total parameters and 13B activated, built for efficient reasoning across a 1M-token context window.
What is Deepseek V4 Flash?
Deepseek V4 Flash is a local model entry from Ollama that Agent Mag tracks for install commands, available tags, modalities, and agent workflow fit. Builders can install it with the Agent Mag CLI and run it through Ollama on their own machine.
DeepSeek-V4-Flash is a preview of the DeepSeek-V4 series, a Mixture-of-Experts model with 284B total parameters and 13B activated, built for efficient reasoning across a 1M-token context window.
| Tag | Size | Context | Input |
|---|---|---|---|
| deepseek-v4-flash:cloud | - | 1M | text |
Related local models
DeepSeek-OCR is a vision-language model that can perform token-efficient OCR.
DeepSeek-V3.2, a model that harmonizes high computational efficiency with superior reasoning and agent performance.
DeepSeek-V3.1-Terminus is a hybrid model that supports both thinking mode and non-thinking mode.
DeepSeek-R1 is a family of open reasoning models with performance approaching that of leading models, such as O3 and Gemini 2.5 Pro.
Related content
Compare pricing, local installs, context windows, and modality filters across the full model catalog.
Find frameworks, SDKs, and infrastructure tools that pair with this model in production workflows.
See Agent Mag coverage of model benchmarks, agent frameworks, and deployment patterns.