DeepSeek V4 Pro has 1.6T total parameters, its largest model by the metric, and V4 Flash has 284B parameters; both models have a context window of 1M tokens (Vincent Chow/South China Morning Post)
DeepSeek released V4 Pro (1.6T total parameters) and V4 Flash (284B parameters), both with a 1M token context window, expanding its model lineup.
Excerpt
<a href="https://www.scmp.com/tech/big-tech/article/3351239/deepseek-releases-next-gen-ai-model-world-leading-efficiency"><img align="RIGHT" border="0" hspace="4" src="http://www.techmeme.com/260424/i5.jpg" vspace="4" /></a>
<p><a href="http://www.techmeme.com/260424/p5#a260424p5" title="Techmeme permalink"><img height="12" src="http://www.techmeme.com/img/pml.png" style="border: none; padding: 0; margin: 0;" width="11" /></a> Vincent Chow / <a href="http://www.scmp.com/">South China Morning Post</a>:<br />
<span style="font-size: 1.3em;"><b><a href="https://www.scmp.com/tech/big-tech/article/3351239/deepseek-releases-next-gen-ai-model-world-leading-efficiency">DeepSeek V4 Pro has 1.6T total parameters, its largest model by the metric, and V4 Flash has 284B parameters; both models have a context window of 1M tokens</a></b></span> — The company claims its much-anticipated V4 model is competitive with top closed-source models from OpenAI and Google DeepMind</p>
Read at source: http://www.techmeme.com/260424/p5#a260424p5