Wednesday, April 29, 2026

DeepSeek V4 falls behind competitors open and closed models

Bad news!

"DeepSeek released weights for its highly anticipated DeepSeek-V4 large language models, which feature 1 million tokens of context and a revised architecture.
The family includes DeepSeek-V4-Pro Preview (1.6 trillion parameters, 49 billion active per token) and DeepSeek-V4-Flash Preview (284 billion parameters, 13 billion active per token).

On Artificial Analysis’ Intelligence Index, among open-weights models, DeepSeek-V4-Pro Preview is behind Moonshot Kimi K2.6 and Xiaomi MiMo-V2.5-Pro, which are tied for first place, and tied for second place with Alibaba Qwen3.6 (all models set to their highest reasoning levels). However, it tops other open-weights models on some of the index’s component benchmarks including GDPval-AA, AA-Omniscience Accuracy, and Humanity’s Last Exam.

Among closed models, its index performance falls behind leading models from Google, OpenAI, Anthropic, and Meta (all set to their highest reasoning levels). DeepSeek, once the standard bearer among China’s open-weights AI developers, now competes with strong rivals. Its new models are optimized to run on Huawei Ascend chips rather than the typical Nvidia GPUs, illustrating the rapid progress of China’s AI chip makers." (Datapoints)


Deepseek V4 is ranked 8th (Source)


No comments: