DeepSeek-V2 Released: Chinese Math Reasoning Surpasses GPT-4o, 236B Parameters Lead with High Efficiency
Chinese AI startup DeepSeek has released DeepSeek-V2, a large language model that outperforms GPT-4o in Chinese math reasoning benchmarks while achieving high efficiency through its MoE architecture with 236B total parameters but only 21B active.