05.06.2025 20:53

DeepSeek R1 Soars to Top 3, Outshining Gemini—An Open-Source Triumph in Coding, Math, and Logic

News image

In a groundbreaking update, DeepSeek R1 has surged into the top three AI models globally, surpassing even Google’s latest Gemini 2.5 Pro.

This milestone, achieved by an open-source model, marks a seismic shift in the AI landscape, particularly in the domains of coding, mathematics, and logical reasoning. Released on May 29, 2025, as DeepSeek-R1-0528, this upgrade from the Chinese AI lab DeepSeek not only challenges proprietary giants but also democratizes access to cutting-edge technology.

DeepSeek R1’s ascent is rooted in its enhanced reasoning capabilities. On the AIME 2025 math test, its accuracy soared from 70% to 87.5%, outpacing Gemini 2.5 Pro and rivaling OpenAI’s o3, according to internal benchmarks.

The model now processes an average of 23,000 tokens per question — nearly double its previous capacity — demonstrating deeper analytical prowess without architectural changes.

In coding, DeepSeek R1 achieved a Codeforces rating of 2,029, outperforming 96.3% of human programmers and closing the gap with o3, while its open-source nature makes it far more cost-effective. General logic also saw gains, with GPQA-Diamond scores rising from 71.5% to 81.0%, cementing its edge over Gemini.

What sets DeepSeek R1 apart is its accessibility. Released under the MIT License, it allows developers worldwide to modify, deploy, and integrate the model freely, a stark contrast to the paywalls of proprietary systems.

Its operational costs are estimated to be 90–96% lower than OpenAI’s o3, making it a game-changer for startups, researchers, and businesses. The model’s reinforcement learning-based architecture, refined through large-scale training, enables it to self-correct and reason step-by-step, offering transparency that proprietary models often lack.

This update also introduced practical features like JSON output support and reduced hallucination rates by 45–50%, enhancing reliability for real-world applications. From solving advanced calculus to generating production-ready code, DeepSeek R1 is proving its mettle across industries.

A distilled version, DeepSeek-R1-0528-Qwen3-8B, further showcases its efficiency, scoring 86% on AIME 2024 while running on minimal resources like an Nvidia H100.


Also read:


DeepSeek R1’s rise challenges the dominance of Western tech giants, raising questions about the future of AI development.

Can open-source models sustain this momentum against the deep pockets of OpenAI and Google?

While its performance is undeniable, concerns linger about data privacy, given DeepSeek’s Chinese origins and potential compliance with local laws.

Nevertheless, DeepSeek R1’s blend of power, affordability, and openness positions it as a beacon for global innovation, proving that the future of AI may not belong to the few, but to the many.


0 comments
Read more