[CL] DeepSeek-Prover-V2: Advancing Formal Mathematical Reasoning via Reinforcement Learning for Subgoal Decomposition Z.Z. Ren, Z Shao, J Song, H Xin... [DeepSeek-AI] (2025) https://t.co/3BlWUrkR5R https://t.co/gsHfgVesGA
Opensource is literally killing it. 5 models released just this week: 1. Microsoft released three Phi-4 reasoning models. New 14B model outperforms OpenAI o1-mini and matches models 5x larger on math and science tasks while running on your laptop. Download from Hugging Face. https://t.co/OZYN4WJYJw
DeepSeek-Prover-V2: A New Era in Automated Mathematical Reasoning https://t.co/0Ddd8HU1aF
Chinese AI start-up DeepSeek has released DeepSeek-Prover V2, an advanced domain-specific AI model designed for automated mathematical reasoning. The model reportedly solves approximately 90% of miniF2F problems, surpasses state-of-the-art performance on the PutnamBench, and successfully addresses formal problems from AIME 24 and 25. DeepSeek-Prover V2 was quietly uploaded to the Hugging Face platform, contributing to the open-source AI ecosystem. Concurrently, Microsoft has launched a series of small language models focused on reasoning, including Phi 4 mini reasoning, Phi 4 reasoning, and Phi 4 reasoning plus. These models utilize a combination of data curation, supervised fine-tuning, and targeted reinforcement learning to achieve strong reasoning capabilities. Microsoft’s new 14-billion parameter Phi-4 reasoning model reportedly outperforms OpenAI’s o1-mini and matches the performance of models five times larger on math and science tasks, while being efficient enough to run on laptops and mobile devices. Both DeepSeek and Microsoft’s releases highlight advancements in AI-driven mathematical and logical reasoning, with reinforcement learning playing a key role in improving model performance.