Microsoft AI has introduced rStar-Math, a system designed to enhance the math reasoning capabilities of small language models (LLMs). With a model size of just 7 billion parameters, rStar-Math demonstrates performance that rivals and occasionally surpasses larger models like OpenAI's o1-preview. The system employs a 'deep thinking' approach, using Monte Carlo Tree Search (MCTS) to explore multiple solution paths and select the most effective one. Through 4 rounds of self-evolution with millions of synthesized solutions for 747k math problems, rStar-Math has shown significant improvements on the MATH benchmark, boosting Qwen2.5-Math-7B from 58.8% to 90.0% and Phi3-mini-3.8B from 41.4% to 86.4%, outperforming o1-preview by 4.5% and 0.9% respectively. Additionally, rStar-Math has achieved notable success on the USA Math Olympiad, solving an average of 53.3% of problems, or 8/15 problems, placing it among the top 20% of high school math students.
HOW TO PROMPT OPENAI'S O1 Ben has written an excellent explanation of how to write prompts for OpenAI's o1 model. I try my best to summarise here, but I encourage you to read the full and detailed article in the quote's thread below: 1. Write Briefs Instead of Prompts:… https://t.co/8EREyJkjW4
[CL] rStar-Math: Small LLMs Can Master Math Reasoning with Self-Evolved Deep Thinking X Guan, L L Zhang, Y Liu, N Shang... [Microsoft Research Asia] (2025) https://t.co/zZQPLbEJek https://t.co/Ny1xbwqCGM
Google has absolutely nailed this one! While I can’t share the results of my testing from Google Native Image output with Gemini 2.0 Flash just yet, I seriously can’t wait for the global rollout so you all can experience it. It’s incredible, beyond amazing. The amount of… https://t.co/VAzyMvhBRL