The Open-R1 initiative has unveiled a series of new open-source AI models, notably the OlympicCoder, which includes two variants: a 7 billion parameter (7B) model and a 32 billion parameter (32B) model. The 7B model has demonstrated superior performance on the International Olympiad in Informatics (IOI) benchmark, outperforming the Claude 3.7 Sonnet model. The 32B variant has excelled against all tested open-weight models, including those significantly larger, marking a notable advancement in competitive programming capabilities. Additionally, the OLMo 2 32B model has been introduced, which is the first fully open model to surpass both GPT-3.5 and GPT-4o mini on a variety of multi-skill benchmarks, while requiring a fraction of the training compute typically needed for such performance. All datasets, training recipes, and benchmarks for these models are being made publicly available, further promoting open-source AI development.
A very exciting day for open-source AI! We're releasing our biggest open source model yet -- OLMo 2 32B -- and it beats the latest GPT 3.5, GPT 4o mini, and leading open weight models like Qwen and Mistral. As usual, all data, weights, code, etc. are available. For a long time,… https://t.co/S0HMk3Ll9o
Announcing OLMo 2 32B: the first fully open model to beat GPT 3.5 & GPT-4o mini on a suite of popular, multi-skill benchmarks. Comparable to best open-weight models, but a fraction of training compute. When you have a good recipe, ✨ magical things happen when you scale it up! https://t.co/z7ovubYFQe
When you have a good recipe, ✨ magical things happen when you scale it up! Announcing OLMo 2 32B: the first fully-open model to beat GPT 3.5 & GPT-4o mini on a suite of popular, multi-skill benchmarks. Comparable to best open-weight models, but fraction of training compute!