
Allen AI has released a new version of their OLMo model, OLMo 1.7-7B, which has shown significant improvements in performance. The model scores 52 on the MMLU, surpassing Llama 2-7B and approaching the performance of Llama 2-13B, with a 24 point increase over previous versions. It also excels on GSM8K, surpassing Llama 2-13B. Key enhancements include a longer context length of 4096 tokens and the use of the new Dolma 1.7 dataset, which has contributed to the improved data quality. The team attributes these advancements to better data management, staged training, and a focus on quality in pretraining procedures.
Allen AI team is moving fast. They're on the ~Pareto frontier~ now, and it's built in the open, so we all get to see how it's done! gratz @mechanicaldirk @soldni @natolambert https://t.co/pEKg6ahJbn
Introducing our best OLMo yet. OLMo 1.7-7B outperforms LLaMa2-7B, approaching LLaMa2-13B at MMLU and GSM8k. High-quality data and staged training are key. I am so proud of our team making such significant improvement in a short period after our first release. https://t.co/9NNwCxAwj6 https://t.co/YkrWiVtGqo
Great to see Allen AI iterating here. OLMo 1.7 is a solid step up, and with fully open data! https://t.co/UAbFNcICnA


