
Alibaba has officially launched Qwen2, an advanced open-source AI model series that includes both base and instruct models in five sizes: 0.5B, 1.5B, 7B, 57B-A14B, and 72B parameters. These models have been trained on data in 27 additional languages and are designed to excel in code and math capabilities. Qwen2-72B, the largest model, outperforms Meta's Llama 3 70B and has a 128K context window. The models are available under the Apache 2.0 license, except for the 72B model. Qwen2 models have achieved impressive results on the Open LLM Leaderboard and MixEval, with the Qwen2-72B-Instruct model leading in several key evaluations, including an MMLU score of nearly 84 and a GSM8K score of 85. The Qwen2-72B-Instruct model is comparable to GPT-4o with an MMLU score of 84.32. The models are also open-sourced with MLX support and can be accessed on Hugging Face. Additionally, Qwen2 can be deployed using SkyPilot.
Absolutely, incredibly inspiring to watch the battle of Alibaba, a Chinese company rivaling GPT-4 with a SOTA open-source model 🤯 Qwen2-72B outperformed other leading open-source SOTA models in 15 benchmarks, including language understanding, language generation, multilingual… https://t.co/tBFnStjvTL https://t.co/waaxujF77h
Qwen2 72B Instruct from @Alibaba_Qwen released yesterday and is topping leaderboards! 🎉 Try it now on the Together API using Qwen/Qwen2-72B-Instruct or through the Playground at: https://t.co/0wUtP8qaYO 🚀 https://t.co/1DUhEgAr6L
The latest version of MLC LLM now supports the newly released model Qwen2! Run it effortlessly on a $100 OrangePi. With Qwen2 0.5B 17.5 tok/s, 1.5B 8.9 tok/s, AI capabilities are more accessible than ever. Explore more at MLC LLM https://t.co/FRLb1NxM7X #MLC #LLM #Qwen2 #OrangePi https://t.co/6aI9bZkGAc https://t.co/fOlaFJZmrY






