
Benchmark results for the Llama 3.1 model series have reportedly been leaked on Azure's GitHub account, showcasing three versions: 405 billion parameters, 70 billion parameters, and 8 billion parameters. The 405 billion parameter model is particularly notable, as it is suggested to outperform the GPT-4o model while being significantly smaller in size. The 70 billion parameter version is claimed to match GPT-4's performance, raising expectations for its efficiency and cost-effectiveness. The authenticity of these benchmarks remains unconfirmed, and discussions surrounding their implications are ongoing within the AI community.
Potential benchmark leaks for a new series of Llama 3.1 models, including a 405 bn param version. Unconfirmed, however the 70Bb one matching GPT-4 levels. Specially, at its size of being 6x smaller. https://t.co/py2hP8ForD
Llama-3 400b benchmarks got leaked on Reddit! And guess what? It outperforms GPT-4o. 🔥 https://t.co/s6tYjPOOGr
Imagine a gpt-4o level model running at 330 tokens/second, 10x cheaper This will happen very soon, with @GroqInc hosting Llama 3.1 Crazy https://t.co/oJ4tgYovfY