Meta has launched its latest AI model, Llama 3.2, which features enhanced multimodal capabilities. The model is now available on various platforms, including SambaNova Cloud and Oracle's OCI Data Science. Additionally, a new course titled 'Introducing Multimodal Llama 3.2' has been released on DeepLearningAI, created in collaboration with Meta and featuring insights from prominent AI experts Andrew Ng and Asangani. The course covers both Llama 3.1 and Llama 3.2, focusing on multimodal prompting and custom tool calling, and is available for free. This development signifies a notable advancement in open-source AI models, which are increasingly competing with commercial counterparts such as GPT-4v and Gemini.
Multi-modal LLM development is accelerating. While commercial models like GPT-4v and Gemini capture attention, open-source models such as LLaVA and NVLM are rapidly closing the gap in performance on public benchmarks. 'A Walkthrough of Nvidia’s Latest Multi-Modal LLM Family' by…
We're very excited to power @AIatMeta's latest course on Llama 3.2 on @DeepLearningAI! All course modules are powered by Together AI's inference! https://t.co/7K1xxSNJBx
📣 New course now available on @DeepLearningAI: Introducing Multimodal Llama 3.2! The course covers both Llama 3.1 & Llama 3.2 and includes detailed rundowns on multimodal prompting, custom tool calling, Llama Stack + more. Take the 1h course for free ⬇️ https://t.co/jA5AKeWfoc