
Google's MediaPipe/Core ML team launched a new Large Language Model (LLM) API for the Gemma 2.5B parameter model, which can run in browsers using WebAI. This development is part of a broader movement towards enhancing machine learning (ML) capabilities directly within web browsers, as demonstrated by other recent advancements. The 🤗 Transformers.js WebGPU Embedding Benchmark was introduced, highlighting the potential for WebGPU to significantly speed up ML models running locally in browsers. Xenovacom's announcement about soon enabling Transformers.js models to utilize a WebGPU backend further emphasizes this trend. Additionally, the recent release of ONNX Runtime with WebGPU support has been noted to make in-browser ML significantly more efficient, with benchmarks showing up to a 40x speed increase. An application developed by Nigel Gebodh allows users to interact with several LLMs, including GoogleAI's Gemma, MistralAI's Mistral, and HuggingFace's Zephyr, showcasing the practical applications of these technologies. Furthermore, Xenovacom demonstrated the capability of running depth estimation with Depth Anything in under 200ms using Transformers.js and WebGPU, indicating the practical benefits of these advancements.
⚡️ Now with WebGPU support! ⚡️ Run depth estimation w/ Depth Anything in under 200ms, thanks to Transformers.js and WebGPU! Try it out yourself! 👇 https://t.co/uwNG3KutRL https://t.co/IO2pBcsWm6 https://t.co/Q6ZNb2avDC
💻 Have a chat with some of the latest #LLMs on the scene! @NigelGebodh's app lets you switch between @GoogleAI's #Gemma, @MistralAI's #Mistral, and @HuggingFace's #Zephyr models. 🎈 Try it out: https://t.co/OC6tIE9ZcL 🛠️ Learn how he built it: https://t.co/m1PJC4X0fN #AI #LLM https://t.co/vfr6HTt9Ka
WebGPU will change ML 🤯 With the recent release of ONNX Runtime with WebGPU, in-browser ML is about to change. We can now fully leverage GPUs to run ML models (think of Phi, SD, etc) entirely in the browser Benchmark in my computer: 40x faster ⚡️ https://t.co/rJDXTTcyBX https://t.co/kMIcFZOmuj


