

Groq Inc., a new player in the AI market, has unveiled its Mixtral model powered by Language Processing Unit (LPU) technology, boasting unprecedented processing speeds of nearly 500 tokens per second (~500 tokens/s). This development marks a significant leap in the generative AI sector, offering near-instantaneous response times and opening up new possibilities for user experience in various applications. The LPU's efficiency is further highlighted by its cost-effectiveness, with a reported price of 27 cents per 1 million tokens, making it a competitive alternative to existing technologies like GPT 3.5. The technology's potential was showcased in a public demo, demonstrating the LPU's ability to generate detailed, factual answers within less than a second, significantly reducing the time spent on data searching. Groq's innovation has been developed by a team including former Google TPU team members, indicating a strong pedigree in AI hardware development. The company's breakthrough has been met with enthusiasm from the AI community, with many noting its potential to revolutionize the way we interact with AI models by significantly reducing latency and making real-time conversations more feasible.
Groq, the new competitor of ChatGPT, has rolled out a Language Processing Unit (LPU). This cutting-edge technology boasts speeds close to 500 tokens per second. It sets a new benchmark for speed and efficiency in digital processing. https://t.co/c87F7RkU5F
Groq going multimodal! when we see their inference speed for llm, for other modality it ll be 💥 https://t.co/VUnWToj0s8
.@GroqInc vs @Google Gemini vs @OpenAI ChatGPT-4 Comparing time to complete answer for a simple code debugging question. Groq wins on speed (10x faster than Gemini, 18x faster than ChatGPT). Gemini wins on quality of answer, though perhaps it went overboard. https://t.co/WqG5t5u1m4