
NVIDIA has announced the availability of new NVIDIA Inference Microservices (NIMs), including Mistral-7B, Mixtral-8x7B, and Mixtral-8x22B. These microservices are designed to provide optimized AI inference for large language models (LLMs). The new NIMs can be downloaded from the NVIDIA API Catalog. NVIDIA NIMs are part of NVIDIA AI Enterprise and offer a streamlined, scalable path for developing and deploying AI-powered enterprise applications. The high performance of Mixtral-8x7B is achieved with NVIDIA H100 Tensor Core GPUs and TensorRT-LLM, as highlighted in a recent NVIDIA Technical Blog.





High-performance GenAI apps with NVIDIA https://t.co/0k1aYz4d8V
New Blog Post from NVIDIA: "Mission NIMpossible: Decoding the Microservices That Accelerate Generative AI" https://t.co/ZjCKKRR0H8 https://t.co/8VX7Waay5w
Unlocking the Power of AI with @nvidia NIM NVIDIA has raised the bar for deploying AI models at scale with NVIDIA NIM (NVIDIA Inference Microservices). Part of NVIDIA AI Enterprise, NIM provides a streamlined, scalable path for developing and deploying AI-powered enterprise… https://t.co/8TNzMnViD8