
NVIDIA has introduced NV-Embed, a generalist embedding model that has achieved the top position on the Massive Text Embedding Benchmark (MTEB) leaderboard. NV-Embed, which utilizes decoder-only large language models (LLMs), has outperformed BERT and T5-based models in general-purpose text embedding tasks. The model incorporates several architectural improvements, including latent attention pooling and a two-stage contrastive instruction-tuning. NV-Embed scored 59.36 on 15 retrieval tasks within the MTEB benchmark, using only publicly available data. This development is expected to enhance the performance of decoder-only LLMs like Mistral-7B, maintaining simplicity and reproducibility. The model was developed by a team including C Lee, R Roy, M Xu, and J Raiman.





Nvidia just released the weights of NV-Embed-v1, the current leader on MTEB. Some additional info below: Base model: Mistral-7B-v0.1 Pooling type: Latent-Attention Embedding dimension: 4096 Max input tokens: 32k https://t.co/mTH5wwxtjO
NV-Embed: NVIDIA’s Groundbreaking Embedding Model Dominates MTEB Benchmarks #DL #AI #ML #DeepLearning #ArtificialIntelligence #MachineLearning #ComputerVision #AutonomousVehicles #NeuroMorphic #Robotics https://t.co/E5SYh5KgJD
🙋🏻♂️@NVIDIAAI just released a cool EMBEDDING model , that's "instruction" based ! that means you can use it with @llama_index instructor embeddings function, or try it out on @huggingface directly with my little @Gradio demo : https://t.co/2FU3MwLwU2 this is the future !