
Salesforce AI Research has introduced xGen-MM, a series of large multimodal models (LMMS). The xGen-MM models, including BLIP3, are state-of-the-art (SOTA) models under 5 billion parameters, excelling in few-shot learning and multimodal benchmarks. These models are designed to be open-source and are available on platforms like Hugging Face. xGen-MM is notable for its performance in both pre-trained and fine-tuned benchmarks, setting a new standard in the field.
XGen-MM: A Series of Large Multimodal Models (LMMS) Developed by Salesforce Al Research Quick read: https://t.co/XPiWQb0xSL Model: https://t.co/16nkXlMOzO #ArtificialIntelligence #ai @SFResearch
Excited to share that our newest multi modal foundation model xGen-mm is out and it’s open source! It’s small (<5B models) and shining in both pre-trained and fine-tuned benchmarks. Check it out 👉@huggingface: https://t.co/SkcXLm9bMu @SFResearch #SalesforceAI #AI #ML
Xmodel-VLM A Simple Baseline for Multimodal Vision Language Model We introduce Xmodel-VLM, a cutting-edge multimodal vision language model. It is designed for efficient deployment on consumer GPU servers. Our work directly confronts a pivotal industry issue by grappling https://t.co/mmbJwgbqAZ
