
OpenVLA, an open-source Vision-Language-Action model, presents a 7-billion parameter model pretrained on 970k robot episodes. It outperforms RT-2-X and Octo, focusing on finetuning and efficient inference, achieving RT-2X performance at 7x fewer parameters. The model aims to change how robots learn new skills by leveraging a combination of Internet-scale vision-language data and diverse robot demonstrations.
OpenVLA is a VLM for robot control, open-source & available for the community: https://t.co/5jmeX2pqN5 Awesome collaboration led by @moo_jin_kim, @KarlPertsch, @siddkaramcheti W.r.t. large-scale robotic learning, this is an important step in making VLAs accessible. A thread 👇 https://t.co/03HChp0j7a
🚨New fully open multi-robot generalist VLA🚨 OpenVLA makes accessible one of the most important paradigms in robotics + AI today, VLAs. - works 0-shot on many robot embodiments - focuses on finetuning and efficient inference - RT-2X performance (!) at 7x fewer params (!!) https://t.co/ieT2BbHsuy
OpenVLA: An Open-Source Vision-Language-Action Model - Presents a 7B open-source vision-language-action model, pretrained on 970k robot episodes from the Open X-Embodiment dataset - Outperforms RT-2-X and Octo proj: https://t.co/wdTFFhAyIK abs: https://t.co/alDZtgK6dQ https://t.co/T3DBlG55QH




