Vision-Language Foundation Models as Effective Robot Imitators
The paper presents RoboFlamingo, a novel framework for robot manipulation using vision-language models (VLMs). The researchers used pre-trained VLMs and fine-tuned them on robotics data, resulting in a system that…
Continue reading