On December 12, during the 2025 Global Developers Pioneer Conference, Omilab proudly presented two significant open-source innovations: VLM-FO1 and OpenTrackVLA. Additionally, it marked the inaugural offline launch of the OmAgent developer community.
VLM-FO1 stands as a plug-and-play framework designed to enhance visual-language models (VLMs). It boasts versatility in supporting a wide array of tasks, encompassing intricate visual reasoning, open-vocabulary object detection, and video object tracking.
On the flip side, OpenTrackVLA emerges as a fully open-source visual-language-action (VLA) system. It seamlessly blends natural language directives, real-time visual perception, and robotic action planning into a cohesive whole.
