Today, SenseTime made an official announcement regarding the open-sourcing of its multimodal autonomous reasoning model, SenseNova-MARS. This model comes in two variants: 8B and 32B. In the core benchmark tests focused on multimodal search and reasoning, the model scored an impressive 69.74, outperforming both Gemini-3-Pro, which scored 69.06, and GPT-5.2, which scored 67.64. SenseNova-MARS stands out as the first Agentic VLM model that facilitates a seamless integration of dynamic visual reasoning with text-image search capabilities. This integration empowers the model to autonomously plan steps, invoke necessary tools, and effectively manage complex tasks, showcasing its 'executive capabilities.'
