Artificial Intelligence

This AI paper introduces VLM-R³: a multimodal framework for region recognition, reasoning and improvement

Multimodal reasoning capabilities can help machines perform tasks such as solving mathematical problems embedded in charts, readings in photos, or explaining scientific charts. The integration of visual and linguistic information makes these systems more closely reflect the human thinking process, making it suitable for tasks that require visual interpretation to be combined with logical progress.

The main challenge in this field is that the current system cannot revisit specific parts of the image while dynamic inference. Traditional models usually start with analyzing the image at once and then perform the rest of the reasoning in plain text. This approach limits the need to revisit the image to confirm details or extract new visual cues during the midterm. These disadvantages are particularly evident in tasks requiring fine-grained spatial awareness, such as identifying small labels in scientific documents or resolving ambiguity in visually complex scenarios.

Some tools and models have been introduced to address this gap, but they often view visual grounding as a one-time operation. For example, existing systems such as Llava-Cot or Qwen2.5-VL provide some visual text integration. Nevertheless, they do not allow repeated and selective query of parts of images based on an evolving reasoning process. Grounding, if the execution is usually static, and lacks flexibility to adapt according to intermediate reasoning steps. Furthermore, these methods do not train the model to determine the importance of a specific image area, resulting in limitations in complex problem solving.

Researchers at Peking University, Alibaba Group and Zeekr Intelligence Technology have launched a model called VLM-R³. The model addresses the challenge by allowing for a more interactive connection between vision and reasoning. It gives the model the ability to determine when visual clarification is needed, determine the exact image area for analysis, and reintegrate this visual content into the inference process. This approach mimics the human problem-solving approach, where one might zoom in on the chart or revisit the paragraph to verify details before making a decision. The structure of this model emphasizes iterative decision making by relying on visual evidence throughout the reasoning process.

To this end, the researchers constructed a dataset called Visuo Language Interleaving Basics (VLIR), which aims to train models through step-by-step interactions between images and text. VLM-R³ combines this data set and operates using a method called Enhanced Policy Optimization (R-GRPO) of Regional Conditions. The training strategy encourages the model to selectively focus on the content of the image, perform transformations such as cropping or scaling, and incorporate these changes into subsequent logical steps. It simulates how humans divert their attention across different visual elements based on their own thoughts. The architecture integrates a pipeline that can cyclic inference with visual inspection in real time, enhancing the system’s ability to interact with visual data during inference.

The results show that performance is strong across multiple benchmarks. On Mathvista, the model reached 70.4%, an increase from 68.2% in baseline. For MathVision, the improvement ranged from 25.1% to 30.2%. On ScienceQA, it improved by 14.3%, reaching 87.9% from 73.6% at baseline. On HallusionBench, the model scored 62.0%, outperforming others like Mulberry, which scored 54.1%. VLM-R³ also showed excellent results for the documentation understanding in DOCVQA, with a score of 96.8%. Comparisons show that even if it uses fewer parameters than Gemini-2 Flash or GPT-4O (GPT-4O), it has competing accuracy, especially in tasks requiring detailed visual analysis and interleaved inference.

This work clearly outlines the problem of how the model handles vision and proposes well-structured solutions in the inference process. By integrating an ongoing method of image analysis, researchers from Alibaba Group, Peking University and Zeekr have come up with a powerful idea of ​​looking again, thinking and perfecting. The proposed framework can significantly improve the accuracy of complex tasks and provide a blueprint for more powerful visually aware AI systems.


Check Paper and Github page. All credits for this study are to the researchers on the project. Also, please feel free to follow us twitter And don’t forget to join us 99K+ ml reddit And subscribe Our newsletter.


Nikhil is an intern consultant at Marktechpost. He is studying for a comprehensive material degree in integrated materials at the Haragpur Indian Technical College. Nikhil is an AI/ML enthusiast and has been studying applications in fields such as biomaterials and biomedical sciences. He has a strong background in materials science, and he is exploring new advancements and creating opportunities for contribution.

Related Articles

Leave a Reply

Your email address will not be published. Required fields are marked *

Back to top button