View a PDF of the paper titled Training Multi-Image Vision Agents via End2End Reinforcement Learning, by Chengqi Dong and 10 other authors
View PDF
HTML (experimental)
Abstract:Recent VLM-based agents aim to replicate OpenAI O3’s “thinking with images” via tool use, but most open-source methods limit input to a single image, falling short on real-world multi-image QA tasks. To address this, we propose IMAgent, an open-source vision agent trained via end-to-end reinforcement learning dedicated for complex multi-image tasks. By leveraging a multi-agent system, we generate challenging and visually-rich multi-image QA pairs to fully activate the tool-use potential of the base VLM. Through manual verification, we obtain MIFG-QA, comprising 10k samples for training and evaluation. With deeper reasoning steps, VLMs may increasingly ignore visual inputs. We therefore develop two specialized tools for visual reflection and confirmation, allowing the model to proactively reallocate its attention to image content during inference. Benefiting from our well-designed action-trajectory two-level mask strategy, IMAgent achieves stable tool use behavior via pure RL training without requiring costly supervised fine-tuning data. Extensive experiments demonstrate that IMAgent maintains strong performance on existing single-image benchmarks while achieving substantial improvements on our proposed multi-image dataset, with our analysis providing actionable insights for the research community. Codes and data will be released soon.
Submission history
From: ChengQi Dong [view email]
[v1]
Fri, 5 Dec 2025 10:02:38 UTC (9,852 KB)
[v2]
Tue, 16 Dec 2025 14:00:19 UTC (10,259 KB)


![[2512.08980] Training Multi-Image Vision Agents via End2End Reinforcement Learning Measuring Intelligence Efficiency of Local AI](https://skytik.cc/wp-content/uploads/2025/11/Measuring-Intelligence-Efficiency-of-Local-AI-768x448.png)