Two New Benchmarks Advance Retrieval-Augmented Generation Research

Researchers introduce RAG-IGBench for evaluating interleaved image-text generation and M4-RAG for multilingual multimodal RAG systems.

Two New Benchmarks Advance Retrieval-Augmented Generation Research

Researchers have introduced two new evaluation frameworks aimed at advancing Retrieval-Augmented Generation (RAG) capabilities in different domains.

RAG-IGBench for Interleaved Generation

According to arXiv paper 2512.05119v1, RAG-IGBench addresses the need for “interleaved image-text generation” in open-domain question answering. The researchers note that “providing user queries with visually enhanced responses can considerably benefit understanding and memory, underscoring the great value of interleaved image-text generation.” The benchmark aims to evaluate RAG systems that can generate responses combining both text and images in an interleaved format.

M4-RAG for Multilingual Multimodal Systems

A separate paper (arXiv 2512.05959v1) introduces M4-RAG, described as “A Massive-Scale Multilingual Multi-Cultural Multimodal RAG.” According to the abstract, while “Vision-language models (VLMs) have achieved strong performance in visual question answering (VQA), yet they remain constrained by static training data.” The M4-RAG framework addresses this limitation by enabling “access to up-to-date” information through retrieval augmentation across multiple languages and cultural contexts.

Both benchmarks aim to push RAG technology beyond text-only applications into more complex multimodal scenarios.