MMMG: a Comprehensive and Reliable Evaluation Suite for Multitask Multimodal Generation
Abstract
MMMG is a comprehensive benchmark for multimodal generation, offering 49 tasks and 937 instructions to align automatic evaluation with human judgment, revealing areas for improvement in reasoning and audio generation.
Automatically evaluating multimodal generation presents a significant challenge, as automated metrics often struggle to align reliably with human evaluation, especially for complex tasks that involve multiple modalities. To address this, we present MMMG, a comprehensive and human-aligned benchmark for multimodal generation across 4 modality combinations (image, audio, interleaved text and image, interleaved text and audio), with a focus on tasks that present significant challenges for generation models, while still enabling reliable automatic evaluation through a combination of models and programs. MMMG encompasses 49 tasks (including 29 newly developed ones), each with a carefully designed evaluation pipeline, and 937 instructions to systematically assess reasoning, controllability, and other key capabilities of multimodal generation models. Extensive validation demonstrates that MMMG is highly aligned with human evaluation, achieving an average agreement of 94.3%. Benchmarking results on 24 multimodal generation models reveal that even though the state-of-the-art model, GPT Image, achieves 78.3% accuracy for image generation, it falls short on multimodal reasoning and interleaved generation. Furthermore, results suggest considerable headroom for improvement in audio generation, highlighting an important direction for future research.
Community
A comprehensive and reliable benchmark for multimodal generation ( image, audio, interleaved text, and image, interleaved text and audio)
✅ Each task has a carefully crafted automatic evaluation pipeline to ensure reliability
✅ Much more aligned with humans than other benchmarks
✅ Comprehensive: 4 modality combination, 49 tasks, 937 instructions
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- LMM4LMM: Benchmarking and Evaluating Large-multimodal Image Generation with LMMs (2025)
- Preliminary Explorations with GPT-4o(mni) Native Image Generation (2025)
- Emerging Properties in Unified Multimodal Pretraining (2025)
- FRABench and GenEval: Scaling Fine-Grained Aspect Evaluation across Tasks, Modalities (2025)
- OCR-Reasoning Benchmark: Unveiling the True Capabilities of MLLMs in Complex Text-Rich Image Reasoning (2025)
- OmniGenBench: A Benchmark for Omnipotent Multimodal Generation across 50+ Tasks (2025)
- MME-Unify: A Comprehensive Benchmark for Unified Multimodal Understanding and Generation Models (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 1
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper