Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

A High-Quality Dataset and Reliable Evaluation for Interleaved Image-Text Generation

About

Recent advancements in Large Multimodal Models (LMMs) have significantly improved multimodal understanding and generation. However, these models still struggle to generate tightly interleaved image-text outputs, primarily due to the limited scale, quality, and instructional richness of current training datasets. To address this, we introduce InterSyn, a dataset that features: (1) large scale, comprising 1.8M multimodal samples; (2) high quality, supported by our proposed Self-Evaluation with Iterative Refinement (SEIR) method for rigorous automated quality refinement; (3) rich instructional diversity, ensured through diverse well-designed question templates, based on human preferences and covering a 3500-topic hierarchy. These characteristics make InterSyn particularly well-suited for training LMMs in interactive image-text generation capabilities. To evaluate the capabilities, we propose SynJudge, a reliable automatic evaluator that aligns closely with human judge and outputs four interpretable scores: Text Content Completeness (TCC), Image Content Completeness (ICC), Image Quality (IQ), and Image-Text Synergy (ITS). These scores are complementary, covering both content and quality as well as cross-modal interaction, thereby forming a comprehensive evaluation framework. Experimental results on InterSyn subsets of up to 200K samples show that 25K-50K already yield substantial improvements, while scaling to 100K/200K brings further gains in TCC, ICC, and especially ITS, highlighting InterSyn's: (1) scalability, as performance consistently improves with more data; (2) efficiency, as significant gains are achievable even with smaller subsets, making it accessible to researchers with varying computational resources.

Yukang Feng, Jianwen Sun, Chuanhao Li, Zizhen Li, Jiaxin Ai, Fanrui Zhang, Yifan Chang, Sizhuo Zhou, Shenglin Zhang, Yu Dai, Kaipeng Zhang• 2025

Related benchmarks

TaskDatasetResultRank
Image Generation EvaluationImage Generation Evaluation IQ
A@198.3
60
Image Generation EvaluationICC
A@197.3
60
Image Generation EvaluationImage Generation Evaluation (ITS)
A@197.4
60
Image Generation EvaluationTCC
A@197.8
60
Interleaved Image-Text GenerationInterSyn (test)
TCC4.41
28
Judge Performance EvaluationSynJudge
TCC (RMSE)0.54
5
Showing 6 of 6 rows

Other info

Follow for update