Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

NExT-OMNI: Towards Any-to-Any Omnimodal Foundation Models with Discrete Flow Matching

About

Next-generation multimodal foundation models capable of any-to-any cross-modal generation and multi-turn interaction will serve as core components of artificial general intelligence systems, playing a pivotal role in human-machine interaction. However, most existing multimodal models remain constrained by autoregressive architectures, whose inherent limitations prevent a balanced integration of understanding and generation capabilities. Although hybrid and decoupling strategies have been explored to address these tasks within unified frameworks separately, their redundant, non-integrated designs limit their applicability to broader scenarios, such as cross-modal retrieval. In this work, we introduce NExT-OMNI, an open-source omnimodal foundation model that achieves unified modeling through discrete flow paradigms. By leveraging metric-induced probability paths and kinetic optimal velocities, NExT-OMNI natively supports any-to-any understanding and generation with enhanced response efficiency, while enabling broader application scenarios through concise unified representations rather than task-decoupled designs. Trained on large-scale interleaved text, image, video, and audio data, NExT-OMNI delivers competitive performance on multimodal generation and understanding benchmarks, while outperforming prior unified models in multi-turn multimodal interaction and cross-modal retrieval, highlighting its architectural advantages as a next-generation multimodal foundation model. To advance further research, we release training details, data protocols, and open-source both the code and model checkpoints.

Run Luo, Xiaobo Xia, Lu Wang, Longze Chen, Renke Shan, Jing Luo, Min Yang, Tat-Seng Chua• 2025

Related benchmarks

TaskDatasetResultRank
Object Hallucination EvaluationPOPE
Accuracy87.4
1455
Multimodal UnderstandingMMBench
Accuracy78.9
637
Multimodal UnderstandingMMMU
Accuracy43.7
437
Multimodal PerceptionMME Perception
Perception Score1.54e+3
79
Text-to-SpeechLibriSpeech clean (test)
WER3.1
66
Image GenerationGenEval 129
Overall Performance Score85
15
Image GenerationDPG-Bench 130
Score84.2
15
Automatic Speech RecognitionLibriSpeech 80 (test-clean)
WER3.1
13
Automatic Speech RecognitionLibriSpeech 80 (test-other)
WER7
10
Showing 9 of 9 rows

Other info

Follow for update