Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ViT-Lens: Towards Omni-modal Representations

About

Aiming to advance AI agents, large foundation models significantly improve reasoning and instruction execution, yet the current focus on vision and language neglects the potential of perceiving diverse modalities in open-world environments. However, the success of data-driven vision and language models is costly or even infeasible to be reproduced for rare modalities. In this paper, we present ViT-Lens-2 that facilitates efficient omni-modal representation learning by perceiving novel modalities with a pretrained ViT and aligning them to a pre-defined space. Specifically, the modality-specific lens is tuned to project any-modal signals to an intermediate embedding space, which are then processed by a strong ViT with pre-trained visual knowledge. The encoded representations are optimized toward aligning with the modal-independent space, pre-defined by off-the-shelf foundation models. ViT-Lens-2 provides a unified solution for representation learning of increasing modalities with two appealing advantages: (i) Unlocking the great potential of pretrained ViTs to novel modalities effectively with efficient data regime; (ii) Enabling emergent downstream capabilities through modality alignment and shared ViT parameters. We tailor ViT-Lens-2 to learn representations for 3D point cloud, depth, audio, tactile and EEG, and set new state-of-the-art results across various understanding tasks, such as zero-shot classification. By seamlessly integrating ViT-Lens-2 into Multimodal Foundation Models, we enable Any-modality to Text and Image Generation in a zero-shot manner. Code and models are available at https://github.com/TencentARC/ViT-Lens.

Weixian Lei, Yixiao Ge, Kun Yi, Jianfeng Zhang, Difei Gao, Dylan Sun, Yuying Ge, Ying Shan, Mike Zheng Shou• 2023

Related benchmarks

TaskDatasetResultRank
Text-to-Video RetrievalMSR-VTT
Recall@137.6
313
3D Object ClassificationModelNet40 (test)--
302
3D Object ClassificationObjaverse-LVIS (test)
Top-1 Accuracy52
95
3D Point Cloud ClassificationScanObjectNN (test)--
92
Audio ClassificationVGG-Sound
Top-1 Accuracy31.7
50
Audio RetrievalAudioCaps
R@114.4
42
Environmental Sound ClassificationESC
Top-1 Acc75.9
28
Audio ClassificationAudioSet
mAP26.7
25
Audio RetrievalClotho
R@18.1
20
Depth Image ClassificationNYU-D
Top-1 Acc68.5
8
Showing 10 of 14 rows

Other info

Code

Follow for update