Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Connecting Joint-Embedding Predictive Architecture with Contrastive Self-supervised Learning

About

In recent advancements in unsupervised visual representation learning, the Joint-Embedding Predictive Architecture (JEPA) has emerged as a significant method for extracting visual features from unlabeled imagery through an innovative masking strategy. Despite its success, two primary limitations have been identified: the inefficacy of Exponential Moving Average (EMA) from I-JEPA in preventing entire collapse and the inadequacy of I-JEPA prediction in accurately learning the mean of patch representations. Addressing these challenges, this study introduces a novel framework, namely C-JEPA (Contrastive-JEPA), which integrates the Image-based Joint-Embedding Predictive Architecture with the Variance-Invariance-Covariance Regularization (VICReg) strategy. This integration is designed to effectively learn the variance/covariance for preventing entire collapse and ensuring invariance in the mean of augmented views, thereby overcoming the identified limitations. Through empirical and theoretical evaluations, our work demonstrates that C-JEPA significantly enhances the stability and quality of visual representation learning. When pre-trained on the ImageNet-1K dataset, C-JEPA exhibits rapid and improved convergence in both linear probing and fine-tuning performance metrics.

Shentong Mo, Shengbang Tong• 2024

Related benchmarks

TaskDatasetResultRank
Semantic segmentationADE20K (val)
mIoU38.68
2731
Semantic segmentationADE20K
mIoU48.7
936
Image ClassificationImageNet-1K
Top-1 Acc86.2
836
Instance SegmentationCOCO
APmask45.3
279
Object DetectionCOCO
AP (Box)50.7
144
Video Object SegmentationDAVIS--
58
Instance SegmentationMS-COCO (val)--
16
Object CountingCLEVR
Count Accuracy86.8
2
Depth PredictionCLEVR
CLEVR Dist Error71.6
2
Showing 9 of 9 rows

Other info

Follow for update