Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Characterizing and overcoming the greedy nature of learning in multi-modal deep neural networks

About

We hypothesize that due to the greedy nature of learning in multi-modal deep neural networks, these models tend to rely on just one modality while under-fitting the other modalities. Such behavior is counter-intuitive and hurts the models' generalization, as we observe empirically. To estimate the model's dependence on each modality, we compute the gain on the accuracy when the model has access to it in addition to another modality. We refer to this gain as the conditional utilization rate. In the experiments, we consistently observe an imbalance in conditional utilization rates between modalities, across multiple tasks and architectures. Since conditional utilization rate cannot be computed efficiently during training, we introduce a proxy for it based on the pace at which the model learns from each modality, which we refer to as the conditional learning speed. We propose an algorithm to balance the conditional learning speeds between modalities during training and demonstrate that it indeed addresses the issue of greedy learning. The proposed algorithm improves the model's generalization on three datasets: Colored MNIST, ModelNet40, and NVIDIA Dynamic Hand Gesture.

Nan Wu, Stanis{\l}aw Jastrz\k{e}bski, Kyunghyun Cho, Krzysztof J. Geras• 2022

Related benchmarks

TaskDatasetResultRank
Action RecognitionUCF-101
Top-1 Acc81.25
147
Emotion RecognitionCH-SIMS
Accuracy (5-Class)55.14
26
Emotion RecognitionMOSEI
Accuracy (7-Class)43.07
26
Emotion RecognitionCH-SIMS 2
Accuracy (5-class)42.55
26
Emotion RecognitionMOSI
Accuracy (7-Class)22.87
26
Multimodal ClassificationSymile
AUROC0.6419
24
Multimodal ClassificationHAIM
AUROC0.7072
24
Emotion RecognitionCREMA-D
Accuracy (6)65.35
23
Multimodal ClassificationINSPECT
AUROC62.33
22
Multimodal ClassificationUKB
AUROC0.794
21
Showing 10 of 14 rows

Other info

Follow for update