Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

BendVLM: Test-Time Debiasing of Vision-Language Embeddings

About

Vision-language model (VLM) embeddings have been shown to encode biases present in their training data, such as societal biases that prescribe negative characteristics to members of various racial and gender identities. VLMs are being quickly adopted for a variety of tasks ranging from few-shot classification to text-guided image generation, making debiasing VLM embeddings crucial. Debiasing approaches that fine-tune the VLM often suffer from catastrophic forgetting. On the other hand, fine-tuning-free methods typically utilize a "one-size-fits-all" approach that assumes that correlation with the spurious attribute can be explained using a single linear direction across all possible inputs. In this work, we propose Bend-VLM, a nonlinear, fine-tuning-free approach for VLM embedding debiasing that tailors the debiasing operation to each unique input. This allows for a more flexible debiasing approach. Additionally, we do not require knowledge of the set of inputs a priori to inference time, making our method more appropriate for online, open-set tasks such as retrieval and text guided image generation.

Walter Gerych, Haoran Zhang, Kimia Hamidieh, Eileen Pan, Maanas Sharma, Thomas Hartvigsen, Marzyeh Ghassemi• 2024

Related benchmarks

TaskDatasetResultRank
Bias Mitigation for Stereotype QueriesUTKFACE Race
KL Divergence0.041
9
Bias Mitigation for Stereotype QueriesUTKFACE Gender
KL Divergence0.004
9
Fair Image RetrievalCelebA (test)
KL Divergence0.011
9
Stereotype Query DebiasingCelebA
KL Divergence0.014
9
DebiasingFAIRFACE Race (test)
KL Divergence0.069
8
DebiasingFAIRFACE Gender (test)
KL Divergence0.006
8
Image CaptioningFairFace (val)
Score (White)0.355
2
Showing 7 of 7 rows

Other info

Code

Follow for update