The University of Auckland
Browse

Bo Pang: CABIN: Debiasing Vision-Language Models Using Backdoor Adjustments

Download (2.24 MB)
Version 2 2025-10-19, 04:56
Version 1 2025-10-13, 00:37
poster
posted on 2025-10-19, 04:56 authored by Bo PangBo Pang, Tingrui Qiao, Caroline WalkerCaroline Walker, Yun Sing KohYun Sing Koh, Chris Cunningham
<p dir="ltr">Vision-language models (VLMs) have demonstrated strong zero-shot inference capabilities but may exhibit stereotypical biases toward certain demographic groups. Consequently, downstream tasks leveraging these models may yield unbalanced performance across different target social groups, potentially reinforcing harmful stereotypes. Mitigating such biases is critical for ensuring fairness in practical applications. Existing debiasing approaches typically rely on curated face-centric datasets for fine-tuning or retraining, risking overfitting and limiting generalizability. To address this issue, we propose a novel framework, CABIN (Causal Adjustment Based INtervention). It leverages a causal framework to identify sensitive attributes in images as confounding factors. Employing a learned mapper, which is trained on general large-scale image-text pairs rather than face-centric datasets, CABIN may use text to adjust sensitive attributes in the image embedding, ensuring independence between these sensitive attributes and image embeddings. This independence enables a backdoor adjustment for unbiased inference without the drawbacks of additional fine-tuning or retraining on narrowly tailored datasets. Through comprehensive experiments and analyses, we demonstrate that CABIN effectively mitigates biases and improves fairness metrics while preserving the zero-shot strengths of VLMs.</p>

History

Publisher

University of Auckland

Usage metrics

    University of Auckland

    Licence

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC