Loading…
Causal Interventional Training for Image Recognition
Deep learning models often fit undesired dataset bias in training. In this paper, we formulate the bias using causal inference , which helps us uncover the ever-elusive causalities among the key factors in training, and thus pursue the desired causal effect without the bias. We start from revisiting...
Saved in:
Published in: | IEEE transactions on multimedia 2023, Vol.25, p.1033-1044 |
---|---|
Main Authors: | , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Deep learning models often fit undesired dataset bias in training. In this paper, we formulate the bias using causal inference , which helps us uncover the ever-elusive causalities among the key factors in training, and thus pursue the desired causal effect without the bias. We start from revisiting the process of building a visual recognition system, and then propose a structural causal model (SCM) for the key variables involved in dataset collection and recognition model: object, common sense, bias, context, and label prediction. Based on the SCM, one can observe that there are "good" and "bad" biases. Intuitively, in the image where a car is driving on a high way in a desert, the "good" bias denoting the common-sense context is the highway, and the "bad" bias accounting for the noisy context factor is the desert. We tackle this problem with a novel causal interventional training ( CIT ) approach, where we control the observed context in each object class. We offer theoretical justifications for CIT and validate it with extensive classification experiments on CIFAR-10, CIFAR-100 and ImageNet, e.g. , surpassing the standard deep neural networks ResNet-34 and ResNet-50, respectively, by 0.95% and 0.70% accuracies on the ImageNet. Our code is open-sourced on the GitHub https://github.com/qinwei-hfut/CIT . |
---|---|
ISSN: | 1520-9210 1941-0077 |
DOI: | 10.1109/TMM.2021.3136717 |