Fair-CDA: Continuous and Directional Augmentation for Group Fairness
DOI:
https://doi.org/10.1609/aaai.v37i8.26183Keywords:
ML: Bias and FairnessAbstract
In this work, we propose Fair-CDA, a fine-grained data augmentation strategy for imposing fairness constraints. We use a feature disentanglement method to extract the features highly related to the sensitive attributes. Then we show that group fairness can be achieved by regularizing the models on transition paths of sensitive features between groups. By adjusting the perturbation strength in the direction of the paths, our proposed augmentation is controllable and auditable. To alleviate the accuracy degradation caused by fairness constraints, we further introduce a calibrated model to impute labels for the augmented data. Our proposed method does not assume any data generative model and ensures good generalization for both accuracy and fairness. Experimental results show that Fair-CDA consistently outperforms state-of-the-art methods on widely-used benchmarks, e.g., Adult, CelebA and MovieLens. Especially, Fair-CDA obtains an 86.3% relative improvement for fairness while maintaining the accuracy on the Adult dataset. Moreover, we evaluate Fair-CDA in an online recommendation system to demonstrate the effectiveness of our method in terms of accuracy and fairness.Downloads
Published
2023-06-26
How to Cite
Sun, R., Zhou, F., Dong, Z., Xie, C., Hong, L., Li, J., Zhang, R., Li, Z., & Li, Z. (2023). Fair-CDA: Continuous and Directional Augmentation for Group Fairness. Proceedings of the AAAI Conference on Artificial Intelligence, 37(8), 9918-9926. https://doi.org/10.1609/aaai.v37i8.26183
Issue
Section
AAAI Technical Track on Machine Learning III