paint-brush
Mathematical Proofs for Fair AI Bias Analysisby@demographic
New Story

Mathematical Proofs for Fair AI Bias Analysis

by DemographicMarch 25th, 2025
Read on Terminal Reader
Read this story w/o Javascript
tldt arrow

Too Long; Didn't Read

This appendix presents formal proofs supporting the findings on inductive bias in DP-based fair learning. Additional results for the CelebA dataset show visualized biases in AI classifiers, with experiments using KDE and MI-based models on ResNet-18.

People Mentioned

Mention Thumbnail
featured image - Mathematical Proofs for Fair AI Bias Analysis
Demographic HackerNoon profile picture
0-item

Abstract and 1 Introduction

2 Related Works

3 Preliminaries

3.1 Fair Supervised Learning and 3.2 Fairness Criteria

3.3 Dependence Measures for Fair Supervised Learning

4 Inductive Biases of DP-based Fair Supervised Learning

4.1 Extending the Theoretical Results to Randomized Prediction Rule

5 A Distributionally Robust Optimization Approach to DP-based Fair Learning

6 Numerical Results

6.1 Experimental Setup

6.2 Inductive Biases of Models trained in DP-based Fair Learning

6.3 DP-based Fair Classification in Heterogeneous Federated Learning

7 Conclusion and References

Appendix A Proofs

Appendix B Additional Results for Image Dataset

Appendix A Proofs

A.1 Proof of Theorem 1


Therefore, for the objective function in Equation (1), we can write the following:




Knowing that TV is a metric distance satisfying the triangle inequality, the above equations show that



Therefore,


A.2 Proof of Theorem 2





A.3 Proof of Theorem 3



Therefore, we can follow the proof of Theorems 1,2 which shows the above inequality leads to the bounds claimed in the theorems.

Appendix B Additional Results for Image Dataset

This part shows the inductive biases of DP-based fair classifier for CelebA dataset, as well as the visualized plots. For the baselines, two fair classifiers are implemented for image fair classification: KDE proposed by [11] and MI proposed by [6], based on ResNet-18 [28].



Figure 5: The results of Figure 2’s experiments for a ResNet-based model on image dataset.




Figure 6: Blond hair samples (Majority, Upper) and Non-blond hair samples (Minority, Lower) in CelebA Dataset predicted by ERM(NN) and MI respectively. The results show that the model has 57.3% and 98.8% negative rates, i.e. prefers to predict all samples being female in Minority, even maintaining almost the same level of accuracy in the whole group.



This paper is available on arxiv under CC BY-NC-SA 4.0 DEED license.


Authors:

(1) Haoyu LEI, Department of Computer Science and Engineering, The Chinese University of Hong Kong (hylei22@cse.cuhk.edu.hk);

(2) Amin Gohari, Department of Information Engineering, The Chinese University of Hong Kong (agohari@ie.cuhk.edu.hk);

(3) Farzan Farnia, Department of Computer Science and Engineering, The Chinese University of Hong Kong (farnia@cse.cuhk.edu.hk).