ML ICML

Testing Group Fairness via Optimal Transport Projections

May 24, 2021
We present a statistical testing framework to detect if a given machine learning classifier fails to satisfy a wide range of group fairness notions. The proposed test is a flexible, interpretable, and statistically rigorous tool for auditing whether exhibited biases are intrinsic to the algorithm or due to the randomness in the data. The statistical challenges, which may arise from multiple impact criteria that define group fairness and which are discontinuous on model parameters, are conveniently tackled by projecting the empirical measure onto the set of group-fair probability models using optimal transport. This statistic is efficiently computed using linear programming and its asymptotic distribution is explicitly obtained. The proposed framework can also be used to test for testing composite fairness hypotheses and fairness with multiple sensitive attributes. The optimal transport testing formulation improves interpretability by characterizing the minimal covariate perturbations that eliminate the bias observed in the audit.

Overall

< 1 minute

Nian Si, Karthyek Murthy, Jose Blanchet, Viet Anh Nguyen

ICML 2021

Share Article

Related publications

ML ICLR Top Tier
February 19, 2024

Nguyen Hung-Quang, Yingjie Lao, Tung Pham, Kok-Seng Wong, Khoa D Doan

CV ML AAAI Top Tier
January 8, 2024

Tran Huynh Ngoc, Dang Minh Nguyen, Tung Pham, Anh Tran

ML AAAI Top Tier
January 8, 2024

Viet Nguyen*, Giang Vu*, Tung Nguyen Thanh, Khoat Than, Toan Tran

ML NeurIPS Top Tier
October 4, 2023

Van-Anh Nguyen, Trung Le, Anh Tuan Bui, Thanh-Toan Do, Dinh Phung