FairMOE: counterfactually-fair mixture of experts with levels of interpretability
Published in Machine Learning, 2024
In this paper, we extend our previous work arguing that interpretability should not be viewed as a binary aspect and that, instead, it should be viewed as a continuous domain-informed notion. Building on our prior work, we leverage the well-known Mixture of Experts architecture with a counterfactual fairness module to ensure the selection of consistently fair experts: FairMOE. We expand on the previous paper with a detailed analysis on the assignment of predictions to gain more insights into the strengths and weaknesses of the individual experts.
Recommended citation: Germino, Joe, Nuno Moniz, and Nitesh V. Chawla. "FairMOE: counterfactually-fair mixture of experts with levels of interpretability." Machine Learning (2024): 1-21. https://doi.org/10.1007/978-3-031-45275-8_23