Re-Evaluating the Machine Learning Pipeline to Improve Fairness and Reliability

Project Description

Fairness in machine learning continues to be a highly relevant issue, with unfair models making headlines on a regular basis. In this project, we re-evaluate the complete machine learning pipeline from the sourcing of data, over the design of ML systems all the way to their implementation with an eye on algorithmic fairness and robustness. We develop new methodologies and collect data to better understand the reliability of findings in the field. We further critically examine the usage and composition of datasets, highlighting gaps and providing recommendations for more sustainable practices. Among other things, we focus on the influence of design decisions, highlighting potential issues of fairness hacking and introducing a new methodology to systematically study and address issues of reliability.

Contact Person

Jan Simson

Publications

  • Simson, J., Pfisterer, F., & Kern, C. (2023). Everything, Everywhere All in One Evaluation: Using Multiverse Analysis to Evaluate the Influence of Model Design Decisions on Algorithmic Fairness (arXiv:2308.16681). arXiv. https://doi.org/10.48550/arXiv.2308.16681