Open Access Open Access  Restricted Access Subscription Access

An Empirical Evaluation of Bias Mitigation Techniques and Their Trade-offs with Predictive Performance in Machine Learning Models

Mission Franklin

Abstract


Machine learning models are increasingly applied in high-stakes domains, where concerns about fairness and accountability have become more critical. Although these models deliver strong predictive performance, they often inherit biases embedded in historical data, which can lead to unequal outcomes across different demographic groups. To address this issue, a variety of bias mitigation techniques have been introduced at different stages of the machine learning pipeline, including pre-processing methods that adjust the data, in-processing approaches that modify learning algorithms, and post-processing techniques that refine model outputs. However, improving fairness is not without challenges, as it may come at the expense of predictive accuracy, creating a trade-off that is not yet fully understood in practice. This study systematically evaluates the effectiveness of different bias mitigation strategies and examines their impact on model performance. Using benchmark datasets with sensitive attributes, baseline models are compared with models incorporating mitigation techniques. Performance is assessed using standard accuracy metrics, while fairness is measured through established statistical indicators. The findings reveal that some mitigation approaches enhance fairness but may reduce accuracy or affect model stability, while others improve generalization. Overall, the study provides valuable insights for selecting appropriate strategies in responsible machine learning applications.


Full Text:

PDF

References


Agarwal, A., Beygelzimer, A., Dudík, M., Langford, J., & Wallach, H. (2018). A reductions approach to fair classification. Proceedings of the 35th International Conference on Machine Learning, 60–69.

Barocas, S., & Selbst, A. D. (2016). Big data’s disparate impact. Columbia Law Review, 116(3), 671–732.

Chouldechova, A. (2017). Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big Data, 5(2), 153–163. https://doi.org/10.1089/big.2016.0047

Dwork, C., Hardt, M., Pitassi, T., Reingold, O., & Zemel, R. (2012). Fairness through awareness. Proceedings of the 3rd Innovations in Theoretical Computer Science Conference, 214–226.

Feldman, M., Friedler, S. A., Moeller, J., Scheidegger, C., & Venkatasubramanian, S. (2015). Certifying and removing disparate impact. Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 259–268. https://doi.org/10.1145/2783258.2783311

Friedler, S. A., Scheidegger, C., Venkatasubramanian, S., Choudhary, S., Hamilton, E. P., & Roth, D. (2019). A comparative study of fairness-enhancing interventions in machine learning. Proceedings of the Conference on Fairness, Accountability, and Transparency, 329–338.

Hardt, M., Price, E., & Srebro, N. (2016). Equality of opportunity in supervised learning. Advances in Neural Information Processing Systems, 29, 3315–3323.

Kamiran, F., & Calders, T. (2012). Data preprocessing techniques for classification without discrimination. Knowledge and Information Systems, 33(1), 1–33. https://doi.org/10.1007/s10115-011-0463-8

Kleinberg, J., Mullainathan, S., & Raghavan, M. (2017). Inherent Trade Offs in the Fair Determination of Risk Scores. In 8th Innovations in Theoretical Computer Science Conference (ITCS 2017) (pp. 43:1–43:23). Schloss Dagstuhl – Leibniz Zentrum für Informatik. https://doi.org/10.4230/LIPIcs.ITCS.2017.43

Mehrabi, N., Morstatter, F., Saxena, N., Lerman, K., & Galstyan, A. (2021). A survey on bias and fairness in machine learning. ACM Computing Surveys, 54(6), 1–35. https://doi.org/10.1145/3457607


Refbacks

  • There are currently no refbacks.