Acta Informatica Pragensia 2024, 13(3), 359-373 | DOI: 10.18267/j.aip.2412797

The Fairness Stitch: A Novel Approach for Neural Network Debiasing

Modar Sulaiman ORCID..., Kallol Roy ORCID...
Institute of Computer Science, University of Tartu, Tartu, Estonia

The pursuit of fairness in machine learning models has become increasingly crucial across various applications, including bank loan approval and face detection. Despite the widespread use of artificial intelligence algorithms, concerns persist regarding biases and discrimination within these models. This study introduces a novel approach, termed “The Fairness Stitch” (TFS), aimed at enhancing fairness in deep learning models by combining model stitching and training jointly, while incorporating fairness constraints. We evaluate the effectiveness of TFS through a comprehensive assessment using two established datasets, CelebA and UTKFace. The evaluation involves a systematic comparison with the existing baseline method, fair deep feature reweighting (FDR). Our analysis demonstrates that TFS achieves a better balance between fairness and performance compared to the baseline method (FDR). Specifically, our method shows significant improvements in mitigating biases while maintaining performance levels. These results underscore the promising potential of TFS in addressing bias-related challenges and promoting equitable outcomes in machine learning models. This research challenges conventional wisdom regarding the efficacy of the last layer in deep learning models for debiasing purposes. The findings suggest that integrating fairness constraints into our proposed framework (TFS) can lead to more effective mitigation of biases and contribute to fairer AI systems.

Keywords: Artificial intelligence; AI bias; Deep learning; Fairness in machine learning; Finetune; Model stitching; Overfitting.

Received: March 13, 2024; Revised: June 9, 2024; Accepted: June 17, 2024; Prepublished online: July 22, 2024; Published: August 22, 2024  Show citation

ACS AIP APA ASA Harvard Chicago Chicago Notes IEEE ISO690 MLA NLM Turabian Vancouver
Sulaiman, M., & Roy, K. (2024). The Fairness Stitch: A Novel Approach for Neural Network Debiasing. Acta Informatica Pragensia13(3), 359-373. doi: 10.18267/j.aip.241
Download citation

References

  1. Bansal, Y., Nakkiran, P., & Barak, B. (2021). Revisiting Model Stitching to Compare Neural Representations. In 35th Conference on Neural Information Processing Systems (NeurIPS 2021). NeurIPS.
  2. Beutel, A., Chen, J., Doshi, T., Qian, H., Wei, L., Wu, Y., Heldt, L., Zhao, Z., Hong, L., Chi, E. H., & Goodrow, C. (2019). Fairness in Recommendation Ranking through Pairwise Comparisons. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, (pp. 2212-2220). ACM. https://doi.org/10.1145/3292500.3330745 Go to original source...
  3. Beutel, A., Chen, J., Doshi, T., Qian, H., Woodruff, A., Luu, C., Kreitmann, P., Bischof, J., & Chi, E. H. (2019). Putting Fairness Principles into Practice. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, (pp. 453-459). ACM. https://doi.org/10.1145/3306618.3314234 Go to original source...
  4. Brodersen, K. H., Ong, C. S., Stephan, K. E., & Buhmann, J. M. (2010). The Balanced Accuracy and Its Posterior Distribution. In 2010 20th International Conference on Pattern Recognition, (pp. 3121-3124). IEEE. https://doi.org/10.1109/ICPR.2010.764 Go to original source...
  5. Caton, S., & Haas, C. (2024). Fairness in Machine Learning: A Survey. ACM Computing Surveys, 56(7), Article 166. https://doi.org/10.1145/3616865 Go to original source...
  6. Cherepanova, V., Nanda, V., Goldblum, M., Dickerson, J. P., & Goldstein, T. (2021). Technical challenges for training fair neural networks. arXiv (Cornell University). https://doi.org/10.48550/arxiv.2102.06764 Go to original source...
  7. Dwork, C., Hardt, M., Pitassi, T., Reingold, O., & Zemel, R. (2012). Fairness Through Awareness. In Proceedings of the 3rd Innovations in Theoretical Computer Science Conference, (pp. 214-226). Springer. https://doi.org/10.1145/2090236.2090255 Go to original source...
  8. Edwards, H., & Storkey, A. (2015). Censoring Representations with an Adversary. arXiv (Cornell University). https://doi.org/10.48550/arxiv.1511.05897 Go to original source...
  9. Fawcett, T. (2004). ROC Graphs: Notes and Practical Considerations for Researchers. Machine Learning, 31(1), 1-38.
  10. Gardner, J., Brooks, C., & Baker, R. (2019). Evaluating the Fairness of Predictive Student Models Through Slicing Analysis. In Proceedings of the 9th International Conference on Learning Analytics & Knowledge, (pp. 225-234). ACM. https://doi.org/10.1145/3303772.3303791 Go to original source...
  11. Goodfellow, I. J., Vinyals, O., & Saxe, A. M. (2014). Qualitatively characterizing neural network optimization problems. arXiv (Cornell University). https://doi.org/10.48550/arxiv.1412.6544 Go to original source...
  12. Hardt, M., Price, E., & Srebro, N. (2016). Equality of Opportunity in Supervised Learning. In 30th Conference on Neural Information Processing Systems (NIPS 2016), (pp. 1-9). NeurIPS.
  13. Hashimoto, T., Srivastava, M., Namkoong, H., & Liang, P. (2018). Fairness Without Demographics in Repeated Loss Minimization. In Proceedings of the 35th International Conference on Machine Learning, (pp. 1929-1938). PMLR.
  14. Jiang, R., Pacchiano, A., Stepleton, T., Jiang, H., & Chiappa, S. (2020). Wasserstein Fair Classification. In Proceedings of the 35th Uncertainty in Artificial Intelligence Conference, (pp. 862-872). PMLR.
  15. Kamishima, T., Akaho, S., & Sakuma, J. (2011). Fairness-aware Learning through Regularization Approach. In 2011 IEEE 11th International Conference on Data Mining Workshops, (pp. 643-650). IEEE. https://doi.org/10.1109/ICDMW.2011.83 Go to original source...
  16. Kearns, M., Neel, S., Roth, A., & Wu, Z.S. (2018). Preventing Fairness Gerrymandering: Auditing and Learning for Subgroup Fairness. In Proceedings of the 35th International Conference on Machine Learning, (pp. 2564-2572). PMLR.
  17. Kirichenko, P., Izmailov, P., & Wilson, A.G. (2023). Last Layer Re-Training Is Sufficient for Robustness to Spurious Correlations. In The Eleventh International Conference on Learning Representations (1-37). ICLR. https://openreview.net/forum?id=Zb6c8A-Fghk
  18. Kumar, A., Raghunathan, A., Jones, R., Ma, T., & Liang, P. (2022). Fine-Tuning can Distort Pretrained Features and Underperform Out-of-Distribution. arXiv (Cornell University). https://doi.org/10.48550/arxiv.2202.10054 Go to original source...
  19. Lahoti, P., Beutel, A., Chen, J., Lee, K., Prost, F., Thain, N., Wang, X., & Chi, E. H. (2020). Fairness without Demographics through Adversarially Reweighted Learning. arXiv (Cornell University). https://doi.org/10.48550/arxiv.2006.13114 Go to original source...
  20. Lee, Y., Chen, A. S., Tajwar, F., Kumar, A., Yao, H., Liang, P., & Finn, C. (2022). Surgical Fine-Tuning improves adaptation to distribution shifts. arXiv (Cornell University). https://doi.org/10.48550/arxiv.2210.11466 Go to original source...
  21. Lenc, K., & Vedaldi, A. (2018). Understanding image representations by measuring their equivariance and equivalence. International Journal of Computer Vision, 127(5), 456-476. https://doi.org/10.1007/s11263-018-1098-y Go to original source...
  22. Liu, Z., Luo, P., Wang, X., & Tang, X. (2015). Deep Learning Face Attributes in the Wild. In Proceedings of the IEEE International Conference on Computer Vision, (pp. 3730-3738). IEEE. https://doi.org/10.1109/iccv.2015.425 Go to original source...
  23. Mao, Y., Deng, Z., Yao, H., Ye, T., Kawaguchi, K., & Zou, J. (2023). Last-Layer Fairness Fine-tuning is Simple and Effective for Neural Networks. arXiv (Cornell University). https://doi.org/10.48550/arxiv.2304.03935 Go to original source...
  24. Narayanan, A. (2018). Translation Tutorial: 21 Fairness Definitions and Their Politics. https://www.youtube.com/watch?v=jIXIuYdnyyk
  25. Padala, M., & Gujar, S. (2020). FNNC: Achieving Fairness through Neural Networks. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence. IJCAI. https://www.ijcai.org/proceedings/2020/0315.pdf Go to original source...
  26. Park, S., Kim, D., Hwang, S., & Byun, H. (2020). README: REpresentation learning by fairness-Aware Disentangling MEthod. arXiv (Cornell University). https://doi.org/10.48550/arxiv.2007.03775 Go to original source...
  27. Parraga, O., More, M. D., Oliveira, C. M., Gavenski, N. S., Kupssinskü, L. S., Medronha, A., Moura, L. V., Simões, G. S., & Barros, R. C. (2023). Fairness in Deep Learning: A survey on vision and language research. ACM Computing Surveys, (in press). https://doi.org/10.1145/3637549 Go to original source...
  28. Rawls, J. (2001). Justice as Fairness: A Restatement. Harvard University Press. Go to original source...
  29. Shwartz-Ziv, R., & Tishby, N. (2017). Opening the Black Box of Deep Neural Networks via Information. arXiv (Cornell University). https://doi.org/10.48550/arXiv.1703.00810 Go to original source...
  30. Wan, M., Zha, D., Liu, N., & Zou, N. (2023). In-Processing Modeling Techniques for Machine Learning Fairness: A survey. ACM Transactions on Knowledge Discovery from Data, 17(3), 1-27. https://doi.org/10.1145/3551390 Go to original source...
  31. Zafar, M. B., Valera, I., Rogriguez, M. G., & Gummadi, K. P. (2019). Fairness Constraints: A Flexible Approach for Fair Classification. Journal of Machine Learning Research, 20(1), 1-42. http://jmlr.org/papers/v20/18-262.html
  32. Zafar, M. B., Valera, I., Rogriguez, M. G., & Gummadi, K. P. (2017). Fairness Constraints: Mechanisms for Fair Classification. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, (pp. 962-970). PMLR.
  33. Zhang, Z., Song, Y., & Qi, H. (2017). Age Progression/Regression by Conditional Adversarial Autoencoder. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, (pp. 5810-5818). IEEE. https://doi.org/10.1109/CVPR.2017.463 Go to original source...

This is an open access article distributed under the terms of the Creative Commons Attribution 4.0 International License (CC BY 4.0), which permits use, distribution, and reproduction in any medium, provided the original publication is properly cited. No use, distribution or reproduction is permitted which does not comply with these terms.