Impact on bias mitigation algorithms to variations in inferred sensitive attribute uncertainty
Concerns about the trustworthiness, fairness, and privacy of AI systems are growing, and strategies for mitigating these concerns are still in their infancy. One approach to improve trustworthiness and fairness in AI systems is to use bias mitigation algorithms. However, most bias mitigation algorit...
Saved in:
| Main Authors: | , |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
Frontiers Media S.A.
2025-03-01
|
| Series: | Frontiers in Artificial Intelligence |
| Subjects: | |
| Online Access: | https://www.frontiersin.org/articles/10.3389/frai.2025.1520330/full |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
| _version_ | 1849705907931316224 |
|---|---|
| author | Yanchen Wang Lisa Singh |
| author_facet | Yanchen Wang Lisa Singh |
| author_sort | Yanchen Wang |
| collection | DOAJ |
| description | Concerns about the trustworthiness, fairness, and privacy of AI systems are growing, and strategies for mitigating these concerns are still in their infancy. One approach to improve trustworthiness and fairness in AI systems is to use bias mitigation algorithms. However, most bias mitigation algorithms require data sets that contain sensitive attribute values to assess the fairness of the algorithm. A growing number of real world data sets do not make sensitive attribute information readily available to researchers. One solution is to infer the missing sensitive attribute information and apply an existing bias mitigation algorithm using this inferred knowledge. While researchers are beginning to explore this question, it is still unclear how robust existing bias mitigation algorithms are to different levels of inference accuracy. This paper explores this question by investigating the impact of different levels of accuracy of the inferred sensitive attribute on the performance of different bias mitigation strategies. We generate variation in sensitive attribute accuracy using both simulation and construction of neural models for the inference task. We then assess the quality of six bias mitigation algorithms that are deployed across different parts of our learning life cycle: pre-processing, in-processing, and post-processing. We find that the disparate impact remover is the least sensitive bias mitigation strategy and that if we apply the bias mitigation algorithms using an inferred sensitive attribute with reasonable accuracy, the fairness scores are higher than the best standard model and the balanced accuracy is similar to that of the standard model. These findings open the door for improving fairness of black box AI systems using some bias mitigation strategies. |
| format | Article |
| id | doaj-art-6ca9136c5d2c430aba14ecf53b2849fc |
| institution | DOAJ |
| issn | 2624-8212 |
| language | English |
| publishDate | 2025-03-01 |
| publisher | Frontiers Media S.A. |
| record_format | Article |
| series | Frontiers in Artificial Intelligence |
| spelling | doaj-art-6ca9136c5d2c430aba14ecf53b2849fc2025-08-20T03:16:21ZengFrontiers Media S.A.Frontiers in Artificial Intelligence2624-82122025-03-01810.3389/frai.2025.15203301520330Impact on bias mitigation algorithms to variations in inferred sensitive attribute uncertaintyYanchen Wang0Lisa Singh1Department of Computer Science, Georgetown University, Washington, DC, United StatesDepartment of Computer Science and School of Public Policy, Georgetown University, Washington, DC, United StatesConcerns about the trustworthiness, fairness, and privacy of AI systems are growing, and strategies for mitigating these concerns are still in their infancy. One approach to improve trustworthiness and fairness in AI systems is to use bias mitigation algorithms. However, most bias mitigation algorithms require data sets that contain sensitive attribute values to assess the fairness of the algorithm. A growing number of real world data sets do not make sensitive attribute information readily available to researchers. One solution is to infer the missing sensitive attribute information and apply an existing bias mitigation algorithm using this inferred knowledge. While researchers are beginning to explore this question, it is still unclear how robust existing bias mitigation algorithms are to different levels of inference accuracy. This paper explores this question by investigating the impact of different levels of accuracy of the inferred sensitive attribute on the performance of different bias mitigation strategies. We generate variation in sensitive attribute accuracy using both simulation and construction of neural models for the inference task. We then assess the quality of six bias mitigation algorithms that are deployed across different parts of our learning life cycle: pre-processing, in-processing, and post-processing. We find that the disparate impact remover is the least sensitive bias mitigation strategy and that if we apply the bias mitigation algorithms using an inferred sensitive attribute with reasonable accuracy, the fairness scores are higher than the best standard model and the balanced accuracy is similar to that of the standard model. These findings open the door for improving fairness of black box AI systems using some bias mitigation strategies.https://www.frontiersin.org/articles/10.3389/frai.2025.1520330/fullinferred sensitive attributemachine learning fairnessbias mitigationdemographic inferencesocial media |
| spellingShingle | Yanchen Wang Lisa Singh Impact on bias mitigation algorithms to variations in inferred sensitive attribute uncertainty Frontiers in Artificial Intelligence inferred sensitive attribute machine learning fairness bias mitigation demographic inference social media |
| title | Impact on bias mitigation algorithms to variations in inferred sensitive attribute uncertainty |
| title_full | Impact on bias mitigation algorithms to variations in inferred sensitive attribute uncertainty |
| title_fullStr | Impact on bias mitigation algorithms to variations in inferred sensitive attribute uncertainty |
| title_full_unstemmed | Impact on bias mitigation algorithms to variations in inferred sensitive attribute uncertainty |
| title_short | Impact on bias mitigation algorithms to variations in inferred sensitive attribute uncertainty |
| title_sort | impact on bias mitigation algorithms to variations in inferred sensitive attribute uncertainty |
| topic | inferred sensitive attribute machine learning fairness bias mitigation demographic inference social media |
| url | https://www.frontiersin.org/articles/10.3389/frai.2025.1520330/full |
| work_keys_str_mv | AT yanchenwang impactonbiasmitigationalgorithmstovariationsininferredsensitiveattributeuncertainty AT lisasingh impactonbiasmitigationalgorithmstovariationsininferredsensitiveattributeuncertainty |