
Machine learning (ML) has become increasingly prevalent in various domains. However, ML algorithms sometimes give unfair outcomes and discrimination against certain groups. Thereby, bias occurs when our results produce a decision that is systematically incorrect. At various phases of the ML pipeline, such as data collection, pre-processing, model selection, and evaluation, these biases appear. Bias reduction methods for ML have been suggested using a variety of techniques. By changing the data or the model itself, adding more fairness constraints, or both, these methods try to lessen bias. The best technique relies on the particular context and application because each technique has advantages and disadvantages. Therefore, in this paper, we present a comprehensive survey of bias mitigation techniques in machine learning (ML) with a focus on in-depth exploration of methods, including adversarial training. We examine the diverse types of bias that can afflict ML systems, elucidate current research trends, and address future challenges. Our discussion encompasses a detailed analysis of pre-processing, in-processing, and post-processing methods, including their respective pros and cons. Moreover, we go beyond qualitative assessments by quantifying the strategies for bias reduction and providing empirical evidence and performance metrics. This paper serves as an invaluable resource for researchers, practitioners, and policymakers seeking to navigate the intricate landscape of bias in ML, offering both a profound understanding of the issue and actionable insights for responsible and effective bias mitigation.
Artificial intelligence, bias, Explainable Artificial Intelligence, Resource (disambiguation), Variety (cybernetics), Adversarial Robustness in Deep Learning Models, Adversarial system, Learning with Noisy Labels in Machine Learning, Data science, Machine Learning, Context (archaeology), Engineering, Artificial Intelligence, Meta-Learning, Machine learning, fairness constraints, FOS: Mathematics, Pathology, Risk analysis (engineering), Biology, Selection bias, pre-processing, Computer network, Statistics, Paleontology, QA75.5-76.95, Computer science, Management science, Programming language, machine learning, Electronic computers. Computer science, Computer Science, Physical Sciences, Data collection, mitigation techniques, Medicine, Pipeline (software), in-processing, Mathematics, Robust Learning
Artificial intelligence, bias, Explainable Artificial Intelligence, Resource (disambiguation), Variety (cybernetics), Adversarial Robustness in Deep Learning Models, Adversarial system, Learning with Noisy Labels in Machine Learning, Data science, Machine Learning, Context (archaeology), Engineering, Artificial Intelligence, Meta-Learning, Machine learning, fairness constraints, FOS: Mathematics, Pathology, Risk analysis (engineering), Biology, Selection bias, pre-processing, Computer network, Statistics, Paleontology, QA75.5-76.95, Computer science, Management science, Programming language, machine learning, Electronic computers. Computer science, Computer Science, Physical Sciences, Data collection, mitigation techniques, Medicine, Pipeline (software), in-processing, Mathematics, Robust Learning
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 31 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Top 10% | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Top 10% | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Top 10% |
