Author: Aadhavan Arkhash Saravanakumar, Machine Learning Intern
Co author: Varatharajah Vaseekaran, Senior Machine Learning Engineer
![]()
In 2018, a healthcare algorithm used across U.S. hospitals was found to prioritize white patients over Black patients for critical care, even when the Black patients were sicker. The model wasn’t explicitly told to discriminate, but it learned patterns from historical data that reflected deep-rooted societal inequities. This is not an isolated case. As Artificial Intelligence (AI) becomes embedded in decisions that affect people’s lives—from determining loan eligibility to predicting criminal recidivism—the risk of reinforcing or amplifying bias is no longer theoretical; it’s a real and urgent problem.
Bias in Machine Learning (ML) models isn’t just a technical oversight; it’s often a mirror of the unequal systems that generated the data. If left unaddressed, these biases can fuel discrimination, erode public trust, and lead to harmful real-world outcomes. In this blog, we delve into the practical aspects of achieving fairness in machine learning, examining the origins of bias, how to detect it, and what can be done to mitigate it effectively.
Bias in ML is not a singular entity but a spectrum of complications that can creep into your data and algorithms.
In simple terms, bias in Artificial Intelligence refers to the phenomenon where the model predictions are influenced by the characteristics of demographic factors, causing the privileged group(s) to be favored over the protected group(s). Privileged groups refer to the demographic groups that are usually given privileges by societal norms, whereas protected groups are the demographic groups that have fewer privileges.
In a high-level manner, bias can be categorized as follows:
These biases tend to impact protected groups severely rather than the privileged groups.
Auditing a Machine Learning model for fairness involves evaluating how model predictions vary across different groups defined by demographic attributes such as race, gender, or age.
Several proven mathematical metrics exist to test for bias, with each being suited for different contexts:
With time, several libraries and tools have been implemented to audit ML models for bias, such as:


The above image shows how different metrics are evaluated based on a specific demographic feature (SEX in this example).
Bias can be mitigated at multiple stages of the Machine Learning lifecycle:
Preprocessing
In-processing
Post-processing
Eg: In credit scoring, the accepted range for disparate impact ratio is 0.8 to 1.25 with an ideal value of 1.0.
Each of these techniques comes with its trade-offs, which may sometimes require sacrificing a small percentage of overall accuracy to ensure no harm is done to protected groups, a trade-off well worth making.
Architecting a fair machine learning system is not a one-time effort but a recurrent process of monitoring and mitigation.
Despite the rapid growth of techniques and solutions for Fair Machine Learning, there remain strong challenges that occur due to social and technical traditions:
Arguably, fairness is not a box to check but a series of ongoing, context-sensitive decisions.
Fairness and bias auditing is both a moral obligation and a technical challenge. It is the responsibility of developers to ensure that our models do not amplify social inequalities. With thoughtful auditing, deliberate mitigation and commitment to ethical best practices, we can build models that serve all users equitably.
While fairness poses potential challenges such as data limitation and tradeoffs, it remains an integral component of ethical AI development. From data collection, preprocessing and model training to deployment and monitoring, fairness must be integrated to depict social responsibility and interdisciplinary awareness.
Insights for this article were drawn from research and thought leadership by PBS NOVA, SAP, Fairlearn, IBM’s AIF360, MathWorks, and academic publications on AI fairness.