If your algorithmic system contains bias, it could be distracting your team and costing you money.
Removing bias is the right thing to do and we know this. But it can also help financially.
To illustrate this, let’s consider insurance claims fraud.
A typical workflow usually runs from data collection, through rules and/or models, scoring, triage, investigation, and decision, with feedback to keep improving the system.
In many cases, triage will narrow the number of claims based on capacity. There’s only so much that we can focus on, so we need to target the claims that have the highest potential for fraud.
This, in turn, is influenced by scoring. A mature program weights the results of each rule or model and adds them up.
If the rules/models are biased, we could be scoring claims incorrectly. And this means that some lower risk claims will be selected for investigation, leaving others, that should be higher risk, out.
We only have capacity to investigate half of the claims we receive.
In practice, we'd be working with hundreds/thousands of claims.
For simplicity, imagine that we have four claims. So we will need to select two for investigation.
This is what we have after we've scored and triaged:
# | Score | Triage Result |
Claim1 | 60 | Investigate |
Claim2 | 50 | Investigate |
Claim3 | 45 | Set aside |
Claim4 | 10 | Set aside |
But there is a problem with our rules that, if accounted for, would reveal this:
# | Score | Triage Result | Problem | Optimal Score | Optimal Result |
Claim1 | 60 | Investigate | 60 | Investigate | |
Claim2 | 50 | Investigate | Driver license type: 10 points | 40 | Set aside |
Claim3 | 45 | Set aside | 45 | Investigate | |
Claim4 | 10 | Set aside | 10 | Set aside |
Claim1 will be included either way.
But Claim2 had 10 points too many added to its score, because it met the criteria for a problematic rule.
In this example, a 10-point rule applied to a claimant without a full/open/regular license. We added this because we identified increased risk of fraud among inexperienced drivers.
There are two potential issues here:
Either of these could be intentional. But “Foreign License” may act as a proxy for race. In Australia, the Racial Discrimination Act 1975 makes it unlawful to discriminate against a person based on their race, colour, country of origin, ethnic origin or immigration status. We should get our legal team to weigh in, if this is the case.
It is more likely that the problem stems from a mistake.
This (mistake) could happen in several ways, such as:
Regardless of the reason, Claim2 should have had 10 fewer points. So, we should be investigating Claim3 instead. It has a higher real likelihood of being fraudulent.
We have not optimised our investigation efforts and may end up settling Claim3. It could be undetected fraud. It wasn’t selected for investigation because another claim was incorrectly included.
If it also turns out that we settled Claim2, because it was cleared by the investigation, we’ve lost money.
Coded bias has directly affected our bottom line.
Disclaimer: The information in this article does not constitute legal advice. It may not be relevant to your circumstances. It was written for specific algorithmic contexts within banks and insurance companies, may not apply to other contexts, and may not be relevant to other types of organisations.