In this previous article, we explored the concept of explainability, its importance, and four challenges.
In this article, we explore some solutions to the first challenge: complex algorithms.
Complex models can operate as "black boxes."
Their internal processes are so complex that even their developers may struggle to explain how inputs lead to specific outputs.
This poses several problems:
Banks and insurers can use several methods to address this problem.
Here are three sets of commonly used methods, each of which work better when cross-functional teams are involved:
Strike a balance between accuracy and simplicity.
Simpler models can be easier to explain, but may not be as accurate as the more complex models.
Consider the performance (accuracy) vs explainability diagram below - Figure 1.
In general, the more complex models (that often produce more accurate results) are harder to explain.
Figure 1: Model explainability vs. model performance comparison, reproduced from Zhang et al. (2021), licensed under CC BY 4.0 [i].
Notes: Not an exact science, the diagram is indicative; other diagrammatic representations may look different. Based on the nature of the research, we have interpreted performance as being synonymous with accuracy/reliability, rather than processing efficiency.
Simplification can include:
If the data going into the model is not relevant, it can make it difficult to determine how decisions are reached.
There are situations where irrelevant data can result in more “accurate” predictions. But if the data is not relevant, then this accuracy gain might be a technical improvement, not a real one.
We also need to be careful not to eliminate relevant data. And this can vary, depending on the use case and the nature of the system.
For both internal data and external data, carefully choose relevant data, and use data for its intended purpose.
There's also a discussion about eliminating protected attributes. We'll explore this in a future article, it is a long story.
These methods work best when combined with proactive complexity management from the previous sections.
There are several techniques that we can use to interpret model processing.
This list is not exhaustive, and what to use will depend on the use case:
In determining which method(s) to use, consider regulatory requirements (e.g., GDPR's "right to explanation") and operational constraints (e.g., computational resources required).
The next article in this series will focus on the second challenge: complicated processes.
[i] Zhang, Y., Liu, D., Cai, W., Chen, H., Heng, P., Li, Q., ... & Wang, S. (2021). Unbox the black-box for the medical explainable AI via multi-modal and multi-centre data fusion: A mini-review, two showcases and beyond. ResearchGate. https://www.researchgate.net/publication/353621660 (CC BY 4.0 https://creativecommons.org/licenses/by/4.0/).
Disclaimer: The information in this article does not constitute legal advice. It may not be relevant to your circumstances. It was written for specific algorithmic contexts within banks and insurance companies, may not apply to other contexts, and may not be relevant to other types of organisations.