WHITEPAPER
Explainable AI: Building trust in business decision-making
Introduction The ethical implications of artificial intelligence (AI) have come to the forefront in recent years. While AI has the potential to revolutionize industries and enhance process efficiency, on the one hand, there are concerns about perpetuating bias and discrimination, infringing on privacy, and making difficult-to-understand decisions on the other. The responsible AI movement seeks to develop ethical and transparent AI systems that align with societal values and respect to human rights. Explainable AI (XAI) has emerged as an essential component of responsible AI, potentially elevating user experience by bolstering confidence in the AI's ability to arrive at optimal decisions. It enables AI systems to provide clear and understandable explanations of their decision-making processes. It empowers stakeholders to understand and trust the reasoning behind the decisions and identify any biases or errors in the system's logic. Furthermore, businesses can leverage XAI to effectively manage, interpret, and trust AI systems while mitigating associated risks with deploying AI technologies.
Why Explainable AI in business decision-making? Business leaders use AI to gain insights into complex data sets, identify trends and patterns, and make predictions that can inform strategy and drive growth. However, these decisions must be explainable and transparent to be trusted by stakeholders, including employees, customers, and investors. Therefore, incorporating explainable AI into business decision-making processes is ethical and necessary for building trust and achieving sustainable success. By providing clear explanations, XAI promotes transparency, fairness, and accountability in AI systems, ultimately leading to better decision-making outcomes for businesses with a competitive advantage.
According to a report by Research and Markets, the global Explainable AI (XAI) market was valued at USD 3.50 billion in 2020 and is projected to reach USD 21.03 billion by 2030, growing at a CAGR of 18.95% from 2021-2030.
© 2023 Fractal Analytics Inc. All rights reserved
01
Integrating RAI and XAI into the Data Science Lifecycle
EDA, Pre-processing and feature engineering and selection
Model development
Model evaluation
Model selection
Business understanding and hypothesis testing
Input data
RAI definition
Data privacy
Data bias
XAI & privacy
Model bias & privacy
Model management
Model accountability
Monitoring
Deployment
Prediction
Bias, XAI & Drifts
Data Privacy
XAI, Prediction bias
Different Components of XAI Framework
A comprehensive XAI ensures that AI systems are transparent, auditable, and fair and helps businesses make better-informed decisions based on AI-generated insights. With this framework, enterprises can confidently understand how AI systems make decisions, the factors behind those decisions, and how to mitigate any associated risks.
Can we explain our data and its features?
Can we explain how agnostic model works?
Feature sensitivity check within exploration Interpretable feature engineering Feature dependency check on target Feature weights calculation on target
Global and local explanations Split and compare quantiles Deep and tree SHAP
Can we explain how specific model works?
Can we explain the risk associated to business?
Gradient-based attribution methods Explanation by simplification GAM plots
Risk monitoring assessment Risk calculation in probabilities/deciles Trade-off between accuracy and interpretability Counterfactuals for decision-making
© 2023 Fractal Analytics Inc. All rights reserved
02
Explainability for business stakeholders
In the business world, interpretability and explainability are crucial for stakeholders to understand machine learning model results and errors. This understanding helps product owners make informed financial decisions. With clear explanations provided by AI and machine learning models, users gain confidence, and developers can justify their models' validity. Transparent modeling also ensures accountability and regulatory compliance for C-suite executives. Additionally, it reduces ambiguity and promotes trust, essential to business success. Incorporating explainability in machine learning algorithms can help mitigate risk and build trust among stakeholders, resulting in the successful adoption and application of AI technologies. To illustrate this point, the below technique based on specific use cases can be utilized. Split & Compare Quantiles is a valuable technique for defining decision thresholds in classification and regression problems. By enabling model evaluation and decision-making, this approach provides a clear understanding of how the model's predictions impact the business objectives, making it useful in the data science toolkit.
Salient Features of SCQ Plot
Model Agnostic
Business Risk Analysis
Quantile-Based Analysis
Data-Agnostic
Effective Decision-Making
Granular Risk Assessment
Post-Deployment Analysis
Customizable Binning
© 2023 Fractal Analytics Inc. All rights reserved
03
Visualizing the invisible: Mitigating business risk with split & compare quantiles (SCQ)
The Split & Compare Quantile Plot is a visualization tool that can aid businesses in decision-making and risk mitigation. This technique enables a comprehensive evaluation of machine learning models across various subsets of data, helping companies optimize their strategies and achieve their goals.
Businesses can take corrective actions and improve their overall performance by identifying areas where their models may be underperforming.
To create a split and compare quantile plot, one can first split the dataset into equal quantiles and then separate the outcomes into favorable and unfavorable categories. For instance, the sample data can be divided into deciles and categorized based on their labels. After dividing the data into equal-sized bins, the percentage of observations in each bin can be computed for both favorable and unfavorable outcomes. This approach is straightforward but efficient, enabling the analysis of a model's performance and facilitating informed business decisions.
Flowchart for Assessing Machine Learning Models with a Split and Compare Quantile Plot
Data Structuring and Grouping
Visualizing the Plot
Providing Explanations
Analyze the plot to determine predicted data/scores in each actual bin
Create a stacked bar plot with actual bins on X-axis and predicted on Y-axis
Gather the actual and predicted values from the trained model
Start
Use the plot to evaluate each prediction granularly
Colour-code bars to differentiate between bins
Group predicted values into equal-sized bins as actual values
Group actual values into bins
Find the percentage of predicted data in each bin of actual data
End
© 2023 Fractal Analytics Inc. All rights reserved
04
Statistical Applications of Split and Compare Quantiles
Salient Features in Action
Model Agnostic
SCQ for Regression Problem
Quantile-based Analysis
Data-Agnostic
Customizable Binning
Post-Deployment Analysis
Split Groups
100.0%
100.0%
100.0%
93.3% 100.0%
100.0%
80.0% 86.7% 93.3% 100.0%
93.3% 100.0%
86.7% 93.3% 100.0%
100.0%
100.0%
100
86.7%
86.7% 80.0%
80.0%
80.0%
66.7% 73.3% 80.0%
80.0%
80
The SCQ plot can provide valuable insights into the accuracy of machine learning models post-deployment when actual values are not available. For instance, consider a prediction of 52000 for which no actual value is available. In this scenario, the Split & Compare Quantile Plot can be used to evaluate the accuracy of the prediction and its associated risks. By examining the plot, we can see that the prediction falls into the bin marked with an orange color quantile in the legend and the 2nd quantile as per the x-axis. Further analysis reveals that in the second quantile, the orange color is accurate only ~30% of the time, while ~13% of the time the prediction falls in value bins that are lower than the actual value, and the remaining ~55% of the time, it falls in the bins above the actual value.
73.3%
60.0% 66.7%
66.7%
60.0% 53.3%
60.0%
60.0%
60.0%
60
53.3%
53.3%
37345.3 - 50909.8 51184.6 - 57559.8 57767.0 - 62276.9 62455.8 - 65108.9 65295.7 - 69573.8 70117.6 - 73683.1 73989.0 - 77836.4 78318.8 - 82656.3 82700.2 - 87736.3 87800.5 - 127222.7
40.0% 46.7%
46.7%
37.5%
40
33.3%
26.7% 20.0%
26.7%
26.7%
25.0%
20.0%
20.0%
20.0%
20
13.3%
6.7% 13.3%
12.5%
6.7%
6.7%
0
Tag
SCQ for classification problem
While machine learning models can be highly accurate, there's inevitably some error associated with them, potentially leading to incorrect predictions. Consider the example of identifying mortgage defaulters in financial services. Even the most sophisticated models can't achieve 100% accuracy and may misidentify non-defaulters as defaulters. In such scenarios, stakeholders can benefit from using Split & Compare Quantile (SCQ) charts, which provide a clear picture of the degree of error associated with a model's predictions. By breaking down the data into deciles, this chart highlights all the labels within each decile, enabling stakeholders to establish the error level the model will likely have at a given threshold.
Assuming a decision threshold of 67% has been identified, which will enable the bank to service 65.25% of customers. Interestingly, this threshold will also result in 11.79% of customers being identified as having a probability of default. In such cases, stakeholders would want to minimize the risk reflected by the model by selecting a threshold that minimizes the error. The first part represents the potential commercial loss resulting from incorrect labeling of defaulters as not likely (false positives), and the second part represents a potential loss of opportunity resulting from false negatives. The stakeholders should use this information to determine the optimal decision boundary that minimizes false positives and negatives. By doing so, the bank can minimize its overall risk exposure while maximizing its potential revenue opportunities.
0000 0000 0000 0000 0000 0000 0000 0000 0000 0000
70
1.18 1.18
9.23
20
1.82 1.09
60
7.92
1.51
8.31
50
15
2.18
8.32
40
2.51
7.49
10
2.41
30
7.56
3.70
6.71
20
5
5.92
10
5.72
4.98
4.80
0
0
© 2023 Fractal Analytics Inc. All rights reserved
05
Business application of split and compare quantiles
Salient Features in Action
Post-Deployment Analysis
Business Risk Analysis
Financial Decision-Making with SCQ
Effective Decision-Making
Granular Risk Assessment
0.001-480 48.0-60.0 60.0-67.0 67.0-73.0 73.0-78.0 78.0-82.0 82.0-86.0 86.0-89.0 89.0-92.5 92.5-86.0
20000
8000
506.27 279.96 307.58 268.13
1580.21
17500
7000
1651.50
Similarly, in the SCQ plot, we can replace probabilities with corresponding monetary values. Doing so allows us to analyze the total monetary value that we will get at a chosen probability threshold. For instance, at a 67% decision threshold, the total monetary transaction value can be estimated at $17288.51K. However, there is a possibility of losing $3508.1K due to model error. In such a scenario, stakeholders would like to select a decision threshold or bins of decision thresholds that minimize the dollar value loss resulting from model errors. By doing so, businesses can optimize their decision-making processes and minimize their overall monetary losses.
611.96
1611.72
15020
6000
667.03
2157.11
12000
5000
845.18
2185.72
10200
4000
856.11
2171.89
7500
3000
1411.31
2122.26
5000
2000
2121.81
1871.97
2114.03
2000
1000
1734.79
0
0
Conclusion
Traditional practices in explainable artificial intelligence (XAI) have typically been limited to basic model explainability and data visualization. However, it's vital to take things one step further, dissect the model and understand the potential risks associated with errors in the model. Surprisingly, a lower error rate doesn't necessarily equal lower financial loss and vice versa. As such, it's crucial not just to analyze the model’s errors statistically but also from a monetary standpoint before making a decision on the threshold. It is important to note that stakeholders may prefer choosing bins of different thresholds instead of a single flat decision boundary. This approach can provide a better trade-off between the monetary value of the model's errors and its accuracy.
Authors
Supriya Panigrahi Consultant, Fractal Dimension
Sray Agarwal
Sakshi Gulyani Consultant, Fractal Dimension
Principal Consultant, Fractal Dimension
© 2023 Fractal Analytics Inc. All rights reserved
06
About Fractal
Fractal is one of the most prominent providers of Artificial Intelligence to Fortune 500® companies. Fractal's vision is to power every human decision in the enterprise, and bring AI, engineering, and design to help the world's most admired companies. Fractal's businesses include Crux Intelligence (AI driven business intelligence), Eugenie.ai (AI for sustainability), Asper.ai (AI for revenue growth management) and Senseforth.ai (conversational AI for sales and customer service). Fractal incubated Qure.ai, a leading player in healthcare AI for detecting Tuberculosis and Lung cancer. Fractal currently has 4000+ employees across 16 global locations, including the United States, UK, Ukraine, India, Singapore, and Australia. Fractal has been recognized as 'Great Workplace' and 'India's Best Workplaces for Women' in the top 100 (large) category by The Great Place to Work® Institute; featured as a leader in Customer Analytics Service Providers Wave™ 2021, Computer Vision Consultancies Wave™ 2020 & Specialized Insights Service Providers Wave™ 2020 by Forrester Research Inc., a leader in Analytics & AI Services Specialists Peak Matrix 2022 by Everest Group and recognized as an 'Honorable Vendor' in 2022 Magic Quadrant™ for data & analytics by Gartner Inc.
For more information, visit fractal.ai
Corporate Headquarters Suite 76J, One World Trade Center, New York, NY 10007
Get in touch
© 2023 Fractal Analytics Inc. All rights reserved
07
Page 1 Page 2 Page 3 Page 4 Page 5 Page 6 Page 7 Page 8Made with FlippingBook - PDF hosting