SlideShare a Scribd company logo
1 of 54
Download to read offline
The Importance
of Model
Fairness and
Interpretability
in AI Systems
Mehrnoosh Sameki
Senior Product Manager
Microsoft Azure ML
Minsoo Thigpen
Product Manager
Microsoft Azure ML
Ehi Nosakhare
Machine Learning Scientist
Microsoft AI Acceleration and
Development Program
Machine Fairness
How to assess AI system’s fairness
and mitigate any observed
unfairness issues
www.aka.ms/Fairlearn-build2020
Article
Resources
➢www.aka.ms/InterpretML-github
➢www.aka.ms/InterpretML-toolkit
➢www.aka.ms/InterpretML-doc
➢www.aka.ms/FairlearnAI
➢www.aka.ms/FairlearnAIDoc
➢www.aka.ms/ResponsibleMLDoc
➢www.aka.ms/AzureResponsibleML
➢www.aka.ms/AzureMLDoc
➢www.aka.ms/AzureMLservice
Agenda
WHY RESPONSIBLE AI?
INTERPRETML
FAIRLEARN TOOLKIT
AI will have a considerable impact on
business and society as a whole
AI impact raises a
host of complex
and challenging
questions
Report
Why Responsible AI?
Microsoft’s AI Principles
Microsoft’s AI Principles
Transparency
AI systems should be explainable
AI systems should have algorithmic
accountability
Machine Learning Interpretability use cases
www.aka.ms/InterpretML-github
Interpretability Toolkit
www.aka.ms/InterpretML-github
Explain Models - www.aka.ms/InterpretML-toolkit
InterpretML Repo - www.aka.ms/InterpretML-github
Interpret
Tabular Data
Additional interpretability
techniques for tabular
extension
www.aka.ms/InterpretML-github github/InterpretML/Interpret-Community
Interpret-
Community
Interpret
Tabular Data
Additional interpretability
techniques for tabular
extension
github/InterpretML/Interpret github/InterpretML/Interpret-Community
Interpret-
Community
AzureML-
Interpret
wrapper
Azure Machine Learning SDK
InterpretML Repo - www.aka.ms/InterpretML-github
Interpret
Tabular Data
Additional interpretability
techniques for tabular
extension
github/InterpretML/Interpret github/InterpretML/Interpret-Community
Interpret-
Community
Interpret-
text-contrib
github/InterpretML/Interpret-text-contrib
Text Data
(Text Classification)
AzureML-
Interpret
wrapper
Azure Machine Learning SDK
InterpretML Repo - www.aka.ms/InterpretML-github
Types of
Models
supported
www.aka.ms/InterpretML-github
Interpretability Approaches
• Given an existing model/system, extract
explanations
• Typically provide approximate explanations
• Examples: LIME, SHAP, Permutation Feature
Importance, Sensitivity Analysis, Influence
Functions, etc.
Tools to explain the system
www.aka.ms/InterpretML-github
SHAP
(SHapley Additive
exPlanations)
• Not a new concept
• Concept based on game theory
• Mathematically sound
• Application in ML relatively new
www.aka.ms/InterpretML-github
SHAP
Modeling scenario:
House pricing
prediction
www.aka.ms/InterpretML-github
How much has each feature contributed
to the prediction compared to the average prediction?
House price
prediction:
€ 300,000
Average house
price prediction
for all apartments
is € 310,000
Delta here is
- €10,000
www.aka.ms/InterpretML-github
Parks contributed
+10k
Cats banned
contributed -50k
50𝑚2
contributed
+10k 2nd floor had net zero
contributions
www.aka.ms/InterpretML-github
SHAP
How do we calculate Shapley values?
• Take your feature of interest (e.g.,
cats-banned) and remove it from the
feature set
• Take the remaining features and
generate all possible coalitions
• Add and remove your feature of
interested to each of the coalitions
and calculate the difference it makes
www.aka.ms/InterpretML-github
SHAP Pros SHAP Cons
Based on a solid theory and distributes the
effects fairly
Computation time: 2k possible coalitions of the
feature values for k features
Contrastive explanations: Instead of comparing a
prediction to the average prediction of the entire
dataset, you could compare it to a subset or even
to a single data point.
Can be misinterpreted
Inclusion of unrealistic data instances when
features are correlated.
www.aka.ms/InterpretML-github
Interpretability Approaches
• Models inherently interpretable
• Typically provide exact explanations
• Examples: Linear Models, Decision Rule Lists,
Explainable Boosting Machines, Shallow
Decision Trees, etc.
Tools to explain the system
www.aka.ms/InterpretML-github
Linear Models
• (Generalized) Linear Models
• Current standard for interpretable
models
• Learns an additive relationship
between data and response:
• y = β0 + β1x1 + β2x2 + ... + βn xn
• βi terms are scalar – single number
that multiplies to each feature value
• Each feature contributes a “score”
that adds up to final output
Example Model :
House Price = 50,000 + 0.2*sq_ft +
10000*bedrooms +
7000*bathrooms
Using this model on a new house:
- 5 Bedrooms
- 3 Bathrooms
- 3000 Sq Feet
- House Price = 50000 + 0.2*3000
+ 10000*5 + 7000 * 3
- House Price = $121,600
www.aka.ms/InterpretML-github
Explainable Boosting Machine
• Generalized Additive Models (with Pairwise
Interactions)
• GAMs have existed since 1980s
• MSR invented new methods of training GAMs for
higher accuracy
• Think of it as a “more expressive” linear model
• Still additive!
• Linear Model: y = β0 + β1x1 + β2x2 + ... + βn xn
• Additive Model: y = f1(x1) + f2(x2) + ... + fn (xn)
• Additive Model with Pairwise Interactions
(EBM):
y = Ʃi fi (xi) + Ʃij fij (xi , xj )
• Full Complexity Model: y = f (x1, ..., xn)
Linear Model can only learn a
straight line!
www.aka.ms/InterpretML-github
Demo of
Interpretability for
Tabular Data
Hardware Performance Dataset - The
Regression goal is to predict the
performance of certain combinations
of hardware parts.
www.aka.ms/InterpretML-github
Demo dataset
https://archive.ics.uci.edu/ml/datasets/Computer+Hardware
Hardware Performance Dataset - The Regression goal is to predict the
performance of certain combinations of hardware parts.
• vendor name: 30
(adviser, amdahl,apollo, basf, bti, burroughs, c.r.d, cambex, cdc, dec, dg, formation, four-phase,
gould, honeywell, hp, ibm, ipl, magnuson, microdata, nas, ncr, nixdorf, perkin-elmer, prime,
siemens, sperry, sratus, wang)
• Model Name: many unique symbols
• MYCT: machine cycle time in nanoseconds (integer)
• MMIN: minimum main memory in kilobytes (integer)
• MMAX: maximum main memory in kilobytes (integer)
• CACH: cache memory in kilobytes (integer)
• CHMIN: minimum channels in units (integer)
• CHMAX: maximum channels in units (integer)
• PRP: published relative performance (integer)
• ERP: estimated relative performance from the original article (integer)
FAIRNESS
More positive outcomes & avoiding
harmful outcomes of AI systems for
groups of people
• Allocation: extends or withholds opportunities, resources, or information.
• Quality of service: whether a system works as well for one person as it does
for another
• Stereotyping: reinforce existing societal stereotypes
• Denigration: actively derogatory or offensive
• Over or under representation: over-represent, under-represent, or even
erase particular groups of people
Crawford et al. 2017
Types of harm
What is
Fairlearn?
A new approach to measuring and mitigating
unfairness in systems that make predictions, serve
users, or make decisions about allocating resources,
opportunities, or information.
www.aka.ms/FairlearnAI
There are many ways that an AI system can behave unfairly.
Fairness in AI
Avoiding negative outcomes of AI systems for different groups of people
A model for screening loan or job application
might be much better at picking good candidates
among white men than among other groups.
A voice recognition system might
fail to work as well for women as it
does for men.
www.aka.ms/FairlearnAI
A toolkit that empowers developers of artificial intelligence systems to
assess their systems' fairness and mitigate any observed fairness issues.
Helps users identify and mitigate unfairness in their machine learning
models with a focus on group fairness.
Automatically
analyze a model’s
predictions
Provide the user with
insights into (un)fairness of
their model’s predictions
Support (algorithmic)
methods to mitigate
unfairness
www.aka.ms/FairlearnAI
• Allocation: extends or withholds opportunities, resources, or information.
• Quality of service: whether a system works as well for one person as it does
for another
• Stereotyping: reinforce existing societal stereotypes
• Denigration: actively derogatory or offensive
• Over or under representation: over-represent, under-represent, or even
erase particular groups of people
Crawford et al. 2017
Types of harm addressed by Fairlearn
www.aka.ms/FairlearnAI
Harm of “allocation”
Example Scenarios: Lending
The data set describes whether each individual repaid the loan or not.
• [Classification] Recommends whether a given individual should get a loan. The
trained model outputs: Yes/Maybe/No decision.
Example Scenarios: School Admissions
The data set describes what was the first-year GPA of each student.
• [Regression] For a given applicant, predicts their GPA at the end of the first year.
The trained model outputs a real-valued prediction that is used as a score to
screen applicants.
www.aka.ms/FairlearnAI
Harm of “quality of service”
Example Scenarios: News Recommendation
The training data indicates what article was presented to which user, whether the user
clicked, and how much time the user spent on the article.
• [Classification] Predict which article to show to each user to optimize click-
through rate. The trained model outputs: Yes/No decision.
Two kinds of (group) fairness: across users (quality of service), or across
publishers/topics of articles (quality of service, but also allocation).
www.aka.ms/FairlearnAI
Assessment
Evaluation
Visualizations
Fairness Criteria
Mitigation Algorithm
Mitigation
• Select a protected attribute
• Select a fairness metric (e.g., disparity in recall, precision, etc
and disparity in selection rate)
Fairlearn Toolkit
• Demographic Parity
• Equalized Odds
• Postprocessing Method
• Reduction Methodwww.aka.ms/FairlearnAI
Fairness assessment
through disparity metrics
Disparity in Performance
• How the model accuracy differs across different buckets of a sensitive feature
(e.g., how accuracy of the model differs for "female" vs. "male" vs. “unspecified"
data points)
Disparity in Selection Rate
• How the model predictions differ across different buckets of a sensitive feature
(e.g., how many "female" individuals have received prediction `approved` on their
loan application in contrast to "male" and “unspecified" data points?).
www.aka.ms/FairlearnAI
Fairness assessment
via Fairlearn visualization dashboard
Let’s see a demo!
www.aka.ms/FairlearnAI
Mitigating Disparities in Ranking from Binary Data
- An example based on the Law School Admissions Council's National Longitudinal Bar
Passage Study
Fairness assessment
via Fairlearn visualization
dashboard
www.aka.ms/FairlearnAI
Demo dataset
Law School Admissions Council's (LSAC) National Longitudinal Bar Passage Study
• The data set contains information about law students collected by LSAC between
1991 and 1997.
• Some of the information is available at the admission time (such as the undergraduate
GPA and LSAT score), and some describes the performance of the students once
admitted.
• We also have access to their self-identified race.
• To simplify this example, we will limit the attention to those self-identified
as black and white (two largest groups) and restrict our attention to two features
(undergraduate GPA and LSAT score).
Assessment
Evaluation
Visualizations
Fairness Criteria
Mitigation Algorithm
Mitigation
• Select a protected attribute
• Select a fairness metric (e.g., disparity in recall, precision, etc
and disparity in selection rate)
Fairlearn Toolkit
• Demographic Parity
• Equalized Odds
• Postprocessing Method
• Reduction Methodwww.aka.ms/FairlearnAI
Demographic parity:
Applicants of each race (gender, ...) have the
same odds of getting approval on their loan applications
Loan approval decision is independent of protected attribute
Equalized odds:
Qualified applicants have the same odds of getting approval on their
loan applications regardless of race (gender, …)
Unqualified applicants have the same odds of getting approval on their
loan applications regardless of race (gender, …)
Fairness Criteria
www.aka.ms/FairlearnAI
Reductions
approach:
Wrapper
around
standard ML
algorithms
Input:
• any standard ML training algorithm (as a
black box)
• data set including sensitive feature
Output:
• a trained model that minimizes error
subject to fairness constraints
Advantages:
• doesn’t need to access the sensitive
feature at test time
• works for a wide range of disparity metrics
• allows extracting the full disparity-accuracy
frontier
Disadvantages:
• requires re-training: the black box is called
10-20 times
www.aka.ms/FairlearnAI
Post
processing:
Picking a fair
threshold rule
Input:
• an existing (already trained) scoring model
• data set including sensitive feature
Output:
• the most accurate among all fair threshold rules (a
separate threshold for each subpopulation)
Advantages:
• simplicity
• no need to re-train the model
Disadvantages:
• requires sensitive feature at test-time
• doesn’t allow trade-offs between disparity and
accuracy
www.aka.ms/FairlearnAI
Unfairness
mitigation
Let’s see a demo!
www.aka.ms/FairlearnAI
Mitigating Disparities in
Ranking from Binary Data
- An example based on the Law
School Admissions Council's
National Longitudinal Bar
Passage Study
Article:
Machine Fairness
How to assess AI system’s fairness
and mitigate any observed
unfairness issues
www.aka.ms/Fairlearn-build2020
Mehrnoosh Sameki
Senior Product Manager
Microsoft Azure ML
Minsoo Thigpen
Product Manager
Microsoft Azure ML
Ehi Nosakhare
Machine Learning Scientist
Microsoft AI Acceleration and
Development Program
Resources
➢www.aka.ms/InterpretML-github
➢www.aka.ms/InterpretML-toolkit
➢www.aka.ms/InterpretML-doc
➢www.aka.ms/FairlearnAI
➢www.aka.ms/FairlearnAIDoc
➢www.aka.ms/ResponsibleMLDoc
➢www.aka.ms/AzureResponsibleML
➢www.aka.ms/AzureMLDoc
➢www.aka.ms/AzureMLservice
Thank You!

More Related Content

What's hot

Explainable AI in Industry (FAT* 2020 Tutorial)
Explainable AI in Industry (FAT* 2020 Tutorial)Explainable AI in Industry (FAT* 2020 Tutorial)
Explainable AI in Industry (FAT* 2020 Tutorial)
Krishnaram Kenthapadi
 
Responsible AI in Industry (ICML 2021 Tutorial)
Responsible AI in Industry (ICML 2021 Tutorial)Responsible AI in Industry (ICML 2021 Tutorial)
Responsible AI in Industry (ICML 2021 Tutorial)
Krishnaram Kenthapadi
 
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)
Krishnaram Kenthapadi
 
Interpretable Machine Learning Using LIME Framework - Kasia Kulma (PhD), Data...
Interpretable Machine Learning Using LIME Framework - Kasia Kulma (PhD), Data...Interpretable Machine Learning Using LIME Framework - Kasia Kulma (PhD), Data...
Interpretable Machine Learning Using LIME Framework - Kasia Kulma (PhD), Data...
Sri Ambati
 
Explainable AI in Industry (WWW 2020 Tutorial)
Explainable AI in Industry (WWW 2020 Tutorial)Explainable AI in Industry (WWW 2020 Tutorial)
Explainable AI in Industry (WWW 2020 Tutorial)
Krishnaram Kenthapadi
 
Explainable AI in Industry (KDD 2019 Tutorial)
Explainable AI in Industry (KDD 2019 Tutorial)Explainable AI in Industry (KDD 2019 Tutorial)
Explainable AI in Industry (KDD 2019 Tutorial)
Krishnaram Kenthapadi
 

What's hot (20)

Explainable AI in Industry (FAT* 2020 Tutorial)
Explainable AI in Industry (FAT* 2020 Tutorial)Explainable AI in Industry (FAT* 2020 Tutorial)
Explainable AI in Industry (FAT* 2020 Tutorial)
 
Fairness-aware Machine Learning: Practical Challenges and Lessons Learned (WW...
Fairness-aware Machine Learning: Practical Challenges and Lessons Learned (WW...Fairness-aware Machine Learning: Practical Challenges and Lessons Learned (WW...
Fairness-aware Machine Learning: Practical Challenges and Lessons Learned (WW...
 
Explainability and bias in AI
Explainability and bias in AIExplainability and bias in AI
Explainability and bias in AI
 
Interpretability of machine learning
Interpretability of machine learningInterpretability of machine learning
Interpretability of machine learning
 
Responsible AI in Industry (ICML 2021 Tutorial)
Responsible AI in Industry (ICML 2021 Tutorial)Responsible AI in Industry (ICML 2021 Tutorial)
Responsible AI in Industry (ICML 2021 Tutorial)
 
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)
 
Explainable AI - making ML and DL models more interpretable
Explainable AI - making ML and DL models more interpretableExplainable AI - making ML and DL models more interpretable
Explainable AI - making ML and DL models more interpretable
 
Measures and mismeasures of algorithmic fairness
Measures and mismeasures of algorithmic fairnessMeasures and mismeasures of algorithmic fairness
Measures and mismeasures of algorithmic fairness
 
Shift AI 2020: How to identify and treat biases in ML Models | Navdeep Sharma...
Shift AI 2020: How to identify and treat biases in ML Models | Navdeep Sharma...Shift AI 2020: How to identify and treat biases in ML Models | Navdeep Sharma...
Shift AI 2020: How to identify and treat biases in ML Models | Navdeep Sharma...
 
Machine Learning Interpretability
Machine Learning InterpretabilityMachine Learning Interpretability
Machine Learning Interpretability
 
Data analytics with python introductory
Data analytics with python introductoryData analytics with python introductory
Data analytics with python introductory
 
Explainable Machine Learning (Explainable ML)
Explainable Machine Learning (Explainable ML)Explainable Machine Learning (Explainable ML)
Explainable Machine Learning (Explainable ML)
 
Interpretable Machine Learning Using LIME Framework - Kasia Kulma (PhD), Data...
Interpretable Machine Learning Using LIME Framework - Kasia Kulma (PhD), Data...Interpretable Machine Learning Using LIME Framework - Kasia Kulma (PhD), Data...
Interpretable Machine Learning Using LIME Framework - Kasia Kulma (PhD), Data...
 
State of AI Report 2023 - Air Street Capital
State of AI Report 2023 - Air Street CapitalState of AI Report 2023 - Air Street Capital
State of AI Report 2023 - Air Street Capital
 
Explainable AI
Explainable AIExplainable AI
Explainable AI
 
Explainable AI in Industry (WWW 2020 Tutorial)
Explainable AI in Industry (WWW 2020 Tutorial)Explainable AI in Industry (WWW 2020 Tutorial)
Explainable AI in Industry (WWW 2020 Tutorial)
 
Explainable AI in Industry (KDD 2019 Tutorial)
Explainable AI in Industry (KDD 2019 Tutorial)Explainable AI in Industry (KDD 2019 Tutorial)
Explainable AI in Industry (KDD 2019 Tutorial)
 
Algorithmic Bias: Challenges and Opportunities for AI in Healthcare
Algorithmic Bias:  Challenges and Opportunities for AI in HealthcareAlgorithmic Bias:  Challenges and Opportunities for AI in Healthcare
Algorithmic Bias: Challenges and Opportunities for AI in Healthcare
 
Shap
ShapShap
Shap
 
Towards Responsible AI - KC.pptx
Towards Responsible AI - KC.pptxTowards Responsible AI - KC.pptx
Towards Responsible AI - KC.pptx
 

Similar to The importance of model fairness and interpretability in AI systems

Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018
Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018
Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018
Sri Ambati
 
Net campus2015 antimomusone
Net campus2015 antimomusoneNet campus2015 antimomusone
Net campus2015 antimomusone
DotNetCampus
 
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!
Sri Ambati
 

Similar to The importance of model fairness and interpretability in AI systems (20)

Spark + AI Summit - The Importance of Model Fairness and Interpretability in ...
Spark + AI Summit - The Importance of Model Fairness and Interpretability in ...Spark + AI Summit - The Importance of Model Fairness and Interpretability in ...
Spark + AI Summit - The Importance of Model Fairness and Interpretability in ...
 
Ai in finance
Ai in financeAi in finance
Ai in finance
 
Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018
Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018
Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018
 
230208 MLOps Getting from Good to Great.pptx
230208 MLOps Getting from Good to Great.pptx230208 MLOps Getting from Good to Great.pptx
230208 MLOps Getting from Good to Great.pptx
 
Human in the loop: Bayesian Rules Enabling Explainable AI
Human in the loop: Bayesian Rules Enabling Explainable AIHuman in the loop: Bayesian Rules Enabling Explainable AI
Human in the loop: Bayesian Rules Enabling Explainable AI
 
Azure Databricks for Data Scientists
Azure Databricks for Data ScientistsAzure Databricks for Data Scientists
Azure Databricks for Data Scientists
 
SESE 2021: Where Systems Engineering meets AI/ML
SESE 2021: Where Systems Engineering meets AI/MLSESE 2021: Where Systems Engineering meets AI/ML
SESE 2021: Where Systems Engineering meets AI/ML
 
Net campus2015 antimomusone
Net campus2015 antimomusoneNet campus2015 antimomusone
Net campus2015 antimomusone
 
PREDICT THE FUTURE , MACHINE LEARNING & BIG DATA
PREDICT THE FUTURE , MACHINE LEARNING & BIG DATAPREDICT THE FUTURE , MACHINE LEARNING & BIG DATA
PREDICT THE FUTURE , MACHINE LEARNING & BIG DATA
 
Machine Learning and AI: Core Methods and Applications
Machine Learning and AI: Core Methods and ApplicationsMachine Learning and AI: Core Methods and Applications
Machine Learning and AI: Core Methods and Applications
 
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!
 
Machine Learning AND Deep Learning for OpenPOWER
Machine Learning AND Deep Learning for OpenPOWERMachine Learning AND Deep Learning for OpenPOWER
Machine Learning AND Deep Learning for OpenPOWER
 
Introduction to Machine Learning with SciKit-Learn
Introduction to Machine Learning with SciKit-LearnIntroduction to Machine Learning with SciKit-Learn
Introduction to Machine Learning with SciKit-Learn
 
Interpretable Machine Learning
Interpretable Machine LearningInterpretable Machine Learning
Interpretable Machine Learning
 
Amazon SageMaker 內建機器學習演算法 (Level 400)
Amazon SageMaker 內建機器學習演算法 (Level 400)Amazon SageMaker 內建機器學習演算法 (Level 400)
Amazon SageMaker 內建機器學習演算法 (Level 400)
 
Big Data Spain 2018: How to build Weighted XGBoost ML model for Imbalance dat...
Big Data Spain 2018: How to build Weighted XGBoost ML model for Imbalance dat...Big Data Spain 2018: How to build Weighted XGBoost ML model for Imbalance dat...
Big Data Spain 2018: How to build Weighted XGBoost ML model for Imbalance dat...
 
AnalyticOps: Lessons Learned Moving Machine-Learning Algorithms to Production...
AnalyticOps: Lessons Learned Moving Machine-Learning Algorithms to Production...AnalyticOps: Lessons Learned Moving Machine-Learning Algorithms to Production...
AnalyticOps: Lessons Learned Moving Machine-Learning Algorithms to Production...
 
Model Evaluation in the land of Deep Learning
Model Evaluation in the land of Deep LearningModel Evaluation in the land of Deep Learning
Model Evaluation in the land of Deep Learning
 
Machine Learning: Considerations for Fairly and Transparently Expanding Acces...
Machine Learning: Considerations for Fairly and Transparently Expanding Acces...Machine Learning: Considerations for Fairly and Transparently Expanding Acces...
Machine Learning: Considerations for Fairly and Transparently Expanding Acces...
 
Machine Learning 2 deep Learning: An Intro
Machine Learning 2 deep Learning: An IntroMachine Learning 2 deep Learning: An Intro
Machine Learning 2 deep Learning: An Intro
 

Recently uploaded

Call Girls In Bellandur ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Bellandur ☎ 7737669865 🥵 Book Your One night StandCall Girls In Bellandur ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Bellandur ☎ 7737669865 🥵 Book Your One night Stand
amitlee9823
 
Call Girls In Doddaballapur Road ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Doddaballapur Road ☎ 7737669865 🥵 Book Your One night StandCall Girls In Doddaballapur Road ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Doddaballapur Road ☎ 7737669865 🥵 Book Your One night Stand
amitlee9823
 
Probability Grade 10 Third Quarter Lessons
Probability Grade 10 Third Quarter LessonsProbability Grade 10 Third Quarter Lessons
Probability Grade 10 Third Quarter Lessons
JoseMangaJr1
 
FESE Capital Markets Fact Sheet 2024 Q1.pdf
FESE Capital Markets Fact Sheet 2024 Q1.pdfFESE Capital Markets Fact Sheet 2024 Q1.pdf
FESE Capital Markets Fact Sheet 2024 Q1.pdf
MarinCaroMartnezBerg
 
Vip Mumbai Call Girls Marol Naka Call On 9920725232 With Body to body massage...
Vip Mumbai Call Girls Marol Naka Call On 9920725232 With Body to body massage...Vip Mumbai Call Girls Marol Naka Call On 9920725232 With Body to body massage...
Vip Mumbai Call Girls Marol Naka Call On 9920725232 With Body to body massage...
amitlee9823
 
Call Girls In Attibele ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Attibele ☎ 7737669865 🥵 Book Your One night StandCall Girls In Attibele ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Attibele ☎ 7737669865 🥵 Book Your One night Stand
amitlee9823
 
Mg Road Call Girls Service: 🍓 7737669865 🍓 High Profile Model Escorts | Banga...
Mg Road Call Girls Service: 🍓 7737669865 🍓 High Profile Model Escorts | Banga...Mg Road Call Girls Service: 🍓 7737669865 🍓 High Profile Model Escorts | Banga...
Mg Road Call Girls Service: 🍓 7737669865 🍓 High Profile Model Escorts | Banga...
amitlee9823
 

Recently uploaded (20)

Call Girls In Bellandur ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Bellandur ☎ 7737669865 🥵 Book Your One night StandCall Girls In Bellandur ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Bellandur ☎ 7737669865 🥵 Book Your One night Stand
 
Call Girls In Doddaballapur Road ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Doddaballapur Road ☎ 7737669865 🥵 Book Your One night StandCall Girls In Doddaballapur Road ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Doddaballapur Road ☎ 7737669865 🥵 Book Your One night Stand
 
Probability Grade 10 Third Quarter Lessons
Probability Grade 10 Third Quarter LessonsProbability Grade 10 Third Quarter Lessons
Probability Grade 10 Third Quarter Lessons
 
Call Girls in Sarai Kale Khan Delhi 💯 Call Us 🔝9205541914 🔝( Delhi) Escorts S...
Call Girls in Sarai Kale Khan Delhi 💯 Call Us 🔝9205541914 🔝( Delhi) Escorts S...Call Girls in Sarai Kale Khan Delhi 💯 Call Us 🔝9205541914 🔝( Delhi) Escorts S...
Call Girls in Sarai Kale Khan Delhi 💯 Call Us 🔝9205541914 🔝( Delhi) Escorts S...
 
BDSM⚡Call Girls in Mandawali Delhi >༒8448380779 Escort Service
BDSM⚡Call Girls in Mandawali Delhi >༒8448380779 Escort ServiceBDSM⚡Call Girls in Mandawali Delhi >༒8448380779 Escort Service
BDSM⚡Call Girls in Mandawali Delhi >༒8448380779 Escort Service
 
FESE Capital Markets Fact Sheet 2024 Q1.pdf
FESE Capital Markets Fact Sheet 2024 Q1.pdfFESE Capital Markets Fact Sheet 2024 Q1.pdf
FESE Capital Markets Fact Sheet 2024 Q1.pdf
 
Discover Why Less is More in B2B Research
Discover Why Less is More in B2B ResearchDiscover Why Less is More in B2B Research
Discover Why Less is More in B2B Research
 
Vip Mumbai Call Girls Marol Naka Call On 9920725232 With Body to body massage...
Vip Mumbai Call Girls Marol Naka Call On 9920725232 With Body to body massage...Vip Mumbai Call Girls Marol Naka Call On 9920725232 With Body to body massage...
Vip Mumbai Call Girls Marol Naka Call On 9920725232 With Body to body massage...
 
Sampling (random) method and Non random.ppt
Sampling (random) method and Non random.pptSampling (random) method and Non random.ppt
Sampling (random) method and Non random.ppt
 
Generative AI on Enterprise Cloud with NiFi and Milvus
Generative AI on Enterprise Cloud with NiFi and MilvusGenerative AI on Enterprise Cloud with NiFi and Milvus
Generative AI on Enterprise Cloud with NiFi and Milvus
 
Mature dropshipping via API with DroFx.pptx
Mature dropshipping via API with DroFx.pptxMature dropshipping via API with DroFx.pptx
Mature dropshipping via API with DroFx.pptx
 
5CL-ADBA,5cladba, Chinese supplier, safety is guaranteed
5CL-ADBA,5cladba, Chinese supplier, safety is guaranteed5CL-ADBA,5cladba, Chinese supplier, safety is guaranteed
5CL-ADBA,5cladba, Chinese supplier, safety is guaranteed
 
Halmar dropshipping via API with DroFx
Halmar  dropshipping  via API with DroFxHalmar  dropshipping  via API with DroFx
Halmar dropshipping via API with DroFx
 
Accredited-Transport-Cooperatives-Jan-2021-Web.pdf
Accredited-Transport-Cooperatives-Jan-2021-Web.pdfAccredited-Transport-Cooperatives-Jan-2021-Web.pdf
Accredited-Transport-Cooperatives-Jan-2021-Web.pdf
 
April 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's AnalysisApril 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's Analysis
 
Call Girls In Attibele ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Attibele ☎ 7737669865 🥵 Book Your One night StandCall Girls In Attibele ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Attibele ☎ 7737669865 🥵 Book Your One night Stand
 
Predicting Loan Approval: A Data Science Project
Predicting Loan Approval: A Data Science ProjectPredicting Loan Approval: A Data Science Project
Predicting Loan Approval: A Data Science Project
 
CebaBaby dropshipping via API with DroFX.pptx
CebaBaby dropshipping via API with DroFX.pptxCebaBaby dropshipping via API with DroFX.pptx
CebaBaby dropshipping via API with DroFX.pptx
 
Mg Road Call Girls Service: 🍓 7737669865 🍓 High Profile Model Escorts | Banga...
Mg Road Call Girls Service: 🍓 7737669865 🍓 High Profile Model Escorts | Banga...Mg Road Call Girls Service: 🍓 7737669865 🍓 High Profile Model Escorts | Banga...
Mg Road Call Girls Service: 🍓 7737669865 🍓 High Profile Model Escorts | Banga...
 
Capstone Project on IBM Data Analytics Program
Capstone Project on IBM Data Analytics ProgramCapstone Project on IBM Data Analytics Program
Capstone Project on IBM Data Analytics Program
 

The importance of model fairness and interpretability in AI systems

  • 1.
  • 2. The Importance of Model Fairness and Interpretability in AI Systems
  • 3. Mehrnoosh Sameki Senior Product Manager Microsoft Azure ML Minsoo Thigpen Product Manager Microsoft Azure ML Ehi Nosakhare Machine Learning Scientist Microsoft AI Acceleration and Development Program
  • 4. Machine Fairness How to assess AI system’s fairness and mitigate any observed unfairness issues www.aka.ms/Fairlearn-build2020 Article
  • 7. AI will have a considerable impact on business and society as a whole
  • 8. AI impact raises a host of complex and challenging questions
  • 12. Transparency AI systems should be explainable AI systems should have algorithmic accountability
  • 13. Machine Learning Interpretability use cases www.aka.ms/InterpretML-github
  • 15. Explain Models - www.aka.ms/InterpretML-toolkit
  • 16. InterpretML Repo - www.aka.ms/InterpretML-github Interpret Tabular Data Additional interpretability techniques for tabular extension www.aka.ms/InterpretML-github github/InterpretML/Interpret-Community Interpret- Community
  • 17. Interpret Tabular Data Additional interpretability techniques for tabular extension github/InterpretML/Interpret github/InterpretML/Interpret-Community Interpret- Community AzureML- Interpret wrapper Azure Machine Learning SDK InterpretML Repo - www.aka.ms/InterpretML-github
  • 18. Interpret Tabular Data Additional interpretability techniques for tabular extension github/InterpretML/Interpret github/InterpretML/Interpret-Community Interpret- Community Interpret- text-contrib github/InterpretML/Interpret-text-contrib Text Data (Text Classification) AzureML- Interpret wrapper Azure Machine Learning SDK InterpretML Repo - www.aka.ms/InterpretML-github
  • 20. Interpretability Approaches • Given an existing model/system, extract explanations • Typically provide approximate explanations • Examples: LIME, SHAP, Permutation Feature Importance, Sensitivity Analysis, Influence Functions, etc. Tools to explain the system www.aka.ms/InterpretML-github
  • 21. SHAP (SHapley Additive exPlanations) • Not a new concept • Concept based on game theory • Mathematically sound • Application in ML relatively new www.aka.ms/InterpretML-github
  • 23. How much has each feature contributed to the prediction compared to the average prediction? House price prediction: € 300,000 Average house price prediction for all apartments is € 310,000 Delta here is - €10,000 www.aka.ms/InterpretML-github
  • 24. Parks contributed +10k Cats banned contributed -50k 50𝑚2 contributed +10k 2nd floor had net zero contributions www.aka.ms/InterpretML-github
  • 25. SHAP How do we calculate Shapley values? • Take your feature of interest (e.g., cats-banned) and remove it from the feature set • Take the remaining features and generate all possible coalitions • Add and remove your feature of interested to each of the coalitions and calculate the difference it makes www.aka.ms/InterpretML-github
  • 26. SHAP Pros SHAP Cons Based on a solid theory and distributes the effects fairly Computation time: 2k possible coalitions of the feature values for k features Contrastive explanations: Instead of comparing a prediction to the average prediction of the entire dataset, you could compare it to a subset or even to a single data point. Can be misinterpreted Inclusion of unrealistic data instances when features are correlated. www.aka.ms/InterpretML-github
  • 27. Interpretability Approaches • Models inherently interpretable • Typically provide exact explanations • Examples: Linear Models, Decision Rule Lists, Explainable Boosting Machines, Shallow Decision Trees, etc. Tools to explain the system www.aka.ms/InterpretML-github
  • 28. Linear Models • (Generalized) Linear Models • Current standard for interpretable models • Learns an additive relationship between data and response: • y = β0 + β1x1 + β2x2 + ... + βn xn • βi terms are scalar – single number that multiplies to each feature value • Each feature contributes a “score” that adds up to final output Example Model : House Price = 50,000 + 0.2*sq_ft + 10000*bedrooms + 7000*bathrooms Using this model on a new house: - 5 Bedrooms - 3 Bathrooms - 3000 Sq Feet - House Price = 50000 + 0.2*3000 + 10000*5 + 7000 * 3 - House Price = $121,600 www.aka.ms/InterpretML-github
  • 29. Explainable Boosting Machine • Generalized Additive Models (with Pairwise Interactions) • GAMs have existed since 1980s • MSR invented new methods of training GAMs for higher accuracy • Think of it as a “more expressive” linear model • Still additive! • Linear Model: y = β0 + β1x1 + β2x2 + ... + βn xn • Additive Model: y = f1(x1) + f2(x2) + ... + fn (xn) • Additive Model with Pairwise Interactions (EBM): y = Ʃi fi (xi) + Ʃij fij (xi , xj ) • Full Complexity Model: y = f (x1, ..., xn) Linear Model can only learn a straight line! www.aka.ms/InterpretML-github
  • 30. Demo of Interpretability for Tabular Data Hardware Performance Dataset - The Regression goal is to predict the performance of certain combinations of hardware parts. www.aka.ms/InterpretML-github
  • 31. Demo dataset https://archive.ics.uci.edu/ml/datasets/Computer+Hardware Hardware Performance Dataset - The Regression goal is to predict the performance of certain combinations of hardware parts. • vendor name: 30 (adviser, amdahl,apollo, basf, bti, burroughs, c.r.d, cambex, cdc, dec, dg, formation, four-phase, gould, honeywell, hp, ibm, ipl, magnuson, microdata, nas, ncr, nixdorf, perkin-elmer, prime, siemens, sperry, sratus, wang) • Model Name: many unique symbols • MYCT: machine cycle time in nanoseconds (integer) • MMIN: minimum main memory in kilobytes (integer) • MMAX: maximum main memory in kilobytes (integer) • CACH: cache memory in kilobytes (integer) • CHMIN: minimum channels in units (integer) • CHMAX: maximum channels in units (integer) • PRP: published relative performance (integer) • ERP: estimated relative performance from the original article (integer)
  • 32.
  • 33. FAIRNESS More positive outcomes & avoiding harmful outcomes of AI systems for groups of people
  • 34. • Allocation: extends or withholds opportunities, resources, or information. • Quality of service: whether a system works as well for one person as it does for another • Stereotyping: reinforce existing societal stereotypes • Denigration: actively derogatory or offensive • Over or under representation: over-represent, under-represent, or even erase particular groups of people Crawford et al. 2017 Types of harm
  • 35. What is Fairlearn? A new approach to measuring and mitigating unfairness in systems that make predictions, serve users, or make decisions about allocating resources, opportunities, or information. www.aka.ms/FairlearnAI
  • 36. There are many ways that an AI system can behave unfairly. Fairness in AI Avoiding negative outcomes of AI systems for different groups of people A model for screening loan or job application might be much better at picking good candidates among white men than among other groups. A voice recognition system might fail to work as well for women as it does for men. www.aka.ms/FairlearnAI
  • 37. A toolkit that empowers developers of artificial intelligence systems to assess their systems' fairness and mitigate any observed fairness issues. Helps users identify and mitigate unfairness in their machine learning models with a focus on group fairness. Automatically analyze a model’s predictions Provide the user with insights into (un)fairness of their model’s predictions Support (algorithmic) methods to mitigate unfairness www.aka.ms/FairlearnAI
  • 38. • Allocation: extends or withholds opportunities, resources, or information. • Quality of service: whether a system works as well for one person as it does for another • Stereotyping: reinforce existing societal stereotypes • Denigration: actively derogatory or offensive • Over or under representation: over-represent, under-represent, or even erase particular groups of people Crawford et al. 2017 Types of harm addressed by Fairlearn www.aka.ms/FairlearnAI
  • 39. Harm of “allocation” Example Scenarios: Lending The data set describes whether each individual repaid the loan or not. • [Classification] Recommends whether a given individual should get a loan. The trained model outputs: Yes/Maybe/No decision. Example Scenarios: School Admissions The data set describes what was the first-year GPA of each student. • [Regression] For a given applicant, predicts their GPA at the end of the first year. The trained model outputs a real-valued prediction that is used as a score to screen applicants. www.aka.ms/FairlearnAI
  • 40. Harm of “quality of service” Example Scenarios: News Recommendation The training data indicates what article was presented to which user, whether the user clicked, and how much time the user spent on the article. • [Classification] Predict which article to show to each user to optimize click- through rate. The trained model outputs: Yes/No decision. Two kinds of (group) fairness: across users (quality of service), or across publishers/topics of articles (quality of service, but also allocation). www.aka.ms/FairlearnAI
  • 41. Assessment Evaluation Visualizations Fairness Criteria Mitigation Algorithm Mitigation • Select a protected attribute • Select a fairness metric (e.g., disparity in recall, precision, etc and disparity in selection rate) Fairlearn Toolkit • Demographic Parity • Equalized Odds • Postprocessing Method • Reduction Methodwww.aka.ms/FairlearnAI
  • 42. Fairness assessment through disparity metrics Disparity in Performance • How the model accuracy differs across different buckets of a sensitive feature (e.g., how accuracy of the model differs for "female" vs. "male" vs. “unspecified" data points) Disparity in Selection Rate • How the model predictions differ across different buckets of a sensitive feature (e.g., how many "female" individuals have received prediction `approved` on their loan application in contrast to "male" and “unspecified" data points?). www.aka.ms/FairlearnAI
  • 43. Fairness assessment via Fairlearn visualization dashboard Let’s see a demo! www.aka.ms/FairlearnAI Mitigating Disparities in Ranking from Binary Data - An example based on the Law School Admissions Council's National Longitudinal Bar Passage Study
  • 44. Fairness assessment via Fairlearn visualization dashboard www.aka.ms/FairlearnAI
  • 45. Demo dataset Law School Admissions Council's (LSAC) National Longitudinal Bar Passage Study • The data set contains information about law students collected by LSAC between 1991 and 1997. • Some of the information is available at the admission time (such as the undergraduate GPA and LSAT score), and some describes the performance of the students once admitted. • We also have access to their self-identified race. • To simplify this example, we will limit the attention to those self-identified as black and white (two largest groups) and restrict our attention to two features (undergraduate GPA and LSAT score).
  • 46. Assessment Evaluation Visualizations Fairness Criteria Mitigation Algorithm Mitigation • Select a protected attribute • Select a fairness metric (e.g., disparity in recall, precision, etc and disparity in selection rate) Fairlearn Toolkit • Demographic Parity • Equalized Odds • Postprocessing Method • Reduction Methodwww.aka.ms/FairlearnAI
  • 47. Demographic parity: Applicants of each race (gender, ...) have the same odds of getting approval on their loan applications Loan approval decision is independent of protected attribute Equalized odds: Qualified applicants have the same odds of getting approval on their loan applications regardless of race (gender, …) Unqualified applicants have the same odds of getting approval on their loan applications regardless of race (gender, …) Fairness Criteria www.aka.ms/FairlearnAI
  • 48. Reductions approach: Wrapper around standard ML algorithms Input: • any standard ML training algorithm (as a black box) • data set including sensitive feature Output: • a trained model that minimizes error subject to fairness constraints Advantages: • doesn’t need to access the sensitive feature at test time • works for a wide range of disparity metrics • allows extracting the full disparity-accuracy frontier Disadvantages: • requires re-training: the black box is called 10-20 times www.aka.ms/FairlearnAI
  • 49. Post processing: Picking a fair threshold rule Input: • an existing (already trained) scoring model • data set including sensitive feature Output: • the most accurate among all fair threshold rules (a separate threshold for each subpopulation) Advantages: • simplicity • no need to re-train the model Disadvantages: • requires sensitive feature at test-time • doesn’t allow trade-offs between disparity and accuracy www.aka.ms/FairlearnAI
  • 50. Unfairness mitigation Let’s see a demo! www.aka.ms/FairlearnAI Mitigating Disparities in Ranking from Binary Data - An example based on the Law School Admissions Council's National Longitudinal Bar Passage Study
  • 51. Article: Machine Fairness How to assess AI system’s fairness and mitigate any observed unfairness issues www.aka.ms/Fairlearn-build2020
  • 52. Mehrnoosh Sameki Senior Product Manager Microsoft Azure ML Minsoo Thigpen Product Manager Microsoft Azure ML Ehi Nosakhare Machine Learning Scientist Microsoft AI Acceleration and Development Program