SlideShare ist ein Scribd-Unternehmen logo
1 von 40
NeurIPS Meetup Japan 2021, Satoshi Hara
Explanation in ML
and Its Reliability
Satoshi Hara
Osaka University
1
NeurIPS Meetup Japan 2021
NeurIPS Meetup Japan 2021, Satoshi Hara
“Explanation” in ML
◼ Most of ML models are highly complex, or “black-box”.
◼ “Explanation in ML”: Obtain some useful information
from the model (in addition to prediction).
2
Preliminary
You are
sick.
Why?
Your XX
score is
too high.
You are
sick.
Why?
???
I don’t
know.
…
XX score is
too high.
Oh…
NeurIPS Meetup Japan 2021, Satoshi Hara
[Typical Explanation 1] Saliency Map
◼ Generate heatmaps where the model has focused on
when making predictions.
3
Preliminary
The outline of zebra
seems to be relevant.
NeurIPS Meetup Japan 2021, Satoshi Hara
[Typical Explanation 2] Similar Examples
◼ Provide some similar examples to the input of interest.
4
These images look similar.
The prediction “Lapwing” will
be correct.
Lapwing
Database
Provide some similar examples
Input
Prediction
Lapwing
Preliminary
NeurIPS Meetup Japan 2021, Satoshi Hara
History of “Explanation”
◼ History of Saliency Map
5
Dawn
2014 2016 2018 2020
2015 2017 2019
Exponential Growth of
Saliency Map Algos
Attack & Manipulation
Sanity Check
[Adebayo+,2018]
GuidedBP
[Springenberg+,2014]
DeepLIFT
[Shrikumar+,2017]
Grad-CAM
[Selvaraju+,2017]
ROAR
[Hooker+,2019]
MoRF/Deletion Metric
[Bach+,2015; Vitali+,2018]
LeRF/Insertion Metric
[Arras+,2017; Vitali+,2018]
Sensitivity
[Kindermans+,2017]
Evaluation Methods
Saliency
[Simonyan+,2014]
IntGrad
[Sundararajan+,2017]
SHAP
[Lundberg+,2017]
LIME
[Ribeiro+,2016]
LRP
[Bach+,2015]
Fairwashing
[Aivodji+,2019]
SmoothGrad
[Smilkov+,2017]
DeepTaylor
[Montavon+,2017]
Occlusion
[Zeiler+,2014]
CAM
[Zhou+,2016]
Manipulation
[Domobrowski+,2019]
The papers on “Explanation”
increased exponentially.
2022 2021 2020 2019 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008
800
700
600
500
400
300
200
100
0
Searched
“Interpretable Machine Learning”
and
“Explainable AI”
on Web of Science
Preliminary
NeurIPS Meetup Japan 2021, Satoshi Hara
History of “Explanation”
◼ History of Saliency Map
6
Dawn
2014 2016 2018 2020
2015 2017 2019
Exponential Growth of
Saliency Map Algos
Attack & Manipulation
Sanity Check
[Adebayo+,2018]
GuidedBP
[Springenberg+,2014]
DeepLIFT
[Shrikumar+,2017]
Grad-CAM
[Selvaraju+,2017]
ROAR
[Hooker+,2019]
MoRF/Deletion Metric
[Bach+,2015; Vitali+,2018]
LeRF/Insertion Metric
[Arras+,2017; Vitali+,2018]
Sensitivity
[Kindermans+,2017]
Evaluation Methods
Saliency
[Simonyan+,2014]
IntGrad
[Sundararajan+,2017]
SHAP
[Lundberg+,2017]
LIME
[Ribeiro+,2016]
LRP
[Bach+,2015]
Fairwashing
[Aivodji+,2019]
SmoothGrad
[Smilkov+,2017]
DeepTaylor
[Montavon+,2017]
Occlusion
[Zeiler+,2014]
CAM
[Zhou+,2016]
Manipulation
[Domobrowski+,2019]
The papers on “Explanation”
increased exponentially.
800
700
600
500
400
300
200
100
0
Searched
“Interpretable Machine Learning”
and
“Explainable AI”
on Web of Science
Reliability of “Explanation” has raised
as a crucial concern.
Are the “Explanation” truly valid?
With “Explanation”, how malicious
we can be?
Preliminary
2022 2021 2020 2019 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008
NeurIPS Meetup Japan 2021, Satoshi Hara
Technical / Social Reliability of “Explanation”
Technical Reliability “Is the explanation valid?”
What we care:
• Do the algorithms output valid “Explanation”?
Research Question:
• How can we evaluate the validity of “Explanation”?
Social Reliability “Does explanation harm the society?”
What we care:
• What will happen if we introduce “Explanation” to society?
Research Question:
• Are there any malicious use cases of “Explanation”?
7
Technical Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Faithfulness & Plausibility of “Explanation”
◼ Faithfulness [Lakkaraju+’19; Jacovi+’20]
• Does “Explanation” reflect the model’s reasoning process?
- Our interest is “How and why the model predicted that way.”
• Any “Explanation” irrelevant to the reasoning process is invalid.
- e.g. “Explanation” outputs something independent of the model.
◼ Plausibility [Lage+’19; Strout+’19]
• Does “Explanation” make sense to the users?
• Any “Explanation” unacceptable by the users is not ideal.
- e.g. Entire program code; Very noisy saliency map.
8
Technical Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Evaluation of “Explanation”
◼ Based on Faithfulness
• Sanity Checks for Saliency Maps, NeurIPS’18.
- Julius Adebayo, Justin Gilmer, Michael Muelly, Ian Goodfellow, Moritz Hardt, Been Kim
• An epoch-making paper by Google Brain.
• Evaluation of Faithfulness for saliency maps.
◼ Based on Plausibility
• Evaluation of Similarity-based Explanations, ICLR’21.
- Kazuaki Hanawa, Sho Yokoi, Satoshi Hara, Kentaro Inui
• Evaluation of Plausibility for similarity-based explanations.
9
NeurIPS Meetup Japan 2021, Satoshi Hara
Evaluation of Saliency Map
◼ Plausibility
• All the maps look more or less plausible.
• Gradient, IntegratedGrad are bit noisy.
◼ Faithfulness?
10
Technical Reliability
The outline of zebra
seems to be relevant.
NeurIPS Meetup Japan 2021, Satoshi Hara
Evaluation of Faithfulness is Not Possible.
◼ Faithfulness
• Does “Explanation” reflect the model’s reasoning process?
◼ Alternative: Sanity Check
• Check the necessary condition for faithful “Explanation”.
◼ Q. What is the necessary condition?
• “Explanation” is model-dependent.
- Any “Explanation” irrelevant to the reasoning process is invalid.
11
Unknown
→ We cannot compare with Ground Truth.
[Remark] Passing Sanity Check alone
does not guarantee faithfulness.
Technical Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Model Parameter Randomization Test
◼ Compare “Explanation” of two models with different
reasoning processes.
• Faithful “Explanation” → Outputs are different.
• Non-Faithful “Explanation” → Outputs can be identical.
12
Satisfies the necessary condition.
Passed the sanity check.
Technical Reliability
[Assumption]
These models have
different reasoning
processes.
Model 1: Fully Trained Model 2: Randomly Initialized
Input “Explanation”
by Algo. 1
“Explanation”
by Algo. 2
“Explanation” by Algo. 1 are different.
“Explanation” by Algo. 2 are identical.
Violates the necessary condition.
Failed the sanity check.
NeurIPS Meetup Japan 2021, Satoshi Hara
Model Parameter Randomization Test
◼ Model 2: DNN with last few layers randomized.
• Saliency Maps of Guided Backprop and Guided GradCAM are
invariant against model randomization.
→ They violate the necessary condition for faithfulness.
13
Model
1
Model
2
[Ref] Sanity Checks for Saliency Maps
Technical Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Evaluation of “Explanation”
◼ Based on Faithfulness
• Sanity Checks for Saliency Maps, NeurIPS’18.
- Julius Adebayo, Justin Gilmer, Michael Muelly, Ian Goodfellow, Moritz Hardt, Been Kim
• An epoch-making paper by Google Brain.
• Evaluation of Faithfulness for saliency maps.
◼ Based on Plausibility
• Evaluation of Similarity-based Explanations, ICLR’21.
- Kazuaki Hanawa, Sho Yokoi, Satoshi Hara, Kentaro Inui
• Evaluation of Plausibility for similarity-based explanations.
14
NeurIPS Meetup Japan 2021, Satoshi Hara
Evaluation of Similarity-based Explanation
◼ Faithfulness
• We can use Model Parameter Randomization Test.
◼ Plausibility?
15
These images look similar.
The prediction “Lapwing” will
be correct.
Lapwing
Database
Provide some similar examples
Input
Prediction
Lapwing
Technical Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Plausibility in Similarity-based Explanation
◼ Example
• Explanation B won’t be acceptable by the users.
- Plausibility of Explanation A > Plausibility of Explanation B
16
Database
frog
Explanation A
Database
truck
Explanation B
frog
Input
Prediction
Technical Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Evaluation of Plausibility is Not Possible.
◼ There is no universal criterion that determines the
acceptability of the users.
◼ Alternative: Sanity Check
• Check the necessary condition for faithful “Plausibility”.
◼ Q. What is the necessary condition?
• Obtained similar instance should belong to the same class.
17
is cat because a similar is cat.
is cat because a similar is dog.
Plausible
Non-Plausible
Identical Class Test
Technical Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Identical Class Test
18
Input
Dot Last Layer
All Layers
Input
Cos Last Layer
All Layers
Input
L2 Dist. Last Layer
All Layers
Influence Function
Relative IF
Fisher Kernel
Dot
Cos
Parameter Grad.
Fraction of Test Instances Passed Identical Class Test
0 0.5 1.0 0 0.5 1.0
(Image Clf.)
CIFAR10
+ CNN
(Text Clf.)
AGNews
+ Bi-LSTM
Cosine similarity of the
parameter gradient
performed almost perfectly.
Technical Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Cosine of Parameter Gradient
• GC 𝑧, 𝑧′ =
∇𝜃ℓ 𝑦,𝑓𝜃 𝑥 ,∇𝜃ℓ 𝑦′,𝑓𝜃 𝑥′
∇𝜃ℓ 𝑦,𝑓𝜃 𝑥 ∇𝜃ℓ 𝑦′,𝑓𝜃 𝑥′
19
Sussex spaniel beer bottle mobile house
Technical Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Technical / Social Reliability of “Explanation”
Technical Reliability “Is the explanation valid?”
What we care:
• Do the algorithms output valid “Explanation”?
Research Question:
• How can we evaluate the validity of “Explanation”?
Social Reliability “Does explanation harm the society?”
What we care:
• What will happen if we introduce “Explanation” to society?
Research Question:
• Are there any malicious use cases of “Explanation”?
20
Social Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Malicious Use Cases of “Explanation”
◼ Q. Are there malicious use cases of “Explanation”?
A. Some may try to deceive people
by providing fake explanations.
◼ Q. When and why fake explanations can be used?
A. Fake explanations can show models better,
e.g., by pretending as if the models are fair.
◼ Q. Why we need to research fake explanations?
Are you evil?
A. We need to know how malicious we can be with fake
explanations. Otherwise, we cannot defend against
possible maliciousness.
21
Social Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Fake “Explanation” for Fairness
◼ Fairness in ML
• Models can be biased towards gender, race, etc.
• Ensuring fairness of the models is crucial nowadays.
◼ What if we cannot detect the use of unfair models?
• Some may use unfair models.
- Unfair models are typically more accurate than the fair ones.
22
Social Reliability
Our model is the most accurate one in this business field.
(because of the use of unfair yet accurate model)
Moreover, our model is fair without any bias.
(by showing fake explanation)
NeurIPS Meetup Japan 2021, Satoshi Hara
Fake “Explanation” for Fairness
◼ Fake “Explanation” by Surrogate Models
• Fairwashing: the risk of rationalization, ICML’19.
- Ulrich Aïvodji, Hiromi Arai, Olivier Fortineau, Sébastien Gambs, Satoshi Hara, Alain Tapp
• Characterizing the risk of fairwashing, NeurIPS’21.
- Ulrich Aïvodji, Hiromi Arai, Sébastien Gambs, Satoshi Hara
◼ Fake “Explanation” by Examples
• Faking Fairness via Stealthily Biased Sampling, AAAI’20.
- Kazuto Fukuchi, Satoshi Hara, Takanori Maehara
◼ Ref.
• It’s Too Easy to Hide Bias in Deep-Learning Systems,
IEEE Spectrum, 2021.
23
NeurIPS Meetup Japan 2021, Satoshi Hara
The risk of “Fairwashing”
◼ Explaining fairness
24
an honest explanation
Your loan application is rejected
because your gender is …
Unfair AI: reject applicants
based on their gender.
Social Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
The risk of “Fairwashing”
◼ Explaining fairness
25
a dishonest explanation
Your loan application is rejected
because your income is low.
Unfair AI: reject applicants
based on their gender.
Social Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
The risk of “Fairwashing”
◼ Explaining fairness
26
Unfair AI: reject applicants
based on their gender.
a dishonest explanation
Your loan application is rejected
because your income is low.
“Fairwashing”
Malicious decision-makers can disclose a fake
explanation to rationalize their unfair decisions.
“Fairwashing”
Social Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
The risk of “Fairwashing”
◼ Explaining fairness
27
Unfair AI: reject applicants
based on their gender.
a dishonest explanation
Your loan application is rejected
because your income is low.
This Study: LaundryML
Possible to systematically generate
fake explanations.
Raise the awareness of the risk of
“Fairwashing”.
“Fairwashing”
Malicious decision-makers can disclose a fake
explanation to rationalize their unfair decisions.
“Fairwashing”
Social Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
◼ The idea
Generate many explanations,
and pick one that is useful for “Fairwashing”.
◼ many explanations
• Use “Model Enumeration” [Hara & Maehara’17; Hara & Ishihata’18]
• Enumerate explanation models.
◼ pick one
• Use fairness metrices such as demographic parity (DP).
• Pick an explanation most faithful to the model, with DP less
than a threshold.
28
LaundryML
Systematically generating fake explanations
The idea
Social Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Result
◼ “Fairwashing” for decisions on Adult dataset
• Feature importance by FairML on “gender” has dropped.
29
A naïve explanation A fake explanation
gender
gender
Social Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Result
◼ “Fairwashing” for decisions on Adult dataset
• Feature importance by FairML on “gender” has dropped.
30
A naïve explanation A false explanation
gender
gender
If
else if
else if
else if
else if
else low-income
then high-income
then low-income
then low-income
then low-income
then high-income
capital gain > 7056
marital = single
education = HS-grad
occupation = other
occupation = white-colloar
Fake Explanation
Social Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Fake “Explanation” for Fairness
◼ Fake “Explanation” by Surrogate Models
• Fairwashing: the risk of rationalization, ICML’19.
- Ulrich Aïvodji, Hiromi Arai, Olivier Fortineau, Sébastien Gambs, Satoshi Hara, Alain Tapp
• Characterizing the risk of fairwashing, NeurIPS’21.
- Ulrich Aïvodji, Hiromi Arai, Sébastien Gambs, Satoshi Hara
◼ Fake “Explanation” by Examples
• Faking Fairness via Stealthily Biased Sampling, AAAI’20.
- Kazuto Fukuchi, Satoshi Hara, Takanori Maehara
◼ Ref.
• It’s Too Easy to Hide Bias in Deep-Learning Systems,
IEEE Spectrum, 2021.
31
NeurIPS Meetup Japan 2021, Satoshi Hara
Fairness Metrics
◼ Quantifying fairness of the models
• Several metrics + toolboxes
- FairML, AI Fairness 360 [Bellamy+’19], Aequitas [Saleiro+’18]
32
AI Fairness 360
Social Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Fake Fairness Metrics
33
Malicious Party
Unfair Model
Service
Fairness Metric
as Evidence
Is this a fake
metric.
No guarantee whether the metrics are
computed appropriately.
Impossible to determine fake or not.
Metric alone is not a valid evidence of fairness.
Social Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Avoiding Fake Fairness Metrics
34
Malicious Party
Unfair Model
Service
Benchmark Data
as Evidence
Fairness metric computed
on the benchmark is fair!
The metric is reproducible
using the benchmark data.
We can avoid fake!
Social Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
(Failed) Avoiding Fake Fairness Metrics
35
Malicious Party
Unfair Model
Service
Benchmark Data
as Evidence
Fairness metric computed
on the benchmark is fair!
The metric is reproducible
using the benchmark data.
We can avoid fake!
The benchmark data can be fake.
Social Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Generating Fake Benchmark
◼ Subsample the benchmark dataset 𝑆
from the original dataset 𝐷.
◼ “Ideal” Fake Benchmark Dataset 𝑆
• Fairness : Fairness metric computed on 𝑆 is fair.
• Stealthiness : The distribution of 𝑆 is close to 𝐷.
36
Benchmark
Fairness
Stealthiness
“Fair” Contingency Table
Original dataset
Social Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
参照用
データ
Goodness-of-Fit Test
Generating Fake Benchmark
◼ Optimization of 𝑆 as LP (Min-Cost Flow)
min𝑆𝑊 𝑆, 𝐷 , s. t. 𝐶 𝑆 = 𝐶𝑇
◼ Detection of fake benchmark using statistical test.
• Min. Distribution diff. ≈ small detection probability
• Rejecting 𝑝 𝑆 = 𝑝(𝐷′) with KS test is probability
at most 𝑂 𝑆 𝛼 × Distribution diff.
37
Stealthiness
(Min. Distribution diff.)
Fairness
(Constraint on Contingency Table)
Reference
Data
Social Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Undetectability of Fake Benchmark
38
Positive Cases in Contingency Table Positive Cases in Contingency Table
Fairness
Metric
(DP)
Distribution
diff.
COMPAS
Positive Cases in Contingency Table Positive Cases in Contingency Table
Fairness
Metric
(DP)
Distribution
diff.
Adult
Random Sampling
Case-Control Sampling
Proposed Sampling
Random Sampling
Case-Control Sampling
Proposed Sampling
Proposed sampling resulted to
fairer metric.
(= achieved fake fairness)
Proposed sampling attained distribution
almost identical to the original distribution.
(= undetectable)
Social Reliability
NeurIPS Meetup Japan 2021, Satoshi Hara
Technical / Social Reliability of “Explanation”
Technical Reliability “Is the explanation valid?”
What we care:
• Do the algorithms output valid “Explanation”?
Research Question:
• How can we evaluate the validity of “Explanation”?
Social Reliability “Does explanation harm the society?”
What we care:
• What will happen if we introduce “Explanation” to society?
Research Question:
• Are there any malicious use cases of “Explanation”?
39
Summary
NeurIPS Meetup Japan 2021, Satoshi Hara
Technical / Social Reliability of “Explanation”
Technical Reliability “Is the explanation valid?”
What we care:
• Do the algorithms output valid “Explanation”?
Research Question:
• How can we evaluate the validity of “Explanation”?
Social Reliability “Does explanation harm the society?”
What we care:
• What will happen if we introduce “Explanation” to society?
Research Question:
• Are there any malicious use cases of “Explanation”?
40
Summary
How can we evaluate the validity of “Explanation”?
Which evaluation is good for which “Explanation”?
When “Explanation” can be used maliciously?
Can we detect malicious use cases?

Weitere ähnliche Inhalte

Was ist angesagt?

SSII2021 [OS2-02] 深層学習におけるデータ拡張の原理と最新動向
SSII2021 [OS2-02] 深層学習におけるデータ拡張の原理と最新動向SSII2021 [OS2-02] 深層学習におけるデータ拡張の原理と最新動向
SSII2021 [OS2-02] 深層学習におけるデータ拡張の原理と最新動向SSII
 
[DL輪読会]相互情報量最大化による表現学習
[DL輪読会]相互情報量最大化による表現学習[DL輪読会]相互情報量最大化による表現学習
[DL輪読会]相互情報量最大化による表現学習Deep Learning JP
 
因果探索: 基本から最近の発展までを概説
因果探索: 基本から最近の発展までを概説因果探索: 基本から最近の発展までを概説
因果探索: 基本から最近の発展までを概説Shiga University, RIKEN
 
IIBMP2016 深層生成モデルによる表現学習
IIBMP2016 深層生成モデルによる表現学習IIBMP2016 深層生成モデルによる表現学習
IIBMP2016 深層生成モデルによる表現学習Preferred Networks
 
[DL輪読会]Understanding Black-box Predictions via Influence Functions
[DL輪読会]Understanding Black-box Predictions via Influence Functions [DL輪読会]Understanding Black-box Predictions via Influence Functions
[DL輪読会]Understanding Black-box Predictions via Influence Functions Deep Learning JP
 
【論文調査】XAI技術の効能を ユーザ実験で評価する研究
【論文調査】XAI技術の効能を ユーザ実験で評価する研究【論文調査】XAI技術の効能を ユーザ実験で評価する研究
【論文調査】XAI技術の効能を ユーザ実験で評価する研究Satoshi Hara
 
機械学習による統計的実験計画(ベイズ最適化を中心に)
機械学習による統計的実験計画(ベイズ最適化を中心に)機械学習による統計的実験計画(ベイズ最適化を中心に)
機械学習による統計的実験計画(ベイズ最適化を中心に)Kota Matsui
 
畳み込みニューラルネットワークの高精度化と高速化
畳み込みニューラルネットワークの高精度化と高速化畳み込みニューラルネットワークの高精度化と高速化
畳み込みニューラルネットワークの高精度化と高速化Yusuke Uchida
 
“機械学習の説明”の信頼性
“機械学習の説明”の信頼性“機械学習の説明”の信頼性
“機械学習の説明”の信頼性Satoshi Hara
 
2014 3 13(テンソル分解の基礎)
2014 3 13(テンソル分解の基礎)2014 3 13(テンソル分解の基礎)
2014 3 13(テンソル分解の基礎)Tatsuya Yokota
 
テーブル・テキスト・画像の反実仮想説明
テーブル・テキスト・画像の反実仮想説明テーブル・テキスト・画像の反実仮想説明
テーブル・テキスト・画像の反実仮想説明tmtm otm
 
【メタサーベイ】数式ドリブン教師あり学習
【メタサーベイ】数式ドリブン教師あり学習【メタサーベイ】数式ドリブン教師あり学習
【メタサーベイ】数式ドリブン教師あり学習cvpaper. challenge
 
【論文紹介】How Powerful are Graph Neural Networks?
【論文紹介】How Powerful are Graph Neural Networks?【論文紹介】How Powerful are Graph Neural Networks?
【論文紹介】How Powerful are Graph Neural Networks?Masanao Ochi
 
Sparse Codingをなるべく数式を使わず理解する(PCAやICAとの関係)
Sparse Codingをなるべく数式を使わず理解する(PCAやICAとの関係)Sparse Codingをなるべく数式を使わず理解する(PCAやICAとの関係)
Sparse Codingをなるべく数式を使わず理解する(PCAやICAとの関係)Teppei Kurita
 
SSII2022 [TS1] Transformerの最前線〜 畳込みニューラルネットワークの先へ 〜
SSII2022 [TS1] Transformerの最前線〜 畳込みニューラルネットワークの先へ 〜SSII2022 [TS1] Transformerの最前線〜 畳込みニューラルネットワークの先へ 〜
SSII2022 [TS1] Transformerの最前線〜 畳込みニューラルネットワークの先へ 〜SSII
 
【DL輪読会】言語以外でのTransformerのまとめ (ViT, Perceiver, Frozen Pretrained Transformer etc)
【DL輪読会】言語以外でのTransformerのまとめ (ViT, Perceiver, Frozen Pretrained Transformer etc)【DL輪読会】言語以外でのTransformerのまとめ (ViT, Perceiver, Frozen Pretrained Transformer etc)
【DL輪読会】言語以外でのTransformerのまとめ (ViT, Perceiver, Frozen Pretrained Transformer etc)Deep Learning JP
 
[DL輪読会]Set Transformer: A Framework for Attention-based Permutation-Invariant...
[DL輪読会]Set Transformer: A Framework for Attention-based Permutation-Invariant...[DL輪読会]Set Transformer: A Framework for Attention-based Permutation-Invariant...
[DL輪読会]Set Transformer: A Framework for Attention-based Permutation-Invariant...Deep Learning JP
 
失敗から学ぶ機械学習応用
失敗から学ぶ機械学習応用失敗から学ぶ機械学習応用
失敗から学ぶ機械学習応用Hiroyuki Masuda
 
【DL輪読会】Scaling Laws for Neural Language Models
【DL輪読会】Scaling Laws for Neural Language Models【DL輪読会】Scaling Laws for Neural Language Models
【DL輪読会】Scaling Laws for Neural Language ModelsDeep Learning JP
 
Active Learning 入門
Active Learning 入門Active Learning 入門
Active Learning 入門Shuyo Nakatani
 

Was ist angesagt? (20)

SSII2021 [OS2-02] 深層学習におけるデータ拡張の原理と最新動向
SSII2021 [OS2-02] 深層学習におけるデータ拡張の原理と最新動向SSII2021 [OS2-02] 深層学習におけるデータ拡張の原理と最新動向
SSII2021 [OS2-02] 深層学習におけるデータ拡張の原理と最新動向
 
[DL輪読会]相互情報量最大化による表現学習
[DL輪読会]相互情報量最大化による表現学習[DL輪読会]相互情報量最大化による表現学習
[DL輪読会]相互情報量最大化による表現学習
 
因果探索: 基本から最近の発展までを概説
因果探索: 基本から最近の発展までを概説因果探索: 基本から最近の発展までを概説
因果探索: 基本から最近の発展までを概説
 
IIBMP2016 深層生成モデルによる表現学習
IIBMP2016 深層生成モデルによる表現学習IIBMP2016 深層生成モデルによる表現学習
IIBMP2016 深層生成モデルによる表現学習
 
[DL輪読会]Understanding Black-box Predictions via Influence Functions
[DL輪読会]Understanding Black-box Predictions via Influence Functions [DL輪読会]Understanding Black-box Predictions via Influence Functions
[DL輪読会]Understanding Black-box Predictions via Influence Functions
 
【論文調査】XAI技術の効能を ユーザ実験で評価する研究
【論文調査】XAI技術の効能を ユーザ実験で評価する研究【論文調査】XAI技術の効能を ユーザ実験で評価する研究
【論文調査】XAI技術の効能を ユーザ実験で評価する研究
 
機械学習による統計的実験計画(ベイズ最適化を中心に)
機械学習による統計的実験計画(ベイズ最適化を中心に)機械学習による統計的実験計画(ベイズ最適化を中心に)
機械学習による統計的実験計画(ベイズ最適化を中心に)
 
畳み込みニューラルネットワークの高精度化と高速化
畳み込みニューラルネットワークの高精度化と高速化畳み込みニューラルネットワークの高精度化と高速化
畳み込みニューラルネットワークの高精度化と高速化
 
“機械学習の説明”の信頼性
“機械学習の説明”の信頼性“機械学習の説明”の信頼性
“機械学習の説明”の信頼性
 
2014 3 13(テンソル分解の基礎)
2014 3 13(テンソル分解の基礎)2014 3 13(テンソル分解の基礎)
2014 3 13(テンソル分解の基礎)
 
テーブル・テキスト・画像の反実仮想説明
テーブル・テキスト・画像の反実仮想説明テーブル・テキスト・画像の反実仮想説明
テーブル・テキスト・画像の反実仮想説明
 
【メタサーベイ】数式ドリブン教師あり学習
【メタサーベイ】数式ドリブン教師あり学習【メタサーベイ】数式ドリブン教師あり学習
【メタサーベイ】数式ドリブン教師あり学習
 
【論文紹介】How Powerful are Graph Neural Networks?
【論文紹介】How Powerful are Graph Neural Networks?【論文紹介】How Powerful are Graph Neural Networks?
【論文紹介】How Powerful are Graph Neural Networks?
 
Sparse Codingをなるべく数式を使わず理解する(PCAやICAとの関係)
Sparse Codingをなるべく数式を使わず理解する(PCAやICAとの関係)Sparse Codingをなるべく数式を使わず理解する(PCAやICAとの関係)
Sparse Codingをなるべく数式を使わず理解する(PCAやICAとの関係)
 
SSII2022 [TS1] Transformerの最前線〜 畳込みニューラルネットワークの先へ 〜
SSII2022 [TS1] Transformerの最前線〜 畳込みニューラルネットワークの先へ 〜SSII2022 [TS1] Transformerの最前線〜 畳込みニューラルネットワークの先へ 〜
SSII2022 [TS1] Transformerの最前線〜 畳込みニューラルネットワークの先へ 〜
 
【DL輪読会】言語以外でのTransformerのまとめ (ViT, Perceiver, Frozen Pretrained Transformer etc)
【DL輪読会】言語以外でのTransformerのまとめ (ViT, Perceiver, Frozen Pretrained Transformer etc)【DL輪読会】言語以外でのTransformerのまとめ (ViT, Perceiver, Frozen Pretrained Transformer etc)
【DL輪読会】言語以外でのTransformerのまとめ (ViT, Perceiver, Frozen Pretrained Transformer etc)
 
[DL輪読会]Set Transformer: A Framework for Attention-based Permutation-Invariant...
[DL輪読会]Set Transformer: A Framework for Attention-based Permutation-Invariant...[DL輪読会]Set Transformer: A Framework for Attention-based Permutation-Invariant...
[DL輪読会]Set Transformer: A Framework for Attention-based Permutation-Invariant...
 
失敗から学ぶ機械学習応用
失敗から学ぶ機械学習応用失敗から学ぶ機械学習応用
失敗から学ぶ機械学習応用
 
【DL輪読会】Scaling Laws for Neural Language Models
【DL輪読会】Scaling Laws for Neural Language Models【DL輪読会】Scaling Laws for Neural Language Models
【DL輪読会】Scaling Laws for Neural Language Models
 
Active Learning 入門
Active Learning 入門Active Learning 入門
Active Learning 入門
 

Ähnlich wie Explanation in Machine Learning and Its Reliability

Distributed processing of large graphs in python
Distributed processing of large graphs in pythonDistributed processing of large graphs in python
Distributed processing of large graphs in pythonJose Quesada (hiring)
 
All in AI: LLM Landscape & RAG in 2024 with Mark Ryan (Google) & Jerry Liu (L...
All in AI: LLM Landscape & RAG in 2024 with Mark Ryan (Google) & Jerry Liu (L...All in AI: LLM Landscape & RAG in 2024 with Mark Ryan (Google) & Jerry Liu (L...
All in AI: LLM Landscape & RAG in 2024 with Mark Ryan (Google) & Jerry Liu (L...Daniel Zivkovic
 
DRONE: A Tool to Detect and Repair Directive Defects in Java APIs Documentation
DRONE: A Tool to Detect and Repair Directive Defects in Java APIs DocumentationDRONE: A Tool to Detect and Repair Directive Defects in Java APIs Documentation
DRONE: A Tool to Detect and Repair Directive Defects in Java APIs DocumentationSebastiano Panichella
 
DutchMLSchool. Logistic Regression, Deepnets, Time Series
DutchMLSchool. Logistic Regression, Deepnets, Time SeriesDutchMLSchool. Logistic Regression, Deepnets, Time Series
DutchMLSchool. Logistic Regression, Deepnets, Time SeriesBigML, Inc
 
Keepler Data Tech | Entendiendo tus propios modelos predictivos
Keepler Data Tech | Entendiendo tus propios modelos predictivosKeepler Data Tech | Entendiendo tus propios modelos predictivos
Keepler Data Tech | Entendiendo tus propios modelos predictivosKeepler Data Tech
 
Workshop Tel Aviv - Graph Data Science
Workshop Tel Aviv - Graph Data ScienceWorkshop Tel Aviv - Graph Data Science
Workshop Tel Aviv - Graph Data ScienceNeo4j
 
Graphs for Ai and ML
Graphs for Ai and MLGraphs for Ai and ML
Graphs for Ai and MLNeo4j
 
Hacking Predictive Modeling - RoadSec 2018
Hacking Predictive Modeling - RoadSec 2018Hacking Predictive Modeling - RoadSec 2018
Hacking Predictive Modeling - RoadSec 2018HJ van Veen
 
Building Large-scale Real-world Recommender Systems - Recsys2012 tutorial
Building Large-scale Real-world Recommender Systems - Recsys2012 tutorialBuilding Large-scale Real-world Recommender Systems - Recsys2012 tutorial
Building Large-scale Real-world Recommender Systems - Recsys2012 tutorialXavier Amatriain
 
Variants of GANs - Jaejun Yoo
Variants of GANs - Jaejun YooVariants of GANs - Jaejun Yoo
Variants of GANs - Jaejun YooJaeJun Yoo
 
[PR12] understanding deep learning requires rethinking generalization
[PR12] understanding deep learning requires rethinking generalization[PR12] understanding deep learning requires rethinking generalization
[PR12] understanding deep learning requires rethinking generalizationJaeJun Yoo
 
Mixed Effects Models - Random Intercepts
Mixed Effects Models - Random InterceptsMixed Effects Models - Random Intercepts
Mixed Effects Models - Random InterceptsScott Fraundorf
 
Neural Nets Deconstructed
Neural Nets DeconstructedNeural Nets Deconstructed
Neural Nets DeconstructedPaul Sterk
 
ntroducing to the Power of Graph Technology
ntroducing to the Power of Graph Technologyntroducing to the Power of Graph Technology
ntroducing to the Power of Graph TechnologyNeo4j
 
Visual geometry with deep learning
Visual geometry with deep learningVisual geometry with deep learning
Visual geometry with deep learningNAVER Engineering
 
Explainability and bias in AI
Explainability and bias in AIExplainability and bias in AI
Explainability and bias in AIBill Liu
 
Explainability for Learning to Rank
Explainability for Learning to RankExplainability for Learning to Rank
Explainability for Learning to RankSease
 

Ähnlich wie Explanation in Machine Learning and Its Reliability (20)

UseR 2017
UseR 2017UseR 2017
UseR 2017
 
Distributed processing of large graphs in python
Distributed processing of large graphs in pythonDistributed processing of large graphs in python
Distributed processing of large graphs in python
 
All in AI: LLM Landscape & RAG in 2024 with Mark Ryan (Google) & Jerry Liu (L...
All in AI: LLM Landscape & RAG in 2024 with Mark Ryan (Google) & Jerry Liu (L...All in AI: LLM Landscape & RAG in 2024 with Mark Ryan (Google) & Jerry Liu (L...
All in AI: LLM Landscape & RAG in 2024 with Mark Ryan (Google) & Jerry Liu (L...
 
DRONE: A Tool to Detect and Repair Directive Defects in Java APIs Documentation
DRONE: A Tool to Detect and Repair Directive Defects in Java APIs DocumentationDRONE: A Tool to Detect and Repair Directive Defects in Java APIs Documentation
DRONE: A Tool to Detect and Repair Directive Defects in Java APIs Documentation
 
DutchMLSchool. Logistic Regression, Deepnets, Time Series
DutchMLSchool. Logistic Regression, Deepnets, Time SeriesDutchMLSchool. Logistic Regression, Deepnets, Time Series
DutchMLSchool. Logistic Regression, Deepnets, Time Series
 
Keepler Data Tech | Entendiendo tus propios modelos predictivos
Keepler Data Tech | Entendiendo tus propios modelos predictivosKeepler Data Tech | Entendiendo tus propios modelos predictivos
Keepler Data Tech | Entendiendo tus propios modelos predictivos
 
Workshop Tel Aviv - Graph Data Science
Workshop Tel Aviv - Graph Data ScienceWorkshop Tel Aviv - Graph Data Science
Workshop Tel Aviv - Graph Data Science
 
Graphs for Ai and ML
Graphs for Ai and MLGraphs for Ai and ML
Graphs for Ai and ML
 
Hacking Predictive Modeling - RoadSec 2018
Hacking Predictive Modeling - RoadSec 2018Hacking Predictive Modeling - RoadSec 2018
Hacking Predictive Modeling - RoadSec 2018
 
Building Large-scale Real-world Recommender Systems - Recsys2012 tutorial
Building Large-scale Real-world Recommender Systems - Recsys2012 tutorialBuilding Large-scale Real-world Recommender Systems - Recsys2012 tutorial
Building Large-scale Real-world Recommender Systems - Recsys2012 tutorial
 
Variants of GANs - Jaejun Yoo
Variants of GANs - Jaejun YooVariants of GANs - Jaejun Yoo
Variants of GANs - Jaejun Yoo
 
[PR12] understanding deep learning requires rethinking generalization
[PR12] understanding deep learning requires rethinking generalization[PR12] understanding deep learning requires rethinking generalization
[PR12] understanding deep learning requires rethinking generalization
 
Mixed Effects Models - Random Intercepts
Mixed Effects Models - Random InterceptsMixed Effects Models - Random Intercepts
Mixed Effects Models - Random Intercepts
 
Neural Nets Deconstructed
Neural Nets DeconstructedNeural Nets Deconstructed
Neural Nets Deconstructed
 
Where Does It Break?
Where Does It Break?Where Does It Break?
Where Does It Break?
 
ntroducing to the Power of Graph Technology
ntroducing to the Power of Graph Technologyntroducing to the Power of Graph Technology
ntroducing to the Power of Graph Technology
 
Visual geometry with deep learning
Visual geometry with deep learningVisual geometry with deep learning
Visual geometry with deep learning
 
Explainable AI
Explainable AIExplainable AI
Explainable AI
 
Explainability and bias in AI
Explainability and bias in AIExplainability and bias in AI
Explainability and bias in AI
 
Explainability for Learning to Rank
Explainability for Learning to RankExplainability for Learning to Rank
Explainability for Learning to Rank
 

Mehr von Satoshi Hara

機械学習で嘘をつく話
機械学習で嘘をつく話機械学習で嘘をつく話
機械学習で嘘をつく話Satoshi Hara
 
機械学習モデルの判断根拠の説明(Ver.2)
機械学習モデルの判断根拠の説明(Ver.2)機械学習モデルの判断根拠の説明(Ver.2)
機械学習モデルの判断根拠の説明(Ver.2)Satoshi Hara
 
異常の定義と推定
異常の定義と推定異常の定義と推定
異常の定義と推定Satoshi Hara
 
Convex Hull Approximation of Nearly Optimal Lasso Solutions
Convex Hull Approximation of Nearly Optimal Lasso SolutionsConvex Hull Approximation of Nearly Optimal Lasso Solutions
Convex Hull Approximation of Nearly Optimal Lasso SolutionsSatoshi Hara
 
Theoretical Linear Convergence of Unfolded ISTA and its Practical Weights and...
Theoretical Linear Convergence of Unfolded ISTA and its Practical Weights and...Theoretical Linear Convergence of Unfolded ISTA and its Practical Weights and...
Theoretical Linear Convergence of Unfolded ISTA and its Practical Weights and...Satoshi Hara
 
Maximally Invariant Data Perturbation as Explanation
Maximally Invariant Data Perturbation as ExplanationMaximally Invariant Data Perturbation as Explanation
Maximally Invariant Data Perturbation as ExplanationSatoshi Hara
 
アンサンブル木モデル解釈のためのモデル簡略化法
アンサンブル木モデル解釈のためのモデル簡略化法アンサンブル木モデル解釈のためのモデル簡略化法
アンサンブル木モデル解釈のためのモデル簡略化法Satoshi Hara
 
機械学習モデルの列挙
機械学習モデルの列挙機械学習モデルの列挙
機械学習モデルの列挙Satoshi Hara
 
KDD'17読み会:Anomaly Detection with Robust Deep Autoencoders
KDD'17読み会:Anomaly Detection with Robust Deep AutoencodersKDD'17読み会:Anomaly Detection with Robust Deep Autoencoders
KDD'17読み会:Anomaly Detection with Robust Deep AutoencodersSatoshi Hara
 
特徴選択のためのLasso解列挙
特徴選択のためのLasso解列挙特徴選択のためのLasso解列挙
特徴選択のためのLasso解列挙Satoshi Hara
 

Mehr von Satoshi Hara (10)

機械学習で嘘をつく話
機械学習で嘘をつく話機械学習で嘘をつく話
機械学習で嘘をつく話
 
機械学習モデルの判断根拠の説明(Ver.2)
機械学習モデルの判断根拠の説明(Ver.2)機械学習モデルの判断根拠の説明(Ver.2)
機械学習モデルの判断根拠の説明(Ver.2)
 
異常の定義と推定
異常の定義と推定異常の定義と推定
異常の定義と推定
 
Convex Hull Approximation of Nearly Optimal Lasso Solutions
Convex Hull Approximation of Nearly Optimal Lasso SolutionsConvex Hull Approximation of Nearly Optimal Lasso Solutions
Convex Hull Approximation of Nearly Optimal Lasso Solutions
 
Theoretical Linear Convergence of Unfolded ISTA and its Practical Weights and...
Theoretical Linear Convergence of Unfolded ISTA and its Practical Weights and...Theoretical Linear Convergence of Unfolded ISTA and its Practical Weights and...
Theoretical Linear Convergence of Unfolded ISTA and its Practical Weights and...
 
Maximally Invariant Data Perturbation as Explanation
Maximally Invariant Data Perturbation as ExplanationMaximally Invariant Data Perturbation as Explanation
Maximally Invariant Data Perturbation as Explanation
 
アンサンブル木モデル解釈のためのモデル簡略化法
アンサンブル木モデル解釈のためのモデル簡略化法アンサンブル木モデル解釈のためのモデル簡略化法
アンサンブル木モデル解釈のためのモデル簡略化法
 
機械学習モデルの列挙
機械学習モデルの列挙機械学習モデルの列挙
機械学習モデルの列挙
 
KDD'17読み会:Anomaly Detection with Robust Deep Autoencoders
KDD'17読み会:Anomaly Detection with Robust Deep AutoencodersKDD'17読み会:Anomaly Detection with Robust Deep Autoencoders
KDD'17読み会:Anomaly Detection with Robust Deep Autoencoders
 
特徴選択のためのLasso解列挙
特徴選択のためのLasso解列挙特徴選択のためのLasso解列挙
特徴選択のためのLasso解列挙
 

Kürzlich hochgeladen

🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘RTylerCroy
 
Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Enterprise Knowledge
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024Rafal Los
 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxMalak Abu Hammad
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountPuma Security, LLC
 
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...Neo4j
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...Martijn de Jong
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonetsnaman860154
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesSinan KOZAK
 
GenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationGenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationMichael W. Hawkins
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsMaria Levchenko
 
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfThe Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfEnterprise Knowledge
 
Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Allon Mureinik
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEarley Information Science
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024The Digital Insurer
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Drew Madelung
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonAnna Loughnan Colquhoun
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationSafe Software
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdfhans926745
 

Kürzlich hochgeladen (20)

🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘
 
Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024
 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptx
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path Mount
 
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonets
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen Frames
 
GenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationGenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day Presentation
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
 
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfThe Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
 
Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf
 

Explanation in Machine Learning and Its Reliability

  • 1. NeurIPS Meetup Japan 2021, Satoshi Hara Explanation in ML and Its Reliability Satoshi Hara Osaka University 1 NeurIPS Meetup Japan 2021
  • 2. NeurIPS Meetup Japan 2021, Satoshi Hara “Explanation” in ML ◼ Most of ML models are highly complex, or “black-box”. ◼ “Explanation in ML”: Obtain some useful information from the model (in addition to prediction). 2 Preliminary You are sick. Why? Your XX score is too high. You are sick. Why? ??? I don’t know. … XX score is too high. Oh…
  • 3. NeurIPS Meetup Japan 2021, Satoshi Hara [Typical Explanation 1] Saliency Map ◼ Generate heatmaps where the model has focused on when making predictions. 3 Preliminary The outline of zebra seems to be relevant.
  • 4. NeurIPS Meetup Japan 2021, Satoshi Hara [Typical Explanation 2] Similar Examples ◼ Provide some similar examples to the input of interest. 4 These images look similar. The prediction “Lapwing” will be correct. Lapwing Database Provide some similar examples Input Prediction Lapwing Preliminary
  • 5. NeurIPS Meetup Japan 2021, Satoshi Hara History of “Explanation” ◼ History of Saliency Map 5 Dawn 2014 2016 2018 2020 2015 2017 2019 Exponential Growth of Saliency Map Algos Attack & Manipulation Sanity Check [Adebayo+,2018] GuidedBP [Springenberg+,2014] DeepLIFT [Shrikumar+,2017] Grad-CAM [Selvaraju+,2017] ROAR [Hooker+,2019] MoRF/Deletion Metric [Bach+,2015; Vitali+,2018] LeRF/Insertion Metric [Arras+,2017; Vitali+,2018] Sensitivity [Kindermans+,2017] Evaluation Methods Saliency [Simonyan+,2014] IntGrad [Sundararajan+,2017] SHAP [Lundberg+,2017] LIME [Ribeiro+,2016] LRP [Bach+,2015] Fairwashing [Aivodji+,2019] SmoothGrad [Smilkov+,2017] DeepTaylor [Montavon+,2017] Occlusion [Zeiler+,2014] CAM [Zhou+,2016] Manipulation [Domobrowski+,2019] The papers on “Explanation” increased exponentially. 2022 2021 2020 2019 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 800 700 600 500 400 300 200 100 0 Searched “Interpretable Machine Learning” and “Explainable AI” on Web of Science Preliminary
  • 6. NeurIPS Meetup Japan 2021, Satoshi Hara History of “Explanation” ◼ History of Saliency Map 6 Dawn 2014 2016 2018 2020 2015 2017 2019 Exponential Growth of Saliency Map Algos Attack & Manipulation Sanity Check [Adebayo+,2018] GuidedBP [Springenberg+,2014] DeepLIFT [Shrikumar+,2017] Grad-CAM [Selvaraju+,2017] ROAR [Hooker+,2019] MoRF/Deletion Metric [Bach+,2015; Vitali+,2018] LeRF/Insertion Metric [Arras+,2017; Vitali+,2018] Sensitivity [Kindermans+,2017] Evaluation Methods Saliency [Simonyan+,2014] IntGrad [Sundararajan+,2017] SHAP [Lundberg+,2017] LIME [Ribeiro+,2016] LRP [Bach+,2015] Fairwashing [Aivodji+,2019] SmoothGrad [Smilkov+,2017] DeepTaylor [Montavon+,2017] Occlusion [Zeiler+,2014] CAM [Zhou+,2016] Manipulation [Domobrowski+,2019] The papers on “Explanation” increased exponentially. 800 700 600 500 400 300 200 100 0 Searched “Interpretable Machine Learning” and “Explainable AI” on Web of Science Reliability of “Explanation” has raised as a crucial concern. Are the “Explanation” truly valid? With “Explanation”, how malicious we can be? Preliminary 2022 2021 2020 2019 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008
  • 7. NeurIPS Meetup Japan 2021, Satoshi Hara Technical / Social Reliability of “Explanation” Technical Reliability “Is the explanation valid?” What we care: • Do the algorithms output valid “Explanation”? Research Question: • How can we evaluate the validity of “Explanation”? Social Reliability “Does explanation harm the society?” What we care: • What will happen if we introduce “Explanation” to society? Research Question: • Are there any malicious use cases of “Explanation”? 7 Technical Reliability
  • 8. NeurIPS Meetup Japan 2021, Satoshi Hara Faithfulness & Plausibility of “Explanation” ◼ Faithfulness [Lakkaraju+’19; Jacovi+’20] • Does “Explanation” reflect the model’s reasoning process? - Our interest is “How and why the model predicted that way.” • Any “Explanation” irrelevant to the reasoning process is invalid. - e.g. “Explanation” outputs something independent of the model. ◼ Plausibility [Lage+’19; Strout+’19] • Does “Explanation” make sense to the users? • Any “Explanation” unacceptable by the users is not ideal. - e.g. Entire program code; Very noisy saliency map. 8 Technical Reliability
  • 9. NeurIPS Meetup Japan 2021, Satoshi Hara Evaluation of “Explanation” ◼ Based on Faithfulness • Sanity Checks for Saliency Maps, NeurIPS’18. - Julius Adebayo, Justin Gilmer, Michael Muelly, Ian Goodfellow, Moritz Hardt, Been Kim • An epoch-making paper by Google Brain. • Evaluation of Faithfulness for saliency maps. ◼ Based on Plausibility • Evaluation of Similarity-based Explanations, ICLR’21. - Kazuaki Hanawa, Sho Yokoi, Satoshi Hara, Kentaro Inui • Evaluation of Plausibility for similarity-based explanations. 9
  • 10. NeurIPS Meetup Japan 2021, Satoshi Hara Evaluation of Saliency Map ◼ Plausibility • All the maps look more or less plausible. • Gradient, IntegratedGrad are bit noisy. ◼ Faithfulness? 10 Technical Reliability The outline of zebra seems to be relevant.
  • 11. NeurIPS Meetup Japan 2021, Satoshi Hara Evaluation of Faithfulness is Not Possible. ◼ Faithfulness • Does “Explanation” reflect the model’s reasoning process? ◼ Alternative: Sanity Check • Check the necessary condition for faithful “Explanation”. ◼ Q. What is the necessary condition? • “Explanation” is model-dependent. - Any “Explanation” irrelevant to the reasoning process is invalid. 11 Unknown → We cannot compare with Ground Truth. [Remark] Passing Sanity Check alone does not guarantee faithfulness. Technical Reliability
  • 12. NeurIPS Meetup Japan 2021, Satoshi Hara Model Parameter Randomization Test ◼ Compare “Explanation” of two models with different reasoning processes. • Faithful “Explanation” → Outputs are different. • Non-Faithful “Explanation” → Outputs can be identical. 12 Satisfies the necessary condition. Passed the sanity check. Technical Reliability [Assumption] These models have different reasoning processes. Model 1: Fully Trained Model 2: Randomly Initialized Input “Explanation” by Algo. 1 “Explanation” by Algo. 2 “Explanation” by Algo. 1 are different. “Explanation” by Algo. 2 are identical. Violates the necessary condition. Failed the sanity check.
  • 13. NeurIPS Meetup Japan 2021, Satoshi Hara Model Parameter Randomization Test ◼ Model 2: DNN with last few layers randomized. • Saliency Maps of Guided Backprop and Guided GradCAM are invariant against model randomization. → They violate the necessary condition for faithfulness. 13 Model 1 Model 2 [Ref] Sanity Checks for Saliency Maps Technical Reliability
  • 14. NeurIPS Meetup Japan 2021, Satoshi Hara Evaluation of “Explanation” ◼ Based on Faithfulness • Sanity Checks for Saliency Maps, NeurIPS’18. - Julius Adebayo, Justin Gilmer, Michael Muelly, Ian Goodfellow, Moritz Hardt, Been Kim • An epoch-making paper by Google Brain. • Evaluation of Faithfulness for saliency maps. ◼ Based on Plausibility • Evaluation of Similarity-based Explanations, ICLR’21. - Kazuaki Hanawa, Sho Yokoi, Satoshi Hara, Kentaro Inui • Evaluation of Plausibility for similarity-based explanations. 14
  • 15. NeurIPS Meetup Japan 2021, Satoshi Hara Evaluation of Similarity-based Explanation ◼ Faithfulness • We can use Model Parameter Randomization Test. ◼ Plausibility? 15 These images look similar. The prediction “Lapwing” will be correct. Lapwing Database Provide some similar examples Input Prediction Lapwing Technical Reliability
  • 16. NeurIPS Meetup Japan 2021, Satoshi Hara Plausibility in Similarity-based Explanation ◼ Example • Explanation B won’t be acceptable by the users. - Plausibility of Explanation A > Plausibility of Explanation B 16 Database frog Explanation A Database truck Explanation B frog Input Prediction Technical Reliability
  • 17. NeurIPS Meetup Japan 2021, Satoshi Hara Evaluation of Plausibility is Not Possible. ◼ There is no universal criterion that determines the acceptability of the users. ◼ Alternative: Sanity Check • Check the necessary condition for faithful “Plausibility”. ◼ Q. What is the necessary condition? • Obtained similar instance should belong to the same class. 17 is cat because a similar is cat. is cat because a similar is dog. Plausible Non-Plausible Identical Class Test Technical Reliability
  • 18. NeurIPS Meetup Japan 2021, Satoshi Hara Identical Class Test 18 Input Dot Last Layer All Layers Input Cos Last Layer All Layers Input L2 Dist. Last Layer All Layers Influence Function Relative IF Fisher Kernel Dot Cos Parameter Grad. Fraction of Test Instances Passed Identical Class Test 0 0.5 1.0 0 0.5 1.0 (Image Clf.) CIFAR10 + CNN (Text Clf.) AGNews + Bi-LSTM Cosine similarity of the parameter gradient performed almost perfectly. Technical Reliability
  • 19. NeurIPS Meetup Japan 2021, Satoshi Hara Cosine of Parameter Gradient • GC 𝑧, 𝑧′ = ∇𝜃ℓ 𝑦,𝑓𝜃 𝑥 ,∇𝜃ℓ 𝑦′,𝑓𝜃 𝑥′ ∇𝜃ℓ 𝑦,𝑓𝜃 𝑥 ∇𝜃ℓ 𝑦′,𝑓𝜃 𝑥′ 19 Sussex spaniel beer bottle mobile house Technical Reliability
  • 20. NeurIPS Meetup Japan 2021, Satoshi Hara Technical / Social Reliability of “Explanation” Technical Reliability “Is the explanation valid?” What we care: • Do the algorithms output valid “Explanation”? Research Question: • How can we evaluate the validity of “Explanation”? Social Reliability “Does explanation harm the society?” What we care: • What will happen if we introduce “Explanation” to society? Research Question: • Are there any malicious use cases of “Explanation”? 20 Social Reliability
  • 21. NeurIPS Meetup Japan 2021, Satoshi Hara Malicious Use Cases of “Explanation” ◼ Q. Are there malicious use cases of “Explanation”? A. Some may try to deceive people by providing fake explanations. ◼ Q. When and why fake explanations can be used? A. Fake explanations can show models better, e.g., by pretending as if the models are fair. ◼ Q. Why we need to research fake explanations? Are you evil? A. We need to know how malicious we can be with fake explanations. Otherwise, we cannot defend against possible maliciousness. 21 Social Reliability
  • 22. NeurIPS Meetup Japan 2021, Satoshi Hara Fake “Explanation” for Fairness ◼ Fairness in ML • Models can be biased towards gender, race, etc. • Ensuring fairness of the models is crucial nowadays. ◼ What if we cannot detect the use of unfair models? • Some may use unfair models. - Unfair models are typically more accurate than the fair ones. 22 Social Reliability Our model is the most accurate one in this business field. (because of the use of unfair yet accurate model) Moreover, our model is fair without any bias. (by showing fake explanation)
  • 23. NeurIPS Meetup Japan 2021, Satoshi Hara Fake “Explanation” for Fairness ◼ Fake “Explanation” by Surrogate Models • Fairwashing: the risk of rationalization, ICML’19. - Ulrich Aïvodji, Hiromi Arai, Olivier Fortineau, Sébastien Gambs, Satoshi Hara, Alain Tapp • Characterizing the risk of fairwashing, NeurIPS’21. - Ulrich Aïvodji, Hiromi Arai, Sébastien Gambs, Satoshi Hara ◼ Fake “Explanation” by Examples • Faking Fairness via Stealthily Biased Sampling, AAAI’20. - Kazuto Fukuchi, Satoshi Hara, Takanori Maehara ◼ Ref. • It’s Too Easy to Hide Bias in Deep-Learning Systems, IEEE Spectrum, 2021. 23
  • 24. NeurIPS Meetup Japan 2021, Satoshi Hara The risk of “Fairwashing” ◼ Explaining fairness 24 an honest explanation Your loan application is rejected because your gender is … Unfair AI: reject applicants based on their gender. Social Reliability
  • 25. NeurIPS Meetup Japan 2021, Satoshi Hara The risk of “Fairwashing” ◼ Explaining fairness 25 a dishonest explanation Your loan application is rejected because your income is low. Unfair AI: reject applicants based on their gender. Social Reliability
  • 26. NeurIPS Meetup Japan 2021, Satoshi Hara The risk of “Fairwashing” ◼ Explaining fairness 26 Unfair AI: reject applicants based on their gender. a dishonest explanation Your loan application is rejected because your income is low. “Fairwashing” Malicious decision-makers can disclose a fake explanation to rationalize their unfair decisions. “Fairwashing” Social Reliability
  • 27. NeurIPS Meetup Japan 2021, Satoshi Hara The risk of “Fairwashing” ◼ Explaining fairness 27 Unfair AI: reject applicants based on their gender. a dishonest explanation Your loan application is rejected because your income is low. This Study: LaundryML Possible to systematically generate fake explanations. Raise the awareness of the risk of “Fairwashing”. “Fairwashing” Malicious decision-makers can disclose a fake explanation to rationalize their unfair decisions. “Fairwashing” Social Reliability
  • 28. NeurIPS Meetup Japan 2021, Satoshi Hara ◼ The idea Generate many explanations, and pick one that is useful for “Fairwashing”. ◼ many explanations • Use “Model Enumeration” [Hara & Maehara’17; Hara & Ishihata’18] • Enumerate explanation models. ◼ pick one • Use fairness metrices such as demographic parity (DP). • Pick an explanation most faithful to the model, with DP less than a threshold. 28 LaundryML Systematically generating fake explanations The idea Social Reliability
  • 29. NeurIPS Meetup Japan 2021, Satoshi Hara Result ◼ “Fairwashing” for decisions on Adult dataset • Feature importance by FairML on “gender” has dropped. 29 A naïve explanation A fake explanation gender gender Social Reliability
  • 30. NeurIPS Meetup Japan 2021, Satoshi Hara Result ◼ “Fairwashing” for decisions on Adult dataset • Feature importance by FairML on “gender” has dropped. 30 A naïve explanation A false explanation gender gender If else if else if else if else if else low-income then high-income then low-income then low-income then low-income then high-income capital gain > 7056 marital = single education = HS-grad occupation = other occupation = white-colloar Fake Explanation Social Reliability
  • 31. NeurIPS Meetup Japan 2021, Satoshi Hara Fake “Explanation” for Fairness ◼ Fake “Explanation” by Surrogate Models • Fairwashing: the risk of rationalization, ICML’19. - Ulrich Aïvodji, Hiromi Arai, Olivier Fortineau, Sébastien Gambs, Satoshi Hara, Alain Tapp • Characterizing the risk of fairwashing, NeurIPS’21. - Ulrich Aïvodji, Hiromi Arai, Sébastien Gambs, Satoshi Hara ◼ Fake “Explanation” by Examples • Faking Fairness via Stealthily Biased Sampling, AAAI’20. - Kazuto Fukuchi, Satoshi Hara, Takanori Maehara ◼ Ref. • It’s Too Easy to Hide Bias in Deep-Learning Systems, IEEE Spectrum, 2021. 31
  • 32. NeurIPS Meetup Japan 2021, Satoshi Hara Fairness Metrics ◼ Quantifying fairness of the models • Several metrics + toolboxes - FairML, AI Fairness 360 [Bellamy+’19], Aequitas [Saleiro+’18] 32 AI Fairness 360 Social Reliability
  • 33. NeurIPS Meetup Japan 2021, Satoshi Hara Fake Fairness Metrics 33 Malicious Party Unfair Model Service Fairness Metric as Evidence Is this a fake metric. No guarantee whether the metrics are computed appropriately. Impossible to determine fake or not. Metric alone is not a valid evidence of fairness. Social Reliability
  • 34. NeurIPS Meetup Japan 2021, Satoshi Hara Avoiding Fake Fairness Metrics 34 Malicious Party Unfair Model Service Benchmark Data as Evidence Fairness metric computed on the benchmark is fair! The metric is reproducible using the benchmark data. We can avoid fake! Social Reliability
  • 35. NeurIPS Meetup Japan 2021, Satoshi Hara (Failed) Avoiding Fake Fairness Metrics 35 Malicious Party Unfair Model Service Benchmark Data as Evidence Fairness metric computed on the benchmark is fair! The metric is reproducible using the benchmark data. We can avoid fake! The benchmark data can be fake. Social Reliability
  • 36. NeurIPS Meetup Japan 2021, Satoshi Hara Generating Fake Benchmark ◼ Subsample the benchmark dataset 𝑆 from the original dataset 𝐷. ◼ “Ideal” Fake Benchmark Dataset 𝑆 • Fairness : Fairness metric computed on 𝑆 is fair. • Stealthiness : The distribution of 𝑆 is close to 𝐷. 36 Benchmark Fairness Stealthiness “Fair” Contingency Table Original dataset Social Reliability
  • 37. NeurIPS Meetup Japan 2021, Satoshi Hara 参照用 データ Goodness-of-Fit Test Generating Fake Benchmark ◼ Optimization of 𝑆 as LP (Min-Cost Flow) min𝑆𝑊 𝑆, 𝐷 , s. t. 𝐶 𝑆 = 𝐶𝑇 ◼ Detection of fake benchmark using statistical test. • Min. Distribution diff. ≈ small detection probability • Rejecting 𝑝 𝑆 = 𝑝(𝐷′) with KS test is probability at most 𝑂 𝑆 𝛼 × Distribution diff. 37 Stealthiness (Min. Distribution diff.) Fairness (Constraint on Contingency Table) Reference Data Social Reliability
  • 38. NeurIPS Meetup Japan 2021, Satoshi Hara Undetectability of Fake Benchmark 38 Positive Cases in Contingency Table Positive Cases in Contingency Table Fairness Metric (DP) Distribution diff. COMPAS Positive Cases in Contingency Table Positive Cases in Contingency Table Fairness Metric (DP) Distribution diff. Adult Random Sampling Case-Control Sampling Proposed Sampling Random Sampling Case-Control Sampling Proposed Sampling Proposed sampling resulted to fairer metric. (= achieved fake fairness) Proposed sampling attained distribution almost identical to the original distribution. (= undetectable) Social Reliability
  • 39. NeurIPS Meetup Japan 2021, Satoshi Hara Technical / Social Reliability of “Explanation” Technical Reliability “Is the explanation valid?” What we care: • Do the algorithms output valid “Explanation”? Research Question: • How can we evaluate the validity of “Explanation”? Social Reliability “Does explanation harm the society?” What we care: • What will happen if we introduce “Explanation” to society? Research Question: • Are there any malicious use cases of “Explanation”? 39 Summary
  • 40. NeurIPS Meetup Japan 2021, Satoshi Hara Technical / Social Reliability of “Explanation” Technical Reliability “Is the explanation valid?” What we care: • Do the algorithms output valid “Explanation”? Research Question: • How can we evaluate the validity of “Explanation”? Social Reliability “Does explanation harm the society?” What we care: • What will happen if we introduce “Explanation” to society? Research Question: • Are there any malicious use cases of “Explanation”? 40 Summary How can we evaluate the validity of “Explanation”? Which evaluation is good for which “Explanation”? When “Explanation” can be used maliciously? Can we detect malicious use cases?