当前位置:网站首页>MIT-6874-Deep Learning in the Life Sciences Week 7
MIT-6874-Deep Learning in the Life Sciences Week 7
2022-07-05 06:00:00 【Wooden girl】
Lecture 05 Interpretable Deep Learning
- Interpretable deep learning
- One 、Intro to Interpretability
- 1a. Interpretability definition: Convert implicit NN information to human-interpretable information
- 1b. Motivation: Verify model works as intended; debug classifier; make discoveries; Right to explanation
- 1c. Ante-hoc (train interpretable model) vs. Post-hoc (interpret complex model; degree of “locality”)
- 2. Interpreting Deep Neural Networks
- Evaluating Attribution Methods
- 3a. Qualitative: Coherence: Attributions should highlight discriminative features / objects of interest
- 3b. Qualitative: Class Sensitivity: Attributions should be sensitive to class labels
- 3c. Quantitative: Sensitivity: Removing feature with high attribution --> large decrease in class probability
- 3d. Quantitative: ROAR & KAR. Low class prob cuz image unseen --> remove pixels, retrain, measure acc. drop
Interpretable deep learning
This section discusses the explicability of deep learning . Model itself means knowledge , Explicability is of great importance to people like deep learning “ Black box model ” for , It is the root of explaining the reason and method why he made such a judgment , It can help the model work in the direction of human expectations . In many scenes , If recommended 、 Medical and other scenarios have great application prospects .
The following is the outline of this class
One 、Intro to Interpretability
1a. Interpretability definition: Convert implicit NN information to human-interpretable information
1b. Motivation: Verify model works as intended; debug classifier; make discoveries; Right to explanation
Why Interpretability?
- 1.Verify that model works as expected: Wrong decisions can be costly and dangerous
- 2. Improve / Debug classifier
- 3. Make new discoveries
- 4.Right to explanation
“Right to be given an explanation for an output of the algorithm”
1c. Ante-hoc (train interpretable model) vs. Post-hoc (interpret complex model; degree of “locality”)
Know good writing : Explainable AI The research of
Ex post explanation VS Self explanation
There are two ways to obtain the interpretability of the model ( Interpretive classification )
- Ante-hoc & Post-hoc
1. Ante-hoc Interpretability —— Prior interpretability ( Model built-in interpretability )
By training a self explanatory model , To get an explanation of the results .
Common interpretable models :
- Naive Bayes
- Linear regression
- Decision tree
- Rule based models
But the complexity of this kind of model is limited , As a result, its fundamental performance is limited
2. Post-hoc Interpretability—— Ex post interpretability
Refers to the black box model , By some means , Reflect his decision logic
Several levels of interpretability :
- Model level explicability :DNN Why does the model decide the decision boundary so
- Interpretability of features : Which features can maximize the activation of the current model
- Towards individual explicability : Explain why this input is so classified
2. Interpreting Deep Neural Networks
2a. Interpreting Models (macroscopic, understand internals) vs. decisions (microscopic, practical applications)
( Course trend ) Several classifications of model interpretability
- Interpreting decisions:
- Attribution method: What attributes determine the current output of the model
- Example-based: What special case leads to the current output of the model
- Interpreting models:
- Representation analysis: The model represents itself
- Data generation: How to use models to generate data
- Example-based: Related cases
DNN interpretability It can be divided into macro and micro levels
Interpreting models It can be divided into the following four aspects , among The analysis of representation can be divided into weight visualization and proxy model
2b. Interpreting Models: Weight visualization, Surrogate model, Activation maximization, Example-based
1. Weight visualization Weight Visualization
Yes CNN Each layer of filter is visualized , To understand what the model is learning at the current level
2. Surrogate model Agent model
Use a simple ,“ Explicable ” Model to “summarize” Model Output , Try to explain “black box” Output .
3. Data Generation / Activation maximization The data generated / Activate maximize
Activate maximize : Find the way to maximize the activation of neurons , Find the input X, Maximize the probability of the model under the current category
Convolution and deconvolution of the model
The initial input is chaotic , With the increase of training layers , Gradually, we can distinguish the characteristics between numbers
advantage : The advantages and disadvantages of this approach
- DNN It can be explained by looking for the input mode that maximizes the output .
- Connecting with data can improve the interpretability of visualization .
4. Example-based
Summary :
- By visualizing the weight of each layer
- Replace models with low accuracy but strong interpretability
- To some extent, we can obtain useful features by maximizing the activation function / Information
- Through effective construction prototype and criticism, Guide model learning to obtain the most useful , Information for distinguishing
2c. Interpreting Decisions:
Example-based
The input training samples have a decisive impact on the results of the model
Attribution Methods: why are gradients noisy?
Give each pixel a causal score , That is, how much the current pixel contributes to the model to this result .
The result of visualizing attribution
The key is , structure ( features ) Saliency map (Saliency Map)
promote saliency map Methods , First : Change of thinking
hypothesis 1: saliency map Is real
- Some pixels randomly distributed in the image are crucial to how the network makes decisions .
- Noise is very important
hypothesis 2: The gradient is discontinuous - DNN Use piecewise linear functions (ReLU Activate ,max-pooling etc. ).
- The mutation jump of importance score on the infinitesimal change of input .
hypothesis 3: - A feature may have a strong impact on a global scale , But it will have a small impact locally
Other attribution methods
Gradient-based Attribution: SmoothGrad, Interior Gradient
Backprop-based Attribution: Deconvolution, Guided Backpropagation
- Observe : Removing more gradients will bring a clearer visual effect
Evaluating Attribution Methods
3a. Qualitative: Coherence: Attributions should highlight discriminative features / objects of interest
- Attribution should be based on distinctive characteristics
3b. Qualitative: Class Sensitivity: Attributions should be sensitive to class labels
- Attribution should be category sensitive
3c. Quantitative: Sensitivity: Removing feature with high attribution --> large decrease in class probability
- Removing features with high attributes will lead to a significant reduction in class probability
3d. Quantitative: ROAR & KAR. Low class prob cuz image unseen --> remove pixels, retrain, measure acc. drop
边栏推荐
- 884. Uncommon words in two sentences
- [practical skills] how to do a good job in technical training?
- LeetCode 0108.将有序数组转换为二叉搜索树 - 数组中值为根,中值左右分别为左右子树
- Graduation project of game mall
- Educational Codeforces Round 107 (Rated for Div. 2) E. Colorings and Dominoes
- 中职网络安全技能竞赛——广西区赛中间件渗透测试教程文章
- 1.15 - 输入输出系统
- Pointnet++ learning
- leetcode-31:下一个排列
- Simple knapsack, queue and stack with deque
猜你喜欢
leetcode-6108:解密消息
Introduction and experience of wazuh open source host security solution
API related to TCP connection
1.14 - 流水线
Spark中groupByKey() 和 reduceByKey() 和combineByKey()
6. Logistic model
数据可视化图表总结(二)
[practical skills] technical management of managers with non-technical background
全国中职网络安全B模块之国赛题远程代码执行渗透测试 //PHPstudy的后门漏洞分析
CF1634 F. Fibonacci Additions
随机推荐
剑指 Offer 58 - II. 左旋转字符串
leetcode-9:回文数
Dichotomy, discretization, etc
Typical use cases for knapsacks, queues, and stacks
[jailhouse article] performance measurements for hypervisors on embedded ARM processors
Transform optimization problems into decision-making problems
6. Logistic model
Brief introduction to tcp/ip protocol stack
Daily question 1342 Number of operations to change the number to 0
LaMDA 不可能觉醒吗?
[jailhouse article] jailhouse hypervisor
QQ电脑版取消转义符输入表情
Cluster script of data warehouse project
【Rust 笔记】13-迭代器(下)
【Rust 笔记】17-并发(上)
Daily question - longest substring without repeated characters
Control Unit 控制部件
剑指 Offer 53 - I. 在排序数组中查找数字 I
每日一题-无重复字符的最长子串
2022年貴州省職業院校技能大賽中職組網絡安全賽項規程