SlideShare ist ein Scribd-Unternehmen logo
1 von 14
Downloaden Sie, um offline zu lesen
PR-12 presentation
NISP: Pruning Networks using Neuron Importance Score Propagation
CVPR2018

Authors: Ruichi Yu et al

Presented by Taesu Kim
Motivation
• Pruning
• Previous approaches
• Focus on single layer or two layers’ statistics
• Greedy pruning
• Entire network is a whole
• Error propagates, especially when network is deep
Motivation
• Entire CNN is a set of feature extractors
• The final responses are the extracted features
• We measured the importance of the neurons across the entire CNN based on
a unified goal
• Minimizing the reconstruction errors of (important) final responses
Approach
• Feature ranking on the final response layer
• NISP: Neuron Importance Score Propagation
• Pruning network using NISP
• Fine-tune the pruned network
NISP: Objective function
• ! , !
• !
• a binary vector ! : neuron prune indicator for the l-th layer
• !
• ! , !
• ! , ! , !
Solution
• The network pruning problem can be formulated as a binary integer program
• Fining the optimal neuron prune indicator s
• It is hard to obtain efficient analytical solutions by directly optimizing the objective
• a sub-optimal solution can be obtained by minimizing the upper bound
• ! !
• Assume the activation function ! is Lipschitz continuous: Identity, ReLU, sigmoid, tanh, etc.
• Lipschitz continuous if ! , ! ,
Solution
• ! , ! , !
• ! , !
• ! ! , ! !
• !
• !
Solution
• !
• !
Solution
• Backpropagate the importance values
Experiments
• Comparison with random pruning and training-from-scratch baseline
• randomly pruning the pre-trained CNN and then fine-tuning
• training a small CNN with the same number of neurons/filters per layer as our pruned model
• !
Experiments
• Feature selection vs. Magnitude of weights
• NISP-FS: using feature selection method in [34]
• NISP-Mag: considering only magnitude of weights
•
[34] Infinite feature selection. G. Roffo et al. ICCV 2015
Experiments
• NISP vs. Layer-by-Layer pruning
•
Experiments
• Comparison with existing methods
[11] Acceleration through elimination of redundant convolutions, M. Figurnov et al, NIPS2016
[20] Compression of deep convolutional neural networks for fast and low power mobile applications,
Y. Kim et al, ICLR 2016
[36] Learning the architecture of deep neural networks, S. Srinivas et al, BMVC 2016
[25] Pruning filters for efficient convnets, H. Li et al, ICLR 2017
[29] Thinnet: A filter level pruning method for deep neural network compression, J.-H. Luo et al ICCV 2017
NISP-A: pruning all conv layers
NISP-B: pruning all conv layers except conv5
NISP-C: pruning all conv layers except conv5, conv4
NISP-D: pruning all conv layers except conv2, conv3, FC6
NISP-x-A: prune 15% filters of each layer
NISP-x-B: prune 25% filters of each layer
Conclusion
• Generic framework for network compression and acceleration based on identifying
the importance levels of neurons
• Neuron importance scores in the layer of interest are obtained by feature ranking
• Formulate the network pruning problem as a binary integer program
• Obtain a closed-form solution to a relaxed version of the formulation
• NISP algorithm propagates the importance to every neuron in the whole network
• It efficiently reduces CNN redundancy and achieves full-network acceleration and
compression

Weitere ähnliche Inhalte

Was ist angesagt?

Was ist angesagt? (20)

[DL輪読会]A closer look at few shot classification
[DL輪読会]A closer look at few shot classification[DL輪読会]A closer look at few shot classification
[DL輪読会]A closer look at few shot classification
 
【DL輪読会】GradMax: Growing Neural Networks using Gradient Information
【DL輪読会】GradMax: Growing Neural Networks using Gradient Information【DL輪読会】GradMax: Growing Neural Networks using Gradient Information
【DL輪読会】GradMax: Growing Neural Networks using Gradient Information
 
[딥논읽] Meta-Transfer Learning for Zero-Shot Super-Resolution paper review
[딥논읽] Meta-Transfer Learning for Zero-Shot Super-Resolution paper review[딥논읽] Meta-Transfer Learning for Zero-Shot Super-Resolution paper review
[딥논읽] Meta-Transfer Learning for Zero-Shot Super-Resolution paper review
 
[DL輪読会]Relational inductive biases, deep learning, and graph networks
[DL輪読会]Relational inductive biases, deep learning, and graph networks[DL輪読会]Relational inductive biases, deep learning, and graph networks
[DL輪読会]Relational inductive biases, deep learning, and graph networks
 
強化学習初心者が強化学習でニューラルネットワークの設計を自動化してみたい
強化学習初心者が強化学習でニューラルネットワークの設計を自動化してみたい強化学習初心者が強化学習でニューラルネットワークの設計を自動化してみたい
強化学習初心者が強化学習でニューラルネットワークの設計を自動化してみたい
 
【DL輪読会】ViT + Self Supervised Learningまとめ
【DL輪読会】ViT + Self Supervised Learningまとめ【DL輪読会】ViT + Self Supervised Learningまとめ
【DL輪読会】ViT + Self Supervised Learningまとめ
 
[DL輪読会]医用画像解析におけるセグメンテーション
[DL輪読会]医用画像解析におけるセグメンテーション[DL輪読会]医用画像解析におけるセグメンテーション
[DL輪読会]医用画像解析におけるセグメンテーション
 
Neural networks for Graph Data NeurIPS2018読み会@PFN
Neural networks for Graph Data NeurIPS2018読み会@PFNNeural networks for Graph Data NeurIPS2018読み会@PFN
Neural networks for Graph Data NeurIPS2018読み会@PFN
 
[DL輪読会]Live-Streaming Fraud Detection: A Heterogeneous Graph Neural Network A...
[DL輪読会]Live-Streaming Fraud Detection: A Heterogeneous Graph Neural Network A...[DL輪読会]Live-Streaming Fraud Detection: A Heterogeneous Graph Neural Network A...
[DL輪読会]Live-Streaming Fraud Detection: A Heterogeneous Graph Neural Network A...
 
SSII2022 [SS2] 少ないデータやラベルを効率的に活用する機械学習技術 〜 足りない情報をどのように補うか?〜
SSII2022 [SS2] 少ないデータやラベルを効率的に活用する機械学習技術 〜 足りない情報をどのように補うか?〜SSII2022 [SS2] 少ないデータやラベルを効率的に活用する機械学習技術 〜 足りない情報をどのように補うか?〜
SSII2022 [SS2] 少ないデータやラベルを効率的に活用する機械学習技術 〜 足りない情報をどのように補うか?〜
 
Coreset+SVM (論文紹介)
Coreset+SVM (論文紹介)Coreset+SVM (論文紹介)
Coreset+SVM (論文紹介)
 
SSII2018TS: 大規模深層学習
SSII2018TS: 大規模深層学習SSII2018TS: 大規模深層学習
SSII2018TS: 大規模深層学習
 
[DL輪読会]Vector-based navigation using grid-like representations in artificial ...
[DL輪読会]Vector-based navigation using grid-like representations in artificial ...[DL輪読会]Vector-based navigation using grid-like representations in artificial ...
[DL輪読会]Vector-based navigation using grid-like representations in artificial ...
 
ResNetの仕組み
ResNetの仕組みResNetの仕組み
ResNetの仕組み
 
論文紹介:End-to-End Object Detection with Transformers
論文紹介:End-to-End Object Detection with Transformers論文紹介:End-to-End Object Detection with Transformers
論文紹介:End-to-End Object Detection with Transformers
 
ArcFace: Additive Angular Margin Loss for Deep Face Recognition
ArcFace: Additive Angular Margin Loss for Deep Face RecognitionArcFace: Additive Angular Margin Loss for Deep Face Recognition
ArcFace: Additive Angular Margin Loss for Deep Face Recognition
 
(DL輪読)Matching Networks for One Shot Learning
(DL輪読)Matching Networks for One Shot Learning(DL輪読)Matching Networks for One Shot Learning
(DL輪読)Matching Networks for One Shot Learning
 
Pruning convolutional neural networks for resource efficient inference
Pruning convolutional neural networks for resource efficient inferencePruning convolutional neural networks for resource efficient inference
Pruning convolutional neural networks for resource efficient inference
 
One shot learning
One shot learningOne shot learning
One shot learning
 
[DL輪読会]Learning to Simulate Complex Physics with Graph Networks
[DL輪読会]Learning to Simulate Complex Physics with Graph Networks[DL輪読会]Learning to Simulate Complex Physics with Graph Networks
[DL輪読会]Learning to Simulate Complex Physics with Graph Networks
 

Ähnlich wie PR12-193 NISP: Pruning Networks using Neural Importance Score Propagation

Deep Learning
Deep LearningDeep Learning
Deep Learning
Pierre de Lacaze
 
Neural network techniques
Neural network techniquesNeural network techniques
Neural network techniques
Vipul Bhargava
 
Forecasting of Sales using Neural network techniques
Forecasting of Sales using Neural network techniquesForecasting of Sales using Neural network techniques
Forecasting of Sales using Neural network techniques
Hitesh Dua
 

Ähnlich wie PR12-193 NISP: Pruning Networks using Neural Importance Score Propagation (20)

Cvpr 2018 papers review (efficient computing)
Cvpr 2018 papers review (efficient computing)Cvpr 2018 papers review (efficient computing)
Cvpr 2018 papers review (efficient computing)
 
DSRLab seminar Introduction to deep learning
DSRLab seminar   Introduction to deep learningDSRLab seminar   Introduction to deep learning
DSRLab seminar Introduction to deep learning
 
Introduction to deep learning
Introduction to deep learningIntroduction to deep learning
Introduction to deep learning
 
04 Deep CNN (Ch_01 to Ch_3).pptx
04 Deep CNN (Ch_01 to Ch_3).pptx04 Deep CNN (Ch_01 to Ch_3).pptx
04 Deep CNN (Ch_01 to Ch_3).pptx
 
Wits presentation 6_28072015
Wits presentation 6_28072015Wits presentation 6_28072015
Wits presentation 6_28072015
 
Deep Learning
Deep LearningDeep Learning
Deep Learning
 
Dp2 ppt by_bikramjit_chowdhury_final
Dp2 ppt by_bikramjit_chowdhury_finalDp2 ppt by_bikramjit_chowdhury_final
Dp2 ppt by_bikramjit_chowdhury_final
 
Neural network techniques
Neural network techniquesNeural network techniques
Neural network techniques
 
Forecasting of Sales using Neural network techniques
Forecasting of Sales using Neural network techniquesForecasting of Sales using Neural network techniques
Forecasting of Sales using Neural network techniques
 
Continuous Control with Deep Reinforcement Learning, lillicrap et al, 2015
Continuous Control with Deep Reinforcement Learning, lillicrap et al, 2015Continuous Control with Deep Reinforcement Learning, lillicrap et al, 2015
Continuous Control with Deep Reinforcement Learning, lillicrap et al, 2015
 
Towards better analysis of deep convolutional neural networks
Towards better analysis of deep convolutional neural networksTowards better analysis of deep convolutional neural networks
Towards better analysis of deep convolutional neural networks
 
Lecture on Deep Learning
Lecture on Deep LearningLecture on Deep Learning
Lecture on Deep Learning
 
Deep learning and image analytics using Python by Dr Sanparit
Deep learning and image analytics using Python by Dr SanparitDeep learning and image analytics using Python by Dr Sanparit
Deep learning and image analytics using Python by Dr Sanparit
 
Deep_Learning_Frameworks_CNTK_PyTorch
Deep_Learning_Frameworks_CNTK_PyTorchDeep_Learning_Frameworks_CNTK_PyTorch
Deep_Learning_Frameworks_CNTK_PyTorch
 
"Quantizing Deep Networks for Efficient Inference at the Edge," a Presentatio...
"Quantizing Deep Networks for Efficient Inference at the Edge," a Presentatio..."Quantizing Deep Networks for Efficient Inference at the Edge," a Presentatio...
"Quantizing Deep Networks for Efficient Inference at the Edge," a Presentatio...
 
A Survey of Convolutional Neural Networks
A Survey of Convolutional Neural NetworksA Survey of Convolutional Neural Networks
A Survey of Convolutional Neural Networks
 
Deep Learning: Evolution of ML from Statistical to Brain-like Computing- Data...
Deep Learning: Evolution of ML from Statistical to Brain-like Computing- Data...Deep Learning: Evolution of ML from Statistical to Brain-like Computing- Data...
Deep Learning: Evolution of ML from Statistical to Brain-like Computing- Data...
 
Transfer Learning and Fine Tuning for Cross Domain Image Classification with ...
Transfer Learning and Fine Tuning for Cross Domain Image Classification with ...Transfer Learning and Fine Tuning for Cross Domain Image Classification with ...
Transfer Learning and Fine Tuning for Cross Domain Image Classification with ...
 
Image Segmentation Using Deep Learning : A survey
Image Segmentation Using Deep Learning : A surveyImage Segmentation Using Deep Learning : A survey
Image Segmentation Using Deep Learning : A survey
 
Handwritten Digit Recognition(Convolutional Neural Network) PPT
Handwritten Digit Recognition(Convolutional Neural Network) PPTHandwritten Digit Recognition(Convolutional Neural Network) PPT
Handwritten Digit Recognition(Convolutional Neural Network) PPT
 

Mehr von Taesu Kim

Mehr von Taesu Kim (6)

PR12-179 M3D-GAN: Multi-Modal Multi-Domain Translation with Universal Attention
PR12-179 M3D-GAN: Multi-Modal Multi-Domain Translation with Universal AttentionPR12-179 M3D-GAN: Multi-Modal Multi-Domain Translation with Universal Attention
PR12-179 M3D-GAN: Multi-Modal Multi-Domain Translation with Universal Attention
 
PR12-165 Few-Shot Adversarial Learning of Realistic Neural Talking Head Models
PR12-165 Few-Shot Adversarial Learning of Realistic Neural Talking Head ModelsPR12-165 Few-Shot Adversarial Learning of Realistic Neural Talking Head Models
PR12-165 Few-Shot Adversarial Learning of Realistic Neural Talking Head Models
 
PR12-151 The Unreasonable Effectiveness of Deep Features as a Perceptual Metric
PR12-151 The Unreasonable Effectiveness of Deep Features as a Perceptual MetricPR12-151 The Unreasonable Effectiveness of Deep Features as a Perceptual Metric
PR12-151 The Unreasonable Effectiveness of Deep Features as a Perceptual Metric
 
PR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networks
PR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networksPR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networks
PR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networks
 
Issues in AI product development and practices in audio applications
Issues in AI product development and practices in audio applicationsIssues in AI product development and practices in audio applications
Issues in AI product development and practices in audio applications
 
PR-043: HyperNetworks
PR-043: HyperNetworksPR-043: HyperNetworks
PR-043: HyperNetworks
 

Kürzlich hochgeladen

Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slide
vu2urc
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI Solutions
Enterprise Knowledge
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
Earley Information Science
 

Kürzlich hochgeladen (20)

Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slide
 
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfThe Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemkeProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI Solutions
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreter
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonets
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024
 
What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?
 

PR12-193 NISP: Pruning Networks using Neural Importance Score Propagation

  • 1. PR-12 presentation NISP: Pruning Networks using Neuron Importance Score Propagation CVPR2018 Authors: Ruichi Yu et al Presented by Taesu Kim
  • 2. Motivation • Pruning • Previous approaches • Focus on single layer or two layers’ statistics • Greedy pruning • Entire network is a whole • Error propagates, especially when network is deep
  • 3. Motivation • Entire CNN is a set of feature extractors • The final responses are the extracted features • We measured the importance of the neurons across the entire CNN based on a unified goal • Minimizing the reconstruction errors of (important) final responses
  • 4. Approach • Feature ranking on the final response layer • NISP: Neuron Importance Score Propagation • Pruning network using NISP • Fine-tune the pruned network
  • 5. NISP: Objective function • ! , ! • ! • a binary vector ! : neuron prune indicator for the l-th layer • ! • ! , ! • ! , ! , !
  • 6. Solution • The network pruning problem can be formulated as a binary integer program • Fining the optimal neuron prune indicator s • It is hard to obtain efficient analytical solutions by directly optimizing the objective • a sub-optimal solution can be obtained by minimizing the upper bound • ! ! • Assume the activation function ! is Lipschitz continuous: Identity, ReLU, sigmoid, tanh, etc. • Lipschitz continuous if ! , ! ,
  • 7. Solution • ! , ! , ! • ! , ! • ! ! , ! ! • ! • !
  • 9. Solution • Backpropagate the importance values
  • 10. Experiments • Comparison with random pruning and training-from-scratch baseline • randomly pruning the pre-trained CNN and then fine-tuning • training a small CNN with the same number of neurons/filters per layer as our pruned model • !
  • 11. Experiments • Feature selection vs. Magnitude of weights • NISP-FS: using feature selection method in [34] • NISP-Mag: considering only magnitude of weights • [34] Infinite feature selection. G. Roffo et al. ICCV 2015
  • 12. Experiments • NISP vs. Layer-by-Layer pruning •
  • 13. Experiments • Comparison with existing methods [11] Acceleration through elimination of redundant convolutions, M. Figurnov et al, NIPS2016 [20] Compression of deep convolutional neural networks for fast and low power mobile applications, Y. Kim et al, ICLR 2016 [36] Learning the architecture of deep neural networks, S. Srinivas et al, BMVC 2016 [25] Pruning filters for efficient convnets, H. Li et al, ICLR 2017 [29] Thinnet: A filter level pruning method for deep neural network compression, J.-H. Luo et al ICCV 2017 NISP-A: pruning all conv layers NISP-B: pruning all conv layers except conv5 NISP-C: pruning all conv layers except conv5, conv4 NISP-D: pruning all conv layers except conv2, conv3, FC6 NISP-x-A: prune 15% filters of each layer NISP-x-B: prune 25% filters of each layer
  • 14. Conclusion • Generic framework for network compression and acceleration based on identifying the importance levels of neurons • Neuron importance scores in the layer of interest are obtained by feature ranking • Formulate the network pruning problem as a binary integer program • Obtain a closed-form solution to a relaxed version of the formulation • NISP algorithm propagates the importance to every neuron in the whole network • It efficiently reduces CNN redundancy and achieves full-network acceleration and compression