Deep Compression/Acceleration(模型压缩加速总结)

版权声明:本文为博主原创文章,未经博主允许不得转载。 https://blog.csdn.net/hw5226349/article/details/84888416

模型压缩论文目录

根据个人理解将模型压缩方面研究分为以下七个方向:

结构structure

[BMVC2018] IGCV3: Interleaved Low-Rank Group Convolutions for Efficient Deep Neural Networks

[CVPR2018] IGCV2: Interleaved Structured Sparse Convolutional Neural Networks

[CVPR2018] MobileNetV2: Inverted Residuals and Linear Bottlenecks

[ECCV2018] ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design

量化quantization

Binarized Neural Networks: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1

[ACM2017] FINN: A Framework for Fast, Scalable Binarized Neural Network Inference

[CVPR2016] DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients

[CVPR2016] XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks

[CVPR2016] Ternary Weight Networks

Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference

[ACM2017] Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations

Two-Step Quantization for Low-bit Neural Networks

剪枝pruning

通道裁剪channel pruning

[NIPS2018] Discrimination-aware Channel Pruning for Deep Neural Networks

[ICCV2017] Channel Pruning for Accelerating Very Deep Neural Networks

[ECCV2018] AMC: AutoML for Model Compression and Acceleration on Mobile Devices

  1. https://www.jiqizhixin.com/articles/AutoML-for-Model-Compression-and-Acceleration-on-Mobile-Devices论文翻译

[ICCV2017] Learning Efficient Convolutional Networks through Network Slimming

[ICLR2018] Rethinking the Smaller-Norm-Less-Informative Assumption in Channel Pruning of Convolution Layers

[CVPR2017] NISP: Pruning Networks using Neuron Importance Score Propagation

[ICCV2017] ThiNet: A Filter Level Pruning Method for Deep Neural Network Compression

稀疏sparsity

SBNet: Sparse Blocks Network for Fast Inference

To Prune, or Not to Prune: Exploring the Efficacy of Pruning for Model Compression

Submanifold Sparse Convolutional Networks

融合fusion

蒸馏distillation

[NIPS2014] Distilling the Knowledge in a Neural Network

综合comprehensive

[ICLR2016] Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding

Model Distillation with Knowledge Transfer from Face Classification to Alignment and Verification

猜你喜欢

转载自blog.csdn.net/hw5226349/article/details/84888416