Note of Compression of Neural Machine Translation Models via Pruning

The problems of NMT Model

I
[Not supported by viewer]
am
[Not supported by viewer]
a
[Not supported by viewer]
student
[Not supported by viewer]
source  language  input
[Not supported by viewer]
-
[Not supported by viewer]
Je
[Not supported by viewer]
suis
[Not supported by viewer]
étudiant
[Not supported by viewer]
target  language  input
[Not supported by viewer]
Je
[Not supported by viewer]
suis
[Not supported by viewer]
étudiant
[Not supported by viewer]
-
[Not supported by viewer]
target  language out put
[Not supported by viewer]
  1. Over-Parameterization
  2. Long running time
  3. Overfitting
  4. Big Storage size

The redundancies of NMT Model

Most important: Higher Layers; Attention and Softmax Weights

redundancy: lower layers; embedding weights;

Traditional Solutions

Optimal Brain Damage (OBD) and Optimal Brain Surgeon(OBS)

Recent Ways

Magnitude based pruning with iterative retraining (based on the magnitude of the repeated pruning and repetitive training) yielded strong results for Convolutional Neural Networks (CNN) performing visual tasks.

Guess you like

Origin www.cnblogs.com/wevolf/p/12105538.html