Note of Compression of Neural Machine Translation Models via Pruning

The problems of NMT Model

I
[Not supported by viewer]
am
[Not supported by viewer]
a
[Not supported by viewer]
student
[Not supported by viewer]
source  language  input
[Not supported by viewer]
-
[Not supported by viewer]
Je
[Not supported by viewer]
suis
[Not supported by viewer]
étudiant
[Not supported by viewer]
target  language  input
[Not supported by viewer]
Je
[Not supported by viewer]
suis
[Not supported by viewer]
étudiant
[Not supported by viewer]
-
[Not supported by viewer]
target  language out put
[Not supported by viewer]
  1. Over-Parameterization
  2. Long running time
  3. Overfitting
  4. Big Storage size

The redundancies of NMT Model

Most important: Higher Layers; Attention and Softmax Weights

redundancy: lower layers; embedding weights;

Traditional Solutions

Optimal Brain Damage (OBD) and Optimal Brain Surgeon(OBS)

Recent Ways

Magnitude based pruning with iterative retraining(基于幅度的剪枝与反复的重复训练)yielded strong results for Convolutional Neural Networks (CNN) performing visual tasks.

猜你喜欢

转载自www.cnblogs.com/wevolf/p/12105538.html
今日推荐