[(Strongly recommended) Li Hongyi 2021/2022 Spring Machine Learning Course] 2022-How to Effectively Use Self-Supervised Models-Data-Efficient &Parameter-Efficient Tuning

insert image description here
Materials: pdf | Video

1. What is self-pretraining

Two training methods:

  1. predict the next word
  2. mask learning

insert image description here

fine-tune: embryonic stem cells, also need fine-tune to exert their power
insert image description here

2. There is a problem

Fine-tuning still requires some data The
insert image description here
model is too large
insert image description here
insert image description here
insert image description here

3. Data-Efficient Fine-tuning: Prompt Tuning (hint, more efficient use of data, such as when the amount of data is small)

insert image description here
insert image description here
insert image description here
insert image description here
The difference between Standard fine-tuning and Standard fine-tuning
insert image description here
Prompt Tuning will perform well in the case of small amount of data
insert image description here

less data

insert image description here
insert image description here
insert image description here

4. Data-Efficient Fine-tuning: Semi-supervised Learning (semi-supervised)

insert image description here
insert image description here
insert image description here
insert image description here
insert image description here

5. zero-shot

insert image description here
insert image description here
insert image description here

6. PLMs Are Gigantic -> Reducing the Number of Parameters (solution for large models)

insert image description here
insert image description here
insert image description here

Parameter-Efficient Fine-tuning

insert image description here
insert image description here
insert image description here

Adapter

insert image description here
insert image description here
insert image description here

LoRA (less parameters than Adapter (feature compression is smaller), faster inference (parallel insertion))

insert image description here
insert image description here
insert image description here
insert image description here
insert image description here

Prefix Tuning (core: fine-tuning is to change hidden representations)

insert image description here
Review Self-attention
insert image description here

insert image description here

insert image description here

Soft Prompting

insert image description here
insert image description here

4 Methods Summary

insert image description here
insert image description here
insert image description here

Early Exit (reduce inference time)

insert image description here
insert image description here
insert image description here
insert image description here
insert image description here

7. Closing Remarks

insert image description here
insert image description here
insert image description here
insert image description here
insert image description here

Guess you like

Origin blog.csdn.net/weixin_43154149/article/details/124370319