[深度学习]从Attention到Transformer到BERT

Jay Alammar用直观直白的方式解释了Attention,Transformer和BERT。并辅以很多生动的图例。

Attention

Visualizing A Neural Machine Translation Model (Mechanics of Seq2seq Models With Attention)

Transformer

The Illustrated Transformer

BERT

The Illustrated BERT, ELMo, and co. (How NLP Cracked Transfer Learning)

其他资料

NLP's ImageNet moment has arrived

猜你喜欢

转载自blog.csdn.net/sinat_33487968/article/details/85231556