La víspera de la revolución LLM de inteligencia artificial: comprender el modelo de procesamiento del lenguaje natural Transformador del origen de ChatGPT en un artículo

 Autor: Zhong Chao, Equipo Ali Group Taobao

 

 

 

 

 

 

 

 

 

 [01] https://web.stanford.edu/~jurafsky/slp3/3.pdf

[02] https://ai.googleblog.com/2017/08/transformer-novel-neural-network.html

[03] "Procesamiento del lenguaje natural: un método basado en un modelo de preentrenamiento" Che Wanxiang et al.

[04] https://cs.stanford.edu/people/karpathy/convnetjs/

[05] https://arxiv.org/abs/1706.03762

[06] https://arxiv.org/abs/1512.03385

[07] https://github.com/Kyubyong/transformer/

[08] http://jalammar.github.io/illustrated-transformer/

[09] https://towardsdatascience.com/this-is-how-to-train-better-transformer-models-d54191299978

[10] "Procesamiento del lenguaje natural en la práctica: aplicación del modelo de preentrenamiento y su comercialización" por Anku A. Patel et al.

[11] https://lilianweng.github.io/posts/2018-06-24-attention/

[12] https://github.com/lilianweng/transformer-tensorflow/

[13] "Predicción de la secuencia espacio-temporal del estado del tráfico vial a corto plazo basada en el aprendizaje profundo" Cui Jianxun

[14] https://www.zhihu.com/question/325839123

[15] https://luweikxy.gitbook.io/machine-learning-notes/self-attention-and-transformer

[16] "Python Deep Learning (2.ª edición)" de Francois Cholet

[17] https://en.wikipedia.org/wiki/Attention_(aprendizaje_máquina)

[18] https://zhuanlan.zhihu.com/p/410776234

[19] https://www.tensorflow.org/tensorboard/get_started

[20] https://paperswithcode.com/method/multi-head-attention

[21] https://zhuanlan.zhihu.com/p/48508221

[22] https://www.joshbelanich.com/self-attention-layer/

[23] https://learning.rasa.com/transformers/kvq/

[24] http://deeplearning.stanford.edu/tutorial/supervised/ConvolutionalNeuralNetwork/

[25] https://zhuanlan.zhihu.com/p/352898810

[26] https://towardsdatascience.com/beautifully-illustrated-nlp-models-from-rnn-to-transformer-80d69faf2109

[27] https://medium.com/analytics-vidhya/understanding-qkv-in-transformer-self-attention-9a5eddaa5960

 

 

Supongo que te gusta

Origin blog.csdn.net/AlibabaTech1024/article/details/129057732
Recomendado
Clasificación