https://www.leiphone.com/news/201709/8tDpwklrKubaecTa.html 雷锋网 完全图解RNN、RNN变体、Seq2Seq、Attention机制
https://www.cnblogs.com/wangduo/p/6773601.html?utm_source=itdadao&utm_medium=referral [译] 理解 LSTM(Long Short-Term Memory, LSTM) 网络
https://spaces.ac.cn/archives/4765/comment-page-1 《Attention is All You Need》浅读(简介+代码)
https://blog.youkuaiyun.com/hahajinbu/article/details/81940355 自然语言处理中的Attention机制总结
http://www.zhuzongkui.top/2018/08/10/competition-summary/ 问题对相似度比赛
https://www.jianshu.com/p/2cf00bb9db34?utm_source=oschina-app Google Colab免费GPU使用攻略
https://blog.youkuaiyun.com/qq_41664845/article/details/84969266 图解Transformer
https://blog.youkuaiyun.com/fkyyly/article/details/86093438 bert代码解析
https://daiwk.github.io/posts/nlp-bert-code-annotated-framework.html bert代码解读
https://www.cnblogs.com/hellcat/p/9735041.html 『计算机视觉』各种Normalization层辨析