2019/01/17
-----
Fig. BERT Applications(圖片來源)。
-----
符號說明:
# basic
// advanced
-----
Paper
# Bert 應用ㄧ 文本分類
Kant, Neel, et al. "Practical Text Classification With Large Pre-Trained Language Models." arXiv preprint arXiv:1812.01207 (2018).
https://arxiv.org/pdf/1812.01207.pdf# Bert 應用二 回答問題
Zhu, Chenguang, Michael Zeng, and
Xuedong Huang. "SDNet: Contextualized Attention-based Deep Network for
Conversational Question Answering." arXiv preprint arXiv:1812.03593 (2018).
https://arxiv.org/pdf/1812.03593.pdf# Bert 應用三 生成文本
Transfer Learning for Style-Specific Text Generation
https://nips2018creativity.github.io/doc/Transfer%20Learning%20for%20Style-Specific%20Text%20Generation.pdf
// Why
Qi, Ye, et al. "When and Why are Pre-trained Word Embeddings Useful for Neural Machine Translation?." arXiv preprint arXiv:1804.06323 (2018).
https://arxiv.org/pdf/1804.06323.pdf// Common Sense
Trinh, Trieu H., and Quoc V. Le. "Do Language Models Have Common Sense?." (2018).
https://openreview.net/pdf?id=rkgfWh0qKX
-----
References
NLP中的语言模型预训练&微调 - CLOUD - CSDN博客
https://blog.csdn.net/muumian123/article/details/84990765
如何应用 BERT :Bidirectional Encoder Representations from Transformers - aliceyangxi1987的博客 - CSDN博客
https://blog.csdn.net/aliceyangxi1987/article/details/84403311
# 用 Bert 硬體開銷還是大。
谷歌终于开源BERT代码:3 亿参数量,机器之心全面解读 _ 机器之心
https://www.jiqizhixin.com/articles/2018-11-01-9
预训练BERT,官方代码发布前他们是这样用TensorFlow解决的 - 知乎
https://zhuanlan.zhihu.com/p/48018623
BERT Explained State of the art language model for NLP
https://towardsdatascience.com/bert-explained-state-of-the-art-language-model-for-nlp-f8b21a9b6270
No comments:
Post a Comment
Note: Only a member of this blog may post a comment.