Monday, December 13, 2021

BERT Applications

BERT Applications

2019/01/17

-----


Fig. BERT Applications(圖片來源)。

-----

符號說明:

# basic
// advanced

-----

Paper

# Bert 應用ㄧ 文本分類
Kant, Neel, et al. "Practical Text Classification With Large Pre-Trained Language Models." arXiv preprint arXiv:1812.01207 (2018).
https://arxiv.org/pdf/1812.01207.pdf

# Bert 應用二 回答問題
Zhu, Chenguang, Michael Zeng, and Xuedong Huang. "SDNet: Contextualized Attention-based Deep Network for Conversational Question Answering." arXiv preprint arXiv:1812.03593 (2018).
https://arxiv.org/pdf/1812.03593.pdf

# Bert 應用三 生成文本
Transfer Learning for Style-Specific Text Generation
https://nips2018creativity.github.io/doc/Transfer%20Learning%20for%20Style-Specific%20Text%20Generation.pdf

// Why
Qi, Ye, et al. "When and Why are Pre-trained Word Embeddings Useful for Neural Machine Translation?." arXiv preprint arXiv:1804.06323 (2018).
https://arxiv.org/pdf/1804.06323.pdf

// Common Sense
Trinh, Trieu H., and Quoc V. Le. "Do Language Models Have Common Sense?." (2018).
https://openreview.net/pdf?id=rkgfWh0qKX

-----

References

NLP中的语言模型预训练&微调 - CLOUD - CSDN博客
https://blog.csdn.net/muumian123/article/details/84990765

如何应用 BERT :Bidirectional Encoder Representations from Transformers - aliceyangxi1987的博客 - CSDN博客
https://blog.csdn.net/aliceyangxi1987/article/details/84403311

# 用 Bert 硬體開銷還是大。
谷歌终于开源BERT代码:3 亿参数量,机器之心全面解读 _ 机器之心
https://www.jiqizhixin.com/articles/2018-11-01-9

预训练BERT,官方代码发布前他们是这样用TensorFlow解决的 - 知乎
https://zhuanlan.zhihu.com/p/48018623

BERT Explained  State of the art language model for NLP
https://towardsdatascience.com/bert-explained-state-of-the-art-language-model-for-nlp-f8b21a9b6270

No comments: