Monday, December 13, 2021

BERT

BERT

2019/01/17

-----


Fig. BERT(圖片來源:Pixabay)。

-----


# BERT

-----

// LeeMeng - 進擊的 BERT:NLP 界的巨人之力與遷移學習

-----


// LeeMeng - 進擊的 BERT:NLP 界的巨人之力與遷移學習

-----


// LeeMeng - 進擊的 BERT:NLP 界的巨人之力與遷移學習

-----


// Add Shine to your Data Science Resume with these 8 Ambitious Projects on GitHub

-----

References

[1] BERT
Devlin, Jacob, et al. "Bert: Pre-training of deep bidirectional transformers for language understanding." arXiv preprint arXiv:1810.04805 (2018).
https://arxiv.org/pdf/1810.04805.pdf

-----

// English

# Illistrated
The Illustrated BERT, ELMo, and co. (How NLP Cracked Transfer Learning) – Jay Alammar – Visualizing machine learning one concept at a time
http://jalammar.github.io/illustrated-bert/

Dissecting BERT Part 1  The Encoder – Dissecting BERT – Medium
https://medium.com/dissecting-bert/dissecting-bert-part-1-d3c3d495cdb3

Dissecting BERT Part 2  BERT Specifics – Dissecting BERT – Medium
https://medium.com/dissecting-bert/dissecting-bert-part2-335ff2ed9c73

Dissecting BERT Appendix  The Decoder – Dissecting BERT – Medium
https://medium.com/dissecting-bert/dissecting-bert-appendix-the-decoder-3b86f66b0e5f

BERT Explained  State of the art language model for NLP
https://towardsdatascience.com/bert-explained-state-of-the-art-language-model-for-nlp-f8b21a9b6270
 
How BERT leverage attention mechanism and transformer to learn word contextual relations
https://towardsdatascience.com/how-bert-leverage-attention-mechanism-and-transformer-to-learn-word-contextual-relations-5bbee1b6dbdb

Deconstructing BERT  Distilling 6 Patterns from 100 Million Parameters
https://www.kdnuggets.com/2019/02/deconstructing-bert-distilling-patterns-100-million-parameters.html

Are BERT Features InterBERTible
https://www.kdnuggets.com/2019/02/bert-features-interbertible.html

Add Shine to your Data Science Resume with these 8 Ambitious Projects on GitHub
https://www.nekxmusic.com/add-shine-to-your-data-science-resume-with-these-8-ambitious-projects-on-github/

-----

// 簡體中文

NLP自然语言处理:文本表示总结 - 下篇(ELMo、Transformer、GPT、BERT)_陈宸的博客-CSDN博客
https://blog.csdn.net/qq_35883464/article/details/100173045

Attention isn’t all you need!BERT的力量之源远不止注意力 - 知乎
https://zhuanlan.zhihu.com/p/58430637 

-----

LeeMeng - 進擊的 BERT:NLP 界的巨人之力與遷移學習
https://leemeng.tw/attack_on_bert_transfer_learning_in_nlp.html

-----

// Code

# PyTorch
GitHub - huggingface_pytorch-pretrained-BERT  📖The Big-&-Extending-Repository-of-Transformers  Pretrained PyTorch models for Google's BERT, OpenAI GPT & GPT-2, Google_CMU Transformer-XL
https://github.com/huggingface/pytorch-pretrained-BERT 

# PyTorch
GitHub - codertimo_BERT-pytorch  Google AI 2018 BERT pytorch implementation
https://github.com/codertimo/BERT-pytorch

No comments:

Post a Comment

Note: Only a member of this blog may post a comment.