arXiv:1810.04805, BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , Authors: Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova
arXiv:1812.03593, SDNet: Contextualized Attention-based Deep Network for Conversational Question Answering, Authors: Chenguang Zhu, Michael Zeng, Xuedong Huang
arXiv:1901.02860, Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context, Authors: Zihang Dai, Zhilin Yang, Yiming Yang, William W. Cohen, Jaime Carbonell, Quoc V. Le and Ruslan Salakhutdinov.
arXiv:1901.04085, Passage Re-ranking with BERT, Authors: Rodrigo Nogueira, Kyunghyun Cho
Github仓库:
官方项目:
google-research/bert, officical TensorFlow code and pre-trained models for BERT , [10053 stars]
除了tensorflow之外的BERT的实现:
codertimo/BERT-pytorch, Google AI 2018 BERT pytorch implementation
huggingface/pytorch-pretrained-BERT, A PyTorch implementation of Google AI's BERT model with script to load Google's pre-trained models , [2422 stars]
Separius/BERT-keras, Keras implementation of BERT with pre-trained weights, [325 stars]
soskek/bert-chainer, Chainer implementation of "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding"
innodatalabs/tbert, PyTorch port of BERT ML model
guotong1988/BERT-tensorflow, BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
dreamgonfly/BERT-pytorch, PyTorch implementation of BERT in "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding"
CyberZHG/keras-bert, Implementation of BERT that could load official pre-trained models for feature extraction and prediction
soskek/bert-chainer, Chainer implementation of "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding"
MaZhiyuanBUAA/bert-tf1.4.0, bert-tf1.4.0
dhlee347/pytorchic-bert, Pytorch Implementation of Google BERT, [106 stars]
kpot/keras-transformer, Keras library for building (Universal) Transformers, facilitating BERT and GPT models, [17 stars]
miroozyx/BERT_with_keras, A Keras version of Google's BERT model, [5 stars]
conda-forge/pytorch-pretrained-bert-feedstock, A conda-smithy repository for pytorch-pretrained-bert. , [0 stars]
BERT的其他资源:
brightmart/bert_language_understanding, Pre-training of Deep Bidirectional Transformers for Language Understanding: pre-train TextCNN, [503 stars]
Pydataman/bert_examples, some examples of bert, run_classifier.py 是基于谷歌bert实现了Quora Insincere Questions Classification二分类比赛。run_ner.py是基于瑞金医院AI大赛 第一赛季数据和bert写的一个命名实体识别。
guotong1988/BERT-chinese, BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding 中文 汉语
zhongyunuestc/bert_multitask, 多任务task
Microsoft/AzureML-BERT, End-to-end walk through for fine-tuning BERT using Azure Machine Learning , [14 stars]
bigboNed3/bert_serving, export bert model for serving, [10 stars]
yoheikikuta/bert-japanese, BERT with SentencePiece for Japanese text. [92 stars]
whqwill/seq2seq-keyphrase-bert, add BERT to encoder part for https://github.com/memray/seq2seq-keyphrase-pytorch, [19 stars]
algteam/bert-examples, bert-demo, [9 stars]
cedrickchee/awesome-bert-nlp, A curated list of NLP resources focused on BERT, attention mechanism, Transformer networks, and transfer learning. [9 stars]
cnfive/cnbert, 中文注释一下bert代码功能, [5 stars]
brightmart/bert_customized, bert with customized features, [20 stars]
yuanxiaosc/BERT_Paper_Chinese_Translation, BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding 论文的中文翻译。Chinese Translation! , [5 stars]
JayYip/bert-multitask-learning, BERT for Multitask Learning, [29 stars]
BERT QA任务:
benywon/ChineseBert, This is a chinese Bert model specific for question answering, [6 stars]
vliu15/BERT, Tensorflow implementation of BERT for QA
matthew-z/R-net, R-net in PyTorch, with BERT and ELMo, [77 stars]
nyu-dl/dl4marco-bert, Passage Re-ranking with BERT, [92 stars]
maksna/bert-fine-tuning-for-chinese-multiclass-classification, use google pre-training model bert to fine-tuning for the chinese multiclass classification
sberbank-ai/ner-bert, BERT-NER (nert-bert) with google bert, [22 stars]
kyzhouhzau/Bert-BiLSTM-CRF, This model base on bert-as-service. Model structure : bert-embedding bilstm crf. , [3 stars]
Hoiy/berserker, Berserker - BERt chineSE woRd toKenizER, Berserker (BERt chineSE woRd toKenizER) is a Chinese tokenizer built on top of Google's BERT model. , [2 stars]
BERT文本生成任务:
asyml/texar, Toolkit for Text Generation and Beyond https://texar.io, Texar is a general-purpose text generation toolkit, has also implemented BERT here for classification, and text generation applications by combining with Texar's other modules. [892 stars]
sakuranew/BERT-AttributeExtraction, USING BERT FOR Attribute Extraction in KnowledgeGraph. fine-tuning and feature extraction. 使用基于bert的微调和特征提取方法来进行知识图谱百度百科人物词条属性抽取。 [10 stars]
BERT visualization toolkit:
jessevig/bertviz, Tool for visualizing BERT's attention, [147 stars]
BERT可视化工具包:
GaoQ1/rasa_nlu_gq, turn natural language into structured data(支持中文,自定义了N种模型,支持不同的场景和任务), [33 stars]
xu-song/bert_as_language_model, bert as language model, fork from https://github.com/google-research/bert, [22 stars]
yuanxiaosc/Deep_dynamic_word_representation, TensorFlow code and pre-trained models for deep dynamic word representation (DDWR). It combines the BERT model and ELMo's deep context word representation., [7 stars]