BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
摘要
This paper introduced BERT (Bidirectional Encoder Representations from Transformers), a pre-trained bidirectional transformer model that achieved state-of-the-art results on eleven NLP tasks. BERT's bidirectional training and fine-tuning approach revolutionized natural language processing.