BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding, Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova, 2019Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Vol. 1DOI: 10.18653/v1/N19-1423 - 这篇开创性论文介绍了BERT模型,该模型显著使用了WordPiece分词。它展示了子词分词在大型Transformer模型中进行语言理解的实际应用和重要性。
Natural Language Processing with Transformers, Lewis Tunstall, Leandro von Werra, Thomas Wolf, 2022 (O'Reilly Media) - 一本实用指南,包含各种分词方法、特殊标记及其在Hugging Face Transformers生态系统中的应用的详细解释和代码示例。