Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing
NB-MLM: Efficient Domain Adaptation of Masked Language Models for Sentiment Analysis
January 2021 • Nikolay Arefyev, Dmitrii Kharchev, Artem Shelmanov
While Masked Language Models (MLM) are pre-trained on massive datasets, the additional training with the MLM objective on domain or task-specific data before fine-tuning for the final task is known to improve the final performance. This is usually referred to as the domain or task adaptation step. However, unlike the initial pre-training, this step is performed for each domain or task individually and is still rather slow, requiring several GPU days compared to several GPU hours required for the final task fine-tu…