State-of-the-art performances for natural language processing tasks are achieved by supervised learning, specifically, fine-tuning pre-trained models such as BERT (Bidirectional Encoder Representation from Transformers). With increasingly accurate models, the size of fine-tuned pre-training corpus is becoming larger and larger. However, very few studies have explored selection corpus. Therefore...