MIHNet: Combining N-gram, Sequential and Global Information for Text Classification
Abstract
Pre-training language model has achieved amazing results in many NLP tasks. In Particular, BERT (Bidirectional Encoder Representations from Transformers) create a new era in NLP tasks. Despite the success, these model perform well at Global Information but weak on n-gram and Sequential information. In this paper, we conduct exhaustive experiments of classical text classification models upon BERT in text classification task and provide a general guide for BERT+ models. Finally, we propose a new text classification model called MIHNet (Multi-dimension Information Integration using Highway network), which integrates Global, n-gram and Sequential information together and get a better performance. Notably, our model obtains new state-of-the-art results on eight widely-studied text classification datasets.
- Publication:
-
Journal of Physics Conference Series
- Pub Date:
- January 2020
- DOI:
- 10.1088/1742-6596/1453/1/012156
- Bibcode:
- 2020JPhCS1453a2156S