Text Classification Based on BERT
Keywords:
Text Classification, BERT Model, Natural Language Processing, Transformer Architecture, Fine-Tuning, Sentiment Analysis, Deep LearningAbstract
Text classification represents a fundamental task in natural language processing, with applications spanning sentiment analysis, topic labeling, and intent detection. This paper explores the application of the Bidirectional Encoder Representations from Transformers (BERT) model, a large-scale pre-trained language model, to advance the state-of-the-art in text classification. We systematically evaluate BERT's ability to capture deep contextualized representations of text, leveraging its transformer-based architecture to understand semantic nuances and syntactic dependencies often missed by traditional methods. Through fine-tuning on multiple benchmark datasets—including IMDB for sentiment classification and AG News for topic categorization—we demonstrate that BERT significantly outperforms previous approaches, achieving accuracy improvements of up to 4.7% over convolutional and recurrent neural network baselines. Additionally, we analyze the impact of different fine-tuning strategies, such as layer-specific learning rates and dynamic token pooling, on classification performance. The study also addresses practical challenges, including computational resource requirements and model interpretability, proposing simplified variants and attention visualization techniques to enhance usability. Our findings affirm BERT’s robustness and versatility as a backbone architecture for text classification tasks, while also highlighting pathways for future optimization in low-resource and real-time application scenarios.
References
Hu Shaoyun, Weng Qingxiong. Research on a word-vector-fusion-based method for architectural text classification [J]. Microcomputer Applications, 2024, 40(02): 18-20+25.
Shen Jinhua, Chen Hongyi, Zhang Gengping, et al. Research on a patent text classification model based on hierarchical classifiers [J]. Journal of Intelligence, 2023, 42(08): 157-163+68.
Xie Liping. Research on Chinese text classification based on convolutional neural networks [J]. Information & Computer (Theory Edition), 2023, 35(20): 94-96.
Jin Gang. Research on automatic long-text classification using a convolutional recurrent neural network based on distributed word-embedding features [J]. Electronic Technology, 2022, 51(06): 52-54.
Wang Daokang, Zhang Wubo. Research on short-text classification based on MacBERT-BiLSTM and attention mechanism [J]. Modern Electronics Technique, 2023, 46(21): 123-128.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2025 Xie Ning

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.
