Please wait a minute...
Data Analysis and Knowledge Discovery
Current Issue | Archive | Adv Search |
The Multi-level Feature Extraction Capsule Network Model for Text Classification Research
YU Ben-gong,ZHU Xiao-Jie,ZHANG Zi-Wei
(School of Management, Hefei University of Technology, Hefei 230009, China) (Key Laboratory of Process Optimization & Intelligent Decision-making, Ministry of Education, Hefei University of Technology, Hefei 230009, China)
Download:
Export: BibTeX | EndNote (RIS)      
Abstract  

[Objective] In order to improve the feature extraction capabilities of the existing shallow text classification models, this paper adopts a structured method to extract text information hierarchically from bottom to top, so as to improve the text classification effect.

[Methods] This paper proposes a text classification model(MFE-CapsNet)based on global and high-level feature acquisition. First, the model uses bidirectional gated recurrent unit (BiGRU) Extract the context information, and introduce the attention coding hidden layer vector to improve the feature extraction ability of the sequence model. We combine the capsule network and use dynamic routing to obtain high-level aggregated local information build an MFE-CapsNet model, and conduct comparative experiments on text classification.

[Results] The experimental results show that the F1 value of the MFE-CapsNet model proposed in this paper reaches 96.21%, 94.17%, and 94.19% on the Chinese data sets in three different fields respectively, Compared with other classification methods, the increase was at least 1.28%, 1.49%, and 0.46% respectively.

[Limitations] The experiment is only verified on three corpora.

[Conclusions] The MFE-CapsNet model uses the improved capsule network algorithm structure to more comprehensively and deeply mine the semantic features of text and improve the performance of text classification.

Key words Text classification      BiGRU      Attention       Capsule network      
Published: 29 March 2021

Cite this article:

YU Ben-gong, ZHU Xiao-Jie, ZHANG Zi-Wei. The Multi-level Feature Extraction Capsule Network Model for Text Classification Research . Data Analysis and Knowledge Discovery, 0, (): 1-.

URL:

http://manu44.magtech.com.cn/Jwk_infotech_wk3/EN/10.11925/infotech.2020.1273     OR     http://manu44.magtech.com.cn/Jwk_infotech_wk3/EN/Y0/V/I/1

[1] Jiang Cuiqing,Wang Xiangxiang,Wang Zhao. Forecasting Car Sales Based on Consumer Attention[J]. 数据分析与知识发现, 2021, 5(1): 128-139.
[2] Yin Haoran,Cao Jinxuan,Cao Luzhe,Wang Guodong. Identifying Emergency Elements Based on BiGRU-AM Model with Extended Semantic Dimension[J]. 数据分析与知识发现, 2020, 4(9): 91-99.
[3] Huang Lu,Zhou Enguo,Li Daifeng. Text Representation Learning Model Based on Attention Mechanism with Task-specific Information[J]. 数据分析与知识发现, 2020, 4(9): 111-122.
[4] Tang Xiaobo,Gao Hexuan. Classification of Health Questions Based on Vector Extension of Keywords[J]. 数据分析与知识发现, 2020, 4(7): 66-75.
[5] Wang Sidi,Hu Guangwei,Yang Siyu,Shi Yun. Automatic Transferring Government Website E-Mails Based on Text Classification[J]. 数据分析与知识发现, 2020, 4(6): 51-59.
[6] Shi Lei,Wang Yi,Cheng Ying,Wei Ruibin. Review of Attention Mechanism in Natural Language Processing[J]. 数据分析与知识发现, 2020, 4(5): 1-14.
[7] Xu Yuemei,Liu Yunwen,Cai Lianqiao. Predicitng Retweets of Government Microblogs with Deep-combined Features[J]. 数据分析与知识发现, 2020, 4(2/3): 18-28.
[8] Xue Fuliang,Liu Lifang. Fine-Grained Sentiment Analysis with CRF and ATAE-LSTM[J]. 数据分析与知识发现, 2020, 4(2/3): 207-213.
[9] Qi Ruihua,Jian Yue,Guo Xu,Guan Jinghua,Yang Mingxin. Sentiment Analysis of Cross-Domain Product Reviews Based on Feature Fusion and Attention Mechanism[J]. 数据分析与知识发现, 2020, 4(12): 85-94.
[10] Xu Tongtong,Sun Huazhi,Ma Chunmei,Jiang Lifen,Liu Yichen. Classification Model for Few-shot Texts Based on Bi-directional Long-term Attention Features[J]. 数据分析与知识发现, 2020, 4(10): 113-123.
[11] Bengong Yu,Yumeng Cao,Yangnan Chen,Ying Yang. Classification of Short Texts Based on nLD-SVM-RF Model[J]. 数据分析与知识发现, 2020, 4(1): 111-120.
[12] Weimin Nie,Yongzhou Chen,Jing Ma. A Text Vector Representation Model Merging Multi-Granularity Information[J]. 数据分析与知识发现, 2019, 3(9): 45-52.
[13] Yunfei Shao,Dongsu Liu. Classifying Short-texts with Class Feature Extension[J]. 数据分析与知识发现, 2019, 3(9): 60-67.
[14] Heran Qin,Liu Liu,Bin Li,Dongbo Wang. Automatic Classification of Ancient Classics with Entity Features[J]. 数据分析与知识发现, 2019, 3(9): 68-76.
[15] Guo Chen,Tianxiang Xu. Sentence Function Recognition Based on Active Learning[J]. 数据分析与知识发现, 2019, 3(8): 53-61.
  Copyright © 2016 Data Analysis and Knowledge Discovery   Tel/Fax:(010)82626611-6626,82624938   E-mail:jishu@mail.las.ac.cn