Analyzing Implicit Discourse Relation with Single Classifier and Multi-Task Network
Wang Hong,Shu Zhan,Gao Yinquan,Tian Wenhong()
School of Information and Software Engineering, University of Electronic Science and Technology of China, Chengdu 610054, China Yangtze Delta Region Institute of University of Electronic Science and Technology of China, Huzhou 313001, China
[Objective] This paper proposes a new method to identify implicit discourse relations based on a single classifier and multi-task learning model. [Methods] First,we modeled the implicit and explicit discourse relationships with the multi-task learning method. Then, we converted the four classification problems to two and trained the single classifier. [Results] We examined our new method with the HIT-CDTB data set. For the corpus with extended and parallel relations, the F1 values reached 0.94 and 0.81 respectively, which were significantly improved with four inter-sentence relations. [Limitations] The performance of our model could be improved with more distributed and expanded datasets. [Conclusions] The proposed method yields the best results with the HIT-CDTB data set. Deleting connectives will add noise to the training set and negatively affect the model’s performance.
王鸿, 舒展, 高印权, 田文洪. 一种单分类器联合多任务网络的隐式句间关系分析方法*[J]. 数据分析与知识发现, 2021, 5(11): 80-88.
Wang Hong, Shu Zhan, Gao Yinquan, Tian Wenhong. Analyzing Implicit Discourse Relation with Single Classifier and Multi-Task Network. Data Analysis and Knowledge Discovery, 2021, 5(11): 80-88.
(Zhang Muyu, Qin Bing, Liu Ting. Chinese Discourse Relation Semantic Taxonomy and Annotation[J]. Journal of Chinese Information Processing, 2014, 28(2):28-36.)
[3]
Miltsakaki E, Prasad R, Joshi A, et al. The Penn Discourse Treebank[C]// Proceedings of the International Conference on Language Resources & Evaluation. 2004: 342-351.
[4]
Lei W Q, Xiang Y X, Wang Y W, et al. Linguistic Properties Matter for Implicit Discourse Relation Recognition: Combining Semantic Interaction, Topic Continuity and Attribution[C]//Proceedings of the 32nd AAAI Conference on Artificial Intelligence. AAAI Press, 2018: 4848-4855.
[5]
Xue N W. Annotating Discourse Connectives in the Chinese Treebank[C]// Proceedings of the ACL Workshop in Frontiers in Annotation II: Pie in the Sky.ACL, 2005: 84-91.
[6]
Graves A. Supervised Sequence Labelling with Recurrent Neural Networks[M]. Berlin, Heidelberg: Springer, 2012.
[7]
Schuster M, Paliwal K K. Bidirectional Recurrent Neural Networks[J]. IEEE Transactions on Signal Processing, 1997, 45(11):2673-2681.
doi: 10.1109/78.650093
[8]
Rönnqvist S, Schenk N, Chiarcos C. A Recurrent Neural Model with Attention for the Recognition of Chinese Implicit Discourse Relations[C]// Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). 2017:256-262.
[9]
Liu Y, Li S J, Zhang X D, et al. Implicit Discourse Relation Classification via Multi-task Neural Networks[C]//Proceedings of the 30th AAAI Conference on Artificial Intelligence. AAAI Press, 2016: 2750-2756.
(Tian Wenhong, Gao Yinquan, Huang Houwen, et al. Implicit Discourse Relation Analysis Based on Multi-Task Bi-LSTM[J]. Journal of Chinese Information Processing, 2019, 33(5):47-53.)
[11]
Liu P F, Qiu X P, Huang X J. Recurrent Neural Network for Text Classification with Multi-Task Learning[C]// Proceedings of the 25th International Joint Conference on Artificial Intelligence. 2016: 2873-2879.
[12]
Liu P F, Qiu X P, Huang X J. Adversarial Multi-Task Learning for Text Classification[C]// Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2017: 1-10.
[13]
Sanh V, Wolf T, Ruder S. A Hierarchical Multi-Task Approach for Learning Embeddings from Semantic Tasks[J]. Proceedings of the AAAI Conference on Artificial Intelligence, 2019, 33:6949-6956.
doi: 10.1609/aaai.v33i01.33016949
[14]
张雅星. 汉语篇章句间关系分析[D]. 太原:山西大学, 2018.
[14]
(Zhang Yaxing. Chinese Discourse Relation Analysis[D]. Taiyuan: Shanxi University, 2018.)
(Zhang Kaili, Deng Dunhua, Li Yuan, et al. Inner-Attention Based Multi-Way Convolutional Neural Network for Relation Recognition in Chinese Compound Sentence[J]. Journal of Chinese Information Processing, 2020, 34(6):9-17, 26.)
(Wan Changxuan, Gan Lixin, Jiang Tengjiao, et al. Chinese Named Entity Implicit Relation Extraction Based on Company Verbs[J]. Chinese Journal of Computers, 2019, 42(12):2795-2820.)
[17]
Gilbert R A, Davis M H, Gaskell M G, et al. The Relationship Between Sentence Comprehension and Lexical-Semantic Retuning[J]. Journal of Memory and Language, 2021, 116:104188.
doi: 10.1016/j.jml.2020.104188
[18]
Cho K, van Merrienboer B, Gulcehre C, et al. Learning Phrase Representations Using RNN Encoder-Decoder for Statistical Machine Translation[C]// Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing. 2014:1724-1734.
[19]
Pennington J, Socher R, Manning C. GloVe: Global Vectors for Word Representation[C]// Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing. 2014: 1532-1543.
(Ji Jianhui, Zhang Muyu, Qin Bing, et al. The Chinese Discourse Parser[J]. Journal of Jiangxi Normal University(Natural Science Edition), 2015, 39(2):124-131.)
[21]
Rutherford A, Xue N W. Improving the Inference of Implicit Discourse Relations via Classifying Explicit Discourse Connectives[C]// Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 2015: 799-808.