|
|
Multi-Round Iterative Retrieval Algorithm for Parsing Question-Answering Process |
Zhou Changshun1,2,Ying Wenhao2(),Zhong Shan2,Gong Shengrong1,2 |
1School of Computer Science and Technology, Soochow University, Suzhou 215008, China 2School of Computer Science and Engineering, Changshu Institute of Technology, Changshu 215500, China |
|
|
Abstract [Objective] This paper designs a retrieval model to explore the interpretability of question-answering tasks. It examines the reasoning processes of these reading comprehension models and improves sentence relevance in traditional unsupervised retrieval algorithms. [Methods] We proposed a new unsupervised retrieval model ISR, which integrated modules of Pearson correlation coefficient, GloVe word embeddings, and IDF weighting. The ISR model conducted fine-grained retrieval of inference sentences through multi-round iterations. [Results] The proposed model’s P, R, and F1 metrics were 2.4%, 1.8%, and 2.1% higher than the MSSwQ model on the MultiRC dataset. Its P, R, and F1 metrics were 4.8%, 2.6%, and 3.7% higher than the MSSwQ on the HotPotQA dataset. [Limitations] There might be excessive matching issues due to the model’s retrieval matching mechanism. [Conclusions] The proposed model improves the accuracy of retrieval inference sentences, which can be effectively applied to the question-answering tasks.
|
Received: 12 February 2023
Published: 04 May 2023
|
|
Fund:National Natural Science Foundation of China(61972059);National Science Foundation for Post-Doctoral Scientists of China(2021M692368) |
Corresponding Authors:
Ying Wenhao,ORCID:0000-0001-5992-5444,E-mail:ywh@cslg.edu.cn。
|
[1] |
黄星宇. 基于ALBERT的医疗问答系统设计与实现[D]. 成都: 电子科技大学, 2022.
|
[1] |
(Huang Xingyu. Design and Implementation of Medical Question Answering System Based on ALBERT[D]. Chengdu: University of Electronic Science and Technology of China, 2022.)
|
[2] |
范亦涵. 面向税务的智能问答系统的设计与实现[D]. 济南: 山东大学, 2019.
|
[2] |
(Fan Yihan. The Design and Implementation of Intelligent Question Answering System for Taxation[D]. Jinan: Shandong University, 2019.)
|
[3] |
黄薇屹. 基于知识图谱的深度法律内容问答模型[D]. 深圳: 中国科学院深圳先进技术研究院, 2020.
|
[3] |
(Huang Weiyi. Deep Neural Networks for Legal Question Answering Based on Knowledge Graph[D]. Shenzhen: Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, 2020.)
|
[4] |
胡月, 周光有. 基于Graph Transformer的知识库问题生成[J]. 中文信息学报, 2022, 36(2): 111-120.
|
[4] |
(Hu Yue, Zhou Guangyou. Question Generation from Knowledge Base with Graph Transformer[J]. Journal of Chinese Information Processing, 2022, 36(2): 111-120.)
|
[5] |
顾迎捷, 桂小林, 李德福, 等. 基于神经网络的机器阅读理解综述[J]. 软件学报, 2020, 31(7): 2095-2126.
|
[5] |
(Gu Yingjie, Gui Xiaolin, Li Defu, et al. Survey of Machine Reading Comprehension Based on Neural Network[J]. Journal of Software, 2020, 31(7): 2095-2126.)
|
[6] |
王保成, 刘利军, 黄青松. 面向中文医疗问答网站的相似问题检索研究[J]. 中文信息学报, 2022, 36(6): 135-145.
|
[6] |
(Wang Baocheng, Liu Lijun, Huang Qingsong. Retrieval of Similar Questions from Chinese Medical Question Answering Website[J]. Journal of Chinese Information Processing, 2022, 36(6): 135-145.)
|
[7] |
赵芸, 刘德喜, 万常选, 等. 检索式自动问答研究综述[J]. 计算机学报, 2021, 44(6): 1214-1232.
|
[7] |
(Zhao Yun, Liu Dexi, Wan Changxuan, et al. Retrieval-Based Automatic Question Answer: A Literature Survey[J]. Chinese Journal of Computers, 2021, 44(6): 1214-1232.)
|
[8] |
杨姗姗, 姜丽芬, 孙华志, 等. 基于时间卷积网络的多项选择机器阅读理解[J]. 计算机工程, 2020, 46(11): 97-103.
doi: 10.19678/j.issn.1000-3428.0055628
|
[8] |
(Yang Shanshan, Jiang Lifen, Sun Huazhi, et al. Multiple Choice Machine Reading Comprehension Based on Temporal Convolutional Network[J]. Computer Engineering, 2020, 46(11): 97-103.)
doi: 10.19678/j.issn.1000-3428.0055628
|
[9] |
Yang Z L, Qi P, Zhang S Z, et al. HotpotQA: A Dataset for Diverse, Explainable Multi-Hop Question Answering[C]// Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. Stroudsburg: Association for Computational Linguistics, 2018: 2369-2380.
|
[10] |
Reddy S, Chen D Q, Manning C D. CoQA: A Conversational Question Answering Challenge[C]// Proceedings of the Transactions of the Association for Computational Linguistics. Stroudsburg: Association for Computational Linguistics, 2019: 249-266.
|
[11] |
Inoue N, Stenetorp P, Inui K. R4C: A Benchmark for Evaluating RC Systems to Get the Right Answer for the Right Reason[C]// Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Stroudsburg: Association for Computational Linguistics, 2020: 6740-6750.
|
[12] |
Ho X, Nguyen A K D, Sugawara S, et al. Constructing a Multi-Hop QA Dataset for Comprehensive Evaluation of Reasoning Steps[C]// Proceedings of the 28th International Conference on Computational Linguistics. Stroudsburg: Association for Computational Linguistics, 2020: 6609-6625.
|
[13] |
Saxena A, Tripathi A, Talukdar P. Improving Multi-Hop Question Answering over Knowledge Graphs Using Knowledge Base Embeddings[C]// Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Stroudsburg: Association for Computational Linguistics, 2020: 4498-4507.
|
[14] |
Lin B Y, Chen X Y, Chen J, et al. KagNet: Knowledge-Aware Graph Networks for Commonsense Reasoning[C]// Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the International Joint Conference on Natural Language Processing. Stroudsburg: Association for Computational Linguistics, 2019: 2829-2839.
|
[15] |
Yasunaga M, Ren H, Bosselut A, et al. QA-GNN: Reasoning with Language Models and Knowledge Graphs for Question Answering[C]// Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies. Stroudsburg: Association for Computational Linguistics, 2021: 535-546.
|
[16] |
Das R, Dhuliawala S, Zaheer M, et al. Go for a Walk and Arrive at the Answer: Reasoning over Paths in Knowledge Bases Using Reinforcement Learning[C]// Proceedings of the 6th International Conference on Learning Representations. Washington: ICLR, 2018.
|
[17] |
Feng Y L, Chen X Y, Lin B Y, et al. Scalable Multi-Hop Relational Reasoning for Knowledge-Aware Question Answering[C]// Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing. Stroudsburg: Association for Computational Linguistics, 2020: 1295-1309.
|
[18] |
Wang P F, Peng N Y, Ilievski F, et al. Connecting the Dots: A Knowledgeable Path Generator for Commonsense Question Answering[C]// Findings of the Association for Computational Linguistics. Stroudsburg: Association for Computational Linguistics, 2020: 4129-4140.
|
[19] |
Yadav V, Bethard S, Surdeanu M. Unsupervised Alignment-Based Iterative Evidence Retrieval for Multi-Hop Question Answering[C]// Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Stroudsburg: Association for Computational Linguistics, 2020: 4514-4525.
|
[20] |
Sydorova A, Poerner N, Roth B. Interpretable Question Answering on Knowledge Bases and Text[C]// Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Stroudsburg: Association for Computational Linguistics, 2019: 4943-4951.
|
[21] |
Chen D Q, Fisch A, Weston J, et al. Reading Wikipedia to Answer Open-Domain Questions[C]// Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics. Stroudsburg: Association for Computational Linguistics, 2017: 1870-1879.
|
[22] |
Pan L M, Chen W H, Xiong W H, et al. Unsupervised Multi-Hop Question Answering by Question Generation[C]// Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Stroudsburg: Association for Computational Linguistics, 2021: 5866-5880.
|
[23] |
Pennington J, Socher R, Manning C D. GloVe: Global Vectors for Word Representation[C]// Proceedings of the Conference on Empirical Methods in Natural Language Processing. Stroudsburg: Association for Computational Linguistics, 2014: 1532-1543.
|
[24] |
杨泉, 孙玉泉. 基于《同义词词林》深度的词义相似度计算研究[J]. 计算机工程与应用, 2020, 56(17): 48-54.
doi: 10.3778/j.issn.1002-8331.2001-0249
|
[24] |
(Yang Quan, Sun Yuquan. Research on Semantic Similarity Calculation Based on Depth of CiLin[J]. Computer Engineering and Applications, 2020, 56(17): 48-54.)
doi: 10.3778/j.issn.1002-8331.2001-0249
|
[25] |
张仰森, 王胜, 魏文杰, 等. 融合语义信息与问题关键信息的多阶段注意力答案选取模型[J]. 计算机学报, 2021, 44(3): 491-507.
|
[25] |
(Zhang Yangsen, Wang Sheng, Wei Wenjie, et al. An Answer Selection Model Based on Multi-Stage Attention Mechanism with Combination of Semantic Information and Key Information of the Question[J]. Chinese Journal of Computers, 2021, 44(3): 491-507.)
|
[26] |
Vikas Y, Steven B, Mihai S. Quick and (not so) Dirty: Unsupervised Selection of Justification Sentences for Multi-Hop Question Answering[C]// Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing. Stroudsburg: Association for Computational Linguistics, 2019: 2578-2589.
|
[27] |
Cui Y M, Liu T, Che W X, et al. ExpMRC: Explainability Evaluation for Machine Reading Comprehension[J]. Heliyon, 2022, 8(4): e09290.
doi: 10.1016/j.heliyon.2022.e09290
|
[28] |
Wang A, Singh A, Michael J, et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding[C]// Proceedings of the EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP. Stroudsburg: Association for Computational Linguistics, 2018: 353-355.
|
[29] |
Trivedi H, Kwon H, Khot T, et al. Repurposing Entailment for Multi-Hop Question Answering Tasks[C]// Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Stroudsburg: Association for Computational Linguistics, 2019: 2948-2958.
|
[30] |
Sun K, Yu D, Yu D, et al. Improving Machine Reading Comprehension with General Reading Strategies[C]// Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Stroudsburg: Association for Computational Linguistics, 2019: 2633-2643.
|
[31] |
Devlin J, Chang M W, Lee K, et al. BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding[C]// Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Stroudsburg: Association for Computational Linguistics, 2019: 4171-4186.
|
[32] |
Liu Y H, Ott M, Goyal N, et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach[OL]. arXiv Preprint, arXiv:1907.11692.
|
[33] |
Yang Z L, Dai Z H, Yang Y M, et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding[C]// Proceedings of the 32nd Conference on Neural Information Processing Systems. Cambridge: MIT Press, 2019: 5753-5763.
|
|
Viewed |
|
|
|
Full text
|
|
|
|
|
Abstract
|
|
|
|
|
Cited |
|
|
|
|
|
Shared |
|
|
|
|
|
Discussed |
|
|
|
|