|
|
Comprehending Texts and Answering Questions Based on Hierarchical Interactive Network |
Cheng Yong1(), Xu Dekuan1, Lv Xueqiang2 |
1School of Chinese Language and Literature, Ludong University, Yantai 264025, China 2School of Computer Science, Beijing University of Information Technology, Beijing 100192, China |
|
|
Abstract [Objective] This paper aims to help computer answer questions accurately based on text comprehension. [Methods] First, we proposed a neural network model based on hirrarchical interaction mechanism. We introduced various human thinking mechanism to build this model, which contained hierarchical processing, content filtering and multi-dimensional attention. Then, we ran the proposed model with dataset from Chinese Machine Reading Comprehension (CMRC) 2017. [Results] The precision of the proposed method on test-set was 0.78, which was better than the best result of other published models. [Limitations] There was no further optimization for the potential answers. [Conclusions] The proposed hierarchical interactive network improves machine’s ability to answer questions based on text comprehension.
|
Received: 24 May 2018
Published: 16 January 2019
|
|
[1] |
郭利敏. 基于卷积神经网络的文献自动分类研究[J]. 图书与情报, 2017(6): 96-103.
doi: 10.11968/tsyqb.1003-6938.2017119
|
[1] |
(Guo Limin.Study of Automatic Classification of Literature Based on Convolution Neural Network[J]. Library and Information, 2017(6): 96-103.)
doi: 10.11968/tsyqb.1003-6938.2017119
|
[2] |
李慧宗, 胡学钢, 杨恒宇,等. 基于LDA的社会化标签综合聚类方法[J]. 情报学报, 2015, 34(2): 146-155.
doi: 10.3772/j.issn.1000-0135.2015.002.004
|
[2] |
(Li Huizong, Hu Xuegang, Yang Hengyu, et al.A Comprehensive Clustering Method of Social Tags Based on LDA[J]. Journal of the China Society for Scientific and Technical Information, 2015, 34(2): 146-155.)
doi: 10.3772/j.issn.1000-0135.2015.002.004
|
[3] |
徐彤阳, 尹凯. 大数据背景下微博语义检索[J]. 情报杂志, 2017, 36(12): 173-179.
|
[3] |
(Xu Tongyang, Yin Kai.Semantic Retrieval of Microblogging in the Background of Large Data[J]. Journal of Intelligence, 2017, 36(12): 173-179.)
|
[4] |
张志昌. 开放域阅读理解关键技术研究[D]. 哈尔滨: 哈尔滨工业大学, 2010.
|
[4] |
(Zhang Zhichang.Key Technologies of Reading Comprehension for Open-Domain[D]. Harbin: Harbin Institute of Technology, 2010.)
|
[5] |
Hermann K M, Kočiský T, Grefenstette E, et al.Teaching Machines to Read and Comprehend[C]// Proceedings of the 28th International Conference on Neural Information Processing Systems. 2015: 1693-1701.
|
[6] |
Kadlec R, Schmid M, Bajgar O, et al.Text Understanding with the Attention Sum Reader Network[C]// Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics. 2016: 908-918.
|
[7] |
Chen D, Bolton J, Manning C D.A Thorough Examination of the CNN/Daily Mail Reading Comprehension Task[C]// Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics. 2016: 2358-2367.
|
[8] |
Dhingra B, Liu H, Yang Z, et al.Gated-Attention Readers for Text Comprehension[C]// Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics. 2017: 1832-1846.
|
[9] |
Sordoni A, Bachman P, Trischler A, et al.Iterative Alternating Neural Attention for Machine Reading[OL]. arXiv Preprint, arXiv: 1606.02245.
|
[10] |
Cui Y, Chen Z, Wei S, et al.Attention-over-Attention Neural Networks for Reading Comprehension[C]// Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics. 2017: 593-602.
|
[11] |
Rajpurkar P, Zhang J, Lopyrev K, et al.SQuAD: 100,000+ Questions for Machine Comprehension of Text[C]// Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. 2016: 2383-2392.
|
[12] |
Wang W, Yang N, Wei F, et al.Gated Self-Matching Networks for Reading Comprehension and Question Answering[C]// Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics. 2017: 189-198.
|
[13] |
Seo M, Kembhavi A, Farhadi A, et al.Bidirectional Attention Flow for Machine Comprehension[OL]. arXiv Preprint, arXiv: 1611.01603.
|
[14] |
Gong Y, Bowman S R.Ruminating Reader: Reasoning with Gated Multi-Hop Attention[OL]. arXiv Preprint, arXiv: 1704.07415.
|
[15] |
Zhang J, Zhu X, Chen Q, et al.Exploring Question Understanding and Adaptation in Neural-network Based Question Answering[OL]. arXiv Preprint, arXiv: 1703.04617.
|
[16] |
Shen Y, Huang P, Gao J, et al.ReasoNet: Learning to Stop Reading in Machine Comprehension[C]// Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2017: 1047-1055.
|
[17] |
Hu M, Peng Y, Huang Z, et al.Mnemonic Reader for Machine Comprehension[OL]. arXiv Preprint, arXiv: 1705.02798.
|
[18] |
Cui Y, Liu T, Chen Z, et al.Dataset for First Evaluation on Chinese Machine Reading Comprehension[OL]. arXiv Preprint, arXiv: 1709.08299.
|
[19] |
顾明远. 教育大辞典[M]. 上海: 上海教育出版社, 1998.
|
[19] |
(Gu Mingyuan.Dictionary of Education[M]. Shanghai: Shanghai Educational Publishing House, 1998.)
|
[20] |
张杰, 魏维. 基于视觉注意力模型的显著性提取[J]. 计算机技术与发展, 2010, 20(11): 109-113.
doi: 10.3969/j.issn.1673-629X.2010.11.027
|
[20] |
(Zhang Jie, Wei Wei.Saliency Extraction Based on Visual Attention Model[J]. Computer Technology and Development, 2010, 20(11): 109-113.)
doi: 10.3969/j.issn.1673-629X.2010.11.027
|
[21] |
张家俊, 宗成庆. 神经网络语言模型在统计机器翻译中的应用[J]. 情报工程, 2017, 3(3): 21-28.
doi: 10.3772/j.issn.2095-915x.2017.03.004
|
[21] |
(Zhang Jiajun, Zong Chengqing.Application of Neural Network Language Model in Statistical Machine Translation[J]. Technology Intelligence Engineering, 2017, 3(3): 21-28.)
doi: 10.3772/j.issn.2095-915x.2017.03.004
|
|
Viewed |
|
|
|
Full text
|
|
|
|
|
Abstract
|
|
|
|
|
Cited |
|
|
|
|
|
Shared |
|
|
|
|
|
Discussed |
|
|
|
|