Please wait a minute...
Data Analysis and Knowledge Discovery  2024, Vol. 8 Issue (5): 151-162    DOI: 10.11925/infotech.2096-3467.2023.0324
Current Issue | Archive | Adv Search |
Constructing Smart Consulting Q&A System Based on Machine Reading Comprehension
Wang Yihu,Bai Haiyan()
Institute of Scientific and Technical Information of China, Beijing 100038, China
Download: PDF (1454 KB)   HTML ( 19
Export: BibTeX | EndNote (RIS)      
Abstract  

[Objective] This paper aims to improve the smart consulting systems to effectively answer academic questions. [Methods] We utilized deep learning, machine reading comprehension, data augmentation, information retrieval, and semantic similarity techniques to construct datasets and an academic knowledge question-answering system. Additionally, we designed a multi-paragraph recall metric to address the characteristics of academic literature and enhance retrieval accuracy with multidimensional features. [Results] Our new model’s ROUGE-L score reached 0.7338, with a question-answering accuracy of 88.65% and a multi-paragraph recall metric accuracy of 88.38%. [Limitations] We only examined the new model with single-domain content, which may limit the system’s performance in dealing with complex issues involving multiple domains. [Conclusions] The deep integration of machine reading comprehension technology with reference services can enhance the efficiency and sharing of academic resources and provide more comprehensive and accurate information support for researchers.

Key wordsDeep Learning      Machine Reading Comprehension      Smart Consulting Services      Q&A Systems     
Received: 12 April 2023      Published: 08 January 2024
ZTFLH:  TP391  
  G252  
Fund:Innovation Research Fund Youth Project of Institute of Scientific and Technical Information of China(QN2023-11)
Corresponding Authors: Bai Haiyan,ORCID:0000-0002-9552-3845,E-mail: bhy@istic.ac.cn。   

Cite this article:

Wang Yihu, Bai Haiyan. Constructing Smart Consulting Q&A System Based on Machine Reading Comprehension. Data Analysis and Knowledge Discovery, 2024, 8(5): 151-162.

URL:

https://manu44.magtech.com.cn/Jwk_infotech_wk3/EN/10.11925/infotech.2096-3467.2023.0324     OR     https://manu44.magtech.com.cn/Jwk_infotech_wk3/EN/Y2024/V8/I5/151

数据集 发布时间 数据来源 问题数据量
WebQA[6] 2016 百度知道 42 000对问题
CMRC 2018[7] 2018 维基百科 20 000对问题
SQuAD-zen[8] 2020 由原始SQuAD数据集翻译 110 000对问题
中医数据集[9] 2020 《黄帝内经翻译版》等文本 13 000对问题
疫情政务数据集[10] 2020 疫情相关政策文档 5 000对问题
Chinese Span-Extraction Datasets
Architecture of the Smart Consulting Q&A System
Example of Question Extraction
问题类型 类型介绍 问题举例 问题数量
事实型问题 此类问题重点关注对事实的提问,例如某种具体病症、药物的介绍,或其性质、成分等。 (1)丙酸倍氯米松是什么
(2)耶尔森菌素存在于哪里
(3)外源性凋亡由什么介导
(4)胎停育常于什么时候发生
1 203
功能型问题 此类问题着重于药物的功效、作用,或其危害、影响。 (1)静息内皮细胞有什么作用
(2)辐射缓和剂用于什么
(3)瘢痕憩室的微创手术特点
(4)钒对胎儿有什么影响
709
原理型问题 此类问题重点关注某种病症或药物的具体原理,或其之间存在的关系。 (1)脂联素如何抑制肝脏炎症改变
(2)为何注射LPS能形成血小板聚集
(3)慢性系统性免疫与癌症的关系
(4)AIS患者发生误吸的主要原因有
571
数据型问题 此类问题包括具体数据问题,例如发病率、灵敏度等;以及归类问题,例如某种病症或药剂分为哪几类等。 (1)肠黏膜屏障可分为几种
(2)OVCF的永久致残率为多少
(3)甲状腺癌有哪几种病理类型
(4)间充质干细胞分为几类
465
Example of Dataset Problem Type
原问题 数据增强备选问题 最终选择
硫酸依替米星肾毒性如何 (1)硫酸依替米星的药理毒性
(2)尿毒症吃硫酸依替米星有用吗
(3)硫酸依替米星片能治肾炎吗
(4)硫酸依替米星片吃了对肾有伤害吗
(5)硫酸依替米星片的药效与功效
14
当散光超过0.75D时患者有什么症状 (1)散光0.75d,有什么症状
(2)散光度数超过0.75d有什么后果
(3)眼睛散光0.75d有问题吗
(4)散光0.75d,严重吗
(5)散光度数是0.75是什么意思
234
钾丢失过多的原因有哪些 (1)钾丢失过多的原因是什么
(2)什么原因导致钾的丢失过多呢
(3)什么是钾丢失过多
(4)钾过多的原因有哪些
(5)钾过多的原因有哪些?
23
Example of Data Augmentation Experiment
Framework of Segment Recall System
Training Results of Multivariate Metrics Machine Learning
Paragraph Recall Results
说明 样例
原始文本 使用语言模型来预测下一个词的probability。
分词文本 使用 语言 模型 来 预测 下 一个 词的 probability。
原始掩码输入 使 用 语 言 [MASK] 型 来[MASK] 测 下 一 个 词 的 pro [MASK]##lity 。
全词掩码输入 使 用 语 言 [MASK] [MASK] 来 [MASK] [MASK] 下 一 个 词 的 [MASK] [MASK] [MASK] 。
Example of Whole Word Mask
Example of Data Pre-Processing
模型选择 ROUGE-1 ROUGE-2 ROUGE-L
RoBERTa 0.693 889 0.629 872 0.709 262
MedBERT 0.689 091 0.624 344 0.704 720
双模型联合 0.718 208 0.650 382 0.733 837
Evaluation Results
类型 原问题 原答案 预测答案
类型1 移植受者会如何造成潜在肝功能损害 受者因基础疾病状态、高强度的放射及化学治疗、复杂用药、移植并发症等造成潜在的肝功能损害。 移植后,受者因基础疾病状态、高强度的放射及化学治疗、复杂用药、移植并发症等造成潜在的肝功能损害。
类型2 蜈蚣的药理作用有什么 蜈蚣具有抗肿瘤、止痉、抗真菌等作用。 蜈蚣具有抗肿瘤、止痉、抗真菌等作用,蜈蚣毒素的主要化学组分有蛋白质、酶、脂肪酸等。
类型3 肠杆菌存在于哪里 肠杆菌素普遍存在于肺炎克雷伯菌中,但由于其能被宿主载脂蛋白2灭活,在感染中几乎不发挥作用。 肠杆菌素普遍存在于肺炎克雷伯菌中。
类型4 lp-pla2有什么功能 Lp-PLA2会产生强化氧化应激反应,损伤血管内膜,并加快动脉粥样硬化的进程,在缺血性脑卒中有着重要的作用。 LP-PLA2在某些情况下具有抗氧化和抗炎功能。
类型5 抗NMDAR抗体脑炎的惊厥症状 在儿童表现最突出,常是儿童就诊的主诉。常成连续发作甚至持续状态,亦可出现亚临床型的癫痫放电。 无法识别。
Example of Training Result Problem Classification
问题类型 RoBERTa MedBERT 双模型联合
类型1 389 376 401
类型2 104 110 109
类型3 72 76 76
类型4 54 60 63
类型5 42 39 12
正确答案总数 565 562 586
总正确率 85.48% 85.02% 88.65%
Training Results
Smart Consulting Q&A System
[1] 周泰冰, 刘文云. 我国公共图书馆数字参考咨询服务比较分析[J]. 情报理论与实践, 2010, 33(12): 84-87.
[1] (Zhou Taibing, Liu Wenyun. Comparative Analysis of Digital Reference Service in Public Libraries in China[J]. Information Studies: Theory & Application, 2010, 33(12): 84-87.)
[2] 刘泽, 邵波, 王怡. 数据驱动下图书馆智慧参考咨询服务模式研究[J]. 情报理论与实践, 2023, 46(5): 176-184.
[2] (Liu Ze, Shao Bo, Wang Yi. Research on a Data-Driven Model for Smart Reference Service in Libraries[J]. Information Studies: Theory & Application, 2023, 46(5): 176-184.)
[3] 刘青华, 谭红英. 国外参考咨询服务形式浅谈[J]. 情报科学, 2002, 20(6): 590-594.
[3] (Liu Qinghua, Tan Hongying. The Reference Service Patterns of Foreign Libraries[J]. Information Science, 2002, 20(6): 590-594.)
[4] Liu S S, Zhang X, Zhang S, et al. Neural Machine Reading Comprehension: Methods and Trends[J]. Applied Sciences, 2019, 9(18): Article No.3698.
[5] Qiu B Y, Chen X, Xu J G, et al. A Survey on Neural Machine Reading Comprehension[OL]. arXiv Preprint, arXiv: 1906.03824.
[6] Li P, Li W, He Z Y, et al. Dataset and Neural Recurrent Sequence Labeling Model for Open-Domain Factoid Question Answering[OL]. arXiv Preprint, arXiv: 1607.06275.
[7] Cui Y M, Liu T, Che W X, et al. A Span-Extraction Dataset for Chinese Machine Reading Comprehension[OL]. arXiv Preprint, arXiv: 1810.07366.
[8] ChineseSquad:中文机器阅读理解数据集[EB/OL]. [2023-05-25]. https://github.com/pluto-junzeng/ChineseSquad.
[8] (ChineseSquad: Chinese Machine Reading Comprehension Dataset[EB/OL]. [2023-05-25]. https://github.com/pluto-junzeng/ChineseSquad.)
[9] 中医文献问题生成数据集[EB/OL]. [2023-05-25]. https://tianchi.aliyun.com/dataset/86895.
[9] (Question Generation Dataset from Texts of Traditional Chinese Medicine[EB/OL]. [2023-05-25]. https://tianchi.aliyun.com/dataset/86895.)
[10] 疫情政务问答助手[EB/OL]. [2023-05-25]. https://www.datafountain.cn/competitions/424/datasets.
[10] (Epidemic Government Q&A Assistant[EB/OL]. [2023-05-25]. https://www.datafountain.cn/competitions/424/datasets.)
[11] Seo M, Kembhavi A, Farhadi A, et al. Bidirectional Attention Flow for Machine Comprehension[OL]. arXiv Preprint, arXiv: 1611.01603.
[12] Yu A W, Dohan D, Luong M T, et al. QANet: Combining Local Convolution with Global Self-Attention for Reading Comprehension[OL]. arXiv Preprint, arXiv: 1804.09541.
[13] Devlin J, Chang M W, Lee K, et al. BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding[OL]. arXiv Preprint, arXiv: 1810.04805.
[14] Liu Y H, Ott M, Goyal N, et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach[OL]. arXiv Preprint, arXiv: 1907.11692.
[15] Lan Z Z, Chen M D, Goodman S, et al. ALBERT: A Lite BERT for Self-Supervised Learning of Language Representations[OL]. arXiv Preprint, arXiv: 1909.11942.
[16] Cui Y M, Che W X, Liu T, et al. Revisiting Pre-Trained Models for Chinese Natural Language Processing[OL]. arXiv Preprint, arXiv: 2004.13922.
[17] Brown T B, Mann B, Ryder N, et al. Language Models are Few-Shot Learners[C]// Proceedings of the 34th International Conference on Neural Information Processing Systems. ACM, 2020: 1877-1901.
[18] Wang S H, Sun Y, Xiang Y, et al. ERNIE 3.0 Titan: Exploring Larger-Scale Knowledge Enhanced Pre-Training for Language Understanding and Generation[OL]. arXiv Preprint, arXiv: 2112.12731.
[19] Zeng W, Ren X Z, Su T, et al. PanGu-α: Large-Scale Autoregressive Pretrained Chinese Language Models with Auto-Parallel Computation[OL]. arXiv Preprint, arXiv: 2104.12369.
[20] Touvron H, Lavril T, Izacard G, et al. LLaMA: Open and Efficient Foundation Language Models[OL]. arXiv Preprint, arXiv: 2302.13971.
[21] 张华平, 李林翰, 李春锦. ChatGPT中文性能测评与风险应对[J]. 数据分析与知识发现, 2023, 7(3): 16-25.
[21] (Zhang Huaping, Li Linhan, Li Chunjin. ChatGPT Performance Evaluation on Chinese Language and Risk Measures[J]. Data Analysis and Knowledge Discovery, 2023, 7(3): 16-25.)
[22] McNeal M, Newyear D. Chatbots: Automating Reference in Public Libraries[M]// Iglesias E. Robots in Academic Libraries:Advancements in Library Automation. Pennsylvania: IGI Global, 2013: 101-114.
[23] Allison D. Chatbots in the Library: Is It Time?[J]. Library Hi Tech, 2012, 30(1): 95-107.
[24] 顾德南. NSTL数字化参考咨询服务初探[J]. 图书情报工作, 2004, 48(1): 19-22.
[24] (Gu Denan. A Discussion on the Digital Reference Services of the National Science and Technology Library[J]. Library and Information Service, 2004, 48(1): 19-22.)
[25] 姚飞, 张成昱, 陈武. 清华智能聊天机器人“小图”的移动应用[J]. 现代图书情报技术, 2014(7/8):120-126.
[25] (Yao Fei, Zhang Chengyu, Chen Wu. The Mobile Application of ‘Xiaotu’—The Smart Talking Robot of Tsinghua University Library[J]. New Technology of Library and Information Service, 2014(7/8):120-126.)
[26] 孙翌, 李鲍, 曲建峰. 图书馆智能化IM咨询机器人的设计与实现[J]. 现代图书情报技术, 2011(5): 88-92.
[26] (Sun Yi, Li Bao, Qu Jianfeng. Design and Implementation of Library Intelligent IM Reference Robot[J]. New Technology of Library and Information Service, 2011(5): 88-92.)
[27] McKie I A S, Narayan B. Enhancing the Academic Library Experience with Chatbots: An Exploration of Research and Implications for Practice[J]. Journal of the Australian Library and Information Association, 2019, 68(3): 268-277.
[28] 王毅, 罗军. 中美图书馆咨询知识库比较研究[J]. 图书情报工作, 2010, 54(17): 40-44.
[28] (Wang Yi, Luo Jun. A Comparative Study on the Library Reference Knowledge Base Between China and U.S.A[J]. Library and Information Service, 2010, 54(17): 40-44.)
[29] 胡潇戈, 戚越, 王玉琦, 等. 面向智能问答的图书馆参考咨询知识库体系设计及构建[J]. 图书情报知识, 2019(5): 101-108.
[29] (Hu Xiaoge, Qi Yue, Wang Yuqi, et al. Design and Construction of Intelligent Question-Answer Knowledge Base for Library Reference and Consultancy[J]. Documentation, Information & Knowledge, 2019(5): 101-108.)
[30] 刘泽, 徐潇洁, 邵波. 基于多策略混合问答系统模型的图书馆咨询机器人的设计与应用[J]. 新世纪图书馆, 2022(5): 43-49.
[30] (Liu Ze, Xu Xiaojie, Shao Bo. Design and Application of Library Consultation Robot Based on Multi-Strategy Mixed Question Answering System Model[J]. New Century Library, 2022(5): 43-49.)
[31] 李文江, 陈诗琴. AIMLBot智能机器人在实时虚拟参考咨询中的应用[J]. 现代图书情报技术, 2012(7/8):127-132.
[31] (Li Wenjiang, Chen Shiqin. Application of AIMLBot Intelligent Robot in Real-Time Virtual Reference Service[J]. New Technology of Library and Information Service, 2012(7/8):127-132.)
[32] 苏剑林. 鱼与熊掌兼得:融合检索和生成的SimBERT模型[EB/OL]. [2023-02-18]. https://kexue.fm/archives/7427.
[32] (Su Jianlin. Have Your Cake and Eat It: A SimBERT Model for Fusion Retrieval and Generation[EB/OL]. [2023-02-18]. https://kexue.fm/archives/7427.)
[33] Bao H B, Dong L, Wei F R, et al. UNILMv2: Pseudo-Masked Language Models for Unified Language Model Pre-Training[C]// Proceedings of the 37th International Conference on Machine Learning. ACM, 2020: 642-652.
[34] Luo R X, Xu J J, Zhang Y, et al. PKUSEG: A Toolkit for Multi-Domain Chinese Word Segmentation[OL]. arXiv Preprint, arXiv: 1906.11455.
[35] Salton G, Buckley C. Term-Weighting Approaches in Automatic Text Retrieval[J]. Information Processing & Management, 1988, 24(5): 513-523.
[36] Robertson S, Zaragoza H. The Probabilistic Relevance Framework: BM25 and Beyond[J]. Foundations and Trends in Information Retrieval, 2009, 3(4): 333-389.
[37] Vaswani A, Shazeer N, Parmar N, et al. Attention is All You Need[C]// Proceedings of the 31st International Conference on Neural Information Processing Systems. ACM, 2017: 6000-6010.
[38] Cui Y M, Che W X, Liu T, et al. Pre-Training with Whole Word Masking for Chinese BERT[J]. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 2021, 29: 3504-3514.
[39] 杨飞洪. 面向中文临床自然语言处理的BERT模型研究[D]. 北京: 北京协和医学院, 2021.
[39] (Yang Feihong. Research on BERT Model for Chinese Clinical Language Processing[D]. Beijing: Peking Union Medical College, 2021.)
[40] Lin C Y. Rouge: A Package for Automatic Evaluation of Summaries[M]// Text Summarization Branches Out. Association for Computational Linguistics, 2004: 74-81.
[1] Ma Zhiyuan, Gao Ying, Zhang Qiang, Zhou Hong, Li Bing, Tao Wan. Research on the Semantic and Structure Fusion-Based Knowledge Graph Completion Model[J]. 数据分析与知识发现, 2024, 8(4): 39-49.
[2] Ruan Guangce, Zhong Jinghan, Zhang Yidi. Review of Term Recognition Studies Based on Deep Learning[J]. 数据分析与知识发现, 2024, 8(4): 64-75.
[3] He Guoxiu, Ren Jiayu, Li Zongyao, Lin Chenxi, Yu Haiyan. Revisiting Deep Learning-based Rumor Detection Models with Interpretable Tools[J]. 数据分析与知识发现, 2024, 8(4): 1-13.
[4] Zhang Zhijian, Xia Sudi, Liu Zhenghao. Seal Recognition and Application Based on Multi-feature Fusion Deep Learning[J]. 数据分析与知识发现, 2024, 8(3): 143-155.
[5] Zhang Xiongtao, Zhu Na, Guo Yuhui. A Survey on Session-Based Recommendation Methods with Graph Neural Network[J]. 数据分析与知识发现, 2024, 8(2): 1-16.
[6] Li Hui, Hu Yaohua, Xu Cunzhen. Personalized Recommendation Algorithm with Review Sentiments and Importance[J]. 数据分析与知识发现, 2024, 8(1): 69-79.
[7] Xiang Zhuoyuan, Chen Hao, Wang Qian, Li Na. Few-Shot Language Understanding Model for Task-Oriented Dialogues[J]. 数据分析与知识发现, 2023, 7(9): 64-77.
[8] Nie Hui, Cai Ruisheng. Online Doctor Recommendation System with Attention Mechanism[J]. 数据分析与知识发现, 2023, 7(8): 138-148.
[9] Li Guangjian, Yuan Yue. Review of Knowledge Elements Extraction in Scientific Literature Based on Deep Learning[J]. 数据分析与知识发现, 2023, 7(7): 1-17.
[10] Wu Jialun, Zhang Ruonan, Kang Wulin, Yuan Puwei. Deep Learning Model of Drug Recommendation Based on Patient Similarity Analysis[J]. 数据分析与知识发现, 2023, 7(6): 148-160.
[11] Wang Xiaofeng, Sun Yujie, Wang Huazhen, Zhang Hengzhang. Construction and Verification of Type-Controllable Question Generation Model Based on Deep Learning and Knowledge Graphs[J]. 数据分析与知识发现, 2023, 7(6): 26-37.
[12] Wang Nan, Wang Qi. Evaluating Student Engagement with Deep Learning[J]. 数据分析与知识发现, 2023, 7(6): 123-133.
[13] Liu Yang, Zhang Wen, Hu Yi, Mao Jin, Huang Fei. Hotel Stock Prediction Based on Multimodal Deep Learning[J]. 数据分析与知识发现, 2023, 7(5): 21-32.
[14] Huang Xuejian, Ma Tinghuai, Wang Gensheng. Detecting Weibo Rumors Based on Hierarchical Semantic Feature Learning Model[J]. 数据分析与知识发现, 2023, 7(5): 81-91.
[15] Wang Yinqiu, Yu Wei, Chen Junpeng. Automatic Question-Answering in Chinese Medical Q & A Community with Knowledge Graph[J]. 数据分析与知识发现, 2023, 7(3): 97-109.
  Copyright © 2016 Data Analysis and Knowledge Discovery   Tel/Fax:(010)82626611-6626,82624938   E-mail:jishu@mail.las.ac.cn