Hai Jiali, Wang Run, Yuan Liangzhi, Zhang Kairui, Deng Wenping, Xiao Yong, Zhou Tao, Chang Kai
[Objective] This paper constructs a retrieval-augmented question-answering (QA) system for Traditional Chinese Medicine (TCM) standards, aiming to provide efficient standard knowledge services and promote the research and application of TCM standardization. [Methods] By comparing the performance of large language models such as BaiChuan, Gemma, and Qwen, we chose GPT-3.5 as the base model. Then, we combined data optimization and retrieval-augmented generation to develop a QA system with semantic analysis, contextual association, and answer-generation capabilities. [Results] On a TCM literature-based question generation dataset, the new system achieved answer relevance precision, recall, and F1 scores of 0.879, 0.839 and 0.857, respectively, as well as contextual relevance scores of 0.838, 0.869, and 0.853. On a TCM standards QA dataset, the system achieved answer relevance scores of 0.871, 0.836 and 0.853, all outperforming baseline models. [Limitations] The system’s intent recognition accuracy still requires further improvement. The scale and granularity of the TCM standards knowledge base need to be expanded and refined. [Conclusions] In response to the practical needs of TCM knowledge services, this study developed a retrieval-augmented QA system for TCM standards. The system can effectively answer various questions related to clinical guidelines, herbal medicine standards, and information standards, covering topics such as treatment principles, syndrome classification, therapeutic methods, and technical specifications, demonstrating its strong practicality and feasibility.