Home Browse Top access

Top access

  • Published in last 1 year
  • In last 2 years
  • In last 3 years
  • All

Please wait a minute...
  • Select all
    |
  • Wang Ruojia, Fan Keming, Liu Zhifeng, Wang Jimin
    Data Analysis and Knowledge Discovery. 2024, 8(8-9): 20-30. https://doi.org/10.11925/infotech.2096-3467.2023.1145
    Abstract (848) PDF (255) HTML (484)   Knowledge map   Save

    [Objective] The study investigates the characteristics of user querying behavior in a generative artificial intelligence (GAI) environment, and evaluates the suitability and effectiveness of GAI technology in search engines. [Methods] Behavioral data were collected through user experiments and questionnaires. Statistical analyses, including the non-parametric Wilcoxon test and the chi-square test, were performed to compare the variations in user search behavior patterns across different search engine environments. [Results] Compared to traditional search engines, the GAI-based search engine shows an average increase of 5.61 characters in query length, an extension of 8.92 seconds in query construction time, and an increase of 1.25 words in task descriptions. In particular, the use of translation and system-following strategies rose to 29.30% and 12.11%, respectively. In addition, users’ subjective satisfaction scores rose by 0.88 points. [Limitations] The study did not examine broader user search behaviors, such as browsing and utilization of search results. [Conclusions] While GAI technology can enhance search engines and improve users’ search experience, it also poses challenges related to high cognitive load, low credibility, and complex interactions.

  • Yu Bengong, Xing Yu, Zhang Shuwen
    Data Analysis and Knowledge Discovery. 2024, 8(11): 22-32. https://doi.org/10.11925/infotech.2096-3467.2023.0746
    Abstract (822) PDF (195) HTML (416)   Knowledge map   Save

    [Objective] To fully extract features from multiple modalities, align and integrate multimodal features, and design downstream tasks, we propose an aspect-based sentiment analysis model of multimodal collaborative contrastive learning (MCCL-ABSA). [Methods] Firstly, on the text side, we utilized the similarity between aspect words and their encoding within sentences. On the image side, the model used the similarity of images encoded in different sequences after random cropping to construct positive and negative samples required for contrastive learning. Secondly, we designed the loss function for contrastive learning tasks to learn more distinguishable feature representation. Finally, we fully integrated text and image features for multimodal aspect-based sentiment analysis while dynamically fine-tuning the encoder by combining contrastive learning tasks. [Results] On the TWITTER-2015 dataset, our model’s accuracy and F1 scores improved by 0.82% and 2.56%, respectively, compared to the baseline model. On the TWITTER-2017 dataset, the highest accuracy and F1 scores were 0.82% and 0.25% higher than the baseline model. [Limitations] We need to examine the model’s generalization on other datasets. [Conclusions] The MCCL-ABSA model effectively improves feature extraction quality, achieves feature integration with a simple and efficient downstream structure, and enhances the efficacy of multimodal sentiment classification.

  • Wang Zhenyu, Zhu Xuefang, Yang Rui
    Data Analysis and Knowledge Discovery. 2025, 9(1): 90-99. https://doi.org/10.11925/infotech.2096-3467.2023.1273
    Abstract (804) PDF (179) HTML (256)   Knowledge map   Save

    [Objective] This paper utilizes large language models (LLMs) to generate high-quality auxiliary knowledge, aiming to improve the performance of multimodal relation extraction. [Methods] We introduced a multimodal similarity detection module to construct multimodal prompt templates, which allow the LLM to integrate visual information and prior knowledge into the generated high-quality auxiliary knowledge. We combined the obtained auxiliary knowledge with the original text and input it into downstream text models to accurately predict entity relationships. [Results] The proposed model outperformed the best-baseline model on the MNRE dataset, achieving 4.09% and 7.84% improvements in accuracy and F1 score. [Limitations] We only examined the proposed model on English datasets. [Conclusions] Comparative experiments and case studies validate the model’s effectiveness in multimodal relation extraction. Our new model provides a direction for applying LLMs to multimodal information extraction tasks in the future.

  • Gao Guangshang
    Data Analysis and Knowledge Discovery. 2024, 8(8-9): 6-19. https://doi.org/10.11925/infotech.2096-3467.2023.0691
    Abstract (796) PDF (272) HTML (443)   Knowledge map   Save

    [Objective] This study explores the explainability mechanisms in explainable recommendation models from the perspectives of embedding and post-processing. [Coverage] Literature searches were conducted in Google Scholar and CNKI using the keywords “explainable recommendation”, “interpretable recommendation” and “explainable AI”. After topic filtering, a total of 64 representative papers on explainability methods were selected and reviewed using the backward snowballing technique. [Methods] From the embedding perspective, the explainability methods for recommendations were studied by analyzing four aspects: knowledge graphs, deep learning, attention mechanisms, and multi-task learning. From the post-processing perspective, the explainability methods were explored by analyzing five aspects: predefined templates, sentence retrieval, natural language generation, reinforcement learning, and knowledge graphs. The explainability methods were compared in detail in terms of their logical reasoning, performance characteristics, and limitations. Finally, the study provided an outlook on the pressing issues that need to be addressed in explainability research. [Results] Explainability can effectively enhance the persuasiveness of recommendation systems, improve the user experience, and is a crucial approach to increase the transparency and trustworthiness of recommendation systems. [Limitations] The study did not address the evaluation metrics for explainability algorithms. [Conclusions] Although existing explainability methods can satisfy the explanation requirements of many applications to a certain extent, there are still numerous challenges in research areas such as conversational interactive explanations and causal explanations.

  • Wang Kuifang, Lyu Lucheng, Sun Wenjun, Wang Yihu, Zhao Yajuan
    Data Analysis and Knowledge Discovery. 2024, 8(8-9): 144-156. https://doi.org/10.11925/infotech.2096-3467.2023.1203
    Abstract (626) PDF (215) HTML (315)   Knowledge map   Save

    [Objective] This paper aims to improve the accuracy of automatic extraction of technical words and function effects of patents. [Methods] First, ChatGPT is used as the Teacher-model, and ChatGLM3 is used as the Student-model. Through knowledge distillation, the training data extracted by ChatGPT are used to fine-tune ChatGLM3, resulting in multiple technical word extraction models and a function word extraction model. These models are performed to extract technical words and function words from the abstract, the first claim, and the technical effect segments of patents, respectively. [Results] Compared to ChatGPT, the fine-tuned technical word extraction models and the function word extraction model show higher accuracy and lower recall rates. The ChatGLM3 fine-tuning model of the first claim has the highest accuracy of 0.734 and F1 values of 0.724, respectively. The accuracy of the function word extraction model reached 0.649, which was higher than the accuracy of the commercial tool’s 0.530. [Limitations] This study needs to be further optimized in the following aspects. The technical field and patent language are single, the amount of verification data is small, and the data cleaning rules are not comprehensive enough. [Conclusions] This research scheme improves the accuracy of large language models in automatically extracting technical effects through knowledge distillation operation. Additionally, this study supports mining cutting-edge innovative and hotspot technologies from patents, facilitating higher quality intelligent patent analysis.

  • Zhang Jing, Gao Zixin, Ding Weijie
    Data Analysis and Knowledge Discovery. 2025, 9(2): 48-58. https://doi.org/10.11925/infotech.2096-3467.2023.1347
    Abstract (554) PDF (105) HTML (105)   Knowledge map   Save

    [Objective] This paper proposes a new model to effectively classify massive police reports. [Methods] We constructed a text classification model based on BERT-DPCNN. Then, we used the BERT pre-trained model to generate word vectors. The model improved the classification performance by optimizing the activation function in the DPCNN model and enhancing the dynamic learning rate. [Results] We conducted comparative experiments between BERT-DPCNN and six other models, including BERT, BERT-CNN, BERT-RCNN, BERT-RNN, BERT-LSTM, and ERNIE. The BERT-DPCNN achieved the best accuracy, recall, and precision. In the binary classification tasks, the accuracy of BERT-DPCNN exceeded 98%. In the eleven-category tasks, the model’s accuracy exceeded 82%. [Limitations] The model has many parameters, and the limited number of experiments calls for further testing. [Conclusions] The new model effectively improves the accuracy of police report classification, providing data support for police departments in analyzing and assessing police incidents.

  • Qin Jian
    Data Analysis and Knowledge Discovery. 2024, 8(8-9): 1-5. https://doi.org/10.11925/infotech.2096-3467.2024.0711
    Abstract (540) PDF (273) HTML (424)   Knowledge map   Save

    The explosive growth of Artificial intelligence (AI) has both positive and negative impacts on many aspects of human society and environment. While academia and industry are optimistic about the positive impacts on enhancing education, environment sustainability, healthcare systems and quality, and transportation of people and goods, experts are concerned about the potential harm and danger resulting from AI’s negative impacts if they are not contained. This essay discusses why it is imperative to emphasize the trustworthiness of AI and current developments in assuring trustworthy AI. The convenience and efficiency brought about by AI applications are embraced by both experts and general public, but how to contain the potential negative impacts and possible harms and dangers that come from ill- and/or even evil-purposed actors is a tremendous, complex challenge. Establishing trustworthy AI is considered as a major approach to fighting and containing the negative impacts of AI. Efforts in trustworthy AI include two broad areas: policies and regulations from governments and research and development (R&D) from academia and industry sectors. The policies and regulations focus on the ethical, legal, and robustness principles to provide guidance for R&D in trustworthy AI. In research publications, a commonly shared view is that trustworthy AI should have properties of reliability, safety, security, privacy, availability, and usability. For different population groups, the requirements for trustworthy AI may vary. One important development in trustworthy AI is the shift from model-centric AI to data-centric AI. The paradigm of data-centric AI emphasizes data quality through systematic design of datasets used for machine learning modeling, which include data design, data sculpting, and data strategies with data policy throughout the whole data design, sculpting, and strategy process. Both policy and technical developments in shaping trustworthy AI and containing the negative impact of AI present many new research and development opportunities for academia and industry.

  • Rang Yuchen, Ma Jing
    Data Analysis and Knowledge Discovery. 2025, 9(1): 100-109. https://doi.org/10.11925/infotech.2096-3467.2023.1130
    Abstract (371) PDF (102) HTML (211)   Knowledge map   Save

    [Objective] To reduce inter-modal differences and strengthen the correlation between modalities, this paper proposes a multimodal alignment sentiment analysis model to accurately capture the sentiment tendencies embedded in multimodal data. [Methods] For the textual modality, the original text data, supplemented with image captions, is processed using the RoBERTa pre-trained model for text feature extraction. We used the Clip Vision Model to extract image features for the image modality. The text and image features are aligned through a multimodal alignment layer based on a Multimodal Transformer to obtain enhanced fused features. Finally, the fused multimodal features are inputted into a multilayer perception for sentiment recognition and classification. [Results] The proposed model achieved an accuracy of 71.78% and an F1 score of 68.97% on the MVSA-Multiple dataset, representing improvements of 1.78% and 0.07%, respectively, over the best-performing baseline model. [Limitations] The model’s performance was not validated using additional datasets. [Conclusions] The proposed model effectively promotes inter-modal fusion, achieves better fusion representations, and enhances sentiment analysis.

  • Song Donghuan, Hu Maodi, Ding Jielan, Qu Zihao, Chang Zhijun, Qian Li
    Data Analysis and Knowledge Discovery. 2025, 9(2): 12-25. https://doi.org/10.11925/infotech.2096-3467.2023.0885
    Abstract (356) PDF (168) HTML (161)   Knowledge map   Save

    [Objective] This study addresses the issue of low classification accuracy in conventional text classification tasks due to factors such as sparse domain-specific training data and significant differences between types. [Methods] We constructed a novel classification model based on the BERT-DPCNN-MMOE framework, integrating the deep pyramid convolutional networks with the multi-gate control unit mechanism. Then, we designed multi-task and transfer learning experiments to validate the effectiveness of the new model against eight well-established and innovative models. [Results] This research independently constructed cross-type multi-task data as the basis for training and testing. The BERT-DPCNN-MMOE model outperformed the other eight baseline models in multi-task and transfer learning experiments, with F1 score improvements exceeding 4.7%. [Limitations] Further research is needed to explore the model’s adaptability to other domains. [Conclusions] The BERT-DPCNN-MMOE model performs better in multi-task and cross-type text classification tasks. It is of significance for future specialized intelligence classification tasks.

  • Li Jiawei, Zhang Shunxiang, Li Shuyu, Duan Wenjie, Wang Yuqing, Deng Jinke
    Data Analysis and Knowledge Discovery. 2024, 8(11): 1-10. https://doi.org/10.11925/infotech.2096-3467.2023.1005
    Abstract (350) PDF (188) HTML (246)   Knowledge map   Save

    [Objective] This paper proposes a Chinese implicit sentiment analysis model based on text graph representation. It fully utilizes external knowledge and context to enhance implicit sentiment text and achieve word-level semantic interaction. [Methods] First, we modeled the target sentence and context as a text graph with words as nodes. Then, we obtained the semantic expansion of the word nodes in the graph through external knowledge linking. Finally, we used the Graph Attention Network to transfer semantic information between the nodes of this text graph. We also obtained the text graph representation through the Readout function. [Results] We evaluated the model on the publicly available implicit sentiment analysis dataset SMP2019-ECISA. Its F1 score reached 78.8%, at least 1.2% higher than the existing model. [Limitations] The size of the generated text graph is related to the length of the text, leading to significant memory and computational overhead for processing long text. [Conclusions] The proposed model uses graph structure to model the relationship between external knowledge, context, and the target sentence at the word level. It effectively represents text semantics and enhances the accuracy of implicit sentiment analysis.

  • Zhao Jiayi, Xu Yuemei, Gu Hanwen
    Data Analysis and Knowledge Discovery. 2024, 8(10): 44-53. https://doi.org/10.11925/infotech.2096-3467.2023.0714
    Abstract (337) PDF (115) HTML (127)   Knowledge map   Save

    [Objective] This study addresses the performance degradation due to catastrophic forgetting when multilingual models handle tasks in new languages. [Methods] We proposed a multilingual sentiment analysis model, mLMs-EWC, based on continual learning. The model incorporates continual learning into multilingual models, enabling it to learn new language features while retaining the linguistic characteristics of previously learned languages. [Results] In continual sentiment analysis experiments involving three languages, the mLMs-EWC model outperformed the Multi-BERT model by approximately 5.0% in French and 4.5% in English tasks. Additionally, the mLMs-EWC model was evaluated on a lightweight distilled model, showing an improvement of up to 24.7% in English tasks. [Limitations] This study focuses on three widely used languages, and further validation is needed to assess the model’s generalization capability to other languages. [Conclusions] The proposed model can alleviate catastrophic forgetting in multilingual sentiment analysis tasks and achieve continual learning on multilingual datasets.

  • Wang Zitong, Li Chenliang
    Data Analysis and Knowledge Discovery. 2025, 9(2): 94-105. https://doi.org/10.11925/infotech.2096-3467.2023.1305
    Abstract (331) PDF (229) HTML (77)   Knowledge map   Save

    [Objective] To more flexibly capture the spatial-temporal features of traffic flow data and achieve more accurate multivariate traffic flow prediction, this paper proposes a Position-Aware Spatial-Temporal Graph Convolutional Network (PASTGCN). [Methods] First, the traffic data’s spatial and periodic temporal position features are represented as explicit position embeddings. Then, based on the spatiotemporal convolutional structure, we incorporated spatial information into the temporal convolutional network for space-aware sequence modeling. Finally, we used static and dynamic dual graph learning methods to capture spatial dependencies. [Results] We conducted experiments on two real-world traffic flow datasets. The PASTGCN model effectively predicted multivariate traffic flows and reduced errors by up to 1.59% compared to existing deep learning models. [Limitations] The experimental datasets are limited, and the proposed graph learning method increased the time complexity. [Conclusions] The PASTGCN model can effectively utilize spatial-temporal position information to achieve more accurate traffic flow prediction.

  • Sun Wenju, Li Qingyong, Zhang Jing, Wang Danyu, Wang Wen, Geng Yangli’ao
    Data Analysis and Knowledge Discovery. 2025, 9(1): 1-30. https://doi.org/10.11925/infotech.2096-3467.2024.0508
    Abstract (329) PDF (158) HTML (277)   Knowledge map   Save

    [Objective] This study comprehensively reviews the advancements in deep incremental learning techniques from the perspective of addressing catastrophic forgetting, aiming to provide references for the research community. [Coverage] Utilizing search terms such as “Incremental Learning”, “Continual Learning”, and “Catastrophic Forgetting”, we retrieved literature from the Web of Science, Google Scholar, DBLP, and CKNI. By reading and organizing the retrieved literature, a total of 105 representative publications were selected. [Methods] The paper begins by defining incremental learning and outlining its problem formulation and inherent challenges. Subsequently, we categorize incremental learning methods into regularization-based, memory-based, and dynamic architecture-based approaches, and review their theoretical underpinnings, advantages and disadvantages in detail. [Results] We evaluated some classical and recent methods in a unified experimental setting. The experimental results demonstrate that regularization-based methods are efficient in application but cannot fully avoid forgetting; memory-based methods are significantly affected by the number of retained exemplars; and dynamic architecture-based methods effectively prevent forgetting but incur additional computational costs. [Limitations] The scope of this review is limited to deep learning approaches, excluding traditional machine learning techniques. [Conclusions] Under optimal conditions, memory-based and dynamic architecture-based strategies tend to outperform regularization-based approaches. However, the increased complexity of these methods may hinder their practical application. Furthermore, current incremental learning methods show suboptimal performance compared to joint training models, marking a critical direction for future research.

  • Li Xiyu, Qian Li, Zhang Zhixiong
    Data Analysis and Knowledge Discovery. 2024, 8(8-9): 200-212. https://doi.org/10.11925/infotech.2096-3467.2023.1148
    Abstract (322) PDF (124) HTML (158)   Knowledge map   Save

    [Objective] This research aims to achieve the automatic quantification of semantic evaluation metrics for scientific papers using large language models, supporting the study of semantic evaluation of scientific literature. [Methods] First, we extracted rhetorical moves related to evaluation metrics from scientific papers with three levels of prompt detail—standard, simplified, and detailed. Then, we compared the effectiveness of these prompts. Third, we fine-tuned a large language model with a small number of annotated samples to develop a model for quantifying semantic evaluation metrics. [Results] Based on the semantic content of the papers, we analyzed the “difficulty of experimental conditions” dimension. The proposed model achieved the best performance. With a training sample size of 100, its Micro-Acc and Fuzzy-Acc reached 0.72 and 0.87, respectively. [Limitations] The experiment only included scientific papers in computer science, and we need to explore the effectiveness of the proposed method across different disciplines was not explored. [Conclusions] The proposed method demonstrates high accuracy and reliability in evaluating scientific papers. Increasing the level of detail in prompts significantly improves the quantification effect. While increasing the number of samples during the fine-tuning stage improves overall performance, the degree of improvement varies across different scoring ranges.

  • Shang Jinling, Zhang Jianyong
    Data Analysis and Knowledge Discovery. 2024, 8(8-9): 122-132. https://doi.org/10.11925/infotech.2096-3467.2023.1212
    Abstract (319) PDF (221) HTML (139)   Knowledge map   Save

    [Objective] This paper addresses the scarcity of query-focused text summarization datasets and explores methods to meet the personalized query needs of researchers. [Methods] Based on ChatGPT and prompt engineering, this paper constructed a generation and self-verification prompt chain. It proposed an automated data annotation framework using large language models like ChatGPT as “data annotators”. Then, we constructed the AMTQFSum dataset, consisting of query-focused summaries of academic conference records in natural language processing. [Results] AMTQFSum demonstrates superior data volume and length distribution. Using the UniEval model, AMTQFSum outperformed existing QFS datasets with an average score improvement of 85% and 33%. We examined the benchmark effectiveness of the AMTQFSum dataset on six classic extractive and abstractive query-focused summarization models. The BART-based model achieved the best results, with ROUGE-1/2/L scores reaching 52.53%, 35.61%, and 44.80%, respectively. [Limitations] The dataset only cover a narrow range of fields. [Conclusions] The large language model data annotation method based on prompt chains provides a feasible solution for automated data annotation. The AMTQFSum dataset provides a foundational resource for for query-focused summarization tasks.

  • Wang Hao, Li Xiaomin, Bu Wenru, Zhao Zibo, Deng Sanhong
    Data Analysis and Knowledge Discovery. 2024, 8(8-9): 179-190. https://doi.org/10.11925/infotech.2096-3467.2023.1194
    Abstract (312) PDF (137) HTML (159)   Knowledge map   Save

    [Objective] In response to the scarcity of annotated data in the current research on entity relationship extraction and graph construction in the field of intangible cultural heritage (ICH), a lightly annotated relationship extraction scheme is proposed. [Methods] Using silk weaving domain texts as the data source, the SREP model is constructed, integrating domain-specific terminology dictionaries and LTP tools for entity recognition. Subsequently, the BERT model is utilized to vectorize the representation of entities and their contextual text, and various clustering algorithms are applied to different feature combinations for relationship extraction experiments to determine the optimal algorithm and feature combination. The Bootstrapping method is then employed for active learning to expand the instances of relationships. Finally, the extracted relationship triples are imported into Gephi to construct a domain-specific knowledge graph. [Results] The experimental results indicate that the K-means algorithm, combining entity intermediate text features with entity type features, achieved the best results in relationship extraction experiments, identifying five types of relationships. During the relationship instance expansion phase, the LR algorithm is more suitable for active learning methods, with an accuracy rate of 0.860, an improvement of 0.105 over the baseline. [Limitations] The effectiveness of the model needs further verification on larger datasets and relationship extraction in different fields. [Conclusions] The model proposed in this study can effectively extract entity relationships from ICH texts and achieve semantic mining and utilization of structured ICH texts, reducing the dependence on annotated data.

  • Li Ying, Li Ming
    Data Analysis and Knowledge Discovery. 2024, 8(10): 89-99. https://doi.org/10.11925/infotech.2096-3467.2023.0683
    Abstract (290) PDF (106) HTML (185)   Knowledge map   Save

    [Objective] This paper proposes a recommendation method for supplementary question-and-answer (Q&A) based on a multi-label, multi-document Q&A classification model enhanced by transfer learning. It aims to identify and recommend supplementary answers in online Q&A communities. [Methods] We introduced new features alongside existing ones to classify the supplementary relationships between questions and answers. Then, we established a transfer learning-enhanced multi-label, multi-document classification model to identify and recommend supplementary answers. [Results] We conducted three meta-tasks on real datasets from the Zhihu community. The proposed method improves precision, recall, and F1 score by 48.29%, 15.75%, and 32.53%, respectively, on average. [Limitations] The method was only applied to health-related Q&A topics in Zhihu and has yet to be validated across different platforms or topics. [Conclusions] The proposed recommendation method effectively recommends supplementary answers. It helps users in Q&A communities obtain more comprehensive answers and promote knowledge utilization within the community.

  • Li Hui, Pang Jingwei
    Data Analysis and Knowledge Discovery. 2024, 8(11): 11-21. https://doi.org/10.11925/infotech.2096-3467.2023.0744
    Abstract (289) PDF (158) HTML (165)   Knowledge map   Save

    [Objective] To effectively utilize information containing audio and video and fully capture the multi-modal interaction among text, image, and audio, this study proposes a multi-modal sentiment analysis model for online users (TIsA) incorporating text, image, and STFT-CNN audio feature extraction. [Methods] First, we separated the video data into audio and image data. Then, we used BERT and BiLSTM to obtain text feature representations and applied STFT to convert audio time-domain signals to the frequency domain. We also utilized CNN to extract audio and image features. Finally, we fused the features from the three modalities. [Results] We conducted empirical research using the “9.5 Luding Earthquake” public sentiment data from Sina Weibo. The proposed TIsA model achieved an accuracy, macro-averaged recall, and macro-averaged F1 score of 96.10%, 96.20%, and 96.10%, respectively, outperforming related baseline models. [Limitations] We should have explored the more profound effects of different fusion strategies on sentiment recognition results. [Conclusions] The proposed TIsA model demonstrates high accuracy in processing audio-containing videos, effectively supporting online public opinion analysis.

  • Zhou Zhigang, Dou Luyao, Li Yi, Bai Zengliang
    Data Analysis and Knowledge Discovery. 2024, 8(12): 52-61. https://doi.org/10.11925/infotech.2096-3467.2023.0883
    Abstract (289) PDF (95) HTML (117)   Knowledge map   Save

    [Objective] This paper identifies potential high-value patents by deeply mining the feature information embedded in patent texts based on bilateral semantics and text sequence features. [Methods] First, we constructed a mixed patent dataset from the fields of amorphous alloys, industrial robots, and gene chips. Then, we employed the BERT word vector model to achieve contextual semantic association and word meaning interpretation of patent texts. Third, we utilized the BiGRU network to extract global text sequence information while CNN captured local text sequence information. Finally, we predicted potential high-value patents by combining “bilateral semantics+global+local” semantic and sequence features. [Results] The proposed BERT-BiGRU-CNN model outperforms existing models and is more suitable for predicting potential high-value patents on a large data scale. Our new model achieves a prediction accuracy of over 35%, about 4% higher than the existing ones. [Limitations] The relationship and integration mechanism between standard essential and high-value patents have yet to be considered, and the algorithm complexity needs further optimization. [Conclusions] The BERT-BiGRU-CNN model performs better in text classification tasks than the CNN model. Our new model improves the prediction accuracy of potentially high-value patents by capturing global and local text sequence features.

  • Cheng Quan, Jiang Shihui, Li Zhuozhuo
    Data Analysis and Knowledge Discovery. 2024, 8(10): 112-124. https://doi.org/10.11925/infotech.2096-3467.2023.0638
    Abstract (280) PDF (89) HTML (117)   Knowledge map   Save

    [Objective] This paper aims to achieve semantic discovery and relation extraction from a large amount of complex user-generated information from an online healthcare platform. [Methods] First, we constructed a semantic discovery model for online health information based on an improved CasRel model. Then, we introduced the ERNIE-Health pre-trained model, which is more suitable for the healthcare domain, into the text encoding layer of the CasRel-based model. Finally, we used a multi-level pointer network in the entity and relation decoding layer to annotate and fuse subject features for relations and object decoding via neural networks. [Results] Compared to the original model, the improved CasRel entity-relation extraction model increased the F1-scores of entity recognition and entity-relation extraction tasks for online health information semantic discovery by 7.62% and 4.87%, respectively. [Limitations] The overall effectiveness of the model still needs to be validated with larger datasets and empirical studies on health information from different disease types. [Conclusions] Three sets of comparative experiments validated the effectiveness of the improved CasRel entity-relation extraction model for online diabetes health information semantic discovery tasks.

  • Zhuang Zhihuang, Xu Xing, Xia Xuewen, Zhang Yinglong, Zhou Xinyu
    Data Analysis and Knowledge Discovery. 2024, 8(8-9): 261-270. https://doi.org/10.11925/infotech.2096-3467.2023.1258
    Abstract (262) PDF (141) HTML (104)   Knowledge map   Save

    [Objective] A neural network model is used to solve the problem of ceramic ware types classification with few samples, and the performance of the model for ceramic ware types classification is improved by using multiscale and attention mechanism optimization. [Methods] A bottleneck structure based on coordinate attention mechanism and multiscale fusion is proposed and applied to the residual network, which innovatively introduces the relationship between scales and ultimately improves the modeling ability of the residual networks in terms of multiscale. [Results] On the public dataset of ceramic ware types images, this model achieves a classification accuracy of 95.71% with only a few samples learning, representing an improvement of 1.01 percentage points over the baseline model ResNet50. In terms of precision, recall, and F1 score metrics, the proposed model outperforms ResNeSt50 by 20.43, 20.53, and 20.52 percentage points, respectively. [Limitations] Although the model’s recognition accuracy and other metrics have increased, the efficiency of inference has decreased, and it would not be suitable for scenarios where rapid ceramic ware classification is required. [Conclusions] The multiscale improvement approach is simple and effective in ceramic ware types classification, and this optimization strategy should be prioritized when performing this type of task or similar humanity data.

  • Shi Xi, Chen Wenjie, Hu Zhengyin, Han Tao, Zhang Kai
    Data Analysis and Knowledge Discovery. 2025, 9(3): 1-15. https://doi.org/10.11925/infotech.2096-3467.2024.0176

    [Objective] This study aims to efficiently extract scientific experiment knowledge and data from academic literature. It constructs a Scientific Experiment Knowledge Graph(SEKG) to provide high-quality data support for knowledge discovery. [Methods] We utilized Event Knowledge Graph technology to uniformly represent and model the complexity, temporality, and integration of knowledge and data in scientific experiments, thereby establishing the schema layer of the SEKG. Large Language Model was employed to enhance the efficiency of knowledge extraction in the data layer, with an empirical analysis conducted on organic solar cells. [Results] By using manual annotation and fine-tuning large language models, we constructed a scientific experiment knowledge graph in the field of organic solar cells. This SEKG comprises 34 types of nodes and 9 types of relationships, totaling 24,348 nodes and 123,642 relations. [Limitations] The data sources were limited to papers and patents. The construction of the SEKG required substantial manual input from experts, highlighting the need for efficiency improvements. Furthermore, fine-grained research procedures and validation rules in subfields were not considered. [Conclusions] The proposed method provides high-quality data support for applications such as experimental protocol recommendations, scientific experiment evolution analysis, and AI for Science, effectively supporting various knowledge discovery scenarios.

  • Jin Qingwen, Li Hurong, Zhang Chen
    Data Analysis and Knowledge Discovery. 2024, 8(12): 101-111. https://doi.org/10.11925/infotech.2096-3467.2023.0892
    Abstract (250) PDF (72) HTML (93)   Knowledge map   Save

    [Objective] This study explores the application of the LIME algorithm and its evolutions in data storytelling, aiming to leverage the explanatory function of data stories. [Methods] We examined the principles, applications, and evolutionary strategies of the LIME algorithm. Based on this theoretical framework, we constructed a data storytelling process assisted by LIME-related algorithms. We collected a partial dataset for cat and dog recognition from the Kaggle platform, and trained an interpretable model with this dataset. Finally, we applied the new data storytelling model to explain image classification performance. [Results] Using an image of a “tabby cat” as the analysis object, the LIME explanation results and storytelling development curve indicated that the important features affecting the prediction results were the M-shaped stripes, black eyes, and pink nose, and the number of key superpixels being 2. [Limitations] Optimization of feature recognition and automated generation of data stories remain challenges. [Conclusions] Applying LIME-related algorithms in the data storytelling helps transform model predictions and explanation results into interpretable stories, better communicating data analysis outcomes.

  • Zhao Jianfei, Chen Ting, Wang Xiaomei, Feng Chong
    Data Analysis and Knowledge Discovery. 2024, 8(8-9): 133-143. https://doi.org/10.11925/infotech.2096-3467.2023.1246
    Abstract (247) PDF (143) HTML (175)   Knowledge map   Save

    [Objective] This paper aims to automate extracting key technical information from complex patent texts and to overcome the dependency on robust domain knowledge annotations in traditional natural language processing models. [Methods] We proposed an unsupervised key information extraction method based on knowledge self-distillation in the large language model. By employing a multiple-role strategy, we conducted a structured analysis of Derwent’s rewritten patent abstracts. This method enhanced the ability of large language models to extract and structurally analyze key content through the knowledge self-distillation strategy. [Results] In the entity and relation extraction tasks, our method’s recall rate reached 95.40% and 51.49%, respectively. The accuracy of the structural analysis format reached 100%. We also achieved an F1-score of 5.01% on the RE-DocRED dataset, a public dataset for the relation triplet extraction task, under unsupervised and zero-shot settings. [Conclusions] The proposed method can effectively extract key information from patent texts without data annotation.

  • Wang Lixiao, Chen Wei, Qiu Hanqi
    Data Analysis and Knowledge Discovery. 2024, 8(8-9): 63-75. https://doi.org/10.11925/infotech.2096-3467.2023.1250
    Abstract (247) PDF (142) HTML (162)   Knowledge map   Save

    [Objective] This paper constructs a weak signal identification model for disruptive technologies based on machine learning. It aims to discover early-stage disruptive technologies and explore their disruptive potential to existing mainstream technologies. [Methods] By summarizing the core characteristics of disruptive technology’s weak signals, we designed a Disruptive Index-Patent (DI-P) based on the patent citation categories. We also constructed historical disruptive technology corpora and designed a weak signal identification model for disruptive technologies based on machine learning. Machine learning models such as Logistic Regression, Gaussian Naive Bayes, Stochastic Gradient Descent, Gradient Boosting Trees, and Random Forests were selected for comprehensive prediction. Finally, we explored the future disruptive paths of technology’s weak signals through link prediction. [Results] We conducted an empirical analysis in hydrogen storage and used the DI-P based on citation categories to obtain historical disruptive technical corpora. Its accuracy and AUC values were better than RDI and DI. By comparing the weak signals of disruptive technologies with high-value patents, we can identify potential future disruptive paths from the perspectives of cost, efficiency, and security. [Limitations] The empirical field is relatively single, data sources are limited to patent data and strategic planning, and the prediction model has limited accuracy. [Conclusions] By combining machine learning models with link prediction methods, we can identify signals of disruptive technologies and their disruption paths with precision and fine granularity.

  • Han Yixiao, Ma Jing
    Data Analysis and Knowledge Discovery. 2024, 8(12): 18-29. https://doi.org/10.11925/infotech.2096-3467.2023.0923
    Abstract (246) PDF (134) HTML (133)   Knowledge map   Save

    [Objective] In response to the challenges that current multimodal emotion models face in feature fusion, resulting in suboptimal accuracy in emotion classification, we propose the RCHFN multimodal emotion classification model. [Methods] We use the CLIP and Chinese-BERT-wwm models to extract image and text features separately while performing unimodal emotion classification concurrently. Then, we use a residual fusion module consisting of merged residual connections and convolution to fuse image and text features to obtain multimodal emotion classification results. Finally, we pass both unimodal and multimodal emotion classification results to a fully connected layer and adjust dynamic weights to obtain the final emotion classification result. [Results] The experimental results show that the RCHFN model achieved sentiment classification accuracies of 81.25% and 79.21% on the Weibo dataset and the Twitter datasets, respectively, with F1 scores of 80.43% and 78.44%, respectively. Compared to other models designed for similar tasks on the same dataset, the model showed an increase in accuracy of 1.79% and 1.79%, along with F1 score improvements of 2.39% and 2.62%, respectively. [Limitations] Further experiments are needed to establish the generalisation of this model to different datasets and its performance on additional modalities. [Conclusions] The RCHFN model proposed in this study effectively addresses the challenges of fusing multimodal discourse features and improving classification accuracy in emotion classification.

  • Teng Fei, Zhang Qi, Qu Jiansheng, Li Haiying, Liu Jiangfeng, Liu Boyu
    Data Analysis and Knowledge Discovery. 2024, 8(11): 33-46. https://doi.org/10.11925/infotech.2096-3467.2023.0767
    Abstract (243) PDF (136) HTML (104)   Knowledge map   Save

    [Objective] This study utilizes big data analytics to identify key and core technologies, improving the accuracy of identification results and providing robust data support for future technological innovation and large-scale applications. [Methods] We proposed a key and core technology identification method using the patent competitiveness index and Doc-LDA topic model based on the definitions of key and core technology concepts. The method distinguished topics by evaluating their strength, topic co-occurrence strength, and effective cohesion constraint coefficient. [Results] Taking new energy vehicles (EVs) as an empirical research example, a total of 10 key and core technologies were identified: fuel cells, solid-state power batteries, high-efficiency high-density motor drive system, lightweight plastic and composite materials, cellular communication, electro-mechatronics integration, multi-gear transmission, vehicle operations, intelligent control, and autonomous driving. Further trend analysis was conducted. [Limitations] Due to the limited granularity of topic refinement, some potential micro-mechanisms have not been fully revealed. [Conclusions] Using the patent competitiveness index and the Doc-LDA topic model provides a comprehensive assessment of the market value and competitive advantage of technologies. The proposed method also enhances the accuracy of technology development trend predictions.

  • Chang Bolin, Yuan Yiguo, Li Bin, Xu Zhixing, Feng Minxuan, Wang Dongbo
    Data Analysis and Knowledge Discovery. 2024, 8(11): 102-113. https://doi.org/10.11925/infotech.2096-3467.2023.0834
    Abstract (234) PDF (127) HTML (102)   Knowledge map   Save

    [Objective] This paper proposes an integrated model incorporating radical information to improve the low accuracy and efficiency of existing automatic word segmentation and part-of-speech tagging for Classical Chinese. [Methods] Based on over 70,000 Chinese characters and their radicals, we constructed a radical vector representation model, Radical2Vector. We combined this model with SikuRoBERTa for representing Classic Chinese texts, forming an integrated BiLSTM-CRF model as the main experimental framework. Additionally, we designed a dual-layer scheme for word segmentation and part-of-speech tagging. Finally, we conducted experiments on the Zuo Zhuan dataset. [Results] The model achieved an F1 score of 95.75% for the word segmentation task and 91.65% for the part-of-speech tagging task. These scores represent 8.71% and 13.88% improvements over the baseline model. [Limitations] The approach only incorporates a single radical for each character and does not utilize other components of the characters. [Conclusions] The proposed model successfully integrates radical information, effectively enhancing the performance of textual representation for Classical Chinese. This model demonstrates exceptional performance in word segmentation and part-of-speech tagging tasks.

  • Wu Shuai, Yang Xiuzhang, He Lin, Gong Zuoquan
    Data Analysis and Knowledge Discovery. 2024, 8(12): 136-148. https://doi.org/10.11925/infotech.2096-3467.2023.1002
    Abstract (232) PDF (80) HTML (88)   Knowledge map   Save

    [Objective] Combining the complex sentence structure features of ancient texts, a method with higher accuracy for identifying entity words in ancient texts was developed to further the development of digital humanities research. [Methods] Trigger words and relative words were used as key feature words to identify entity words, and a sentence pattern template was designed. Based on the characteristics of ancient texts, a Bert-BiLSTM-MHA-CRF model was constructed. The fusion of syntactic features and the Bert-BiLSTM-MHA-CRF model was used to achieve deep and fine-grained entity recognition of ancient texts. [Results] The F1 Score of this method is 0.88 on the conventional annotated test data set, 0.83 on the small sample annotated test data set, 0.79 (The Book of Songs), 0.81 (Master Lü’s Spring and Autumn Annals) and 0.85 (Discourses of the States) on the transfer learning test data set. [Limitations] In the design of syntactic feature templates, only single ancient books are used as feature templates. Semantic information mining does not take into account the structural features of characters such as phonetic symbols and radicals in ancient texts. [Conclusions] In small sample annotation and transfer learning experiments, this method can also achieve accurate named entity recognition of ancient texts, providing high quality corpus data for digital humanities research.

  • Zhu Yujing, Chen Fang, Wang Xuezhao
    Data Analysis and Knowledge Discovery. 2024, 8(10): 1-13. https://doi.org/10.11925/infotech.2096-3467.2023.0699
    Abstract (227) PDF (130) HTML (168)   Knowledge map   Save

    [Objective] In response to Western technology export controls on China, this study proposes a method for identifying critical core technologies by mapping the U.S. Commerce Control List (CCL) to a patent-based dual-layer network. The goal is to provide a reference for selecting and prioritizing technology breakthrough directions. [Methods] The study integrates the CCL and patent data to build a dual-layer network consisting of a CCL-related network and a weighted patent citation network. We used a community detection algorithm to identify technology clusters in both layers and calculated the semantic similarity of inter-layer clusters to achieve automatic mapping. Using Word2Vec and the n-gram method, we extracted keywords from each cluster to represent technical topics. Finally, we identified the patent clusters with the highest similarity to the CCL clusters as critical core technologies. [Results] Empirical results in industrial software demonstrate that this method identifies 12 distinct patent clusters with the highest similarity to the CCL clusters, all of which have a similarity of over 0.85. They involve integrated circuit IP cores, precision measurement, process control, motion control, and turbine detection. Literature research has verified them as key core technologies in industrial software. [Limitations] The study only focused on industrial software for empirical research. The technical approach can be improved, and the identification results require further interpretation and analysis. [Conclusions] The proposed method efficiently and accurately identifies key core technology at a micro-level, features a high degree of automation, and is highly readable, providing significant practical application value.

  • He Jun, Yu Jianjun, Rong Xiaohui
    Data Analysis and Knowledge Discovery. 2024, 8(10): 136-145. https://doi.org/10.11925/infotech.2096-3467.2023.0645
    Abstract (226) PDF (298) HTML (119)   Knowledge map   Save

    [Objective] This paper aims to ensure the objectivity, timeliness, and accuracy of the overall budget performance evaluation of research institutions, and to improve the efficiency of performance evaluation work. [Methods] We proposed a method for predicting research institutions’ overall budget performance evaluation based on LightGBM. Our method integrates various data from scientific research management information systems. It uses machine learning algorithms to analyze and predict the overall budget performance evaluation results by correlating research inputs and outputs with performance. [Results] In the application of the overall budget performance evaluation of research institutions, the accuracy of the proposed method reached 94.12%. The human resources required for the budget performance evaluation process were reduced from 10 people to 5, and the time cost was shortened from 38 days to about 10 days. [Limitations] Some performance evaluation indicators are subjective and difficult to quantify using business data from scientific research management information systems. [Conclusions] The proposed method has excellent performance in predicting overall budget performance evaluation results. It reduces the fairness issues due to subjective evaluation, and saves the human resources and time costs in budget performance evaluation, thus improving their efficiency.

  • Chen Ting, Ding Honghao, Zhou Haoyu, Wu Jiang
    Data Analysis and Knowledge Discovery. 2025, 9(2): 159-171. https://doi.org/10.11925/infotech.2096-3467.2023.1424
    Abstract (225) PDF (97) HTML (113)   Knowledge map   Save

    [Objective] This study explores the impacts of bullet-screen(danmu)content and behavioral characteristics on consumers purchasing behavior in live-streaming e-commerce, as well as the moderating effect of host-product relevance. [Methods] First, we retrieved the bullet-screen data from the Douyin platform and the consumer data from the Huitun platform based on the Elaboration Likelihood Model. Then, we studied the impacts of bullet-screen content characteristics (central route) and behavior characteristics (peripheral route) on consumer purchasing behavior with text mining and zero-inflated negative binomial regression. We also discussed the moderating effect of host-product relevance with grouping regression. [Results] Information richness, social interaction degree and number of bullet-screen comments positively impact purchasing behavior. The emotional polarity of bullet screen comments exhibits an inverted U-shaped effect on purchasing behavior. Compared with live streaming rooms with low host-product relevance, those with high host-product relevance have broader positive impacts on purchase behavior. [Limitations] We only investigated the bullet-screen data from a single live-streaming e-commerce platform. [Conclusions] This study examines the factors influencing consumers’ actual purchasing behavior from the perspective of bullet-screen comments. It provides insights for improving communication between merchants and consumers in live-streaming e-commerce, ultimately enhancing sales performance.

  • Yu Bengong, Cao Chengwei
    Data Analysis and Knowledge Discovery. 2024, 8(10): 54-65. https://doi.org/10.11925/infotech.2096-3467.2023.0722
    Abstract (221) PDF (104) HTML (77)   Knowledge map   Save

    [Objective] This paper aims to address the problem in current aspect-based sentiment analysis research, where the use of sentiment knowledge to enhance syntactic dependency graphs overlooks syntactic reachability and positional relationships between words and does not adequately extract semantic information. [Methods] We proposed an aspect-based sentiment analysis model based on a position-weighted reachability matrix and multi-space semantic information extraction. First, we used a reachability matrix to incorporate syntactic reachability relationships between words into the syntactic dependency graph, and we employed position-weighting to adjust the matrix to enhance contextual feature extraction. Then, we integrated the sentiment features with the enhanced dependency graph to extract aspect word features. Third, we use the multi-head self-attention mechanism combined with a graph convolutional network (GCN) to learn contextual semantic information from multiple feature spaces. Finally, we fused feature vectors containing positional information, syntactic information, affective knowledge, and semantic information for sentiment polarity classification. [Results] Compared to the best-performing models, the proposed model improved accuracy on the Lap14, Rest14, and Rest15 datasets by 1.00%, 1.25%, and 0.76%. When using BERT, the PRM-GCN- BERT model’s accuracy on the Lap14, Rest14, Rest15, and Rest16 datasets increased by 0.50%, 0.22%, 1.98%, and 0.31%. [Limitations] The proposed model was not applied to Chinese or other language datasets. [Conclusions] The proposed model enhances feature aggregation in graph convolutional networks, improves contextual feature extraction, and boosts semantic learning effectiveness, thereby significantly improving the accuracy of aspect-based sentiment analysis.

  • Wen Tingxin, Bai Yunhe
    Data Analysis and Knowledge Discovery. 2024, 8(12): 86-100. https://doi.org/10.11925/infotech.2096-3467.2023.0881
    Abstract (221) PDF (94) HTML (105)   Knowledge map   Save

    [Objective] This study proposes an interpretable model for the interaction quality of fake news groups based on RF-GA-XGBoost and SHAP. Our model mitigates the negative impacts of fake news by leveraging the interaction quality of social media user groups and accurately identifies the causes and mechanisms of positive interactions. [Methods] First, we retrieved 500 fake news articles and 7,029 comments from the Weibo21 dataset. Then, we assessed the fake news groups’ interaction quality across three dimensions: content, form, and comment sentiment. Third, we extracted fake news text features from these dimensions. Fourth, we used the sequential forward search strategy of random forest to extract the optimal feature subset of fake news text. We constructed a prediction model for group interaction quality based on GA-XGBoost, and compared its performance with other mainstream machine learning algorithms such as LR, SVM, and XGBoost. Finally, the SHAP model provides causal explanations for the impact of important features on the group interaction quality. [Results] Our model’s F1-score and AUC values are over 86%, outperforming the comparison models across six performance metrics. Additionally, features such as the number of content characters, words, and negative sentiment words in fake news text significantly influence the interaction quality of social media groups. [Limitations] This paper does not conduct multi-feature interaction interpretation analysis or explore the early high-quality group interaction patterns based on timestamps. [Conclusions] The proposed model accurately identifies the ways in which different features impact group interaction quality, providing effective decision-making support for social media platforms to improve their operational strategies and functional designs.

  • Zhu Xiang, Zhang Yunqiu, Sun Shaodan, Zhang Liman
    Data Analysis and Knowledge Discovery. 2024, 8(12): 125-135. https://doi.org/10.11925/infotech.2096-3467.2023.0869
    Abstract (220) PDF (109) HTML (108)   Knowledge map   Save

    [Objective] This paper proposes a drug knowledge discovery method that fuses meta-path features of heterogeneous knowledge network to improve the performance of drug knowledge discovery. [Methods] Based on different meta-paths connecting drug and target entity in heterogeneous knowledge network, the HeteSim algorithm is used to calculate the multi-dimensional semantic similarity of drug-target entity. These meta-path features are fused with drug similarity and target entity similarity features as feature inputs for machine learning models to achieve drug knowledge discovery. [Results] The drug heterogeneous knowledge network contains 12,015 nodes and 1,895,445 edges. Taking drug-target relation prediction as an example, the 21-dimensional HeteSim features between drug and target were calculated. The AUC value of this method achieved the highest value on the three machine learning models (XGBoost=0.993, RF=0.990, SVM=0.975). The accuracy, precision and F-value of this method are also higher than those of the other two comparison methods. Through literature search of 20 prediction results, it is found that some prediction results can be supported by evidence in previous literature. [Limitations] Although PU learning strategy is used to reduce the influence of sample imbalance, some results will still be distorted. [Conclusions] The drug knowledge discovery method proposed in this study has certain progressiveness and effectiveness, and has certain theoretical and methodological reference significance.

  • Xu Haoshuai, Hong Liang, Hou Wenjun
    Data Analysis and Knowledge Discovery. 2024, 8(10): 66-76. https://doi.org/10.11925/infotech.2096-3467.2023.0973
    Abstract (219) PDF (116) HTML (104)   Knowledge map   Save

    [Objective] This paper addresses the challenge of constructing label mapping in prompt learning-based relation extraction methods when labeled data is scarce. [Methods] The proposed approach enhances prompt effectiveness by injecting relational semantics into the prompt template. Data augmentation is performed through prompt ensemble, and an instance-level attention mechanism is used to extract important features during the prototype construction process. [Results] On the public FewRel dataset, the accuracy of the proposed method surpasses the baseline model by 2.13%, 0.55%, 1.40%, and 2.91% in four few-shot test scenarios, respectively. [Limitations] The method does not utilize learnable virtual prompt templates in constructing prompt templates, and there is still room for improvement in the representation of answer words. [Conclusions] The proposed method effectively mitigates the problem of limited information and insufficient accuracy in prototype construction under few-shot scenarios, improving the model’s accuracy in few-shot relation extraction tasks.

  • Shi Bin, Wang Hao, Liu Maolin, Deng Sanhong
    Data Analysis and Knowledge Discovery. 2024, 8(10): 146-158. https://doi.org/10.11925/infotech.2096-3467.2023.0688
    Abstract (219) PDF (110) HTML (135)   Knowledge map   Save

    [Objective] This study aims to construct a Chinese Ceramic Image Description Model (CCI-ClipCap) to provide technical support for ceramic culture research and digital preservation. [Methods] Based on ClipCap, the prompt paradigm is introduced to improve the model’s understanding of cross-modal data, enabling automatic description of ceramic images. Additionally, we proposed a text similarity evaluation method tailored for structured textual representation. [Results] The CCI-ClipCap model improved the multi-modal fusion process with the prompt paradigm, effectively extracting information from ceramic images and generating accurate textual descriptions. Compared to baseline models, the Bleu and Rouge values increased by 0.04 and 0.14, respectively. [Limitations] The data used originated from the British Museum collections, not native Chinese datasets. This single-source data may affect the model’s performance. [Conclusions] The CCI-ClipCap model generates text with rich levels of expression, demonstrating a soild understanding of ceramic knowledge and exhibiting high professionalism.

  • Du Jialin, Wang Xizi, Hu Guangwei
    Data Analysis and Knowledge Discovery. 2024, 8(11): 59-71. https://doi.org/10.11925/infotech.2096-3467.2023.0778
    Abstract (219) PDF (152) HTML (102)   Knowledge map   Save

    [Objective] This study investigates the factors influencing public satisfaction with government-citizen interaction platforms. We constructed an analysis model for factors affecting public satisfaction. [Methods] We extracted micro-level variables from the leadership mailbox corpus, which were combined with macroeconomic variables to establish a public satisfaction analysis model using the Gradient Boosting Decision Tree (GBDT) method. We also eliminated less influential variables with SHAP analysis to optimize the model. [Results] The proposed model outperformed comparison models across accuracy, recall, precision, and F1-score. Key features affecting public satisfaction with the leadership mailbox include GDP growth rate, PCDI growth rate, CPI growth rate, message topic, message type, and response mode. [Limitations] The study did not explore a broader range of influencing factors or more extensive government-citizen interaction scenarios. [Conclusions] The new model optimizes the variable selection process and visualizes how each feature influences the level, direction, and manner of public satisfaction with government responses. The model is a data-driven tool for administrative decision-making.

  • Zhou Shu, Wang Hao, Shi Guoliang, Shi Bin, Qiu Jingwen
    Data Analysis and Knowledge Discovery. 2024, 8(8-9): 240-250. https://doi.org/10.11925/infotech.2096-3467.2023.1117
    Abstract (218) PDF (101) HTML (95)   Knowledge map   Save

    [Objective] This paper constructs an interactive matching model to study the inaccurate responses, insufficient precision in returning multiple results, and polysemy problems in multi-turn dialogue in the intelligent question-answering system in the financial sector. [Methods] We proposed a multi-granularity and multi-attention interaction matching model (MGMAI) based on BERT. MGMAI included a preprocessing layer, a representation layer, an attention interaction layer, a semantic aggregation layer, and a dialog selection layer. This model focused on the key information in dialogues and utilized such information to achieve efficient dialog matching. [Results] The MGMAI model was applied to two open multi-turn dialogue datasets for training and validation. The model was fine-tuned on financial data. Experimental results showed that MGMAI outperformed the DCM model by 0.019, 0.010, and 0.007 on the R10@1, R10@2, and R10@5 metrics. [Limitations] This model was only tested in intelligent question-answering systems with financial data. We did not validate its generalization ability in other fields. [Conclusions] The MGMAI model can effectively improve the accuracy of multi-turn dialogues and handle ambiguous issues facing intelligent question-answering systems in the financial sector. It shows potential application value and room for improvement.

  • Gao Yuan, Li Chongyang, Qu Boting, Jiao Mengyun
    Data Analysis and Knowledge Discovery. 2025, 9(4): 158-169. https://doi.org/10.11925/infotech.2096-3467.2024.0784
    Abstract (218) PDF (50) HTML (59)   Knowledge map   Save

    [Objective] This paper aims to advance the research on urban tourism flow network structure, and to address the issues of inaccurate point-of-interest recognition and distorted visiting sequence in current tourist journey reconstruction methods based on travelogue texts. [Methods] This paper proposes a method based on a large language model for reconstructing tourist journeys, and explores the structural characteristics of urban tourism flow networks by combining it with social network analysis methods. [Results] The proposed method for reconstructing tourist journey achieves a precision of 94.00% and a recall of 87.78% in POI recognition, significantly outperforming the statistics-based Conditional Random Fields (CRF) method. The reconstructed journey shows a similarity of 83.81% to the actual journey. [Limitations] Tourist journey reconstruction effects depend to a certain extent on the training effects of the Prompts of the large language model. [Conclusions] The conclusions drawn align with public perception and current research findings when taking Xi’an as a case study, demonstrating the accuracy and versatility of the proposed tourist journey reconstruction method.