[Objective] This paper aims to quickly generate real-time promotional book summaries and reduce the consumption of workforce and resources. [Methods] First, we constructed a dataset with the crawled book information based on prompt learning. Then, we used data enhancement and keyword extraction to increase information and generated the primary promotion language with the T5 PEGASUS. When the number of book reviews reaches the threshold, the summary of the book reviews will also be added. [Results] Compared with the optimal baseline model, the Rouge-1、Rouge-2、and Rouge-L scores of the proposed model were improved by 29.0%, 37.6%, and 31.9%, respectively. Adding the summary of book reviews can reflect the interests of users. [Conclusions] The proposed model could generate summaries based on the characteristics of the book corpus and has practical value.
李岱峰, 林凯欣, 李栩婷. 基于提示学习与T5 PEGASUS的图书宣传自动摘要生成器*[J]. 数据分析与知识发现, 2023, 7(3): 121-130.
Li Daifeng, Lin Kaixin, Li Xuting. Identifying Named Entities of Adverse Drug Reaction with Adversarial Transfer Learning. Data Analysis and Knowledge Discovery, 2023, 7(3): 121-130.
( Zhang Ling. Library Promotion Under New Media Environment Technologies[J]. Library and Information Service, 2015, 59(1): 78-81.)
doi: 10.13266/j.issn.0252-3116.2015.01.010
[2]
张琼. 加强图书网络营销的途径[J]. 科技与出版, 2012(6): 78-79.
[2]
Zhang Qiong. Ways to Strengthen the Network Marketing of Books[J]. Science-Technology & Publication, 2012(6): 78-79.)
[3]
Carbonell J, Goldstein J. The Use of MMR, Diversity-Based Reranking for Reordering Documents and Producing Summaries[C]// Proceedings of the 21st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval. ACM, 1998: 335-336.
( Wang Xuxiang, Han Bin, Gao Rui, et al. Automatic Extraction of Text Summarization Based on Improved TextRank[J]. Computer Applications and Software, 2021, 38(6): 155-160.)
[5]
Louis A. A Bayesian Method to Incorporate Background Knowledge During Automatic Text Summarization[C]// Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics. 2014: 333-338.
[6]
Yin W P, Pei Y L. Optimizing Sentence Modeling and Selection for Document Summarization[C]// Proceedings of the 24th International Conference on Artificial Intelligence. ACM, 2015: 1383-1389.
[7]
Singh A, Gupta M, Varma V. Unity in Diversity: Learning Distributed Heterogeneous Sentence Representation for Extractive Summarization[OL]. arXiv Preprint, arXiv: 1912.11688.
( Zhao Hong. A Survey of Deep Learning Methods for Abstractive Text Summarization[J]. Journal of the China Society for Scientific and Technical Information, 2020, 39(3): 330-344.)
( Ding Jianli, Li Yang, Wang Jialiang. Short Text Automatic Summarization Method Based on Dual Encoder[J]. Journal of Computer Applications, 2019, 39(12): 3476-3481.)
doi: 10.11772/j.issn.1001-9081.2019050800
( Deng Weibin, Li Yunbo, Zhang Yiming, et al. An Abstractive Text Summarization Method Combining BERT and Convolutional Gating Unit[J]. Control and Decision, 2023, 38(1): 152-160.)
( Zhang Chengzhi, Tong Tiantian, Zhou Qingqing. Automatic Summarization of Book Reviews Based on Fine-Grained Review Mining[J]. Journal of the China Society for Scientific and Technical Information, 2021, 40(2): 163-172.)
( Guo Yanqing, Zhao Rui, Kong Xiangwei, et al. News-Summarization Extraction Method Based on Weighted Event Elements Strategy[J]. Computer Science, 2016, 43(1): 237-241.)
doi: 10.11896/j.issn.1002-137X.2016.01.051
( Liu Biyang, Sun Rui, Ji Donghong. Weibo-Oriented News Summarization Based on Matrix Factorization and Submodular Maximization[J]. Application Research of Computers, 2017, 34(10): 2892-2896.)
( Zhao Danning, Mu Dongmei, Bai Sen. Automatically Extracting Structural Elements of Sci-Tech Literature Abstracts Based on Deep Learning[J]. Data Analysis and Knowledge Discovery, 2021, 5(7): 70-80.)
[16]
刘争. 公共图书馆图书宣传工作的困惑与策略[J]. 办公室业务, 2019(12): 57-58.
[16]
Liu Zheng. Puzzlement and Strategy of Public Library’s Book Publicity Work[J]. Office Operations, 2019(12): 57-58.)
[17]
Liu P F, Yuan W Z, Fu J L, et al. Pre-Train, Prompt, and Predict: A Systematic Survey of Prompting Methods in Natural Language Processing. arXiv Preprint, arXiv: 2107.13586.
[18]
Petroni F, Rocktäschel T, Lewis P, et al. Language Models as Knowledge Bases?[OL]. arXiv Preprint, arXiv: 1909.01066.
[19]
Brown T B, Mann B, Ryder N, et al. Language Models are Few-Shot Learners[OL]. arXiv Preprint, arXiv: 2005.14165.
[20]
Wei J, Zou K. EDA: Easy Data Augmentation Techniques for Boosting Performance on Text Classification Tasks[OL]. arXiv Preprint, arXiv: 1901.11196.
( Li Wei, Yan Xiaodong, Xie Xiaoqing. An Improved TextRank for Tibetan Summarization[J]. Journal of Chinese Information Processing, 2020, 34(9): 36-43.)
[22]
Raffel C, Shazeer N, Roberts A, et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer[OL]. arXiv Preprint, arXiv: 1910.10683.
[23]
Zhang J Q, Zhao Y, Saleh M, et al. PEGASUS: Pre-Training with Extracted Gap-Sentences for Abstractive Summarization[C]// Proceedings of the 37th International Conference on Machine Learning. ACM, 2020: 11328-11339.
[24]
Yang A, Liu K, Liu J, et al. Adaptations of ROUGE and BLEU to Better Evaluate Machine Reading Comprehension Task[C]// Proceedings of the 2018 Workshop on Machine Reading for Question Answering. 2018: 98-104.
[25]
Yang A, Liu K, Liu J, et al. Adaptations of ROUGE and BLEU to Better Evaluate Machine Reading Comprehension Task[C]// Proceedings of the 2018 Workshop on Machine Reading for Question Answering. 2018: 98-104.
[26]
Mihalcea R, Tarau P. TextRank: Bringing Order into Text[C]// Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing. 2004: 404-411.
[27]
Zhang Y G, Dong L L, Li S H. Extraction of Elements of Protest Based on BERT Model and TextTeaser Improved Algorithm[J]. Journal of Physics: Conference Series, 2021, 1955(1): 012107.
doi: 10.1088/1742-6596/1955/1/012107
[28]
Carbonell J, Goldstein J. The Use of MMR, Diversity-Based Reranking for Reordering Documents and Producing Summaries[C]// Proceedings of the 21st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval. ACM, 1998: 335-336.
[29]
Jadhav A, Rajan V. Extractive Summarization with SWAP-NET: Sentences and Words from Alternating Pointer Networks[C]// Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics. 2018: 142-151.
[30]
Liu Y. Fine-Tune BERT for Extractive Summarization[OL]. arXiv Preprint, arXiv: 1903.10318.