Please wait a minute...
Advanced Search
数据分析与知识发现  2024, Vol. 8 Issue (1): 69-79     https://doi.org/10.11925/infotech.2096-3467.2022.1270
  研究论文 本期目录 | 过刊浏览 | 高级检索 |
考虑评论情感表达力及其重要性的个性化推荐算法*
李慧,胡耀华(),徐存真
西安电子科技大学经济与管理学院 西安 710119
Personalized Recommendation Algorithm with Review Sentiments and Importance
Li Hui,Hu Yaohua(),Xu Cunzhen
School of Economics and Management, Xidian University, Xi’an 710119, China
全文: PDF (946 KB)   HTML ( 18
输出: BibTeX | EndNote (RIS)      
摘要 

【目的】 针对数据稀疏性问题,为探索情感表达对用户特征表示的影响,提出一种考虑评论情感表达力及其重要性的个性化推荐算法。【方法】 使用BERT预训练语言模型获取评论文本的向量表示,通过Bi-GRU网络学习其中的语义特征,分别采用情感权重和注意力机制为评论向量分配权重,最后利用DeepFM算法对用户特征和产品特征进行深度交互,预测用户对产品的评分值。【结果】 在Amazon Product Data数据集上的实验结果表明,所提算法比基线算法在RMSE和MAE指标上最多可降低24.43%和31.44%;使用情感权重为用户评论加权的方法相比于注意力机制,在RMSE与MAE指标上最多可降低2.59%和3.89%。【局限】 所用情感分析方法无法表现用户对产品不同属性的情感倾向。【结论】 所提算法考虑了情感对用户特征表达的影响,能够提高推荐准确性。

服务
把本文推荐给朋友
加入引用管理器
E-mail Alert
RSS
作者相关文章
李慧
胡耀华
徐存真
关键词 推荐算法深度学习情感分析注意力机制    
Abstract

[Objective] To address the data sparsity issue and explore the impacts of emotional expression on user feature learning, this paper proposes a personalized recommendation algorithm based on sentiment and the importance of online reviews. [Methods] First, we used the BERT pre-trained language model to generate the vector representation of review texts. Then, we fed them into a Bi-GRU network to learn their semantic features. We also assigned weights to the review vector using sentiment weights and attention mechanisms. Finally, we utilized the DeepFM algorithm for deep interaction between user and product features to predict the user’s rating of the products. [Results] We examined the proposed model with the Amazon product data dataset. Our model reduced the RMSE and MAE metrics by up to 24.43% and 31.44% compared to the baseline models. Compared with the attention mechanism, our method reduced the RMSE and MAE metrics by up to 2.59% and 3.89%. [Limitations] The sentiment analysis method cannot represent the users’ emotional tendencies towards the different attributes of the product. [Conclusions] The proposed method considers the influence of user sentiment on user feature expression, improving the recommendation accuracy.

Key wordsRecommendation Algorithm    Deep Learning    Sentiment Analysis    Attention Mechanism
收稿日期: 2022-12-01      出版日期: 2023-09-12
ZTFLH:  G203  
  TP391  
基金资助:*国家自然科学基金项目(71203173)
通讯作者: 胡耀华,ORCID:0000-0001-9306-0974,E-mail:huyaohua@stu.xidian.edu.cn。   
引用本文:   
李慧, 胡耀华, 徐存真. 考虑评论情感表达力及其重要性的个性化推荐算法*[J]. 数据分析与知识发现, 2024, 8(1): 69-79.
Li Hui, Hu Yaohua, Xu Cunzhen. Personalized Recommendation Algorithm with Review Sentiments and Importance. Data Analysis and Knowledge Discovery, 2024, 8(1): 69-79.
链接本文:  
https://manu44.magtech.com.cn/Jwk_infotech_wk3/CN/10.11925/infotech.2096-3467.2022.1270      或      https://manu44.magtech.com.cn/Jwk_infotech_wk3/CN/Y2024/V8/I1/69
Fig.1  PREEI算法框架
数据集 用户数 产品数 评论数 稀疏性
MI 1 429 900 10 261 99.20%
AIV 5 130 1 685 37 126 99.57%
OP 4 905 2 420 53 258 99.55%
THI 16 638 10 217 134 476 99.92%
VG 24 303 10 672 231 780 99.91%
Table 1  实验数据集的基本统计信息
算法 MI AIV OP THI VG
RMSE MAE RMSE MAE RMSE MAE RMSE MAE RMSE MAE
PMF 0.995 0.846 1.150 0.935 0.991 0.805 1.108 0.920 1.204 0.962
ConvMF+ 0.942 0.713 1.033 0.771 0.877 0.655 1.020 0.748 1.083 0.818
DeepCoNN 0.897 0.797 1.004 0.792 0.870 0.738 0.961 0.760 1.129 0.910
NARRE 0.966 0.687 0.971 0.714 0.858 0.629 0.977 0.703 1.062 0.806
CARL 0.836 0.609 0.965 0.716 0.861 0.645 0.980 0.742 1.087 0.862
ATN 0.918 0.692 0.986 0.747 0.844 0.644 0.954 0.712 1.077 0.849
PREEI 0.822 0.587 0.869 0.641 0.836 0.644 0.890 0.642 1.010 0.768
Table 2  各算法推荐准确性比较
算法 MI AIV OP THI VG
RMSE MAE RMSE MAE RMSE MAE RMSE MAE RMSE MAE
PMF 17.39 30.61 24.43 31.44 15.64 20.00 19.68 30.22 16.11 20.17
ConvMF+ 12.74 17.67 15.88 16.86 4.68 1.68 12.75 14.17 6.74 6.11
DeepCoNN 8.36 26.35 13.45 19.07 3.91 12.74 7.39 15.53 10.54 15.60
NARRE 14.91 14.56 10.50 10.22 2.56 -2.38 8.90 8.68 4.90 4.71
CARL 1.79 3.45 9.85 10.47 2.90 0.16 9.18 13.48 7.17 10.70
ATN 10.46 15.17 11.87 14.19 0.95 0.00 6.71 9.83 6.22 9.54
Table 3  PREEI算法相较基线算法在推荐准确性上的提升程度(%)
Fig.2  隐藏层数与Dropout比率对算法性能的影响
模型 MI AIV OP THI VG
RMSE MAE RMSE MAE RMSE MAE RMSE MAE RMSE MAE
PREEI 0.822 0.587 0.869 0.641 0.836 0.644 0.890 0.642 1.010 0.768
PREEI-A 0.847 0.597 0.870 0.642 0.844 0.659 0.890 0.650 1.012 0.777
PREEI-D 0.877 0.639 0.890 0.686 0.859 0.664 0.898 0.673 1.028 0.803
PREEI-W 0.886 0.675 0.983 0.761 0.919 0.737 1.018 0.785 1.111 0.874
Table 4  变体模型推荐准确性比较
用户 评论编号 评论 情感得分&权重
用户1 a For any sound techs that like to make things simple and easy to work,these xlr jumper cables are the way to go,and the quality is impressive. I used them for a new set of wireless mike to jump into my 16 channel board and was it so easy,my mikes are color coded and my cable fitted right in with the same color code thus making it easy to identify from afar and easy to control on the mixing board. 情感得分=1.473
情感权重=0.225
b If you are have problems knowing what and where your systems are terminating,these are a must for you. I have mastched them with my microphones and it is easy for controlling and identifying. 情感得分=-0.070
情感权重=0.048
用户2 a Love my assortment of picks and the price was right. I like that their is different gauges,because I don’t always like using the same so this is great with the three different gauges to choose from in all these nice colours but I only received 47 instead of 48 picks but whose counting,I'm still happy with my picks and would purchase from seller again,thanks. 情感得分=2.768
情感权重=0.505
b Fast shipping,great price and most of all great stings,very satisfied would buy these strings again and would recommend to a friend. 情感得分=0.363
情感权重=0.046
Table 5  部分用户情感权重案例
[1] 丁浩, 胡广伟, 王婷, 等. 基于时序漂移的潜在因子模型推荐方法[J]. 数据分析与知识发现, 2022, 6(10): 1-8.
[1] (Ding Hao, Hu Guangwei, Wang Ting, et al. Recommendation Method for Potential Factor Model Based on Time Series Drift[J]. Data Analysis and Knowledge Discovery, 2022, 6(10): 1-8.)
[2] 马莹雪, 甘明鑫, 肖克峻. 融合标签和内容信息的矩阵分解推荐方法[J]. 数据分析与知识发现, 2021, 5(5): 71-82.
[2] (Ma Yingxue, Gan Mingxin, Xiao Kejun. A Matrix Factorization Recommendation Method with Tags and Contents[J]. Data Analysis and Knowledge Discovery, 2021, 5(5): 71-82.)
[3] 易明, 刘明, 冯翠翠. 融合异质信息网络表示学习的跨领域推荐研究[J]. 情报学报, 2022, 41(4): 337-349.
[3] (Yi Ming, Liu Ming, Feng Cuicui. Cross-Domain Recommendation Combining Heterogeneous Information Network Representation Learning[J]. Journal of the China Society for Scientific and Technical Information, 2022, 41(4): 337-349.)
[4] Liu D H, Li J, Du B, et al. A Hybrid Neural Network Approach to Combine Textual Information and Rating Information for Item Recommendation[J]. Knowledge and Information Systems, 2021, 63(3): 621-646.
doi: 10.1007/s10115-020-01528-2
[5] 李伟卿, 池毛毛, 王伟军. 面向用户长短期偏好调节的可解释个性化推荐方法研究[J]. 图书情报工作, 2021, 65(12): 101-111.
doi: 10.13266/j.issn.0252-3116.2021.12.010
[5] (Li Weiqing, Chi Maomao, Wang Weijun. Explainable Personalized Recommendation Method Based on Adjustment of Users’ Long-and Short-Term Preferences[J]. Library and Information Service, 2021, 65(12): 101-111.)
doi: 10.13266/j.issn.0252-3116.2021.12.010
[6] 聂卉, 邱以菲. 融合用户兴趣及评论效用的评论信息推荐[J]. 图书情报工作, 2021, 65(10): 68-78.
doi: 10.13266/j.issn.0252-3116.2021.10.008
[6] (Nie Hui, Qiu Yifei. Integrating User Interests with Review Helpfulness for Review Recommendation[J]. Library and Information Service, 2021, 65(10): 68-78.)
doi: 10.13266/j.issn.0252-3116.2021.10.008
[7] Mikolov T, Sutskever I, Chen K, et al. Distributed Representations of Words and Phrases and Their Compositionality[C]// Proceedings of the 26th International Conference on Neural Information Processing Systems. 2013.
[8] Pennington J, Socher R, Manning C. GloVe: Global Vectors for Word Representation[C]// Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing. 2014: 1532-1543.
[9] Devlin J, Chang M W, Lee K, et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding[C]// Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies. 2019: 4171-4186.
[10] Wang C, Blei D M. Collaborative Topic Modeling for Recommending Scientific Articles[C]// Proceedings of the 17th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2011: 448-456.
[11] McAuley J, Leskovec J. Hidden Factors and Hidden Topics: Understanding Rating Dimensions with Review Text[C]// Proceedings of the 7th ACM Conference on Recommender Systems. 2013: 165-172.
[12] Bao Y, Fang H, Zhang J. TopicMF: Simultaneously Exploiting Ratings and Reviews for Recommendation[C]// Proceedings of the 28th AAAI Conference on Artificial Intelligence. 2014: 2-8.
[13] Tan Y Z, Zhang M, Liu Y Q, et al. Rating-Boosted Latent Topics: Understanding Users and Items with Ratings and Reviews[C]// Proceedings of the 25th International Joint Conference on Artificial Intelligence. 2016: 2640-2646.
[14] Zhang H J, Ji Y Z, Li J X, et al. A Triple Wing Harmonium Model for Movie Recommendation[J]. IEEE Transactions on Industrial Informatics, 2016, 12(1): 231-239.
doi: 10.1109/TII.2015.2475218
[15] Kim D, Park C, Oh J, et al. Convolutional Matrix Factorization for Document Context-Aware Recommendation[C]// Proceedings of the 10th ACM Conference on Recommender Systems. 2016: 233-240.
[16] Salakhutdinov R, Mnih A. Probabilistic Matrix Factorization[C]// Proceedings of the 20th International Conference on Neural Information Processing Systems. 2007: 1257-1264.
[17] Dong X, Yu L, Wu Z H, et al. A Hybrid Collaborative Filtering Model with Deep Structure for Recommender Systems[C]// Proceedings of the 31st AAAI Conference on Artificial Intelligence. 2017: 1309-1315.
[18] Zheng L, Noroozi V, Yu P S. Joint Deep Modeling of Users and Items Using Reviews for Recommendation[C]// Proceedings of the 10th ACM International Conference on Web Search and Data Mining. 2017: 425-434.
[19] Li D, Liu H, Zhang Z L, et al. CARM: Confidence-Aware Recommender Model via Review Representation Learning and Historical Rating Behavior in the Online Platforms[J]. Neurocomputing, 2021, 455(C): 283-296.
[20] Shuai J, Zhang K, Wu L, et al. A Review-Aware Graph Contrastive Learning Framework for Recommendation[C]// Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval. 2022: 1283-1293.
[21] Seo S, Huang J, Yang H, et al. Interpretable Convolutional Neural Networks with Dual Local and Global Attention for Review Rating Prediction[C]// Proceedings of the 11th ACM Conference on Recommender Systems. 2017: 297-305.
[22] Chen C, Zhang M, Liu Y Q, et al. Neural Attentional Rating Regression with Review-Level Explanations[C]// Proceedings of the 2018 World Wide Web Conference. 2018: 1583-1592.
[23] Lu Y C, Dong R H, Smyth B. Coevolutionary Recommendation Model: Mutual Learning Between Ratings and Reviews[C]// Proceedings of the 2018 World Wide Web Conference. 2018: 773-782.
[24] Cho K, van Merrienboer B, Gulcehre C, et al. Learning Phrase Representations Using RNN Encoder-Decoder for Statistical Machine Translation[C]// Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing. 2014: 1724-1734.
[25] Koren Y, Bell R, Volinsky C. Matrix Factorization Techniques for Recommender Systems[J]. Computer, 2009, 42(8): 30-37.
[26] Wu L B, Quan C, Li C L, et al. A Context-Aware User-Item Representation Learning for Item Recommendation[J]. ACM Transactions on Information Systems, 2019, 37(2): 22.
[27] Liu H, Zheng C, Li D T C, et al. EDMF: Efficient Deep Matrix Factorization with Review Feature Learning for Industrial Recommender System[J]. IEEE Transactions on Industrial Informatics, 2022, 18(7): 4361-4371.
doi: 10.1109/TII.2021.3128240
[28] Liu Y, Miyazaki J. Knowledge-Aware Attentional Neural Network for Review-Based Movie Recommendation with Explanations[J]. Neural Computing and Applications, 2023, 35(3): 2717-2735.
doi: 10.1007/s00521-022-07689-1
[29] Liu H, Qiao H Z, Shi X Y, et al. Aspect-Aware Asymmetric Representation Learning Network for Review-Based Recommendation[C]// Proceedings of 2022 International Joint Conference on Neural Networks. IEEE, 2022: 1-8.
[30] Forman C, Ghose A, Wiesenfeld B. Examining the Relationship Between Reviews and Sales: The Role of Reviewer Identity Disclosure in Electronic Markets[J]. Information Systems Research, 2008, 19(3): 291-313.
doi: 10.1287/isre.1080.0193
[31] 彭敏, 席俊杰, 代心媛, 等. 基于情感分析和LDA主题模型的协同过滤推荐算法[J]. 中文信息学报, 2017, 31(2): 194-203.
[31] (Peng Min, Xi Junjie, Dai Xinyuan, et al. Collaborative Filtering Recommendation Based on Sentiment Analysis and LDA Topic Model[J]. Journal of Chinese Information Processing, 2017, 31(2): 194-203.)
[32] Shen R P, Zhang H R, Yu H, et al. Sentiment Based Matrix Factorization with Reliability for Recommendation[J]. Expert Systems with Applications, 2019, 135: 249-258.
doi: 10.1016/j.eswa.2019.06.001
[33] Yang C, Chen X H, Liu L, et al. Leveraging Semantic Features for Recommendation: Sentence-Level Emotion Analysis[J]. Information Processing & Management, 2021, 58(3): 102543.
doi: 10.1016/j.ipm.2021.102543
[34] Cai Y, Ke W M, Cui E, et al. A Deep Recommendation Model of Cross-Grained Sentiments of User Reviews and Ratings[J]. Information Processing & Management, 2022, 59(2): 102842.
doi: 10.1016/j.ipm.2021.102842
[35] 李浩君, 吕韵, 汪旭辉, 等. 融入情感分析的多层交互深度推荐模型研究[J]. 数据分析与知识发现, 2023, 7(3): 43-57.
[35] (Li Haojun, Lv Yun, Wang Xuhui, et al. A Deep Recommendation Model with Multi-Layer Interaction and Sentiment Analysis[J]. Data Analysis and Knowledge Discovery, 2023, 7(3): 43-57.)
[36] Wang P P, Li L, Wang R, et al. Learning Persona-Driven Personalized Sentimental Representation for Review-Based Recommendation[J]. Expert Systems with Applications, 2022, 203: 117317.
doi: 10.1016/j.eswa.2022.117317
[37] Baccianella S, Esuli A, Sebastiani F. SentiWordNet 3.0: An Enhanced Lexical Resource for Sentiment Analysis and Opinion Mining[C]// Proceedings of the 7th International Conference on Language Resources and Evaluation. 2010: 2200-2204.
[38] Guo H F, Tang R M, Ye Y M, et al. DeepFM: A Factorization-Machine Based Neural Network for CTR Prediction[C]// Proceedings of the 26th International Joint Conference on Artificial Intelligence. 2017: 1725-1731.
[39] Tan J T, Xu S Y, Ge Y Q, et al. Counterfactual Explainable Recommendation[C]// Proceedings of the 30th ACM International Conference on Information & Knowledge Management. 2021: 1784-1793.
[40] Manjunatha Adithya R, Sandhya K, Sachin G M, et al. Feature Validated Sentiment Recommendation of Hotel Reviews[M]// Soft Computing for Security Applications:Proceedings of ICSCS 2022. Singapore: Springer Nature Singapore, 2022: 413-422.
[41] Kingma D P, Ba J L. Adam: A Method for Stochastic Optimization[C]// Proceedings of the 3rd International Conference on Learning Representations. 2015.
[42] Srivastava N, Hinton G, Krizhevsky A, et al. Dropout: A Simple Way to Prevent Neural Networks from Overfitting[J]. The Journal of Machine Learning Research, 2014, 15(1): 1929-1958.
[1] 李雪莲, 王碧, 李立鑫, 韩迪轩. 融合抽象语义表示和依存语法的方面级情感分析*[J]. 数据分析与知识发现, 2024, 8(1): 55-68.
[2] 向卓元, 陈浩, 王倩, 李娜. 面向任务型对话的小样本语言理解模型研究*[J]. 数据分析与知识发现, 2023, 7(9): 64-77.
[3] 何丽, 杨美华, 刘璐瑶. 融合SPO语义和句法信息的事件检测方法*[J]. 数据分析与知识发现, 2023, 7(9): 114-124.
[4] 韩普, 顾亮, 叶东宇, 陈文祺. 基于多任务和迁移学习的中文医学文献实体识别研究*[J]. 数据分析与知识发现, 2023, 7(9): 136-145.
[5] 聂卉, 蔡瑞昇. 引入注意力机制的在线问诊推荐研究*[J]. 数据分析与知识发现, 2023, 7(8): 138-148.
[6] 李广建, 袁钺. 基于深度学习的科技文献知识单元抽取研究综述[J]. 数据分析与知识发现, 2023, 7(7): 1-17.
[7] 王楠, 王淇. 基于深度学习的学生课堂专注度测评方法*[J]. 数据分析与知识发现, 2023, 7(6): 123-133.
[8] 汪晓凤, 孙雨洁, 王华珍, 张恒彰. 融合深度学习和知识图谱的类型可控问句生成模型构建及验证*[J]. 数据分析与知识发现, 2023, 7(6): 26-37.
[9] 吴佳伦, 张若楠, 康武林, 袁普卫. 基于患者相似性分析的药物推荐深度学习模型研究*[J]. 数据分析与知识发现, 2023, 7(6): 148-160.
[10] 徐康, 余胜男, 陈蕾, 王传栋. 基于语言学知识增强的自监督式图卷积网络的事件关系抽取方法*[J]. 数据分析与知识发现, 2023, 7(5): 92-104.
[11] 刘洋, 张雯, 胡毅, 毛进, 黄菲. 基于多模态深度学习的酒店股票预测*[J]. 数据分析与知识发现, 2023, 7(5): 21-32.
[12] 黄学坚, 马廷淮, 王根生. 基于分层语义特征学习模型的微博谣言事件检测*[J]. 数据分析与知识发现, 2023, 7(5): 81-91.
[13] 闫尚义, 王靖亚, 刘晓文, 崔雨萌, 陶知众, 张晓帆. 基于多头自注意力池化与多粒度特征交互融合的微博情感分析*[J]. 数据分析与知识发现, 2023, 7(4): 32-45.
[14] 张昱, 张海军, 刘雅情, 梁科晋, 王月阳. 基于双向掩码注意力机制的多模态情感分析*[J]. 数据分析与知识发现, 2023, 7(4): 46-55.
[15] 韩普, 仲雨乐, 陆豪杰, 马诗雯. 基于对抗性迁移学习的药品不良反应实体识别研究*[J]. 数据分析与知识发现, 2023, 7(3): 131-141.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
版权所有 © 2015 《数据分析与知识发现》编辑部
地址:北京市海淀区中关村北四环西路33号 邮编:100190
电话/传真:(010)82626611-6626,82624938
E-mail:jishu@mail.las.ac.cn