Please wait a minute...
Advanced Search
数据分析与知识发现  2023, Vol. 7 Issue (2): 108-118     https://doi.org/10.11925/infotech.2096.3467.2022.1083
  研究论文 本期目录 | 过刊浏览 | 高级检索 |
基于多层交互注意力机制的商品属性抽取*
苏明星,吴厚月,李健,黄菊,张顺香()
安徽理工大学计算机科学与工程学院 淮南 232001
AEMIA:Extracting Commodity Attributes Based on Multi-level Interactive Attention Mechanism
Su Mingxing,Wu Houyue,Li Jian,Huang Ju,Zhang Shunxiang()
School of Computer Science and Engineering, Anhui University of Science & Technology, Huainan 232001, China
全文: PDF (1146 KB)   HTML ( 22
输出: BibTeX | EndNote (RIS)      
摘要 

目的】 提升模型对文本结构特征和文本特征间关联性的感知,充分挖掘文本内在语义,深层次指导抽取任务。【方法】 对文本、句法和词性进行特征抽取,得到各自的特征;将不同的特征进行融合,获得完备的文本结构特征;再设计一个多层交互注意力机制,该机制聚焦于文本结构特征和文本特征之间的深层关联,并采用双线性融合策略,以保证信息的完整性;最后,通过常用的分类器进行属性抽取。【结果】 在公开的数据集上,所提模型的属性抽取准确率相比于已有模型至少提高了1.2个百分点。【局限】 所提模型对隐式属性词感知迟钝,句子中出现三个以上隐式属性词,模型的性能将大幅度降低。【结论】 在显式的商品属性词抽取任务中,建模文本结构特征与文本特征间关联性的方法可以有效提高属性抽取的准确率。

服务
把本文推荐给朋友
加入引用管理器
E-mail Alert
RSS
作者相关文章
苏明星
吴厚月
李健
黄菊
张顺香
关键词 属性抽取交互注意力机制依存关系BiGRUBERT    
Abstract

[Objective] This paper develops a new model to improve the perception of structural features and correlation between text features, aiming to fully explore the internal semantics and extract attributes. [Methods] First, we extracted the features of text, syntax and part of speech. Then, we merged different features to obtain complete text structure features. Third, we designed a multi-layer interactive attention mechanism, which focuses on the deep correlation between text structural features and text features. Fourth, we adopted bilinear fusion strategy to ensure the information integrity. Finally, we extracted attributes with common classifiers. [Results] We examined the new model with publicly available data sets, and found its extraction accuracy was at least 1.2 percentage point higher than that of the existing methods. [Limitations] The model was insensitive to implicit attribute words, and the performance of the model will be greatly reduced with more than three implicit attribute words in the sentence. [Conclusions] The proposed method can effectively improve the accuracy of commodity attributes extraction.

Key wordsAttribute Extraction    Interactive Attention Mechanism    Dependency Relationship    BiGRU    BERT
收稿日期: 2022-10-17      出版日期: 2023-03-28
ZTFLH:  TP391  
基金资助:*国家自然科学基金项目(62076006);安徽省属高校协同创新项目的研究成果之一(GXXT-2021-008)
通讯作者: 常志军,ORCID:0000-0001-9211-8599,E-mail: changzj@mail.las.ac.cn。   
引用本文:   
苏明星, 吴厚月, 李健, 黄菊, 张顺香. 基于多层交互注意力机制的商品属性抽取*[J]. 数据分析与知识发现, 2023, 7(2): 108-118.
Su Mingxing, Wu Houyue, Li Jian, Huang Ju, Zhang Shunxiang. AEMIA:Extracting Commodity Attributes Based on Multi-level Interactive Attention Mechanism. Data Analysis and Knowledge Discovery, 2023, 7(2): 108-118.
链接本文:  
https://manu44.magtech.com.cn/Jwk_infotech_wk3/CN/10.11925/infotech.2096.3467.2022.1083      或      https://manu44.magtech.com.cn/Jwk_infotech_wk3/CN/Y2023/V7/I2/108
Fig.1  商品属性抽取模型AEMIA
Fig.2  基于依赖关系的上下文提取示例
数据集 训练集 测试集
评论语句 属性词 评论语句 属性词
L-14 3 041 2 358 800 654
R-14 3 045 3 693 800 1 134
R-15 1 315 1 192 685 678
R-16 2 000 1 743 676 622
Table 1  实验数据集
项目 配置
OS Windows10
CPU Intel core Tmi7-8700kcpu@3.7GHz
GPU Tesla V100 (16GB)
Python 3.6
PyTorch 1.6
Memory 64GB
Table 2  实验环境
参数
BERT模型 BERT-basehttps://github.com/google-research/bert.
CRF模型 CRF++0.58http://taku910.github.io/crfpp/.
Transformer层数 12
BERT的隐藏层维度 768
BERT中的注意力头数 12
BatchSize 64
Dropout 0.5
Learning rate 1 × 10 - 5
GRU的隐藏层维度 128
优化器 Adam
多层交互注意力的层数 5
Table 3  AEMIA模型参数设置
模型 L-14 R-14 R-15 R-16
Acc/% F1/% Acc/% F1/% Acc/% F1/% Acc/% F1/%
对比实验组 A1 CRF 78.14 74.01 84.06 82.32 72.10 66.54 74.71 69.67
A2 BiGRU+CRF 79.57 76.92 85.73 82.12 73.22 68.12 75.21 71.21
A3 DE-CNN 84.38 81.16 86.14 83.27 74.15 70.18 77.27 74.27
A4 CIA-CRF 83.19 79.12 86.03 84.09 75.26 70.61 78.09 73.21
A5 DepREm-CRF 84.62 81.09 87.21 83.14 77.73 71.86 79.16 74.24
消融实验组 A6 AEMIA(MIA) 83.02 80.11 86.18 82.10 76.69 69.82 78.96 73.23
A7 AEMIA(POS) 84.73 79.05 87.41 83.07 77.03 70.12 80.27 74.92
A8 AEMIA(DE) 83.82 79.71 87.28 83.19 77.91 70.04 80.18 74.88
A9 AEMIA 86.01 81.20 88.93 84.18 79.12 71.05 81.67 75.03
Table 4  实验结果数据
Fig.3  交互注意力层数对准确率的影响
模型 P
A1 CRF 0.081 2
A2 BiGRU+CRF 0.041 2
A3 DE-CNN 0.045 2
A4 CIA-CRF 0.024 7
A5 DepREm-CRF 0.037 8
A9 AEMIA 0.017 5
Table 5  显著性得分
[1] 张严, 李天瑞. 面向评论的方面级情感分析综述[J]. 计算机科学, 2020, 47(6): 194-200.
doi: 10.11896/jsjkx.200200127
[1] (Zhang Yan,Li Tianrui. Review of Comment-Oriented Aspect-Based Sentiment Analysis[J]. Computer Science, 2020, 47(6): 194-200.)
doi: 10.11896/jsjkx.200200127
[2] Hu M Q, Liu B. Mining and Summarizing Customer Reviews[C]// Proceedings of the 10th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2004: 168-177.
[3] Blair-Goldensohn S, Hannan K, McDonald R, et al. Building a Sentiment Summarizer for Local Service Reviews[C]// Proceedings of WWW Workshop on NLP in the Information Explosion Era. 2008: 339-348.
[4] Poria S, Cambria E, Ku L W, et al. A Rule-Based Approach to Aspect Extraction from Product Reviews[C]// Proceedings of the 2nd Workshop on Natural Language Processing for Social Media. 2014: 28-37.
[5] Rana T A, Cheah Y N. A Two-Fold Rule-Based Model for Aspect Extraction[J]. Expert Systems with Applications, 2017, 89: 273-285.
doi: 10.1016/j.eswa.2017.07.047
[6] Wang B, Wang H F. Bootstrapping Both Product Features and Opinion Words from Chinese Customer Reviews with Cross-Inducing[C]// Proceedings of the 3rd International Joint Conference on Natural Language Processing. 2008: 289-295.
[7] Titov I, McDonald R. Modeling Online Reviews with Multi-grain Topic Models[C]// Proceedings of the 17th International Conference on World Wide Web. 2008: 111-120.
[8] Jakob N, Gurevych I. Extracting Opinion Targets in a Single- and Cross-Domain Setting with Conditional Random Fields[C]// Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing. 2010: 1035-1045.
[9] Hamdan H, Bellot P, Bechet F. Lsislif: CRF and Logistic Regression for Opinion Target Extraction and Sentiment Polarity Analysis[C]// Proceedings of the 9th International Workshop on Semantic Evaluation. 2015: 753-758.
[10] Gupta D K, Reddy K S, Shweta, et al. PSO-Asent: Feature Selection Using Particle Swarm Optimization for Aspect Based Sentiment Analysis[C]// Proceedings of International Conference on Applications of Natural Language to Information Systems. 2015: 220-233.
[11] 彭春艳, 张晖, 包玲玉, 等. 基于条件随机域的生物命名实体识别[J]. 计算机工程, 2009, 35(22): 197-199.
[11] (Peng Chunyan, Zhang Hui, Bao Lingyu, et al. Biological Named Entity Recognition Based on Conditional Random Fields[J]. Computer Engineering, 2009, 35(22): 197-199.)
[12] Toh Z, Wang W T. DLIREC: Aspect Term Extraction and Term Polarity Classification System[C]// Proceedings of the 8th International Workshop on Semantic Evaluation. 2014: 235-240.
[13] Poria S, Cambria E, Gelbukh A. Aspect Extraction for Opinion Mining with a Deep Convolutional Neural Network[J]. Knowledge-Based Systems, 2016, 108: 42-49.
doi: 10.1016/j.knosys.2016.06.009
[14] 苏丰龙, 谢庆华, 邱继远, 等. 基于深度学习的领域实体属性词聚类抽取研究[J]. 微型机与应用, 2016, 35(1): 53-55, 59.
[14] (Su Fenglong, Xie Qinghua, Qiu Jiyuan, et al. Study on Word Clustering for Attribute Extraction Based on Deep Learning[J]. Microcomputer & Its Applications, 2016, 35(1): 53-55, 59.)
[15] Li X, Lam W. Deep Multi-task Learning for Aspect Term Extraction with Memory Interaction[C]// Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing. 2017: 2886-2892.
[16] 王仁武, 张文慧. 基于深度学习的隐性评价对象识别方法[J]. 计算机工程, 2019, 45(8): 315-320.
[16] (Wang Renwu, Zhang Wenhui. Implicit Evaluation Object Recognition Method Based on Deep Learning[J]. Computer Engineering, 2019, 45(8): 315-320.)
[17] Yin Y C, Wang C G, Zhang M. PoD: Positional Dependency-Based Word Embedding for Aspect Term Extraction[C]// Proceedings of the 28th International Conference on Computational Linguistics. 2020: 1714-1719.
[18] Phan M H, Ogunbona P O. Modelling Context and Syntactical Features for Aspect-Based Sentiment Analysis[C]// Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 2020: 3211-3220.
[19] Devlin J, Chang M W, Lee K, et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding[C]// Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies, Volume 1 (Long and Short Papers). 2019: 4171-4186.
[20] Cho K, van Merriënboer B, Gulcehre C, et al. Learning Phrase Representations Using RNN Encoder-Decoder for Statistical Machine Translation[C]// Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing. 2014: 1724-1734.
[21] Ma R X, Wang K, Qiu T, et al. Feature-Based Compositing Memory Networks for Aspect-Based Sentiment Classification in Social Internet of Things[J]. Future Generation Computer Systems, 2019, 92(C): 879-888.
[22] Liu G, Guo J B. Bidirectional LSTM with Attention Mechanism and Convolutional Layer for Text Classification[J]. Neurocomputing, 2019, 337: 325-338.
doi: 10.1016/j.neucom.2019.01.078
[23] Tenenbaum J B, Freeman W T. Separating Style and Content with Bilinear Models[J]. Neural Computation, 2000, 12(6): 1247-1283.
pmid: 10935711
[24] Kim J H, On K W, Lim W, et al. Hadamard Product for Low-rank Bilinear Pooling[OL]. arXiv Preprint, arXiv: 1610.04325.
[25] Liu Y. Fine-tune BERT for Extractive Summarization[C]// Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing. 2019: 3546-3551.
[26] Toh Z Q, Su J. NLANGPat SemEval-2016 Task 5: Improving Aspect Based Sentiment Analysis Using Neural Network Features[C]// Proceedings of the 10th International Workshop on Semantic Evaluation. 2016: 282-288.
[27] Xu H, Liu B, Shu L, et al. Double Embeddings and CNN-Based Sequence Labeling for Aspect Extraction[C]// Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2:Short Papers). 2018: 592-598.
[28] 尉桢楷, 程梦, 周夏冰, 等. 基于类卷积交互式注意力机制的属性抽取研究[J]. 计算机研究与发展, 2020, 57(11): 2456-2466.
[28] (Yu Zhenkai, Cheng Meng, Zhou Xiabing, et al. Convolutional Interactive Attention Mechanism for Aspect Extraction[J]. Journal of Computer Research and Development, 2020, 57(11): 2456-2466.)
[29] 李成梁, 赵中英, 李超, 等. 基于依存关系嵌入与条件随机场的商品属性抽取方法[J]. 数据分析与知识发现, 2020, 4(5): 54-65.
[29] (Li Chengliang, Zhao Zhongying, Li Chao, et al. Extracting Product Properties with Dependency Relationship Embedding and Conditional Random Field[J]. Data Analysis and Knowledge Discovery, 2020, 4(5): 54-65.)
[30] Dror R, Baumer G, Shlomov S, et al. The Hitchhiker’s Guide to Testing Statistical Significance in Natural Language Processing[C]// Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1:Long Papers). 2018: 1383-1392.
[1] 赵一鸣, 潘沛, 毛进. 基于任务知识融合与文本数据增强的医学信息查询意图强度识别研究*[J]. 数据分析与知识发现, 2023, 7(2): 38-47.
[2] 王宇飞, 张智雄, 赵旸, 张梦婷, 李雪思. 中文科技论文标题自动生成系统的设计与实现*[J]. 数据分析与知识发现, 2023, 7(2): 61-71.
[3] 张思阳, 魏苏波, 孙争艳, 张顺香, 朱广丽, 吴厚月. 基于多标签Seq2Seq模型的情绪-原因对提取模型*[J]. 数据分析与知识发现, 2023, 7(2): 86-96.
[4] 施运梅, 袁博, 张乐, 吕学强. IMTS:融合图像与文本语义的虚假评论检测方法*[J]. 数据分析与知识发现, 2022, 6(8): 84-96.
[5] 吴江, 刘涛, 刘洋. 在线社区用户画像及自我呈现主题挖掘——以网易云音乐社区为例*[J]. 数据分析与知识发现, 2022, 6(7): 56-69.
[6] 郑洁, 黄辉, 秦永彬. 一种融合法律知识的相似案例匹配模型*[J]. 数据分析与知识发现, 2022, 6(7): 99-106.
[7] 潘慧萍, 李宝安, 张乐, 吕学强. 基于多特征融合的政府工作报告关键词提取研究*[J]. 数据分析与知识发现, 2022, 6(5): 54-63.
[8] 肖悦珺, 李红莲, 张乐, 吕学强, 游新冬. 特征融合的中文专利文本分类方法研究*[J]. 数据分析与知识发现, 2022, 6(4): 49-59.
[9] 杨林, 黄晓硕, 王嘉阳, 丁玲玲, 李子孝, 李姣. 基于BERT-TextCNN的临床试验疾病亚型识别研究*[J]. 数据分析与知识发现, 2022, 6(4): 69-81.
[10] 郭航程, 何彦青, 兰天, 吴振峰, 董诚. 基于Paragraph-BERT-CRF的科技论文摘要语步功能信息识别方法研究*[J]. 数据分析与知识发现, 2022, 6(2/3): 298-307.
[11] 丁晟春, 游伟静, 王小英. 基于属性词补全的武器装备属性抽取研究*[J]. 数据分析与知识发现, 2022, 6(2/3): 289-297.
[12] 张云秋, 汪洋, 李博诚. 基于RoBERTa-wwm动态融合模型的中文电子病历命名实体识别*[J]. 数据分析与知识发现, 2022, 6(2/3): 242-250.
[13] 王永生, 王昊, 虞为, 周泽聿. 融合结构和内容的方志文本人物关系抽取方法*[J]. 数据分析与知识发现, 2022, 6(2/3): 318-328.
[14] 商丽丽, 唐华云, 王延昭, 左美云. 在线评论可行动信息识别研究*[J]. 数据分析与知识发现, 2022, 6(12): 1-12.
[15] 严冬梅, 何雯馨, 陈智. 融合情感特征的基于RoBERTa-TCN的股价预测研究[J]. 数据分析与知识发现, 2022, 6(12): 123-134.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
版权所有 © 2015 《数据分析与知识发现》编辑部
地址:北京市海淀区中关村北四环西路33号 邮编:100190
电话/传真:(010)82626611-6626,82624938
E-mail:jishu@mail.las.ac.cn