Please wait a minute...
Advanced Search
数据分析与知识发现  2022, Vol. 6 Issue (4): 120-129     https://doi.org/10.11925/infotech.2096-3467.2021.0884
  研究论文 本期目录 | 过刊浏览 | 高级检索 |
基于Transformer和图卷积神经网络的隐喻识别*
郭樊容,黄孝喜(),王荣波,谌志群,胡创,谢一敏,司博宇
杭州电子科技大学认知与智能计算研究所 杭州 310018
Identifying Metaphor with Transformer and Graph Convolutional Network
Guo Fanrong,Huang Xiaoxi(),Wang Rongbo,Chen Zhiqun,Hu Chuang,Xie Yimin,Si Boyu
Institute of Cognitive and Intelligent Computing, Hangzhou Dianzi University, Hangzhou 310018, China
全文: PDF (864 KB)   HTML ( 40
输出: BibTeX | EndNote (RIS)      
摘要 

【目的】 提出一种基于图卷积神经网络和Transformer的隐喻识别模型,既能有效识别单词隐喻,又能解决多个词语共同引发的隐喻表达。【方法】 利用图卷积神经网络提取句法依存树中的句法结构信息,联合从Transformer结构学习的深层语义表示,通过SoftMax计算目标词隐喻表达的概率。【结果】 在英文数据集UVA VERB 和 UVA ALLPOS上F1值分别提高1.9% 和1.7%,TOEFL VERB和 TOEFL ALLPOS上F1值分别提高1.1%和1.9%;在中文数据集CCL上F1值提高1.2%。【局限】 如果句子中存在歧义或者指代信息不明确的现象,则不能有效识别句子中的隐喻现象。【结论】 图卷积神经网络和句法依存树的引入确实能在一定程度上丰富目标词语义信息,提高单词和多词隐喻的识别效果。

服务
把本文推荐给朋友
加入引用管理器
E-mail Alert
RSS
作者相关文章
郭樊容
黄孝喜
王荣波
谌志群
胡创
谢一敏
司博宇
关键词 隐喻识别图卷积神经网络依存分析Transformer    
Abstract

[Objective] This paper proposes a metaphor identification model based on graph convolutional neural network and Transformer, aiming to effectively find metaphor expressions with multiple words. [Methods] We used the graph convolutional neural network to extract the structure information from the syntactic dependency tree. Then, we combined the structure with deep semantic representation by the Transformer. Finally, we calculated the probability of metaphorical expression for the target words through SoftMax. [Results] Compared with the existing algorithms, the F1 values of our model increased by 1.9% and 1.7% on UVA VERB and UVA ALL POS datasets. The F1 values were also improved by 1.1% and 1.9% on TOEFL VERB and TOEFL ALL POS. The F1 value increased by 1.2% on the Chinese data CCL. [Limitations] If there is ambiguity or ambiguous referential information in the sentence, our model will not effectively identify the metaphor expressions. [Conclusions] Graph convolutional network and syntactic dependency tree can enrich the semantics of target words, which improves the recognition of single and multi-word metaphors.

Key wordsMetaphor Identification    Graph Convolutional Neural Network    Syntactic Dependency    Transformer
收稿日期: 2021-08-23      出版日期: 2022-05-12
ZTFLH:  TP391  
基金资助:*教育部人文社会科学研究规划基金项目(18YJA740016);国家社会科学基金重大规划项目(18ZDA290)
通讯作者: 黄孝喜,ORCID:0000-0003-4483-3664     E-mail: huangxx@hdu.edu.cn
引用本文:   
郭樊容, 黄孝喜, 王荣波, 谌志群, 胡创, 谢一敏, 司博宇. 基于Transformer和图卷积神经网络的隐喻识别*[J]. 数据分析与知识发现, 2022, 6(4): 120-129.
Guo Fanrong, Huang Xiaoxi, Wang Rongbo, Chen Zhiqun, Hu Chuang, Xie Yimin, Si Boyu. Identifying Metaphor with Transformer and Graph Convolutional Network. Data Analysis and Knowledge Discovery, 2022, 6(4): 120-129.
链接本文:  
https://manu44.magtech.com.cn/Jwk_infotech_wk3/CN/10.11925/infotech.2096-3467.2021.0884      或      https://manu44.magtech.com.cn/Jwk_infotech_wk3/CN/Y2022/V6/I4/120
Fig.1  基于Transformer和图卷积神经网络的隐喻识别模型
Fig.2  句子依存结构
实验超参数 参数值
句子长度 128
RoBERTa Layers 12
RoBERTa Attention Heads 12
Mutil-Head Attention Heads 8
Batch Size 16
迭代次数 3
窗口范围(k) 2
图卷积神经网络的层数 2
第一层图卷积神经网络的隐藏层维度 512
第二层图卷积神经网络的隐藏层维度 256
Table 1  实验超参数设置
方法 数据集 P R F1
VUA ALLPOS 0.608 0.700 0.651
VUA VERB 0.600 0.763 0.672
VUA ALLPOS 0.716 0.736 0.726
VUA VERB 0.682 0.713 0.697
VUA ALLPOS 0.730 0.757 0.743
VUA VERB 0.693 0.723 0.708
VUA ALLPOS 0.746 0.715 0.730
VUA VERB 0.761 0.781 0.771
VUA ALLPOS 0.756 0.783 0.769
VUA VERB 0.789 0.819 0.804
本文 VUA ALLPOS 0.777 0.795 0.786
VUA VERB 0.812 0.834 0.823
Table 2  在VUA数据集上的实验结果
方法 数据集 P R F1
TOEFL ALLPOS 0.709 0.697 0.703
TOEFL VERB 0.731 0.707 0.719
TOEFL ALLPOS 0.695 0.735 0.715
TOEFL VERB 0.733 0.766 0.749
本文 TOEFL ALLPOS 0.725 0.742 0.734
TOEFL VERB 0.760 0.764 0.762
Table 3  在TOEFL数据集上的实验结果
方法 P R F1
hqu - - 0.833
faun - - 0.831
YNU-HPCC - - 0.831
MITLAB - - 0.827
prism - - 0.821
0.881 0.896 0.888
本文 0.889 0.912 0.900
Table 4  在CCL数据集上的实验结果
模型 数据集 P R F1
Ours VUA ALLPOS 0.777 0.795 0.786
VUA VERB 0.812 0.834 0.823
-LOCAL VUA ALLPOS 0.761 0.787 0.773
VUA VERB 0.805 0.812 0.808
-GCN VUA ALLPOS 0.740 0.767 0.753
VUA VERB 0.764 0.813 0.785
Table 5  模块消融实验结果
Fig.3  图卷积神经网络隐藏层参数影响
[1] Aggarwal S, Singh R. Metaphor Detection Using Deep Contextualized Word Embeddings[OL]. arXiv Preprint, arXiv:2009.12565.
[2] Wu C H, Wu F Z, Chen Y B, et al. Neural Metaphor Detecting with CNN-LSTM Model[C]//Proceedings of the 2018 Workshop on Figurative Language Processing. Stroudsburg, PA, USA: Association for Computational Linguistics, 2018: 110-114.
[3] Dankers V, Malhotra K, Kudva G, et al. Being Neighbourly: Neural Metaphor Identification in Discourse[C]//Proceedings of the 2nd Workshop on Figurative Language Processing. Stroudsburg, PA, USA: Association for Computational Linguistics, 2020: 227-234.
[4] Gong H, Bhat S P, Viswanath P. Geometry of Compositionality[OL]. arXiv Preprint, arXiv: 1611.09799.
[5] Leong C W B, Klebanov B B, Shutova E. A Report on the 2018 VUA Metaphor Detection Shared Task[C]//Proceedings of the 2018 Workshop on Figurative Language Processing. Stroudsburg, PA, USA: Association for Computational Linguistics, 2018: 56-66.
[6] Neuman Y, Assaf D, Cohen Y, et al. Metaphor Identification in Large Texts Corpora[J]. PLoS One, 2013, 8(4):e62343.
doi: 10.1371/journal.pone.0062343
[7] Shutova E, Lin S, Korhonen A. Metaphor Identification Using Verb and Noun Clustering[C]//Proceedings of the 23rd International Conference on Computational Linguistics. Stroudsburg, PA, USA: Association for Computational Linguistics, 2010: 1002-1010.
[8] Klebanov B B, Leong C W, Flor M. Supervised Word-Level Metaphor Detection: Experiments with Concreteness and Reweighting of Examples[C]//Proceedings of the 3rd Workshop on Metaphor in NLP. Stroudsburg, PA, USA: Association for Computational Linguistics, 2015: 11-20.
[9] Gao G, Choi E, Choi Y, et al. Neural Metaphor Detection in Context[C]//Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. Stroudsburg, PA, USA: Association for Computational Linguistics, 2018: 607-613.
[10] Mao R, Lin C H, Guerin F. End-to-End Sequential Metaphor Identification Inspired by Linguistic Theories[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Stroudsburg, PA, USA: Association for Computational Linguistics, 2019: 3888-3898.
[11] Group P. MIP: A Method for Identifying Metaphorically Used Words in Discourse[J]. Metaphor and Symbol, 2007, 22(1):1-39.
doi: 10.1080/10926480709336752
[12] Wilks Y, Fass D. The Preference Semantics Family[J]. Computers & Mathematics with Applications, 1992, 23(2-5):205-221.
doi: 10.1016/0898-1221(92)90141-4
[13] Pennington J, Socher R, Manning C. GloVe: Global Vectors for Word Representation[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). Stroudsburg, PA, USA: Association for Computational Linguistics, 2014: 1532-1543.
[14] Peters M E, Neumann M, Iyyer M, et al. Deep Contextualized Word Representations[OL]. arXiv Preprint, arXiv:1802.05365.
[15] 苏传东, 黄孝喜, 王荣波, 等. 基于词嵌入融合和循环神经网络的中英文隐喻识别[J]. 数据分析与知识发现, 2020, 4(4):91-99.
[15] ( Su Chuandong, Huang Xiaoxi, Wang Rongbo, et al. Identifying Chinese/English Metaphors with Word Embedding and Recurrent Neural Network[J]. Data Analysis and Knowledge Discovery, 2020, 4(4):91-99.)
[16] Devlin J, Chang M, Lee K, et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding[OL]. arXiv Preprint, arXiv:1810.04805.
[17] 张冬瑜, 崔紫娟, 李映夏, 等. 基于Transformer和BERT的名词隐喻识别[J]. 数据分析与知识发现, 2020, 4(4):100-108.
[17] ( Zhang Dongyu, Cui Zijuan, Li Yingxia, et al. Identifying Noun Metaphors with Transformer and BERT[J]. Data Analysis and Knowledge Discovery, 2020, 4(4):100-108.)
[18] Li S Q, Zeng J J, Zhang J H, et al. ALBERT-BiLSTM for Sequential Metaphor Detection[C]//Proceedings of the 2nd Workshop on Figurative Language Processing. Stroudsburg, PA, USA: Association for Computational Linguistics, 2020: 110-115.
[19] Gong H Y, Gupta K, Jain A, et al. IlliniMet: Illinois System for Metaphor Detection with Contextual and Linguistic Information[C]//Proceedings of the 2nd Workshop on Figurative Language Processing. Stroudsburg, PA, USA: Association for Computational Linguistics, 2020: 146-153.
[20] Liu Y H, Ott M, Goyal N, et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach[OL]. arXiv Preprint, arXiv: 1907. 11692.
[21] Liu X, Luo Z C, Huang H Y. Jointly Multiple Events Extraction via Attention-based Graph Information Aggregation[C]//Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. 2020:1247-1256.
[22] Nguyen T, Grishman R. Graph Convolutional Networks with Argument-Aware Pooling for Event Detection[C]//Proceedings of the 2019 AAAI Conference on Artificial Intelligence. 2019:7370-7377.
[23] Song L, Wang Z, Yu M, et al. Exploring Graph-Structured Passage Representation for Multi-Hop Reading Comprehension with Graph Neural Networks[OL]. arXiv Preprint, arXiv: 1809. 02040.
[24] Su C D, Fukumoto F, Huang X X, et al. DeepMet: A Reading Comprehension Paradigm for Token-Level Metaphor Detection[C]//Proceedings of the 2nd Workshop on Figurative Language Processing. Stroudsburg, PA, USA: Association for Computational Linguistics, 2020: 30-39.
[25] Vaswani A, Shazeer N, Parmar N, et al. Attention is All You Need[C]//Proceedings of the 31st Conference on Neural Information Processing Systems. 2017: 5998-6008.
[26] Leong C W B, Klebanov B B, Hamill C, et al. A Report on the 2020 VUA and TOEFL Metaphor Detection Shared Task[C]//Proceedings of the 2nd Workshop on Figurative Language Processing. Stroudsburg, PA, USA: Association for Computational Linguistics, 2020: 18-29.
[27] 朱嘉莹, 王荣波, 黄孝喜, 等. 基于Bi-LSTM的多层面隐喻识别方法[J]. 大连理工大学学报, 2020, 60(2):209-215.
[27] ( Zhu Jiaying, Wang Rongbo, Huang Xiaoxi, et al. Multi-Level Metaphor Detection Method Based on Bi-LSTM[J]. Journal of Dalian University of Technology, 2020, 60(2):209-215.)
[1] 范涛,王昊,吴鹏. 基于图卷积神经网络和依存句法分析的网民负面情感分析研究*[J]. 数据分析与知识发现, 2021, 5(9): 97-106.
[2] 张冬瑜,崔紫娟,李映夏,张伟,林鸿飞. 基于Transformer和BERT的名词隐喻识别*[J]. 数据分析与知识发现, 2020, 4(4): 100-108.
[3] 苏传东,黄孝喜,王荣波,谌志群,毛君钰,朱嘉莹,潘宇豪. 基于词嵌入融合和循环神经网络的中英文隐喻识别*[J]. 数据分析与知识发现, 2020, 4(4): 91-99.
[4] 黄孝喜, 李晗雨, 王荣波, 王小华, 谌志群. 基于卷积神经网络与SVM分类器的隐喻识别*[J]. 数据分析与知识发现, 2018, 2(10): 77-83.
[5] 黄孝喜, 张华, 陆蓓, 王荣波, 吴铤. 一种基于词语抽象度的汉语隐喻识别方法[J]. 现代图书情报技术, 2015, 31(4): 34-40.
[6] 石晶,张丽娟. 利用知网扩展内部-外部算法[J]. 现代图书情报技术, 2009, 25(7-8): 54-58.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
版权所有 © 2015 《数据分析与知识发现》编辑部
地址:北京市海淀区中关村北四环西路33号 邮编:100190
电话/传真:(010)82626611-6626,82624938
E-mail:jishu@mail.las.ac.cn