Please wait a minute...
Advanced Search
数据分析与知识发现  2022, Vol. 6 Issue (10): 68-78     https://doi.org/10.11925/infotech.2096-3467.2022.0009
  研究论文 本期目录 | 过刊浏览 | 高级检索 |
基于BERT和互信息的金融风险逻辑关系量化方法
贾明华1,2(),王秀利1,3
1中央财经大学信息学院 北京 102206
2北京大学图书馆 北京 100871
3国家金融安全教育部工程研究中心 北京 102206
Quantifying Logical Relations of Financial Risks with BERT and Mutual Information
Jia Minghua1,2(),Wang Xiuli1,3
1School of Information, Central University of Finance and Economics, Beijing 102206, China
2Peking University Library, Beijing 100871, China
3Engineering Research Center of State Financial Security, Ministry of Education, Beijing 102206, China
全文: PDF (1229 KB)   HTML ( 22
输出: BibTeX | EndNote (RIS)      
摘要 

【目的】 通过量化金融风险逻辑关系防控金融风险,同时处理金融事件词频量化不可靠问题。【方法】 提出一种结合领域知识的基于BERT和互信息的金融风险逻辑关系量化分析方法,并在通用数据集COPA和金融领域数据集上进行关系量化。【结果】 基于BERT和互信息能够有效解决词频量化不可靠问题,该方法在金融风险逻辑关系量化的准确率达到80.1%,较对比方法提升了3.1%~37.4%。【局限】 仅考虑了金融领域的语料,在非金融等其他语料上的效果有待检验。【结论】 所提方法能够揭示金融风险事件的演化路径,改善金融风险逻辑关系量化的效果。

服务
把本文推荐给朋友
加入引用管理器
E-mail Alert
RSS
作者相关文章
贾明华
王秀利
关键词 金融风险关系量化领域知识BERT互信息    
Abstract

[Objective] This paper tries to prevent and control financial risks by quantifying their logical relationship, which also improve the reliability of processing word frequency of financial events. [Methods] We proposed a quantitative analysis method for the logical relation of financial risks based on BERT and mutual information combined with domain knowledge. Then, we quantified the relations with COPA and financial data sets. [Results] The proposed model effectively addressed the issue of unreliable quantization of word frequency. Its accuracy reached 80.1%, which was 3.1%~37.4% higher than the benchmark models. [Limitations] More research is needed to examine our new model with non-financial and other corpora. [Conclusions] Our new method can reveal the evolutionary path of financial risk events and improve the effect quantitative presentation of their logical relationship.

Key wordsFinancial Risk    Relationship Quantization    Domain Knowledge    BERT    Mutual Information
收稿日期: 2022-01-05      出版日期: 2022-11-16
ZTFLH:  TP391  
通讯作者: 贾明华,ORCID:0000-0003-0859-7502      E-mail: 2020212349@email.cufe.edu.cn
引用本文:   
贾明华, 王秀利. 基于BERT和互信息的金融风险逻辑关系量化方法[J]. 数据分析与知识发现, 2022, 6(10): 68-78.
Jia Minghua, Wang Xiuli. Quantifying Logical Relations of Financial Risks with BERT and Mutual Information. Data Analysis and Knowledge Discovery, 2022, 6(10): 68-78.
链接本文:  
https://manu44.magtech.com.cn/Jwk_infotech_wk3/CN/10.11925/infotech.2096-3467.2022.0009      或      https://manu44.magtech.com.cn/Jwk_infotech_wk3/CN/Y2022/V6/I10/68
模型 核心思想 细分模型
BERT[17] 采用Transformer编码器,包含编码器(Encoder)和解码器(Decoder)两部分 BERT-base-uncased
BERT-large-uncased
XLNet[18] 改进Transformer结构为Transformer-XL XLNet-base-cased
XLNet-base-cased
RoBERTa[19] 沿用BERT基础模型,优化掩藏语言模型[20] RoBERTa-base
RoBERTa-large
ERNIE[21-22] 优化掩藏语言模型和相邻句预测 ERNIE(Baidu)
ERNIE(Tsinghua)
ALBERT[23] 优化相邻句预测 ALBERT-base
ALBERT-large
Table 1  常见BERT模型对比
Fig.1  基于BERT和互信息的事件关系量化模型
Fig. 2  BERT预训练语言模型
Fig. 3  BERT的Embedding表示
Fig.4  Transformer编码单元
类型 文本内容
Premise: The man broke his toe. What was the CAUSE of this?
Alternative 1: He got a hole in his sock.
Alternative 2: He dropped a hammer on his foot.
Premise: I tipped the bottle. What happened as a RESULT?
Alternative 1: The liquid in the bottle froze.
Alternative 2: The liquid in the bottle poured out.
Premise: I knocked on my neighbor's door. What happened as a RESULT?
Alternative 1: My neighbor invited me in.
Alternative 2: My neighbor left his house.
Table 2  COPA数据示例
编号 抽象主题事件 泛化事件数 结果事件数 原因事件数
E1 货币超发 3 10 10
E2 股市大跌 3 10 10
E3 美联储加息 3 10 10
E4 人民币升值 3 10 10
E5 人民币贬值 3 10 10
E6 中美贸易摩擦 3 10 3
E7 英国脱欧 3 10 1
E8 股市上涨 3 10 10
Table 3  抽象主题事件
编号 一因多果 由果溯因
方法A 方法B 方法A 方法B
E1 0.73 1.00 0.59 1.00
E2 0.73 1.00 0.75 1.00
E3 0.95 1.00 0.92 1.00
E4 0.79 1.00 0.85 1.00
E5 0.88 1.00 0.79 1.00
E6 0.86 1.00 0.37 1.00
E7 0.59 1.00 0.04 1.00
E8 0.58 1.00 0.69 1.00
Table 4  关系量化值结果对比
Fig.5  一因多果关系量化值分布
Fig. 6  由果溯因关系量化值分布
模型 参数量 模型
层数
隐层
大小
批大小 词表
大小
BERT-base-uncased 108M 12 768 16 30 522
BERT-large-uncased 334M 24 1 024 4 30 522
RoBERTa-base 123M 12 768 16 30 522
RoBERTa-large 355M 24 1 024 4 50 265
ALBERT-base 12M 12 768 32 30 000
ALBERT-large 18M 24 1 024 12 30 522
Table 5  BERT模型参数设置
实验方法 Test Set Dev Set Dev + Test
协方差* 50.2% 49.0% 49.6%
共现频率 50.0% 51.8% 50.9%
互信息 57.8% 58.8% 58.3%
BERT-base-uncased+PMI 58.2% 62.0% 60.1%
BERT-large-uncased+PMI 71.6% 68.6% 70.1%
RoBERTa-base+PMI 71.4% 76.8% 74.1%
RoBERTa-large+PMI 68.8% 70.6% 69.7%
ALBERT-base+PMI 57.6% 58.4% 58.0%
ALBERT-large+PMI 78.8% 81.4% 80.1%
Table 6  在COPA数据集上关系量化推理任务的准确率
[1] Singhal A. Introducing the Knowledge Graph: Things, Not Strings[EB/OL]. (2012-05-16). [2020-03-01]. https://www.blog.google/products/search/introducing-knowledge-graph-things-not/.
[2] 刘宗田, 黄美丽, 周文, 等. 面向事件的本体研究[J]. 计算机科学, 2009, 36(11): 189-192.
[2] (Liu Zongtian, Huang Meili, Zhou Wen, et al. Research on Event-oriented Ontology Model[J]. Computer Science, 2009, 36(11): 189-192.)
[3] Lee S. Simulation Modeling with Event Graphs[J]. Communications of the ACM, 1983, 26(11): 957-963.
doi: 10.1145/182.358460
[4] Buss A H. Modeling with Event Graphs[C]// Proceedings of the 28th Conference on Winter Simulation. 1996: 153-160.
[5] Yang C C, Shi X D. Discovering Event Evolution Graphs from Newswires[C]// Proceedings of the 15th International Conference on World Wide Web. 2006: 945-946.
[6] Yang C C, Shi X D, Wei C P. Discovering Event Evolution Graphs from News Corpora[J]. IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans, 2009, 39(4): 850-863.
doi: 10.1109/TSMCA.2009.2015885
[7] Li Z Y, Zhao S D, Ding X, et al. EEG: Knowledge Base for Event Evolutionary Principles and Patterns[C]// Proceedings of the 6th National Conference on Social Media Processing. 2017: 40-52.
[8] Li Z Y, Ding X, Liu T. Constructing Narrative Event Evolutionary Graph for Script Event Prediction[C]// Proceedings of the 27th International Joint Conference on Artificial Intelligence. 2018: 4201-4207.
[9] Ding X, Li Z Y, Liu T, et al. ELG: An Event Logic Graph[OL].arXiv Preprint, arXiv: 1907.08015.
[10] 胡扬, 闫宏飞, 陈翀. 面向金融知识图谱的实体和关系联合抽取算法[J]. 重庆理工大学学报(自然科学), 2020, 34(5): 139-149.
[10] (Hu Yang, Yan Hongfei, Chen Chong. Joint Entity and Relation Extraction for Constructing Financial Knowledge Graph[J]. Journal of Chongqing University of Technology (Natural Science), 2020, 34(5): 139-149.)
[11] 李江龙, 吕学强, 周建设, 等. 金融领域的事件句抽取[J]. 计算机应用研究, 2017, 34(10) : 2915-2918.
[11] (Li Jianglong, Lyu Xueqiang, Zhou Jianshe, et al. Event Sentence Extraction in Financial Field[J]. Application Research of Computers, 2017, 34(10): 2915-2918.)
[12] Quinlan J R. C4.5: Programs for Machine Learning[M]. San Mateo, CA: Morgan Kaufmann, 1993.
[13] 程兴国, 肖南峰. 词类共现频率的MapReduce并行生成方法[J]. 重庆理工大学学报(自然科学), 2013, 27(11): 53-57.
[13] (Cheng Xingguo, Xiao Nanfeng. Parallel Implementation for Co-occurrence Statistics with MapReduce Model[J]. Journal of Chongqing University of Technology (Natural Science), 2013, 27(11): 53-57.)
[14] 钟茂生, 刘慧, 刘磊. 词汇间语义相关关系量化计算方法[J]. 中文信息学报, 2009, 23(2): 115-122.
[14] (Zhong Maosheng, Liu Hui, Liu Lei. Method of Semantic Relevance Relation Measurement Between Words[J]. Journal of Chinese Information Processing, 2009, 23(2): 115-122.)
[15] 黄进, 阮彤, 蒋锐权. 基于SVM结合依存句法的金融领域舆情分析[J]. 计算机工程与应用, 2015, 51(23): 230-235.
[15] (Huang Jin, Ruan Tong, Jiang Ruiquan. Sentiment Analysis in Financial Domain Based on SVM with Dependency Syntax[J]. Computer Engineering and Applications, 2015, 51(23): 230-235.)
[16] 张洪宽, 宋晖, 王舒怡, 等. 基于BERT的端到端中文篇章事件抽取[C]// 第19届中国计算语言学大会论文集. 2020: 390-401.
[16] (Zhang Hongkuan, Song Hui, Wang Shuiyi, et al. A BERT-Based End-to-End Model for Chinese Document-level Event Extraction[C]// Proceedings of the 19th Chinese National Conference on Computational Linguistics. 2020: 390-401.)
[17] Devlin J, Chang M W, Lee K, et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding[C]// Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies, 2019: 4171-4186.
[18] Yang Z L, Dai Z H, Yang Y M, et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding[C]// Proceedings of the 33rd International Conference on Neural Information Processing Systems. 2019: 5753-5763.
[19] Liu Y H, Ott M, Goyal N, et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach[OL]. arXiv Preprint, arXiv: 1907.11692.
[20] 刘欢, 张智雄, 王宇飞. BERT模型的主要优化改进方法研究综述[J]. 数据分析与知识发现, 2021, 5(1): 3-15.
[20] (Liu Huan, Zhang Zhixiong, Wang Yufei. A Review on Main Optimization Methods of BERT[J]. Data Analysis and Knowledge Discovery, 2021, 5(1): 3-15.)
[21] Sun Y, Wang S H, Li Y K, et al. ERNIE: Enhanced Representation Through Knowledge Integration[OL]. arXiv Preprint, arXiv: 1904.09223.
[22] Zhang Z Y, Han X, Liu Z Y, et al. ERNIE: Enhanced Language Representation with Informative Entities[C]// Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 2019: 1441-1451.
[23] Lan Z Z, Chen M D, Goodman S, et al. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations[C]// Proceedings of the 8th International Conference on Learning Representations. 2020: 1-17.
[24] Do Q X, Chan Y S, Roth D. Minimally Supervised Event Causality Identification[C]// Proceedings of the Conference on Empirical Methods in Natural Language Processing. 2011: 294-303.
[25] Hashimoto C, Torisawa K, Kloetzer J, et al. Generating Event Causality Hypotheses through Semantic Relations[C]// Proceedings of the 29th AAAI Conference on Artificial Intelligence. 2015: 2396-2403.
[26] Luo Z Y, Sha Y C, Zhu K Q, et al. Commonsense Causal Reasoning Between Short Texts[C]// Proceedings of the 15th International Conference on Principles of Knowledge Representation and Reasoning. 2016: 421-430.
[27] Staliūnaitė I, Gorinski P J, Iacobacci I. Improving Commonsense Causal Reasoning by Adversarial Training and Data Augmentation[C]// Proceedings of the 35th Conference on Innovative Applications of Artificial Intelligence. 2021: 13834-13842.
[28] Sharma A, Kiciman E. Causal Inference and Counterfactual Reasoning[C]// Proceedings of the 7th ACM IKDD CoDS and 25th COMAD. 2020: 369-370.
[29] Han M Y, Wang Y L.Doing Good or Doing Right? Exploring the Weakness of Commonsense Causal Reasoning Models[C]// Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing. 2021: 151-157.
[30] 张东东, 彭敦陆. ENT-BERT:结合BERT和实体信息的实体关系分类模型[J]. 小型微型计算机系统, 2020, 41(12): 2557-2562.
[30] (Zhang Dongdong, Peng Dunlu. ENT-BERT: Entity Relation Classification Model Combining BERT and Entity Information[J]. Journal of Chinese Computer Systems, 2020, 41(12): 2557-2562.)
[31] Bai C Y, Pan L M, Luo S L, et al. Joint Extraction of Entities and Relations by a Novel End-to-End Model with a Double-Pointer Module[J]. Neurocomputing, 2020, 377: 325-333.
doi: 10.1016/j.neucom.2019.09.097
[1] 施运梅, 袁博, 张乐, 吕学强. IMTS:融合图像与文本语义的虚假评论检测方法*[J]. 数据分析与知识发现, 2022, 6(8): 84-96.
[2] 吴江, 刘涛, 刘洋. 在线社区用户画像及自我呈现主题挖掘——以网易云音乐社区为例*[J]. 数据分析与知识发现, 2022, 6(7): 56-69.
[3] 郑洁, 黄辉, 秦永彬. 一种融合法律知识的相似案例匹配模型*[J]. 数据分析与知识发现, 2022, 6(7): 99-106.
[4] 景慎旗, 赵又霖. 基于医学领域知识和远程监督的医学实体关系抽取研究*[J]. 数据分析与知识发现, 2022, 6(6): 105-114.
[5] 潘慧萍, 李宝安, 张乐, 吕学强. 基于多特征融合的政府工作报告关键词提取研究*[J]. 数据分析与知识发现, 2022, 6(5): 54-63.
[6] 肖悦珺, 李红莲, 张乐, 吕学强, 游新冬. 特征融合的中文专利文本分类方法研究*[J]. 数据分析与知识发现, 2022, 6(4): 49-59.
[7] 杨林, 黄晓硕, 王嘉阳, 丁玲玲, 李子孝, 李姣. 基于BERT-TextCNN的临床试验疾病亚型识别研究*[J]. 数据分析与知识发现, 2022, 6(4): 69-81.
[8] 郭航程, 何彦青, 兰天, 吴振峰, 董诚. 基于Paragraph-BERT-CRF的科技论文摘要语步功能信息识别方法研究*[J]. 数据分析与知识发现, 2022, 6(2/3): 298-307.
[9] 张云秋, 汪洋, 李博诚. 基于RoBERTa-wwm动态融合模型的中文电子病历命名实体识别*[J]. 数据分析与知识发现, 2022, 6(2/3): 242-250.
[10] 王永生, 王昊, 虞为, 周泽聿. 融合结构和内容的方志文本人物关系抽取方法*[J]. 数据分析与知识发现, 2022, 6(2/3): 318-328.
[11] 谢星雨, 余本功. 基于MFFMB的电商评论文本分类研究*[J]. 数据分析与知识发现, 2022, 6(1): 101-112.
[12] 周泽聿,王昊,赵梓博,李跃艳,张小琴. 融合关联信息的GCN文本分类模型构建及其应用研究*[J]. 数据分析与知识发现, 2021, 5(9): 31-41.
[13] 陈杰,马静,李晓峰. 融合预训练模型文本特征的短文本分类方法*[J]. 数据分析与知识发现, 2021, 5(9): 21-30.
[14] 马江微, 吕学强, 游新冬, 肖刚, 韩君妹. 融合BERT与关系位置特征的军事领域关系抽取方法*[J]. 数据分析与知识发现, 2021, 5(8): 1-12.
[15] 李文娜, 张智雄. 基于联合语义表示的不同知识库中的实体对齐方法研究*[J]. 数据分析与知识发现, 2021, 5(7): 1-9.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
版权所有 © 2015 《数据分析与知识发现》编辑部
地址:北京市海淀区中关村北四环西路33号 邮编:100190
电话/传真:(010)82626611-6626,82624938
E-mail:jishu@mail.las.ac.cn