|
|
Causal Relation Extraction of Uyghur Events Based on Bidirectional Long Short-term Memory Model |
TIAN Shengwei① ZHOU Xingfa① YU Long② FENG Guanjun③ Aishan WUMAIER④ LI Pu⑤ |
①(School of Software, Xinjiang University, Urumqi 830046, China)
②(Net Center, Xinjiang University, Urumqi 830046, China)
③(College of Humanities, Xinjiang University, Urumqi 830046, China)
④(School of Software, Xinjiang University, Urumqi 830046, China)
⑤(School of Languages, Xinjiang University, Urumqi 830046, China) |
|
|
Abstract Since the traditional events causal relation has the disadvantages of small recognition coverage, a method for causal relation extraction of Uyghur events is presented based on Bidirectional Long Short-Term Memory (BiLSTM) model. In order to make full use of the event structure information, 10 characteristics of the Uyghur events structure information are extracted based on the study of the events causal relationship and Uyghur language features; At the same time, the word embedding is introduced as the input of BiLSTM to extract the deep semantic features of the Uyghur events and Batch Normalization (BN) algorithm is usded to accelerate the convergence of BiLSTM. Finally, concatenating these two kinds of features as the input of the softmax classifier to extract the Uyghur events causal relations. This method is used in the causal relation extraction of Uyghur events, and the results show that the precision rate, the recall rate and F value can reach 89.19 %, 83.19% and 86.09 %, indicating the effectiveness and practicability of the method of causal relation extraction of Uyghur events.
|
Received: 02 May 2017
Published: 14 September 2017
|
|
Fund:The National Natural Science Foundation of China (61662074, 61563051, 61262064), The Key Project of National Natural Science Foundation of China (61331011), Xinjiang Uygur Autonomous Region Scientific and Technological Personnel Training Project (QN2016YX0051) |
Corresponding Authors:
YU Long
E-mail: yul_xju@163.com
|
|
|
|
[1] |
SHA L, LI S J, CHANG B B, et al. Joint learning teplates and slots for event schema induction[C]. Proceedings of the North American Chapter of the Association for Computational Linguistics-Human Language Technologies, California, USA, 2016: 428-434. doi: 10.18653/v1/N16-1049.
|
[2] |
CHEN C and NG V. Joint modeling for chinese event extraction with rich linguistic features[C]. Proceedings of COLING 2012, Mumbai, India, 2012: 529-544.
|
[3] |
CHEN Y, XU L, LIU K, et al. Event extraction via dynamic multi-pooling convolutional neural networks[C]. Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, Beijing, China, 2015: 167-176. doi: 10.3115/v1/P15-1017.
|
[4] |
ZHANG Y J, LIU Z T, and ZHOU W, Event recognition based on deep learning in Chinese texts[J]. Plos One, 2016, 11(8): e0160147. doi: 10.1371/journal.pone.0160147.
|
[5] |
CHANG C Y, TENG Z Y and ZHANG Y. Expectation- regulated neural model for event mention extraction[C]. Proceedings of NAACL-HLT, California, USA, 2016: 400-410. doi: 10.18653/v1/N16-1045.
|
[6] |
干红华, 潘云鹤. 一种基于事件的因果关系的结构分析方法[J]. 模式识别与人工智能, 2003, 16(1): 56-62. doi: 10. 3969/j.issn.1003-6059.2003.01.011.
|
|
Gan Honghua and Pan Yunhe. A new analysis of the structure of event causation[J]. Pattern Recognition and Artiflcial Intelligence, 2003, 16(1): 56-62. doi: 10.3969/j.issn. 1003-6059.2003.01.011.
|
[7] |
GIRJU R and MOLDOVAN D. Text mining for causal relations[C]. Proceedings of the 15th International Florida Artiflcial Intelligence Research Society Conference, Florida, USA, 2002: 360-364.
|
[8] |
MARCU D and ECHIHABI A. An unsupervised approach to recognizing discourse relations[C]. Proceedings of the 40th Annual Meeting on Association for Computational Linguistics, Philadelphia, USA, 2002: 368-375. doi: 10.3115/ 1073083.1073145.
|
[9] |
SORGENTE A, VETTIGLI G, and MELE F. Automatic extraction of cause-effect relations in natural language text[C]. Proceedings of the 13th Conference of the Italian Association for Artificial Intellgence, Rome, Italian, 2013: 37-48.
|
[10] |
付剑锋, 刘宗田, 刘炜, 等. 基于层叠条件随机场的事件因果关系抽取[J]. 模式识别与人工智能, 2011, 24(4): 567-573. doi: 10.3969/j.issn.1003-6059.2011.04.016.
|
|
FU Jianfeng, LIU Zongtian, LIU Wei, et al. Event causal relation extraction based on cascaded conditional random flelds[J]. Pattern Recognition and Artiflcial Intelligence, 2011, 24(4): 567-573. doi: 10.3969/j.issn.1003-6059.2011.04.016.
|
[11] |
钟军, 禹龙, 田生伟, 等. 基于双层模型的维语突发事件因果关系抽取[J]. 自动化学报, 2014, 40(4): 771-779. doi: 10.3724/ SP.J.1004.2013.00771.
|
|
Zhong Jun, Yu Long, Tian Shengwei, et al. Causal relation extraction of uyghur emergency events based on cascaded model[J]. Acta Automatica Sinica, 2014, 40(4): 771-779. doi: 10.3724/SP.J.1004.2013.00771.
|
[12] |
TANG D, QIN B, FENG X, et al. Effective LSTMs for target-dependent sentiment classification[C]. Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, Osaka, Japan, 2016: 3298-3307.
|
[13] |
ZHOU P, SHI W, TIAN J, et al. Attention-based bidirectional long short-term memory networks for relation classification[C]. Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Berlin, Germany, 2016: 207-212. doi: 10.18653/v1/P16-2034.
|
[14] |
贺宇, 潘达, 付国宏. 基于自动编码特征的汉语解释性意见句识别[J]. 北京大学学报(自然科学版), 2015, 51(2): 235-240. doi: 10.13209/J.0479-8023.2015.041.
|
|
HE Yu, PAN Da and FU Guohong. Chinese explanatory opinionated sentence recognition based on autoEncoding features[J]. Acta Scientiarum Naturalium Universitatis Pekinensis, 2015, 51(2): 235-240. doi: 10.13209/J.0479-8023. 2015.041.
|
[15] |
MIKOLOV T, SUTSKEVER I, CHEN K, et al. Distributed representations of words and phrases and their compositionality[C]. Proceedings of Advances in Neural Information Processing Systems, Vancouve, Canada, 2013: 3111-3119.
|
[16] |
FENG X C, HUANG L F, TANG D Y, et al. A language-independent neural network for event detection[C]. Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Berlin, Germany, 2016: 66-71. doi: 10.18653/v1/P16-2011.
|
[17] |
SUTSKEVER I, VINYALS O, and LE Q V. Sequence to sequence learning with neural networks[C]. Proceedings of Advances in Neural Information Processing Systems, Quebec, Canada, 2014: 3104-3112.
|
[18] |
IOFFE S and SZEGEDY C. Batch normalization: Accelerating deep network training by reducing internal covariate shift[C]. Proceedings of the 32th International Conference on Machine Learning, Lille, France, 2015: 448-456.
|
|
|
|