Abstract:Since the traditional events causal relation has the disadvantages of small recognition coverage, a method for causal relation extraction of Uyghur events is presented based on Bidirectional Long Short-Term Memory (BiLSTM) model. In order to make full use of the event structure information, 10 characteristics of the Uyghur events structure information are extracted based on the study of the events causal relationship and Uyghur language features; At the same time, the word embedding is introduced as the input of BiLSTM to extract the deep semantic features of the Uyghur events and Batch Normalization (BN) algorithm is usded to accelerate the convergence of BiLSTM. Finally, concatenating these two kinds of features as the input of the softmax classifier to extract the Uyghur events causal relations. This method is used in the causal relation extraction of Uyghur events, and the results show that the precision rate, the recall rate and F value can reach 89.19 %, 83.19% and 86.09 %, indicating the effectiveness and practicability of the method of causal relation extraction of Uyghur events.
SHA L, LI S J, CHANG B B, et al. Joint learning teplates and slots for event schema induction[C]. Proceedings of the North American Chapter of the Association for Computational Linguistics-Human Language Technologies, California, USA, 2016: 428-434. doi: 10.18653/v1/N16-1049.
[2]
CHEN C and NG V. Joint modeling for chinese event extraction with rich linguistic features[C]. Proceedings of COLING 2012, Mumbai, India, 2012: 529-544.
[3]
CHEN Y, XU L, LIU K, et al. Event extraction via dynamic multi-pooling convolutional neural networks[C]. Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, Beijing, China, 2015: 167-176. doi: 10.3115/v1/P15-1017.
[4]
ZHANG Y J, LIU Z T, and ZHOU W, Event recognition based on deep learning in Chinese texts[J]. Plos One, 2016, 11(8): e0160147. doi: 10.1371/journal.pone.0160147.
[5]
CHANG C Y, TENG Z Y and ZHANG Y. Expectation- regulated neural model for event mention extraction[C]. Proceedings of NAACL-HLT, California, USA, 2016: 400-410. doi: 10.18653/v1/N16-1045.
Gan Honghua and Pan Yunhe. A new analysis of the structure of event causation[J]. Pattern Recognition and Artiflcial Intelligence, 2003, 16(1): 56-62. doi: 10.3969/j.issn. 1003-6059.2003.01.011.
[7]
GIRJU R and MOLDOVAN D. Text mining for causal relations[C]. Proceedings of the 15th International Florida Artiflcial Intelligence Research Society Conference, Florida, USA, 2002: 360-364.
[8]
MARCU D and ECHIHABI A. An unsupervised approach to recognizing discourse relations[C]. Proceedings of the 40th Annual Meeting on Association for Computational Linguistics, Philadelphia, USA, 2002: 368-375. doi: 10.3115/ 1073083.1073145.
[9]
SORGENTE A, VETTIGLI G, and MELE F. Automatic extraction of cause-effect relations in natural language text[C]. Proceedings of the 13th Conference of the Italian Association for Artificial Intellgence, Rome, Italian, 2013: 37-48.
FU Jianfeng, LIU Zongtian, LIU Wei, et al. Event causal relation extraction based on cascaded conditional random flelds[J]. Pattern Recognition and Artiflcial Intelligence, 2011, 24(4): 567-573. doi: 10.3969/j.issn.1003-6059.2011.04.016.
Zhong Jun, Yu Long, Tian Shengwei, et al. Causal relation extraction of uyghur emergency events based on cascaded model[J]. Acta Automatica Sinica, 2014, 40(4): 771-779. doi: 10.3724/SP.J.1004.2013.00771.
[12]
TANG D, QIN B, FENG X, et al. Effective LSTMs for target-dependent sentiment classification[C]. Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, Osaka, Japan, 2016: 3298-3307.
[13]
ZHOU P, SHI W, TIAN J, et al. Attention-based bidirectional long short-term memory networks for relation classification[C]. Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Berlin, Germany, 2016: 207-212. doi: 10.18653/v1/P16-2034.
HE Yu, PAN Da and FU Guohong. Chinese explanatory opinionated sentence recognition based on autoEncoding features[J]. Acta Scientiarum Naturalium Universitatis Pekinensis, 2015, 51(2): 235-240. doi: 10.13209/J.0479-8023. 2015.041.
[15]
MIKOLOV T, SUTSKEVER I, CHEN K, et al. Distributed representations of words and phrases and their compositionality[C]. Proceedings of Advances in Neural Information Processing Systems, Vancouve, Canada, 2013: 3111-3119.
[16]
FENG X C, HUANG L F, TANG D Y, et al. A language-independent neural network for event detection[C]. Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Berlin, Germany, 2016: 66-71. doi: 10.18653/v1/P16-2011.
[17]
SUTSKEVER I, VINYALS O, and LE Q V. Sequence to sequence learning with neural networks[C]. Proceedings of Advances in Neural Information Processing Systems, Quebec, Canada, 2014: 3104-3112.
[18]
IOFFE S and SZEGEDY C. Batch normalization: Accelerating deep network training by reducing internal covariate shift[C]. Proceedings of the 32th International Conference on Machine Learning, Lille, France, 2015: 448-456.