Author: |
江明潔 Chiang, Ming-Chieh |
---|---|
Thesis Title: |
以遠程監督式學習從中文文本進行關係自動擷取 Automatic Relation Extraction from a Chinese Corpus through Distant-supervised Learning |
Advisor: |
柯佳伶
Koh, Jia-Ling |
Degree: |
碩士 Master |
Department: |
資訊工程學系 Department of Computer Science and Information Engineering |
Thesis Publication Year: | 2020 |
Academic Year: | 108 |
Language: | 中文 |
Number of pages: | 66 |
Keywords (in Chinese): | 關係擷取 、雙向長短期記憶模型 、多維注意力機制 、回饋學習機制 |
Keywords (in English): | relation extraction, bi-LSTM model, multi-level structured attention, feedback for learning |
DOI URL: | http://doi.org/10.6345/NTNU202000093 |
Thesis Type: | Academic thesis/ dissertation |
Reference times: | Clicks: 120 Downloads: 30 |
Share: |
School Collection Retrieve National Library Collection Retrieve Error Report |
本論文研究從中文文本進行關係擷取,以類神經網路架構為基礎,採用遠程監督式學習的概念,預測文本句子中是否具有特定關係,並擷取出句子中具有此特定關係的實體詞配對。本論文將詞嵌入向量和詞性嵌入向量作為模型輸入特徵,分別訓練關係偵測模型和鑑別模型,前者使用具有時序性的雙向長短期記憶網路模型,用來預測文本句子中是否具有特定關係,並在模型中使用多維注意力機制,針對符合某特定關係的句子們找出句子中相對重要的字作為候選實體詞;後者使用實體詞配對之向量差,利用鑑別模型輸出該實體詞配對是否具有某特定關係。經過上述兩個模型得到的結果,透過回饋學習機制,增加關係偵測模型的訓練資料,並調整關係偵測模型的訓練參數以提升關係分類效果。
In this paper, we study the problem of relation extraction from a Chinese corpus through distant-supervised learning. We constructed two models based on the recurrent neural networks to solve the problem. The two models use the word embedding and POS embedding as inputs. The first one is the relation detection model, which detects the relation of a sentence and selects the candidate entity words with multi-level structured (2-D matrix) attention mechanism. The candidate entity words will be combined to be entity pairs, which are inputted to the discriminative model. The second one is the discriminative model, which uses the vector difference of an entity pair to determine if an entity pair satisfies a relation. The results of the discriminative model can find more entity pairs of relations. These pairs can be used as additional training data of the relation detection model to improve the performance of the relation detection model through the feedback for learning.
[1] E. Agichtein, L. Gravano. (2000). Snowball: Extracting Relations from Large Plain-Text Collections. In Proceedings of the 5th ACM International Conference on Digital Libraries.
[2] A. Bordes, N. Usunier, A. Garcia-Dur´an. (2013). Translating Embeddings for Modeling Multi-relational Data. In Proceedings of the 2013 Neural Information Processing Systems Conference. (NIPS 2013)
[3] S. Brin. (1998). Extracting patterns and relations from the World-Wide Web. In Proceedings of the 1998 International Workshop on the Web and Databases (WebDB’98)
[4] B. Chiang. (2018). Automatic Detection of User’s Query Intentions for Community Question Answering. In Department of Computer Science and Information Engineering, National Taiwan Normal University.
[5] R. Hoffmann, C. Zhang, X. Ling, L. Zettlemoyer, and D.S. Weld. (2011). Knowledge-Based Weak Supervision for Information Extraction of Overlapping Relations. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics.
[6] N. Kalchbrenner, E. Grefenstette, and P. Blunsom. (2014). A Convolutional Neural Network for Modelling Sentences. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics.
[7] Y. Kim. (2014). Convolutional Neural Networks for Sentence Classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP).
[8] D. P. Kingma, and M. Welling. (2014). Stochastic Gradient VB and the Variational Auto-encoder. In Proceedings of the 2nd International Conference on Learning Representations (ICLR).
[9] Z. Linz, M. Feng, C. Nogueira dos Santos, M. Yu, B. Xiang, B. Zhou & Y. Bengiozy. (2017). A Structured Self-Attentive Sentence Embedding. In Proceedings of the 5th International Conference on Learning Representations (ICLR).
[10] Y. Lin, S. Shen, Z. Liu, H. Luan, M. Sun. (2016). Neural Relation Extraction with Selective Attention over Instances. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics.
[11] M.Luong, H.Pham, and C.Manning. (2015). Effective Approach to Attention-based Neural Machine. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP).
[12] T. Mikolov, K. Chen, G. Corrado, J. Dean. (2013). Efficient Estimation of Word Representations in Vector Space. In arXiv:1301.3781 [cs.CL].
[13] M. Mintz, S. Bills, R. Snow, and D. Jurafsky. (2009). Distant supervision for relation extraction without labeled data. In Proceedings of the 47th Annual Meeting of the ACL and the 4th IJCNLP of the AFNLP.
[14] M. Qu, X. Ren, Y. Zhang , and J. Han. (2018). Weakly-supervised Relation Extraction by Pattern-enhanced Embedding Learning. In Proceedings of the 27th International Conference on World Wide Web (WWW).
[15] S. Riedel, L. Yao, and A. McCallum. (2010). Modeling Relations and Their Mentions without Labeled Text. In Proceedings of ECML PKDD.
[16] D. Zelenko, C. Aone, A. Richardella. (2003). Kernel Methods for Relation Extraction. In the Journal of Machine Learning Research.
[17] D. Zeng, K. Liu, Y. Chen, and J. Zhao. (2015). Distant Supervision for Relation Extraction via Piecewise Convolutional Neural Networks. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP).
[18] C. Zhang, Y. Li, N. Du, W. Fan, P. S. Yu. (2018). On the Generative Discovery of Structured Medical Knowledge. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD).
[19] D. Zhang, D. Wang. (2015). Relation Classification via Recurrent Neural Network. arXiv:1508.01006.
[20] P. Zhou, W. Shi, J. Tian, Z. Qi, B. Li, H. Hao, B. Xu. (2016). Attention-Based Bidirectional Long Short-Term Memory Networks for Relation Classification. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics.
[21] G. Zhou, J. Su, J. Zhang, M. Zhang. (2005). Exploring Various Knowledge in Relation Extraction. In Proceedings of the 43rd Annual Meeting of the ACL.