簡易檢索 / 詳目顯示

研究生: 王皓平
Wang, Hao-Ping
論文名稱: 運用類神經網路方法分析基於面向的情感極性分類
Aspect-based Sentiment Polarity Classification using Neural Network Methods
指導教授: 侯文娟
Hou, Wen-Juan
口試委員: 郭俊桔
Kuo, June-Jei
方瓊瑤
Fang, Chiung-Yao
侯文娟
Hou, Wen-Juan
口試日期: 2022/06/20
學位類別: 碩士
Master
系所名稱: 資訊工程學系
Department of Computer Science and Information Engineering
論文出版年: 2022
畢業學年度: 110
語文別: 中文
論文頁數: 48
中文關鍵詞: 自然語言處理雙向長短期記憶自注意力機制情感分析編碼器深度學習
英文關鍵詞: Natural language processing, Bi-directional Long Short-Term Memory, Self-attention mechanism, sentiment analysis, encoder, deep learning
研究方法: 主題分析比較研究文件分析法
DOI URL: http://doi.org/10.6345/NTNU202200793
論文種類: 學術論文
相關次數: 點閱:137下載:8
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 隨著時代以及科技技術的成長,人們不像過去一樣,需要查看報紙、購買雜誌、詢問左右鄰居的情報才能知道自己想要得知的資訊。在科技技術的成長下,不管是餐廳的評價、筆記型電腦的實用程度,大部分的人們都可以使用網際網路來查看是否有所想要的資訊。
    本論文使用的資料集由SemEval-2014 Task 4官方所提供,並且含有四項子任務:(一) Aspect term extraction、(二) Aspect term polarity、(三) Aspect category detection、(四) Aspect category polarity,本論文進行第二項子任務研究,判斷出句子中的面向詞是正面、負面或中立,評估方式採用Accuracy,並且與當年競賽結果相比較。
    本論文實驗方法將資料先進行前處理並且轉成詞向量作為輸入的來源,以及將極性做情感標籤,並且使用Bi-LSTM (Bi-directional Long Short-Term Memory)、Self-attention(自注意力機制)及使用Two-level encoding對資料進行訓練。
    最後去比對每種不同模型的準確率,結果顯示Two-level encoding預測準確率餐廳達82%,筆記型電腦則達78%。

    With the development of the times and technology, people are not the same as in the past, that they need to check newspapers, buy magazines, and ask their neighbors for information to know what they want to know. With the development of technology, most people can use the Internet to check whether they have the information they want, no matter it is restaurant reviews or the practicality of laptops.

    The dataset used in this study is officially provided by SemEval-2014 Task 4, which contains four subtasks: (1) Aspect term extraction, (2) Aspect term polarity, (3) Aspect category detection, (4) Aspect category polarity, This paper conducts the second sub-task study to judge whether the term words in the sentence are positive, negative or neutral.

    The experimental method of this study pre-processes the data and converts it into the word vector as the source of input, and uses polarity as emotional label. Then Bi-LSTM (Bi-directional Long Short-Term Memory), Self-attention and Two-level encoding models are used to train the data.

    Finally, to compare the accuracy of each different model, the results show that using the two-level encoding method, the prediction accuracy rate is 82% for restaurants, and 78% for laptops.

    第一章 緒論 1 第一節 研究動機 1 第二節 研究目的 2 第三節 論文架構 3 第二章 文獻探討 4 第一節 自然語言處理工具包(NLTK) 4 第二節 情感分析 5 第三節 長短期記憶模型 7 第四節 自注意力機制 8 第五節 條件隨機場 11 第三章 實驗方法與步驟 14 第一節 實驗資料 14 第二節 評估指標 15 第三節 實驗流程 16 第四節 資料整理 18 第五節 去除標點符號及停用詞 20 第六節 情感標籤 23 第七節 轉換為向量 24 第八節 雙向長短期記憶 28 第九節 Encoder-Decoder 29 第四章 實驗結果 32 第一節 開發環境 32 第二節 實驗結果 33 第三節 分析與討論 40 第五章 結論與未來展望 43 第一節 結論 43 第二節 未來研究方向 44 參考文獻 45

    Cortes, C., & Vapnik, V. (1995). Support-vector networks. Machine learning, 20(3), 273-297.

    Gao, Z., Feng, A., Song, X., & Wu, X. (2019). Target-dependent sentiment classification with BERT. Ieee Access, 7, 154290-154299.

    Hochreiter, S., & Schmidhuber, J. (1997). Long short-term memory. Neural computation, 9(8), 1735-1780.

    Kiritchenko, S., Zhu, X., Cherry, C., & Mohammad, S. (2014, August). Nrc-canada-2014: Detecting aspects and sentiment in customer reviews. In Proceedings of the 8th international workshop on semantic evaluation (SemEval 2014) (pp. 437-442).

    Lafferty, J., McCallum, A., & Pereira, F. C. (2001). Conditional random fields: Probabilistic models for segmenting and labeling sequence data.

    Lample, G., Ballesteros, M., Subramanian, S., Kawakami, K., & Dyer, C. (2016). Neural architectures for named entity recognition. arXiv preprint arXiv:1603.01360.

    Mikolov, T., Chen, K., Corrado, G., & Dean, J. (2013). Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781.

    Mikolov, T., Sutskever, I., Chen, K., Corrado, G. S., & Dean, J. (2013). Distributed representations of words and phrases and their compositionality. Advances in neural information processing systems, 26.

    Nguyen, H. T., & Le Nguyen, M. (2018, November). Effective attention networks for aspect-level sentiment classification. In 2018 10th International Conference on Knowledge and Systems Engineering (KSE) (pp. 25-30). IEEE.

    Pennington, J., Socher, R., & Manning, C. D. (2014, October). Glove: Global vectors for word representation. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP) (pp. 1532-1543).

    Pontiki, M., Galanis, D., Pavlopoulos, J., Papageorgiou, H., Androutsopoulos, I., & Manandhar, S. (2014). SemEval-2014 Task 4: Aspect based sentiment analysis. In Proceedings of the International Workshop on Semantic Evaluation, SemEval ’14, 27–35.

    Tang, D., Qin, B., Feng, X., & Liu, T. (2015). Effective LSTMs for target-dependent sentiment classification. arXiv preprint arXiv:1512.01100.

    Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... & Polosukhin, I. (2017). Attention is all you need. Advances in neural information processing systems, 30.

    Varsamopoulos, S., Bertels, K., & Almudever, C. G. (2018). Designing neural network based decoders for surface codes. arXiv preprint arXiv:1811.12456.

    Wagner, J., Arora, P., Cortes, S., Barman, U., Bogdanova, D., Foster, J., & Tounsi, L. (2014). Dcu: Aspect-based polarity classification for semeval task 4.

    Wallach, H. M. (2004). Conditional random fields: An introduction. Technical Reports (CIS), 22.

    Wagner, W. (2010). Steven bird, ewan klein and edward loper: Natural language processing with python, analyzing text with the natural language toolkit. Language Resources and Evaluation, 44(4), 421-424.

    Xie, J., Chen, B., Gu, X., Liang, F., & Xu, X. (2019). Self-attention-based BiLSTM model for short text fine-grained sentiment classification. IEEE Access, 7, 180558-180570.

    Xiang, J., Qiu, Z., Hao, Q., & Cao, H. (2020). Multi-time scale wind speed prediction based on WT-bi-LSTM. In MATEC Web of Conferences (Vol. 309, p. 05011). EDP Sciences.

    Zaremba, W., Sutskever, I., & Vinyals, O. (2014). Recurrent neural network regularization. arXiv preprint arXiv:1409.2329.

    Zhang, Y., Jin, R., & Zhou, Z. H. (2010). Understanding bag-of-words model: a statistical framework. International Journal of Machine Learning and Cybernetics, 1(1), 43-52.

    下載圖示
    QR CODE