簡易檢索 / 詳目顯示

研究生: 姚秉均
Yao, Bing-Jun
論文名稱: 基於圖神經網路之假新聞偵測研究
Fake News Detection based on Graph Neural Networks
指導教授: 陳柏琳
Chen, Berlin
口試委員: 曾厚強
Tseng, Hou-Chiang
陳冠宇
Chen, Kuan-Yu
洪志偉
Hung, Jeih-Weih
陳柏琳
Chen, Berlin
口試日期: 2023/07/21
學位類別: 碩士
Master
系所名稱: 資訊工程學系
Department of Computer Science and Information Engineering
論文出版年: 2023
畢業學年度: 111
語文別: 中文
論文頁數: 39
中文關鍵詞: 假新聞知識庫自然語言處理類神經網路
英文關鍵詞: Fake News, Knowledge Base, Natural Language Processing, Neural Network
研究方法: 實驗設計法
DOI URL: http://doi.org/10.6345/NTNU202301268
論文種類: 學術論文
相關次數: 點閱:109下載:9
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 在現今互聯網時代,隨著網路技術不斷發展,使得我們在閱覽資訊上也越來越便利,但與此同時,假新聞也藉著技術發展的順風車變得更容易生產傳播以及造成影響。所以本人便打算做和假新聞辨識相關的研究,找到較好的假新聞辨識的方法並提升假新聞辨識的準確率。由於看到CompareNet這一圖神經網路模型對於假新聞的辨識相較其他基礎方法效果更好。因此本研究是以CompareNet這一研究為基礎,基於LUN(Labeled Unreliable News Dataset)語料庫中的LUN-train語料庫創建了一個含有普通名詞、複數名詞、專有名詞、動詞、形容詞、副詞的知識庫,並將該知識庫和CompareNet這一研究相結合,使用LUN語料庫中的LUN-train語料庫來訓練模型、使用SLN(Satirical and Legitimate News Database)以及LUN語料庫中的LUN-test語料庫來對模型進行測試,提升假新聞辨識的準確率。

    In the current internet era, with the continuous development of internet technology, accessing information has become increasingly convenient. However, at the same time, fake news has taken advantage of this technological advancement, making it easier to produce and spread, causing significant impacts. Therefore, I intend to conduct research related to fake news detection to find better methods for identifying fake news and improve the accuracy of fake news detection.
    After observing that the CompareNet graph neural network model has shown better results in fake news detection compared to other baseline methods, my work is based on the CompareNet study. I created a knowledge base containing common nouns, plural nouns, proper nouns, verbs, adjectives, and adverbs based on the LUN-train corpus (Labeled Unreliable News Dataset). Then, my work integrated this knowledge base with the CompareNet research. Trained the model using the LUN-train corpus and tested it using the SLN (Satirical and Legitimate News Database) and the LUN-test corpus from the LUN dataset to enhance the accuracy of fake news detection.

    目錄 第一章 前言............................................................................................... 1 1.1緒論 ....................................................................................................................... 1 1.2研究動機與目的 ..................................................................................................... 2 1.3研究概述以及論文結構 .......................................................................................... 2 第二章 相關研究 ...................................................................................... 4 2.1 假新聞的定義以及偵測方式 ................................................................................. 4 2.2 圖神經網路 ........................................................................................................... 8 2.2.1 CompareNet: .................................................................................................. 8 2.2.2 基於圖的神經網路模型: ............................................................................. 12 2.3 知識庫與知識表達 .............................................................................................. 12 第三章 研究方法 .................................................................................... 14 3.1 研究流程 ............................................................................................................. 14 3.2 資料處理 ............................................................................................................. 15 3.3 知識庫與分類模型 .............................................................................................. 16 3.4 研究限制 ............................................................................................................. 17 第四章 實驗結果與分析 ........................................................................ 18 4.1 資料處理 ............................................................................................................. 18 4.1.1 斷詞處理 ...................................................................................................... 18 4.1.2 去不需用字及取詞語 .................................................................................... 19 4.1.3 取詞性 .......................................................................................................... 19 4.2 模型訓練分析 ..................................................................................................... 20 4.2.1 知識庫結構 ................................................................................................... 20 4.2.2 知識庫文字雲 ............................................................................................... 21 4.2.3 模型分析 ...................................................................................................... 29 4.2.4 方法比較 ...................................................................................................... 32 4.3 小結 .................................................................................................................... 34 第五章 結論與未來展望 ........................................................................ 36 參考文獻 ................................................................................................... 37

    [1] Hunt Allcott and Matthew Gentzkow. Social media and fake news in the 2016 election. Journal of economic perspectives, 31(2):211–236, 2017.
    [2] Clint Burfoot and Timothy Baldwin. Automatic satire detection: Are you having a laugh? In Proceedings of the ACL-IJCNLP 2009 conference short papers, pages 161–164, 2009.
    [3] Claire Wardle. https://firstdraftnews.org/articles/fake-news-complicated/ , 2017.
    [4] IFLA. How to Spot Fake News, 2017.
    [5] Hannah Rashkin, Eunsol Choi, Jin Yea Jang, Svitlana Volkova, Yejin Choi. Truth of Varying Shades: Analyzing Language in Fake News and Political Fact-Checking. Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pages 2931–2937, 2017.
    [6] Zhiwei Jin, Juan Cao, Yongdong Zhang, and Jiebo Luo. News verification by exploiting conflicting social viewpoints in microblogs. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, pages 2972–2978, 2016.
    [7] Lianwei Wu, Yuan Rao, Haolin Jin, Ambreen Nazir, and Ling Sun. Different absorption from the same sharing: Sifted multi-task learning for fake news detection. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, pages 4643–4652, 2019.
    [8] Jiawei Zhang, Bowen Dong, and Philip S. Yu. Fakedetector: Effective fake news detection with deep diffusive neural network. In In Proceedings of the 36th IEEE International Conference on Data Engineering, pages 1826–1829. 2020.
    [9] A´ lvaro Ibrain Rodr´ıguez and Lara Lloret Iglesias. Fake news detection using deep learning. CoRR, abs/1910.03496, 2019.
    [10] Vaibhav Vaibhav, Raghuram Mandyam, and Eduard Hovy. Do sentence interactions matter? leveraging sentence level representations for fake news classification. In Proceedings of the Thirteenth Workshop on Graph-Based Methods for Natural Language Processing (TextGraphs-13), pages 134–139, 2019.
    [11] Jeff Z. Pan, Siyana Pavlova, Chenxi Li, Ningxi Li, Yangmei Li, and Jinshuo Liu. Content based fake news detection using knowledge graphs. In The Semantic Web - ISWC 2018 - 17th International Semantic Web Conference, volume 11136, pages 669–683, 2018.
    [12] William Yang Wang. “liar, liar pants on fire”: A new benchmark dataset for fake news detection. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 422–426, 2017.
    [13] Dhruv Khattar, Jaipal Singh Goud, Manish Gupta, and Vasudeva Varma. Mvae: Multimodal variational autoencoder for fake news detection. In The World Wide Web Conference, page 2915–2921, 2019.
    [14] Youze Wang, Shengsheng Qian, Jun Hu, Quan Fang, and Changsheng Xu. Fake news detection via knowledge-driven multimodal graph convolutional networks. In Proceedings of the 2020 International Conference on Multimedia Retrieval, pages 540–547, 2020.
    [15] Yoon Kim. Convolutional neural networks for sentence classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1746–1751, 2014.
    [16] Sepp Hochreiter and J¨urgen Schmidhuber. Long short-term memory. Neural computation, 9(8):1735–1780, 1997.
    [17] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics, 2019.
    [18] Thomas N. Kipf and Max Welling. Semisupervised classification with graph convolutional networks. In International Conference on Learning Representations (ICLR), 2017.
    [19] Petar Velikovi, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Li, and Yoshua Bengio. Graph attention networks. In International Conference on Learning Representations, 2018.
    [20] Hannah Rashkin, Eunsol Choi, Jin Yea Jang, Svitlana Volkova, and Yejin Choi. Truth of varying shades: Analyzing language in fake news and political fact-checking. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pages 2931–2937, 2017.
    [21] David M Blei, Andrew Y Ng, and Michael I Jordan. Latent dirichlet allocation. Journal of machineLearning research, 3(Jan): 993–1022, 2003.
    [22] https://sobigdata.d4science.org/group/tagme/ , 2010.
    [23] Linmei Hu, Tianchi Yang, Chuan Shi, Houye Ji, and Xiaoli Li. Heterogeneous graph attention networks for semi-supervised short text classification. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, pages 4821–4830, 2019.
    [24] Antoine Bordes, Nicolas Usunier, Alberto Garcia-Duran, Jason Weston, and Oksana Yakhnenko. Translating embeddings for modeling multirelational data. In Advances in Neural Information Processing Systems 26, pages 2787–2795, 2013.
    [25] Vaibhav Vaibhav, Raghuram Mandyam, and Eduard Hovy. Do sentence interactions matter? leveraging sentence level representations for fake news classification. In Proceedings of the Thirteenth Workshop on Graph-Based Methods for Natural Language Processing (TextGraphs-13), pages 134–139, 2019.
    [26] Mohammad Azad¬, Igor Chikalov, Mikhail Moshkov. Representation of Knowledge by Decision Trees for Decision Tables with Multiple Decisions. Procedia Computer Science 176 (2020) pages 653–659, 2020.
    [27] Asta Margien˙e, Simona Ramanauskait˙e 1, Justas Nugaras 2 and Pavel Stefanoviˇc, “Automated Transformation from Competency List to Tree: Way to Competency-Based Adaptive Knowledge E-Evaluation,” Applied Sciences 12(3), DOI:10.3390/app12031582, 2022.
    [28] W.J. Wilbur, K.J.J.o.i.s. Sirotkin. “The automatic identification of stop words,” 18 (1), pp. 45-55, 1992.
    [29] Victoria Rubin, Niall Conroy, Yimin Chen, and Sarah Cornwell. 2016. Fake news or truth? using satirical cues to detect potentially misleading news. In Proceedings of the Second Workshop on Computational Approaches to Deception Detection, pages 7–17, 2016.
    [30] Neutrino3316. https://github.com/fxsjy/jieba , 2020.

    下載圖示
    QR CODE