研究生: |
簡文浩 Jian, Wen-Hau |
---|---|
論文名稱: |
基於3D人臉辨識之擴增實境技術改善臉盲症社交輔助系統 A Social Assistance System for Augmented Reality Technology to Redound Face Blindness with 3D Face Recognition |
指導教授: |
陳美勇
Chen, Wei-Yung |
口試委員: |
蘇順豐
Su, Shun-Feng 練光祐 Lian, Kuang-Yow 方瓊瑤 Fang, Chiung-Yao 陳美勇 Chen, Mei-Yung |
口試日期: | 2021/07/22 |
學位類別: |
碩士 Master |
系所名稱: |
機電工程學系 Department of Mechatronic Engineering |
論文出版年: | 2021 |
畢業學年度: | 109 |
語文別: | 中文 |
論文頁數: | 104 |
中文關鍵詞: | 臉盲症 、卷積神經網絡 、3D 可變模型 、人臉辨識 、擴增實境 |
英文關鍵詞: | Prosopagnosia, Convolutional Neural Network, 3D Morphable Model, Face Recognition, Augmented Reality |
研究方法: | 實驗設計法 |
DOI URL: | http://doi.org/10.6345/NTNU202101188 |
論文種類: | 學術論文 |
相關次數: | 點閱:140 下載:12 |
分享至: |
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
本論文目標為開發一套 AR(Augmented Reality,擴增實境)眼鏡輔助系 統,協助臉盲症患者在社交上、對於生活中他人的辨識。本研究主要貢獻為 提出以三維人臉模型作為人臉辨識之資料擴增基礎,滿足對於臉盲症患者實 際社交情境之實用性,並且將各軟體與硬體平台之優勢進行系統整合與設 計,實現可讓患者立即投入使用之社交輔助工具。具體架構包含以下內容: 第一,以結構光技術(Structured Light)結合立體視覺攝影機,經由 Structured Light 與2D RGB 輸入,2D 資料通過深度神經網路(Deep Neural Network)進行 人臉的提取,並確認三維空間中人臉之座標,運用深度學習將3D 點雲資訊 和2D 影像進行實時三維人臉密集重建,並取得人臉正面、側面 等7個角度 之人臉資訊,提高人臉對於側臉與大動態辨識的準確度。第二,藉由第一部 分產生之人臉資訊,輸入卷積神經網路進行運算,卷積神經網路以輸出128 維之特徵向量取代傳統高維分類器作為人臉特徵依據,將計算之特徵向量與 系統內 SQL(Structured Query Language)資料庫,進行歐式距離計算並比對, 取得最小歐式距離並對應該人臉的姓名資料。第三,將人臉標籤資訊、空間 中之人臉座標點,藉由相機模型投影,實現 AR 眼鏡中顯示即時人臉辨識標 籤及人臉 Bounding Box。本論文希望臉盲症患者在戴上 AR 眼鏡後,AR 眼 鏡能夠即時從環境中掃描人臉,並從既有資料庫之中,辨別出對應之身分, 將該人的位置與人名標註至 AR 眼鏡中,幫助臉盲症患者能夠辨別出生活中 每個人之身分,不因認不出臉而產生困惑,突破社交上的阻礙,降低而因社 交上之阻礙,導致產生自閉症等心理疾病之可能性。
The paper aims to develop an AR (Augmented Reality) eyewear system to help people with face blindness to recognize others in their life socially. This specific framework consists of the following: First, Structured Light technology is used to combine stereoscopic cameras with 2D RGB input through Structured Light, 2D data is extracted from the face through Deep Neural Network, and the coordinates of the face are identified in 3D space . In addition, the 3D point cloud information and 2D images are reconstructed bvin real time by deep learning, and the face information is obtained from seven angles, including front and side faces, to improve the accuracy of face recognition for side faces and large movements. Second, the face information generated in the first part is input to the convolutional neural network for computation. The convolutional neural network replaces the traditional high-dimensional classifier with a 128-dimensional feature vector as the basis for face characteristics. The minimum European distance is obtained and the name of the face is matched. This paper hopes that the AR glasses can instantly scan the face from the environment and identify the corresponding identity from the existing database, and label the location and name of the person into the AR glasses after the face blindness patient wears the AR glasses. The AR glasses help people with face blindness to identify everyone in their lives, so that they will not be confused by their inability to recognize faces, and to break through social barriers and reduce the possibility of autism and other psychological disorders caused by social barriers.
[1] S. L. Corrow, K. A. Dalrymple and J. JS. Barton, “Prosopagnosia: current perspectives,” Eye and brain, vol. 8, pp. 165-175, Sep. 2016.
[2] I. Kennerknecht, N. Plümpe, S. Edwards and R. Raman, “Hereditary prosopagnosia (HPA): the first report outside the Caucasian population,” Journal of human genetics, vol. 52, pp. 230-236, Feb. 2007.
[3] I. Kennerknecht, N. Y. Ho and C. N. Wong, “Prevalence of hereditary prosopagnosia (HPA) in Hong Kong Chinese population,” American journal of medical genetics Part A, vol. 146A, no. 22, pp. 2863-2870, Nov. 2008.
[4] “New Collaboration between IKEA and Apple,” IKEA, 7 June 2017, about.ikea.com/en/newsroom/2017/06/07/new-collaboration-between-ikea-and-apple.
[5] “Slimmed-down F-35 Gen III Helmet to Be Introduced Sooner.” FlightGlobal, 30 Mar. 2016, www.flightglobal.com/slimmed-down-f-35-gen-iii-helmet-to-be-introduced-sooner/120125.article.
[6] “bsolutely real: Virtual and augmented reality open new avenues in the BMW Group Production System,” BMWGROUP, 25 Apr. 2019, www.press.bmwgroup.com/argentina/article/detail/T0295123ES/absolutamente-real:-la-realidad-virtual-y-aumentada-abre-nuevas-v%C3%ADas-en-el-sistema-de-producci%C3%B3n-de-bmw-group?language=es.
[7] “一次了解人臉辨識的各種應用【2021 最新版】,” FaceMe, 28 Jan. 2021, tw.cyberlink.com/faceme/insights/articles/236/how_is_facial_recognition_used_in_2021.
[8] I. Masi, Y. Wu, T. Hassner and P. Natarajan, “Deep Face Recognition: A Survey,” in 2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI), Parana, 2018, pp. 471-478.
[9] C. Szegedy et al., “Going deeper with convolutions,” in 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, 2015, pp. 1-9.
[10] F. Schroff, D. Kalenichenko and J. Philbin, “FaceNet: A unified embedding for face recognition and clustering,” in 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, 2015, pp. 815-823.
[11] V. Blanz and T. Vetter, “A morphable model for the synthesis of 3D faces,” in SIGGRAPH 99: 26th International Conference on Computer Graphics and Interactive Techniques, Los Angeles, CA, July 1999, pp.187-194.
[12] X. Zhu, Z. Lei, X. Liu, H. Shi and S. Z. Li, "Face Alignment Across Large Poses: A 3D Solution," 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, 2016, pp.146-155.
[13] X. Zhu, X. Liu, Z. Lei and S. Z. Li, "Face Alignment in Full Pose Range: A 3D Total Solution," in IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 41, no. 1, pp. 78-92, 1 Jan. 2019
[14] “Depth Camera D435i,” Intel REALSENSE, www.intelrealsense.com.
[15] “Jetson Nano,” NVIDIA Developer, www. developer.nvidia.com.
[16] “Dream Glass 4K,” Dream Glass, www. dreamworldvision.com.