簡易檢索 / 詳目顯示

研究生: 章皓鈞
Chang, Hao-Chun
論文名稱: 基於深度學習之路面破損檢測
Road Crack Detection Based On Deep Learning
指導教授: 吳順德
Wu, Shuen-De
口試委員: 呂有勝 劉益宏 吳順德
口試日期: 2021/07/29
學位類別: 碩士
Master
系所名稱: 機電工程學系
Department of Mechatronic Engineering
論文出版年: 2021
畢業學年度: 109
語文別: 中文
論文頁數: 44
中文關鍵詞: 路面破損偵測深度學習影像處理
英文關鍵詞: Road crack detection, Deep learning, Image processing
DOI URL: http://doi.org/10.6345/NTNU202101717
論文種類: 學術論文
相關次數: 點閱:73下載:5
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 目前國內的道路維護方式多為定期派遣工程車檢測以及依賴人民的通報,而為了盡早的發現道路損壞並進行修復,本研究運用Mask R-CNN深度學習之方式建立道路破損辨識模型。透過Mask R-CNN深度學習演算法,以及運用python 、OpenCV撰寫進行道路破損檢測與資料整合,持續的分析模型數據並根據結果進行再訓練。利用路面破損辨識模型檢測出路面上的龜裂、裂縫、補綻、變形以及坑洞,並在龜裂、裂縫、補綻、變形達到86%以上的召回率,精確率除了裂縫、坑洞之外有82%以上,此外對檢測出來的破損範圍進行面積計算,為日後養護維修提供面積的量化指標,進而輔助人力巡查作業。

    At present, domestic road maintenance methods are mostly dispatching construction vehicles for inspection and relying on people’s notifications. In order to discover and repair road crack as soon as possible, this study uses Mask R-CNN deep learning to establish a road crack identification model. Through the Mask R-CNN deep learning algorithm, and the use of python and OpenCV to write for road crack detection and data integration, continuous analysis of model data and retraining based on the results. The road surface crack identification model is used to detect alligator crack, linear crack, patch, deformation and pothole on the road, and the recall rate of alligator crack, linear crack, patch and deformation is over 86%. Except for linear crack and pothole, the accuracy rate is more than 82%. In addition, the area of the detected damage area is calculated to provide a quantitative indicator of the area for future maintenance and repair, and then assist manpower inspection operations.

    摘要 i Abstract ii 誌謝 iii 目錄 v 表目錄 vii 圖目錄 viii 第一章 緒論 1 1.1 前言 1 1.2 研究動機與目標 2 1.3 文獻探討 3 1.3.1 二階段的物件偵測(two-stage object detection) 5 1.3.2 一階段的物件偵測(one-stage object detection) 6 1.4 研究方法 7 1.5 論文章節介紹 8 第二章 定義破損與資料蒐集 9 2.1 破損定義 10 2.2 蒐集影像並標註破損 15 第三章 Mask R-CNN模型訓練 17 3.1 卷積神經網路概論 17 3.1.1 機器學習介紹 17 3.1.2 卷積神經網路演進 19 3.1.3 卷積神經網路(Convolutional Neural Network, CNN) 22 3.2 R-CNN ( Regions with Convolutional Neural Network) 24 3.2.1 非極大值抑制(Non-Maximum Suppression,NMS) 25 3.3 Fast R-CNN 26 3.3.1 ROI pooling 27 3.4 Faster R-CNN 28 3.4.1 Region Proposal Networks (RPN) 28 3.5 Mask R-CNN 29 3.6 模型評估 32 第四章 破損面積與座標計算 33 4.1 透視變換(Perspective Transformation) 34 4.2 破損座標計算 36 第五章 實驗過程與結果 38 5.1 訓練結果 38 5.2 道路破損資料彙整 39 第六章 結論 41 6.1 結論 41 6.2 未來展望 41 參考文獻 43

    [1] R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature hierarchies for accurate object detection and semantic segmentation,” in 2014 IEEE Conference on Computer Vision and Pattern Recognition, pp. 580–587, June 2014.
    [2] R. Girshick, “Fast r-cnn,” in 2015 IEEE International Conference on Computer Vision (ICCV), pp. 1440–1448, Dec 2015.
    [3] S. Ren, K. He, R. Girshick, and J. Sun, “Faster r-cnn: Towards real-time object detection with region proposal networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39, pp. 1137–1149, June 2017.
    [4] J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real-time object detection,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 779–788, June 2016.
    [5] K. He, G. Gkioxari, P. Dollar, and R. Girshick, “Mask R-CNN,” CoRR abs/1703.06870 (2017).
    [6] Labelme, [Online], Available: https://github.com/wkentaro/labelme
    [7] OpenCV, [Online],Available: https://opencv.org
    [8] 內政部營建署, “市區道路管理維護與技術規範手冊研究”,2002
    [9] K. Fukushima and S. Miyake, “Neocognitron: A self-organizing neural network model for a mechanism of visual pattern recognition,” in Competition and Cooperation in Neural Nets, Springer, pp. 267–285, 1982.
    [10] D.H. Hubel and T.N. Wiesel, “Receptive fields and functional architecture of monkey striate cortex,” J. Physiol. (1968) 215–243.
    [11] K. Fukushima and S. Miyake, “Neocognitron: a self-organizing neural network model for a mechanism of visual pattern recognition,” in: Competition and Cooperation in Neural Nets, 1982, pp. 267–285.
    [12] Y. LeCun, L. Bottou, Y. Bengio and P. Haffner, “Gradient-based learning applied to document recognition,” Proc. IEEE 86 (11) (1998) 2278–2324.
    [13] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A. C. Berg, and L. Fei-Fei, “Imagenet large scale visual recognition challenge,” Int. J. Conflict Violence (IJCV) 115 (3) (2015) 211–252.
    [14] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke and A. Rabinovich, “Going deeper with convolutions,” in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015, pp. 1–9.
    [15] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in: Proceedings of the International Conference on Learning Representations (ICLR), 2015.
    [16] K. He, X. Zhang, S. Ren and J. Sun, “Deep residual learning for image recognition,” in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 770–778.
    [17] Image,[Online],Available:https://engineering.matterport.com/7c761e238b46
    [18] J. R. Uijlings, K. E. van de Sande, T. Gevers, and A. W. Smeulders, “Selective search for object recognition,” IJCV, 2013. 2.
    [19] B. Liu, Z. F. Hao and X. W. Yang, “Nesting Support Vector Machine for Multi-Classification,” ICMLC 2005, pp. 4220-4225, August 2005.
    [20] Neubeck A, Gool L V. Efficient Non-Maximum Suppression,”International Conference on Pattern Recognition.” IEEE Computer Society, 2006:850-855
    [21] R. Rothe, M. Guillaumin, and L. Van Gool, “Non-maximum suppression for object detection by passing messages between windows" In Proceedings of the Asian Conference on Computer Vision (ACCV), Singapore, 1–5 November 2014.
    [22] S. Qiao, L.C. Chen, A. Yuille, “DetectoRS: Detecting Objects with Recursive Feature Pyramid and Switchable Atrous Convolution” IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 10213-10224
    [23] L. Tingting, C. Xiaojie, L. Yudong, W. Yongtao, T. Zhi, C. Wei, C. Jingdong and . Haibin, “CBNetV2: A Composite Backbone Network Architecture for Object Detection” arXiv:2107.00420
    [24] SuperDataScience, CNN結構示意圖, [Online], Available : SuperDataScience:Convolutional Neural Networks (CNN): Step 3 – Flattening, Aug 2018
    [25] jiongnima, Mask R-CNN結構圖, [Online], https://blog.csdn.net/jiongnima/article/details/79094159

    下載圖示
    QR CODE