Computer vision for detecting building pathologies has interested researchers for quite some time. Vision-based crack detection is a non-destructive assessment technique, which can be useful especially for Cultural Heritage (CH) where strict regulations apply and, even simple, interventions are not permitted. Recently, shallow and deep machine learning architectures applied on various types of imagery are gaining ground. In this article a crack detection methodology for stone masonry walls is presented. In the proposed approach, crack detection is approached as an unsupervised anomaly detection problem on RGB (Red Green Blue) image patches. Towards this direction, some of the most popular state of the art CNN (Convolutional Neural Network) architectures are deployed and modified to binary classify the images or image patches by predicting a specific class for the tested imagery; 'Crack' or 'No crack', and detect and localize those cracks on the RGB imagery with high accuracy. Testing of the model was performed on various test sites and random images retrieved from the internet and collected by the authors and results suggested the high performance of specific networks compared to the rest, considering also the small numbers of epochs required for training. Those results met the accuracy delivered by more complex and computationally heavy approaches, requiring a large amount of data for training. Source code is available on GitHub https://github.com/pagraf/Crack-detection while datasets are available on Zenodo https://doi.org/10.5281/zenodo.6516913 .
翻译:摘要:建筑病害检测的计算机视觉技术已经引起研究人员相当大的兴趣。基于视觉的裂纹检测是一种非破坏性的评估技术,特别是对于文化遗产,该技术非常有用,因为存在严格的规定,即使是简单的干预也是不被允许的。最近,应用于各种类型图像的浅层和深层机器学习架构正在流行。本文提出了石质砌体墙壁的裂纹检测方法。在所提出的方法中,裂纹检测被视为RGB(红绿蓝)图像补丁上的无监督异常检测问题。为此,部署并修改了一些最流行的现有卷积神经网络架构,通过预测被测试图像或图像补丁的特定类别来进行二进制分类;“有裂纹”或“无裂纹”,并高精度地检测和定位这些裂纹在RGB图像上。模型测试在各种测试现场和从互联网收集的随机图像上进行了测试,结果表明与其余模型相比,特定网络提供了高性能,同时考虑到所需的训练时期较短。这些结果达到了通过更复杂和计算量较大的方法实现的准确性,这些方法需要大量训练数据。源代码可在GitHub上获得https://github.com/pagraf/Crack-detection,而数据集可在Zenodo上获得https://doi.org/10.5281/zenodo.6516913 。