分割
计算机科学
人工智能
卷积神经网络
编码器
模式识别(心理学)
残余物
图像分割
像素
人工神经网络
计算机视觉
算法
操作系统
作者
Chaoqun Xiang,Jingjing Guo,Ran Cao,Lu Deng
标识
DOI:10.1016/j.autcon.2023.104894
摘要
The performance of crack segmentation is influenced by complex scenes, including irregularly shaped cracks, complex image backgrounds, and limitations in acquiring global contextual information. To alleviate the influence of these factors, a dual-encoder network fusing transformers and convolutional neural networks (DTrC-Net) is proposed in this study. The structure of the DTrC-Net was designed to capture both the local features and global contextual information of crack images. To enhance feature fusion between the adjacent and codec layers, a feature fusion module and a residual path module were also added to the network. Through a series of comparative experiments, DTrC-Net was found to generate better predictions than other state-of-the-art segmentation networks, with the highest precision (75.60%), recall (78.86%), F1-score (76.44%), and intersection over union (64.30%) on the Crack3238 dataset. Moreover, a fast processing speed of 78 frames per second was achieved using the DTrC-Net with an image size of 256 × 256 pixels. Overall, it was found that the proposed DTrC-Net outperformed other advanced networks in terms of accuracy in crack segmentation and demonstrated superior generalizability in complex scenes.
科研通智能强力驱动
Strongly Powered by AbleSci AI