
@article{ref1,
title="Based on cross-scale fusion attention mechanism network for semantic segmentation for street scenes",
journal="Frontiers in neurorobotics",
year="2023",
author="Ye, Xin and Gao, Lang and Chen, Jichen and Lei, Mingyue",
volume="17",
number="",
pages="e1204418-e1204418",
abstract="Semantic segmentation, which is a fundamental task in computer vision. Every pixel will have a specific semantic class assigned to it through semantic segmentation methods. Embedded systems and mobile devices are difficult to deploy high-accuracy segmentation algorithms. Despite the rapid development of semantic segmentation, the balance between speed and accuracy must be improved. As a solution to the above problems, we created a cross-scale fusion attention mechanism network called CFANet, which fuses feature maps from different scales. We first design a novel efficient residual module (ERM), which applies both dilation convolution and factorized convolution. Our CFANet is mainly constructed from ERM. Subsequently, we designed a new multi-branch channel attention mechanism (MCAM) to refine the feature maps at different levels. Experiment results show that CFANet achieved 70.6% mean intersection over union (mIoU) and 67.7% mIoU on Cityscapes and CamVid datasets, respectively, with inference speeds of 118 FPS and 105 FPS on NVIDIA RTX2080Ti GPU cards with 0.84M parameters.<p /> <p>Language: en</p>",
language="en",
issn="1662-5218",
doi="10.3389/fnbot.2023.1204418",
url="http://dx.doi.org/10.3389/fnbot.2023.1204418"
}