TY - JOUR
T1 - CN4SRSS
T2 - Combined network for super-resolution reconstruction and semantic segmentation in frontal-viewing camera images of vehicle
AU - Ryu, Kyung Bong
AU - Kang, Seon Jong
AU - Jeong, Seong In
AU - Jeong, Min Su
AU - Park, Kang Ryoung
N1 - Publisher Copyright:
© 2023 The Authors
PY - 2024/4
Y1 - 2024/4
N2 - Recently, the importance of semantic segmentation research for scene understanding in frontal viewing camera images of autonomous vehicles has increased. The existing state-of-the-art (SOTA) methods for semantic segmentation exhibit high accuracy for high-resolution images and low-resolution (LR) images without degradation factors of blur and noise. Owing to the nature of vehicles, the need is increasing for the pre-judgment of emergencies through the accurate semantic segmentation of LR images with the degradation factors acquired by low-cost camera at far distance. However, no research exists on super-resolution reconstruction (SR)-based semantic segmentation of LR images with degradation factors. Therefore, this study proposes a novel combined network for a super-resolution reconstruction and semantic segmentation (CN4SRSS) framework based on attention and re-focus network (ARNet), which exhibits low computational cost and high semantic segmentation accuracy. The experimental results using LR image datasets based on CamVid and Minicity datasets, which are open databases, show that the semantic segmentation accuracy (pixel accuracy) based on the proposed CN4SRSS and DeepLab v3 + is 93.14% and 89.48%, respectively. Particularly, the proposed method shows higher accuracy when compared to the SOTA methods. Furthermore, the proposed method has been confirmed that requires lower computational cost in terms of the number of parameters, memory usage, number of multi-adds calculation, and floating-point operations per second (FLOPs) than the SOTA methods.
AB - Recently, the importance of semantic segmentation research for scene understanding in frontal viewing camera images of autonomous vehicles has increased. The existing state-of-the-art (SOTA) methods for semantic segmentation exhibit high accuracy for high-resolution images and low-resolution (LR) images without degradation factors of blur and noise. Owing to the nature of vehicles, the need is increasing for the pre-judgment of emergencies through the accurate semantic segmentation of LR images with the degradation factors acquired by low-cost camera at far distance. However, no research exists on super-resolution reconstruction (SR)-based semantic segmentation of LR images with degradation factors. Therefore, this study proposes a novel combined network for a super-resolution reconstruction and semantic segmentation (CN4SRSS) framework based on attention and re-focus network (ARNet), which exhibits low computational cost and high semantic segmentation accuracy. The experimental results using LR image datasets based on CamVid and Minicity datasets, which are open databases, show that the semantic segmentation accuracy (pixel accuracy) based on the proposed CN4SRSS and DeepLab v3 + is 93.14% and 89.48%, respectively. Particularly, the proposed method shows higher accuracy when compared to the SOTA methods. Furthermore, the proposed method has been confirmed that requires lower computational cost in terms of the number of parameters, memory usage, number of multi-adds calculation, and floating-point operations per second (FLOPs) than the SOTA methods.
KW - ARNet
KW - CN4SRSS
KW - Frontal viewing camera image
KW - Semantic segmentation
KW - Super-resolution reconstruction
UR - http://www.scopus.com/inward/record.url?scp=85179844817&partnerID=8YFLogxK
U2 - 10.1016/j.engappai.2023.107673
DO - 10.1016/j.engappai.2023.107673
M3 - Article
AN - SCOPUS:85179844817
SN - 0952-1976
VL - 130
JO - Engineering Applications of Artificial Intelligence
JF - Engineering Applications of Artificial Intelligence
M1 - 107673
ER -