Accelerator-Aware Fast Spatial Feature Network for Real-Time Semantic Segmentation

Semantic segmentation is performed to understand an image at the pixel level; it is widely used in the field of autonomous driving. In recent years, deep neural networks achieve good accuracy performance; however, there exist few models that have a good trade-off between high accuracy and low infere...

Full description

Bibliographic Details
Main Authors: Minjong Kim, Byungjae Park, Suyoung Chi
Format: Article
Language:English
Published: IEEE 2020-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9296217/
id doaj-ea8566e62bc749708235114ca12c5c50
record_format Article
spelling doaj-ea8566e62bc749708235114ca12c5c502021-03-30T04:20:12ZengIEEEIEEE Access2169-35362020-01-01822652422653710.1109/ACCESS.2020.30451479296217Accelerator-Aware Fast Spatial Feature Network for Real-Time Semantic SegmentationMinjong Kim0https://orcid.org/0000-0002-6823-2137Byungjae Park1https://orcid.org/0000-0002-8952-0736Suyoung Chi2https://orcid.org/0000-0002-8811-6934Artificial Intelligence Laboratory, Electronics and Telecommunications Research Institute, Daejeon, South KoreaSchool of Mechanical Engineering, Korea University of Technology and Education, Cheonan, South KoreaArtificial Intelligence Laboratory, Electronics and Telecommunications Research Institute, Daejeon, South KoreaSemantic segmentation is performed to understand an image at the pixel level; it is widely used in the field of autonomous driving. In recent years, deep neural networks achieve good accuracy performance; however, there exist few models that have a good trade-off between high accuracy and low inference time. In this paper, we propose a fast spatial feature network (FSFNet), an optimized lightweight semantic segmentation model using an accelerator, offering high performance as well as faster inference speed than current methods. FSFNet employs the FSF and MRA modules. The FSF module has three different types of subset modules to extract spatial features efficiently. They are designed in consideration of the size of the spatial domain. The multi-resolution aggregation module combines features that are extracted at different resolutions to reconstruct the segmentation image accurately. Our approach is able to run at over 203 FPS at full resolution (1024×2048) in a single NVIDIA 1080Ti GPU, and obtains a result of 69.13% mIoU on the Cityscapes test dataset. Compared with existing models in real-time semantic segmentation, our proposed model retains remarkable accuracy while having high FPS that is over 30% faster than the state-of-the-art model. The experimental results proved that our model is an ideal approach for the Cityscapes dataset. The code is publicly available at: https://github.com/computervision8/FSFNet.https://ieeexplore.ieee.org/document/9296217/Real-timehigh-resolutionsemantic segmentationconvolutional neural networkCityscapes dataset
collection DOAJ
language English
format Article
sources DOAJ
author Minjong Kim
Byungjae Park
Suyoung Chi
spellingShingle Minjong Kim
Byungjae Park
Suyoung Chi
Accelerator-Aware Fast Spatial Feature Network for Real-Time Semantic Segmentation
IEEE Access
Real-time
high-resolution
semantic segmentation
convolutional neural network
Cityscapes dataset
author_facet Minjong Kim
Byungjae Park
Suyoung Chi
author_sort Minjong Kim
title Accelerator-Aware Fast Spatial Feature Network for Real-Time Semantic Segmentation
title_short Accelerator-Aware Fast Spatial Feature Network for Real-Time Semantic Segmentation
title_full Accelerator-Aware Fast Spatial Feature Network for Real-Time Semantic Segmentation
title_fullStr Accelerator-Aware Fast Spatial Feature Network for Real-Time Semantic Segmentation
title_full_unstemmed Accelerator-Aware Fast Spatial Feature Network for Real-Time Semantic Segmentation
title_sort accelerator-aware fast spatial feature network for real-time semantic segmentation
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2020-01-01
description Semantic segmentation is performed to understand an image at the pixel level; it is widely used in the field of autonomous driving. In recent years, deep neural networks achieve good accuracy performance; however, there exist few models that have a good trade-off between high accuracy and low inference time. In this paper, we propose a fast spatial feature network (FSFNet), an optimized lightweight semantic segmentation model using an accelerator, offering high performance as well as faster inference speed than current methods. FSFNet employs the FSF and MRA modules. The FSF module has three different types of subset modules to extract spatial features efficiently. They are designed in consideration of the size of the spatial domain. The multi-resolution aggregation module combines features that are extracted at different resolutions to reconstruct the segmentation image accurately. Our approach is able to run at over 203 FPS at full resolution (1024×2048) in a single NVIDIA 1080Ti GPU, and obtains a result of 69.13% mIoU on the Cityscapes test dataset. Compared with existing models in real-time semantic segmentation, our proposed model retains remarkable accuracy while having high FPS that is over 30% faster than the state-of-the-art model. The experimental results proved that our model is an ideal approach for the Cityscapes dataset. The code is publicly available at: https://github.com/computervision8/FSFNet.
topic Real-time
high-resolution
semantic segmentation
convolutional neural network
Cityscapes dataset
url https://ieeexplore.ieee.org/document/9296217/
work_keys_str_mv AT minjongkim acceleratorawarefastspatialfeaturenetworkforrealtimesemanticsegmentation
AT byungjaepark acceleratorawarefastspatialfeaturenetworkforrealtimesemanticsegmentation
AT suyoungchi acceleratorawarefastspatialfeaturenetworkforrealtimesemanticsegmentation
_version_ 1724182035080151040