Accelerator-Aware Fast Spatial Feature Network for Real-Time Semantic Segmentation
Semantic segmentation is performed to understand an image at the pixel level; it is widely used in the field of autonomous driving. In recent years, deep neural networks achieve good accuracy performance; however, there exist few models that have a good trade-off between high accuracy and low infere...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2020-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9296217/ |
id |
doaj-ea8566e62bc749708235114ca12c5c50 |
---|---|
record_format |
Article |
spelling |
doaj-ea8566e62bc749708235114ca12c5c502021-03-30T04:20:12ZengIEEEIEEE Access2169-35362020-01-01822652422653710.1109/ACCESS.2020.30451479296217Accelerator-Aware Fast Spatial Feature Network for Real-Time Semantic SegmentationMinjong Kim0https://orcid.org/0000-0002-6823-2137Byungjae Park1https://orcid.org/0000-0002-8952-0736Suyoung Chi2https://orcid.org/0000-0002-8811-6934Artificial Intelligence Laboratory, Electronics and Telecommunications Research Institute, Daejeon, South KoreaSchool of Mechanical Engineering, Korea University of Technology and Education, Cheonan, South KoreaArtificial Intelligence Laboratory, Electronics and Telecommunications Research Institute, Daejeon, South KoreaSemantic segmentation is performed to understand an image at the pixel level; it is widely used in the field of autonomous driving. In recent years, deep neural networks achieve good accuracy performance; however, there exist few models that have a good trade-off between high accuracy and low inference time. In this paper, we propose a fast spatial feature network (FSFNet), an optimized lightweight semantic segmentation model using an accelerator, offering high performance as well as faster inference speed than current methods. FSFNet employs the FSF and MRA modules. The FSF module has three different types of subset modules to extract spatial features efficiently. They are designed in consideration of the size of the spatial domain. The multi-resolution aggregation module combines features that are extracted at different resolutions to reconstruct the segmentation image accurately. Our approach is able to run at over 203 FPS at full resolution (1024×2048) in a single NVIDIA 1080Ti GPU, and obtains a result of 69.13% mIoU on the Cityscapes test dataset. Compared with existing models in real-time semantic segmentation, our proposed model retains remarkable accuracy while having high FPS that is over 30% faster than the state-of-the-art model. The experimental results proved that our model is an ideal approach for the Cityscapes dataset. The code is publicly available at: https://github.com/computervision8/FSFNet.https://ieeexplore.ieee.org/document/9296217/Real-timehigh-resolutionsemantic segmentationconvolutional neural networkCityscapes dataset |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Minjong Kim Byungjae Park Suyoung Chi |
spellingShingle |
Minjong Kim Byungjae Park Suyoung Chi Accelerator-Aware Fast Spatial Feature Network for Real-Time Semantic Segmentation IEEE Access Real-time high-resolution semantic segmentation convolutional neural network Cityscapes dataset |
author_facet |
Minjong Kim Byungjae Park Suyoung Chi |
author_sort |
Minjong Kim |
title |
Accelerator-Aware Fast Spatial Feature Network for Real-Time Semantic Segmentation |
title_short |
Accelerator-Aware Fast Spatial Feature Network for Real-Time Semantic Segmentation |
title_full |
Accelerator-Aware Fast Spatial Feature Network for Real-Time Semantic Segmentation |
title_fullStr |
Accelerator-Aware Fast Spatial Feature Network for Real-Time Semantic Segmentation |
title_full_unstemmed |
Accelerator-Aware Fast Spatial Feature Network for Real-Time Semantic Segmentation |
title_sort |
accelerator-aware fast spatial feature network for real-time semantic segmentation |
publisher |
IEEE |
series |
IEEE Access |
issn |
2169-3536 |
publishDate |
2020-01-01 |
description |
Semantic segmentation is performed to understand an image at the pixel level; it is widely used in the field of autonomous driving. In recent years, deep neural networks achieve good accuracy performance; however, there exist few models that have a good trade-off between high accuracy and low inference time. In this paper, we propose a fast spatial feature network (FSFNet), an optimized lightweight semantic segmentation model using an accelerator, offering high performance as well as faster inference speed than current methods. FSFNet employs the FSF and MRA modules. The FSF module has three different types of subset modules to extract spatial features efficiently. They are designed in consideration of the size of the spatial domain. The multi-resolution aggregation module combines features that are extracted at different resolutions to reconstruct the segmentation image accurately. Our approach is able to run at over 203 FPS at full resolution (1024×2048) in a single NVIDIA 1080Ti GPU, and obtains a result of 69.13% mIoU on the Cityscapes test dataset. Compared with existing models in real-time semantic segmentation, our proposed model retains remarkable accuracy while having high FPS that is over 30% faster than the state-of-the-art model. The experimental results proved that our model is an ideal approach for the Cityscapes dataset. The code is publicly available at: https://github.com/computervision8/FSFNet. |
topic |
Real-time high-resolution semantic segmentation convolutional neural network Cityscapes dataset |
url |
https://ieeexplore.ieee.org/document/9296217/ |
work_keys_str_mv |
AT minjongkim acceleratorawarefastspatialfeaturenetworkforrealtimesemanticsegmentation AT byungjaepark acceleratorawarefastspatialfeaturenetworkforrealtimesemanticsegmentation AT suyoungchi acceleratorawarefastspatialfeaturenetworkforrealtimesemanticsegmentation |
_version_ |
1724182035080151040 |