EPYNET: Efficient Pyramidal Network for Clothing Segmentation

Soft biometrics traits extracted from a human body, including the type of clothes, hair color, and accessories, are useful information used for people tracking and identification. Semantic segmentation of these traits from images is still a challenge for researchers because of the huge variety of cl...

Full description

Bibliographic Details
Main Authors: Andrei De Souza Inacio, Heitor Silverio Lopes
Format: Article
Language:English
Published: IEEE 2020-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9222020/
id doaj-673f34a730d4494b8982d26a693ca2e2
record_format Article
spelling doaj-673f34a730d4494b8982d26a693ca2e22021-03-30T04:41:47ZengIEEEIEEE Access2169-35362020-01-01818788218789210.1109/ACCESS.2020.30308599222020EPYNET: Efficient Pyramidal Network for Clothing SegmentationAndrei De Souza Inacio0https://orcid.org/0000-0002-3559-5935Heitor Silverio Lopes1Graduate Program in Electrical Engineering and Industrial Informatics, Federal University of Technology – Paraná, Curitiba, BrazilGraduate Program in Electrical Engineering and Industrial Informatics, Federal University of Technology – Paraná, Curitiba, BrazilSoft biometrics traits extracted from a human body, including the type of clothes, hair color, and accessories, are useful information used for people tracking and identification. Semantic segmentation of these traits from images is still a challenge for researchers because of the huge variety of clothing styles, layering, shapes, and colors. To tackle these issues, we proposed EPYNET, a framework for clothing segmentation. EPYNET is based on the Single Shot MultiBox Detector (SSD) and the Feature Pyramid Network (FPN) with the EfficientNet model as the backbone. The framework also integrates data augmentation methods and noise reduction techniques to increase the accuracy of the segmentation. We also propose a new dataset named UTFPR-SBD3, consisting of 4,500 manually annotated images into 18 classes of objects, plus the background. Unlike available public datasets with imbalanced class distributions, the UTFPR-SBD3 has, at least, 100 instances per class to minimize the training difficulty of deep learning models. We introduced a new measure of dataset imbalance, motivated by the difficulty in comparing different datasets for clothing segmentation. With such a measure, it is possible to detect the influence of the background, classes with small items, or classes with a too high or too low number of instances. Experimental results on UTFPR-SBD3 show the effectiveness of EPYNET, outperforming the state-of-art methods for clothing segmentation on public datasets. Based on these results, we believe that the proposed approach can be potentially useful for many real-world applications related to soft biometrics, people surveillance, image description, clothes recommendation, and others.https://ieeexplore.ieee.org/document/9222020/Soft biometricsclothing segmentationcomputer visiondeep learning
collection DOAJ
language English
format Article
sources DOAJ
author Andrei De Souza Inacio
Heitor Silverio Lopes
spellingShingle Andrei De Souza Inacio
Heitor Silverio Lopes
EPYNET: Efficient Pyramidal Network for Clothing Segmentation
IEEE Access
Soft biometrics
clothing segmentation
computer vision
deep learning
author_facet Andrei De Souza Inacio
Heitor Silverio Lopes
author_sort Andrei De Souza Inacio
title EPYNET: Efficient Pyramidal Network for Clothing Segmentation
title_short EPYNET: Efficient Pyramidal Network for Clothing Segmentation
title_full EPYNET: Efficient Pyramidal Network for Clothing Segmentation
title_fullStr EPYNET: Efficient Pyramidal Network for Clothing Segmentation
title_full_unstemmed EPYNET: Efficient Pyramidal Network for Clothing Segmentation
title_sort epynet: efficient pyramidal network for clothing segmentation
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2020-01-01
description Soft biometrics traits extracted from a human body, including the type of clothes, hair color, and accessories, are useful information used for people tracking and identification. Semantic segmentation of these traits from images is still a challenge for researchers because of the huge variety of clothing styles, layering, shapes, and colors. To tackle these issues, we proposed EPYNET, a framework for clothing segmentation. EPYNET is based on the Single Shot MultiBox Detector (SSD) and the Feature Pyramid Network (FPN) with the EfficientNet model as the backbone. The framework also integrates data augmentation methods and noise reduction techniques to increase the accuracy of the segmentation. We also propose a new dataset named UTFPR-SBD3, consisting of 4,500 manually annotated images into 18 classes of objects, plus the background. Unlike available public datasets with imbalanced class distributions, the UTFPR-SBD3 has, at least, 100 instances per class to minimize the training difficulty of deep learning models. We introduced a new measure of dataset imbalance, motivated by the difficulty in comparing different datasets for clothing segmentation. With such a measure, it is possible to detect the influence of the background, classes with small items, or classes with a too high or too low number of instances. Experimental results on UTFPR-SBD3 show the effectiveness of EPYNET, outperforming the state-of-art methods for clothing segmentation on public datasets. Based on these results, we believe that the proposed approach can be potentially useful for many real-world applications related to soft biometrics, people surveillance, image description, clothes recommendation, and others.
topic Soft biometrics
clothing segmentation
computer vision
deep learning
url https://ieeexplore.ieee.org/document/9222020/
work_keys_str_mv AT andreidesouzainacio epynetefficientpyramidalnetworkforclothingsegmentation
AT heitorsilveriolopes epynetefficientpyramidalnetworkforclothingsegmentation
_version_ 1724181327066955776