Unsupervised Transformation Network Based on GANs for Target-Domain Oriented Image Translation

Image-to-image translation usually refers to the task of translating an input image from the source domain to the target domain while preserving the structure in the source domain. Recently, generative adversarial networks (GANs) using paired images for this task have made great progress. However, p...

Full description

Bibliographic Details
Main Authors: Hongwei Ge, Yao Yao, Zheng Chen, Liang Sun
Format: Article
Language:English
Published: IEEE 2018-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/8492399/
id doaj-acf5d9b39d8d4a44901f5a2586e0eb40
record_format Article
spelling doaj-acf5d9b39d8d4a44901f5a2586e0eb402021-03-29T21:21:47ZengIEEEIEEE Access2169-35362018-01-016613426135010.1109/ACCESS.2018.28760968492399Unsupervised Transformation Network Based on GANs for Target-Domain Oriented Image TranslationHongwei Ge0https://orcid.org/0000-0002-8937-1515Yao Yao1Zheng Chen2Liang Sun3https://orcid.org/0000-0003-2794-8654College of Computer Science and Technology, Dalian University of Technology, Dalian, ChinaCollege of Computer Science and Technology, Dalian University of Technology, Dalian, ChinaDepartment of Computer Science and Engineering, Washington University in St. Louis, St. Louis, MO, USACollege of Computer Science and Technology, Dalian University of Technology, Dalian, ChinaImage-to-image translation usually refers to the task of translating an input image from the source domain to the target domain while preserving the structure in the source domain. Recently, generative adversarial networks (GANs) using paired images for this task have made great progress. However, paired training data will not be available for many tasks. In this paper, a GAN-based unsupervised transformation network (UTN-GAN) is proposed for image-to-image translation. Importantly, UTN-GAN employs hierarchical representations and weight-sharing mechanism to translate images from the source domain to the target domain without paired images. We employ two groups of unsupervised GANs to generate the images in different domains first, and then discriminate them. In UTN-GAN, an auto-encoder reconstruction network is designed to extract the hierarchical representations of the images in the source domain by minimizing the reconstruction loss. In particular, the high-level representations (semantics) are shared with a translation network to guarantee that the input image and the output image are paired up in the different domains. All network structures are trained together by using a joint loss function. The experimental studies in qualitative and quantitative aspects on several image translation tasks show that the proposed algorithm is effective and competitive compared with some state-of-the-art algorithms.https://ieeexplore.ieee.org/document/8492399/Generative adversarial networksunsupervised learningstyle transferimage-to-image translation
collection DOAJ
language English
format Article
sources DOAJ
author Hongwei Ge
Yao Yao
Zheng Chen
Liang Sun
spellingShingle Hongwei Ge
Yao Yao
Zheng Chen
Liang Sun
Unsupervised Transformation Network Based on GANs for Target-Domain Oriented Image Translation
IEEE Access
Generative adversarial networks
unsupervised learning
style transfer
image-to-image translation
author_facet Hongwei Ge
Yao Yao
Zheng Chen
Liang Sun
author_sort Hongwei Ge
title Unsupervised Transformation Network Based on GANs for Target-Domain Oriented Image Translation
title_short Unsupervised Transformation Network Based on GANs for Target-Domain Oriented Image Translation
title_full Unsupervised Transformation Network Based on GANs for Target-Domain Oriented Image Translation
title_fullStr Unsupervised Transformation Network Based on GANs for Target-Domain Oriented Image Translation
title_full_unstemmed Unsupervised Transformation Network Based on GANs for Target-Domain Oriented Image Translation
title_sort unsupervised transformation network based on gans for target-domain oriented image translation
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2018-01-01
description Image-to-image translation usually refers to the task of translating an input image from the source domain to the target domain while preserving the structure in the source domain. Recently, generative adversarial networks (GANs) using paired images for this task have made great progress. However, paired training data will not be available for many tasks. In this paper, a GAN-based unsupervised transformation network (UTN-GAN) is proposed for image-to-image translation. Importantly, UTN-GAN employs hierarchical representations and weight-sharing mechanism to translate images from the source domain to the target domain without paired images. We employ two groups of unsupervised GANs to generate the images in different domains first, and then discriminate them. In UTN-GAN, an auto-encoder reconstruction network is designed to extract the hierarchical representations of the images in the source domain by minimizing the reconstruction loss. In particular, the high-level representations (semantics) are shared with a translation network to guarantee that the input image and the output image are paired up in the different domains. All network structures are trained together by using a joint loss function. The experimental studies in qualitative and quantitative aspects on several image translation tasks show that the proposed algorithm is effective and competitive compared with some state-of-the-art algorithms.
topic Generative adversarial networks
unsupervised learning
style transfer
image-to-image translation
url https://ieeexplore.ieee.org/document/8492399/
work_keys_str_mv AT hongweige unsupervisedtransformationnetworkbasedongansfortargetdomainorientedimagetranslation
AT yaoyao unsupervisedtransformationnetworkbasedongansfortargetdomainorientedimagetranslation
AT zhengchen unsupervisedtransformationnetworkbasedongansfortargetdomainorientedimagetranslation
AT liangsun unsupervisedtransformationnetworkbasedongansfortargetdomainorientedimagetranslation
_version_ 1724193118789566464