Improvement in Classification Performance Based on Target Vector Modification for All-Transfer Deep Learning

This paper proposes a target vector modification method for the all-transfer deep learning (ATDL) method. Deep neural networks (DNNs) have been used widely in many applications; however, the DNN has been known to be problematic when large amounts of training data are not available. Transfer learning...

Full description

Bibliographic Details
Main Authors: Yoshihide Sawada, Yoshikuni Sato, Toru Nakada, Shunta Yamaguchi, Kei Ujimoto, Nobuhiro Hayashi
Format: Article
Language:English
Published: MDPI AG 2019-01-01
Series:Applied Sciences
Subjects:
Online Access:http://www.mdpi.com/2076-3417/9/1/128
Description
Summary:This paper proposes a target vector modification method for the all-transfer deep learning (ATDL) method. Deep neural networks (DNNs) have been used widely in many applications; however, the DNN has been known to be problematic when large amounts of training data are not available. Transfer learning can provide a solution to this problem. Previous methods regularize all layers, including the output layer, by estimating the relation vectors, which are then used instead of one-hot target vectors of the target domain. These vectors are estimated by averaging the target domain data of each target domain label in the output space. This method improves the classification performance, but it does not consider the relation between the relation vectors. From this point of view, we propose a relation vector modification based on constrained pairwise repulsive forces. High pairwise repulsive forces provide large distances between the relation vectors. In addition, the risk of divergence is mitigated by the constraint based on distributions of the output vectors of the target domain data. We apply our method to two simulation experiments and a disease classification using two-dimensional electrophoresis images. The experimental results show that reusing all layers through our estimation method is effective, especially for a significantly small number of the target domain data.
ISSN:2076-3417