Deep Attention-Guided Hashing

With the rapid growth of multimedia data (e.g., image, audio, and video) on the Web, the learning-based hashing techniques, such as deep supervised hashing, have proven to be very efficient for large-scale multimedia search. The recent successes seen in the learning-based hashing methods are largely...

Full description

Bibliographic Details
Main Authors: Zhan Yang, Osolo Ian Raymond, Wuqing Sun, Jun Long
Format: Article
Language:English
Published: IEEE 2019-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/8607076/
id doaj-38135cdd5d0f48e98108bdd0fbad3e2e
record_format Article
spelling doaj-38135cdd5d0f48e98108bdd0fbad3e2e2021-03-29T22:01:41ZengIEEEIEEE Access2169-35362019-01-017112091122110.1109/ACCESS.2019.28918948607076Deep Attention-Guided HashingZhan Yang0https://orcid.org/0000-0002-6336-0228Osolo Ian Raymond1https://orcid.org/0000-0001-9891-6161Wuqing Sun2Jun Long3School of Information Science and Engineering, Central South University, Changsha, ChinaSchool of Information Science and Engineering, Central South University, Changsha, ChinaSchool of Information Science and Engineering, Central South University, Changsha, ChinaSchool of Information Science and Engineering, Central South University, Changsha, ChinaWith the rapid growth of multimedia data (e.g., image, audio, and video) on the Web, the learning-based hashing techniques, such as deep supervised hashing, have proven to be very efficient for large-scale multimedia search. The recent successes seen in the learning-based hashing methods are largely due to the success of the deep learning-based hashing methods. However, there are some limitations to the previous learning-based hashing methods (e.g., the learned hash codes containing repetitive and highly correlated information). In this paper, we propose a novel learning-based hashing method, named deep attention-guided hashing (DAgH). DAgH is implemented using two stream frameworks. The core idea is to use the guided hash codes which are generated by the hashing network of the first stream framework (called the first hashing network) to guide the training of the hashing network of the second stream framework (called the second hashing network). Specifically, in the first network, it leverages an attention network and hashing network to generate the attention-guided hash codes from the original images. The loss function we propose contains two components: the semantic loss and the attention loss. The attention loss is used to punish the attention network to obtain the salient region from pairs of images; in the second network, these attention-guided hash codes are used to guide the training of the second hashing network (i.e., these codes are treated as supervised labels to train the second network). By doing this, DAgH can make full use of the most critical information contained in images to guide the second hashing network in order to learn efficient hash codes in a true end-to-end fashion. Results from our experiments demonstrate that DAgH can generate high-quality hash codes and it outperforms the current state-of-the-art methods on three benchmark datasets: CIFAR-10, NUS-WIDE, and ImageNet.https://ieeexplore.ieee.org/document/8607076/Supervised learning-based hashingattention-guided strategy
collection DOAJ
language English
format Article
sources DOAJ
author Zhan Yang
Osolo Ian Raymond
Wuqing Sun
Jun Long
spellingShingle Zhan Yang
Osolo Ian Raymond
Wuqing Sun
Jun Long
Deep Attention-Guided Hashing
IEEE Access
Supervised learning-based hashing
attention-guided strategy
author_facet Zhan Yang
Osolo Ian Raymond
Wuqing Sun
Jun Long
author_sort Zhan Yang
title Deep Attention-Guided Hashing
title_short Deep Attention-Guided Hashing
title_full Deep Attention-Guided Hashing
title_fullStr Deep Attention-Guided Hashing
title_full_unstemmed Deep Attention-Guided Hashing
title_sort deep attention-guided hashing
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2019-01-01
description With the rapid growth of multimedia data (e.g., image, audio, and video) on the Web, the learning-based hashing techniques, such as deep supervised hashing, have proven to be very efficient for large-scale multimedia search. The recent successes seen in the learning-based hashing methods are largely due to the success of the deep learning-based hashing methods. However, there are some limitations to the previous learning-based hashing methods (e.g., the learned hash codes containing repetitive and highly correlated information). In this paper, we propose a novel learning-based hashing method, named deep attention-guided hashing (DAgH). DAgH is implemented using two stream frameworks. The core idea is to use the guided hash codes which are generated by the hashing network of the first stream framework (called the first hashing network) to guide the training of the hashing network of the second stream framework (called the second hashing network). Specifically, in the first network, it leverages an attention network and hashing network to generate the attention-guided hash codes from the original images. The loss function we propose contains two components: the semantic loss and the attention loss. The attention loss is used to punish the attention network to obtain the salient region from pairs of images; in the second network, these attention-guided hash codes are used to guide the training of the second hashing network (i.e., these codes are treated as supervised labels to train the second network). By doing this, DAgH can make full use of the most critical information contained in images to guide the second hashing network in order to learn efficient hash codes in a true end-to-end fashion. Results from our experiments demonstrate that DAgH can generate high-quality hash codes and it outperforms the current state-of-the-art methods on three benchmark datasets: CIFAR-10, NUS-WIDE, and ImageNet.
topic Supervised learning-based hashing
attention-guided strategy
url https://ieeexplore.ieee.org/document/8607076/
work_keys_str_mv AT zhanyang deepattentionguidedhashing
AT osoloianraymond deepattentionguidedhashing
AT wuqingsun deepattentionguidedhashing
AT junlong deepattentionguidedhashing
_version_ 1724192259344171008