Deep Attention Residual Hashing

Yang LI  Zhuang MIAO  Ming HE  Yafei ZHANG  Hang LI  

IEICE TRANSACTIONS on Fundamentals of Electronics, Communications and Computer Sciences   Vol.E101-A   No.3   pp.654-657
Publication Date: 2018/03/01
Online ISSN: 1745-1337
DOI: 10.1587/transfun.E101.A.654
Type of Manuscript: LETTER
Category: Image
hashing,  residual network,  attention layer,  image retrieval,  

Full Text: PDF>>
Buy this Article

How to represent images into highly compact binary codes is a critical issue in many computer vision tasks. Existing deep hashing methods typically focus on designing loss function by using pairwise or triplet labels. However, these methods ignore the attention mechanism in the human visual system. In this letter, we propose a novel Deep Attention Residual Hashing (DARH) method, which directly learns hash codes based on a simple pointwise classification loss function. Compared to previous methods, our method does not need to generate all possible pairwise or triplet labels from the training dataset. Specifically, we develop a new type of attention layer which can learn human eye fixation and significantly improves the representation ability of hash codes. In addition, we embedded the attention layer into the residual network to simultaneously learn discriminative image features and hash codes in an end-to-end manner. Extensive experiments on standard benchmarks demonstrate that our method preserves the instance-level similarity and outperforms state-of-the-art deep hashing methods in the image retrieval application.