Paper-Reading

Deep Hashing

tags: Deep Learning, Computer Vision, Image Retrieval, Hashing, Survey

Hashing BaseLine: https://github.com/willard-yuan/hashing-baseline-for-image-retrieval

Some representative papers about deep hashing

Recent Papers

  1. (CNNH) Supervised Hashing via Image Representation Learning [paper][code][slide]
    Rongkai Xia, Yan Pan, Hanjiang Lai, Cong Liu, and Shuicheng Yan. [AAAI], 2014
    • (NINH) Simultaneous Feature Learning and Hash Coding with Deep Neural Networks [paper]
      Hanjiang Lai, Yan Pan, Ye Liu, and Shuicheng Yan. [CVPR], 2015
    • (DRSDH) Bit-Scalable Deep Hashing With Regularized Similarity Learning for Image Retrieval and Person Re-Identification [paper][code]
      Ruimao Zhang, Liang Lin, Rui Zhang, Wangmeng Zuo, and Lei Zhang. [TIP], 2015
    • Convolutional Neural Networks for Text Hashing [paper]
      Jiaming Xu, PengWang, Guanhua Tian, Bo Xu, Jun Zhao, Fangyuan Wang, Hongwei Hao. [IJCAI], 2015
    • (DSRH) Deep Semantic Ranking Based Hashing for Multi-Label Image Retrieval [paper][code]
      Fang Zhao, Yongzhen Huang, Liang Wang, and Tieniu Tan. [CVPR], 2015
    • (DH) Deep Hashing for Compact Binary Codes Learning [paper]
      Venice Erin Liong, Jiwen Lu, Gang Wang, Pierre Moulin, and Jie Zhou. [CVPR], 2015
    • Deep Learning of Binary Hash Codes for Fast Image Retrieval [paper][code][questions]
      Kevin Lin, Huei-Fang Yang, Jen-Hao Hsiao, and Chu-Song Chen. [CVPRW], 2015
    • (DPSH) Feature Learning based Deep Supervised Hashing with Pairwise Labels [paper][code]
    • Deep Learning to Hash with Multiple Representations [paper]
      Yoonseop Kang, Saehoon Kim, Seungjin Choi. [ACMMM], 2012
    • Inductive Transfer Deep Hashing for Image Retrieval[paper]
      Xinyu Ou, Lingyu Yan, Hefei Ling∗ , Cong Liu, Maolin Liu
    • A Deep Hashing Learning Network [paper]

Details

CNNH-stage1

2.In the second stage, the idea is to simultaneously learn a good feature representation for the input images as well as a set of hash functions, via a deep convolutional network tailored to the learned hash codes in \(H\) and optionally the discrete class labels of the images. (Using Alexnet)

CNNH

triplets DRSCH

  1. Deep Hash Functions: \(\displaystyle h(x; w) = sign(w^T[f_a(x); f_b(x)])\)
  2. Semantic Ranking Supervision: (preserve multilevel semantic structure)various evaluation criteria can be used to measure the consistency of the rankings predicted by hash functions, such as the Normalized Discounted Cu- mulative Gain (NDCG) score: \(\displaystyle NDCG@p=\frac{1}{Z}\sum\limits_{i=1}^{p}\frac{2^{r_i}-1}{\log(1+i)}\), where \(p\) is the truncated position in a ranking list, \(Z\) is a normalization constant to ensure that the NDCG score for the correct ranking is one, and \(r_i\) is the similarity level of the \(i\)-th database point in the ranking list.
  3. Optimization with Surrogate Loss:
    Given a query \(q\) and a ranking list \(\{x_i\}^M_{i=1}\) for \(q\), we can define a ranking loss on a set of triplets of hash codes as follows: \(\displaystyle L_\omega(h(q),\{h(x_i)\}^M_{i=1})=\sum\limits_{i=1}^M\sum\limits_{j:r_j<r_i}\omega(r_i, r_j)[d_H(h(q), h(x_i)) - d_H(h(q), h(x_j))+\rho]_+\). According to NDCD, weight \(\omega(r_i, r_j)= \frac{2^{r_i}-2^{r_j}}{Z}\)
    The objective function can be given by the empirical loss subject to some regularization:
\[\mathcal{F}(W)=\sum\limits_{q\in\mathcal{D}, \{x_i\}_{i=1}^M\subset\mathcal{D}}L_{\omega} ( h(q;W) , \{h(x_i;W) \} ^M_{i=1}) + \frac{\alpha}{2} || mean_{q}(h(q;W))||_2^2 + \frac{\beta}{2}||W||_2^2\]

And calculate derivative values.

DSRH DSRH-2