Loading…
Adaptive multi-bit quantization for hashing
Recently, hashing methods which try to solve similarity-preserving approximate nearest search problem have obtained widely applications in various fields such as content-based image retrieval, object recognition and pose estimation. However, how to learn effective hash codes to describe the similari...
Saved in:
Published in: | Neurocomputing (Amsterdam) 2015-03, Vol.151, p.319-326 |
---|---|
Main Authors: | , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Recently, hashing methods which try to solve similarity-preserving approximate nearest search problem have obtained widely applications in various fields such as content-based image retrieval, object recognition and pose estimation. However, how to learn effective hash codes to describe the similarities in the large-scale database still remains as a NP-hard problem. Besides, a fatal problem lying in the existing hashing methods is that they usually threshold the real values to binary codes using single-bit quantization (SBQ) at the highest point density, which may destroy the data structure seriously. Due to this problem, double-bit quantization (DBQ) is proposed to solve the problem of SBQ by adaptively learning thresholds to quantize the real values to two bits, and achieves impressive results. However, one problem in DBQ is that it neglects the amount of the information contained in different data dimensions. In this paper, we propose a multi-bit quantization method based on bit allocation to quantize each projected dimension with variable bit numbers. Besides, different from existing methods of choosing threshold, we propose an incomplete coding manner by clustering to generate binary codes. Experiments on two large datasets demonstrate the feasibility of our method.
•We propose an adaptive multi-bit quantization for hashing by fully considering the data distribution.•Our method incorporates bit allocation, incomplete binary encoding and clustering-based quantization into a joint framework.•Experiments on real datasets demonstrate that our AMBQ method outperforms most of the state-of-art methods. |
---|---|
ISSN: | 0925-2312 1872-8286 |
DOI: | 10.1016/j.neucom.2014.09.033 |