RelaHash: deep hashing with relative position

PVT Minh, NDD Viet, NT Son, BN Anh, J Jaafar - IEEE Access, 2023 - ieeexplore.ieee.org
IEEE Access, 2023ieeexplore.ieee.org
Deep hashing has been widely used as a solution to encoding binary hash code for
approximating nearest neighbor problem. It has been showing superior performance in
terms of its ability to index high-level features by learning compact binary code. Many recent
state-of-the-art deep hashing methods often use multiple loss terms at once, thus introducing
optimization difficulty and may result in sub-optimal hash codes. OrthoHash was proposed to
replace those losses with just a single loss function. However, the quantization error …
Deep hashing has been widely used as a solution to encoding binary hash code for approximating nearest neighbor problem. It has been showing superior performance in terms of its ability to index high-level features by learning compact binary code. Many recent state-of-the-art deep hashing methods often use multiple loss terms at once, thus introducing optimization difficulty and may result in sub-optimal hash codes. OrthoHash was proposed to replace those losses with just a single loss function. However, the quantization error minimization problem in OrthoHash is still not addressed effectively. In this paper, we take one step further - propose a single-loss model that can effectively minimize the quantization error without explicit loss terms. Specifically, we introduce a new way to measure the similarity between the relaxed codes with centroids, called relative similarity. The relative similarity is the similarity between the relative position representation of continuous codes and the normalized centroids. The resulting model outperforms many state-of-the-art deep hashing models on popular benchmark datasets.
ieeexplore.ieee.org
Showing the best result for this search. See all results