Sparse-Inductive Generative Adversarial Hashing for Nearest Neighbor
Search
- URL: http://arxiv.org/abs/2306.06928v1
- Date: Mon, 12 Jun 2023 08:07:23 GMT
- Title: Sparse-Inductive Generative Adversarial Hashing for Nearest Neighbor
Search
- Authors: Hong Liu
- Abstract summary: We propose a novel unsupervised hashing method, termed Sparsity-Induced Generative Adversarial Hashing (SiGAH)
SiGAH encodes large-scale high-scale high-dimensional features into binary codes, which solves the two problems through a generative adversarial training framework.
Experimental results on four benchmarks, i.e. Tiny100K, GIST1M, Deep1M, and MNIST, have shown that the proposed SiGAH has superior performance over state-of-the-art approaches.
- Score: 8.020530603813416
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised hashing has received extensive research focus on the past
decade, which typically aims at preserving a predefined metric (i.e. Euclidean
metric) in the Hamming space. To this end, the encoding functions of the
existing hashing are typically quasi-isometric, which devote to reducing the
quantization loss from the target metric space to the discrete Hamming space.
However, it is indeed problematic to directly minimize such error, since such
mentioned two metric spaces are heterogeneous, and the quasi-isometric mapping
is non-linear. The former leads to inconsistent feature distributions, while
the latter leads to problematic optimization issues. In this paper, we propose
a novel unsupervised hashing method, termed Sparsity-Induced Generative
Adversarial Hashing (SiGAH), to encode large-scale high-dimensional features
into binary codes, which well solves the two problems through a generative
adversarial training framework. Instead of minimizing the quantization loss,
our key innovation lies in enforcing the learned Hamming space to have similar
data distribution to the target metric space via a generative model. In
particular, we formulate a ReLU-based neural network as a generator to output
binary codes and an MSE-loss based auto-encoder network as a discriminator,
upon which a generative adversarial learning is carried out to train hash
functions. Furthermore, to generate the synthetic features from the hash codes,
a compressed sensing procedure is introduced into the generative model, which
enforces the reconstruction boundary of binary codes to be consistent with that
of original features. Finally, such generative adversarial framework can be
trained via the Adam optimizer. Experimental results on four benchmarks, i.e.,
Tiny100K, GIST1M, Deep1M, and MNIST, have shown that the proposed SiGAH has
superior performance over the state-of-the-art approaches.
Related papers
- Hamming Encoder: Mining Discriminative k-mers for Discrete Sequence
Classification [1.6693049653540362]
Existing pattern-based methods measure the discriminative power of each feature individually during the mining process.
It is difficult to ensure the overall discriminative performance after converting sequences into feature vectors.
We propose Hamming, which utilizes a binarized 1D-convolutional neural network (1DCNN) architecture to mine discriminative k-mer sets.
arXiv Detail & Related papers (2023-10-16T12:03:27Z) - Improving Dual-Encoder Training through Dynamic Indexes for Negative
Mining [61.09807522366773]
We introduce an algorithm that approximates the softmax with provable bounds and that dynamically maintains the tree.
In our study on datasets with over twenty million targets, our approach cuts error by half in relation to oracle brute-force negative mining.
arXiv Detail & Related papers (2023-03-27T15:18:32Z) - Towards Better Out-of-Distribution Generalization of Neural Algorithmic
Reasoning Tasks [51.8723187709964]
We study the OOD generalization of neural algorithmic reasoning tasks.
The goal is to learn an algorithm from input-output pairs using deep neural networks.
arXiv Detail & Related papers (2022-11-01T18:33:20Z) - Asymmetric Scalable Cross-modal Hashing [51.309905690367835]
Cross-modal hashing is a successful method to solve large-scale multimedia retrieval issue.
We propose a novel Asymmetric Scalable Cross-Modal Hashing (ASCMH) to address these issues.
Our ASCMH outperforms the state-of-the-art cross-modal hashing methods in terms of accuracy and efficiency.
arXiv Detail & Related papers (2022-07-26T04:38:47Z) - HHF: Hashing-guided Hinge Function for Deep Hashing Retrieval [14.35219963508551]
latent codes extracted by textbfDeep textbfNeural textbfNetwork (DNN) will inevitably lose semantic information during the binarization process.
textbfHashing-guided textbfHinge textbfFunction (HHF) is proposed to avoid such conflict.
In detail, we carefully design a specific inflection point, which relies on the hash bit length and category numbers to balance metric learning and quantization learning.
arXiv Detail & Related papers (2021-12-04T03:16:42Z) - CIMON: Towards High-quality Hash Codes [63.37321228830102]
We propose a new method named textbfComprehensive stextbfImilarity textbfMining and ctextbfOnsistency leartextbfNing (CIMON)
First, we use global refinement and similarity statistical distribution to obtain reliable and smooth guidance. Second, both semantic and contrastive consistency learning are introduced to derive both disturb-invariant and discriminative hash codes.
arXiv Detail & Related papers (2020-10-15T14:47:14Z) - Self-Supervised Bernoulli Autoencoders for Semi-Supervised Hashing [1.8899300124593648]
This paper investigates the robustness of hashing methods based on variational autoencoders to the lack of supervision.
We propose a novel supervision method in which the model uses its label distribution predictions to implement the pairwise objective.
Our experiments show that both methods can significantly increase the hash codes' quality.
arXiv Detail & Related papers (2020-07-17T07:47:10Z) - Generative Semantic Hashing Enhanced via Boltzmann Machines [61.688380278649056]
Existing generative-hashing methods mostly assume a factorized form for the posterior distribution.
We propose to employ the distribution of Boltzmann machine as the retrievalal posterior.
We show that by effectively modeling correlations among different bits within a hash code, our model can achieve significant performance gains.
arXiv Detail & Related papers (2020-06-16T01:23:39Z) - Error-Corrected Margin-Based Deep Cross-Modal Hashing for Facial Image
Retrieval [26.706148476396105]
Cross-modal hashing facilitates mapping of heterogeneous multimedia data into a common Hamming space.
We propose a novel cross-modal hashing architecture-deep neural decoder cross-modal hashing (DNDCMH)
arXiv Detail & Related papers (2020-04-03T08:20:08Z) - Learning to Hash with Graph Neural Networks for Recommender Systems [103.82479899868191]
Graph representation learning has attracted much attention in supporting high quality candidate search at scale.
Despite its effectiveness in learning embedding vectors for objects in the user-item interaction network, the computational costs to infer users' preferences in continuous embedding space are tremendous.
We propose a simple yet effective discrete representation learning framework to jointly learn continuous and discrete codes.
arXiv Detail & Related papers (2020-03-04T06:59:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.