Loading…

Data-Focus Proxy Hashing

Cross-modal hashing approaches are designed to transform data from disparate modalities into a shared Hamming space while preserving the semantic similarity relationships between the modalities. Generally, most supervised cross-modal hashing methods treat the entire data as input. However, the total...

Full description

Saved in:
Bibliographic Details
Main Authors: Ji, Wenjin, Mao, Xianling, Liu, Jinyu, Tu, Rongcheng, Huang, Heyan
Format: Conference Proceeding
Language:English
Subjects:
Online Access:Request full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Cross-modal hashing approaches are designed to transform data from disparate modalities into a shared Hamming space while preserving the semantic similarity relationships between the modalities. Generally, most supervised cross-modal hashing methods treat the entire data as input. However, the total information within the data encompasses both vital semantic details and extraneous background noise. Such background information may introduce noise and undermine the preservation of semantic similarity relationships between modalities. To address this issue, we propose a novel Data-Focus Proxy Hashing (DFPH) for cross-modal retrieval, which emphasizes important data information. Our approach begins by training an intermediary network designed to produce a unique proxy hash code specific to each individual category. Then, by incorporating attention mechanisms, and vital textual information, we first design a novel modality-specific hashing network, including an image-focused hash codes generator and a text-focused hash codes generator. Subsequently, we utilize the hashing loss function to train the hashing network under the supervision of these proxy hash codes. Extensive experiments on two benchmarks demonstrate that our proposed DFPH outperforms existing baselines in cross-modal retrieval tasks. Our code is available at https://github.com/JWJ990626/DFPH.
ISSN:2768-1904
DOI:10.1109/CSCWD61410.2024.10580005