Preserved in Portico This version is not peer-reviewed
Self-Attention and Adversary Guided Hashing Network for Cross-Modal Retrieval
: Received: 16 September 2020 / Approved: 18 September 2020 / Online: 18 September 2020 (04:16:58 CEST)
A peer-reviewed article of this Preprint also exists.
Journal reference: Computers & Electrical Engineering 2021
Recently deep cross-modal hashing networks have received increasing interests due to its superior query efficiency and low storage cost. However, most of existing methods concentrate less on hash representations learning part, which means the semantic information of data cannot be fully used. Furthermore, they may neglect the high-ranking relevance and consistency of hash codes. To solve these problems, we propose a Self-Attention and Adversary Guided Hashing Network (SAAGHN). Specifically, it employs self-attention mechanism in hash representations learning part to extract rich semantic relevance information. Meanwhile, in order to keep invariability of hash codes, adversarial learning is adopted in the hash codes learning part. In addition, to generate higher-ranking hash codes and avoid local minima early, a new batch semi-hard cosine triplet loss and a cosine quantization loss are proposed. Extensive experiments on two benchmark datasets have shown that SAAGHN outperforms other baselines and achieves the state-of-the-art performance.
adversarial learning; deep cross-modal hashing; self-attention mechanism
MATHEMATICS & COMPUTER SCIENCE, Artificial Intelligence & Robotics
This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
We encourage comments and feedback from a broad range of readers. See criteria for comments and our diversity statement.