• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
搜索

Author:

Zhou, Suqing (Zhou, Suqing.) | Han, Yu (Han, Yu.) | Chen, Ning (Chen, Ning.) | Huang, Siyu (Huang, Siyu.) | Igorevich, Kostromitin Konstantin (Igorevich, Kostromitin Konstantin.) | Luo, Jia (Luo, Jia.) | Zhang, Peiying (Zhang, Peiying.)

Indexed by:

EI SCIE

Abstract:

Cross-modal hashing retrieval has attracted extensive attention due to its low storage requirements as well as high retrieval efficiency. In particular, how to more fully exploit the correlation of different modality data and generate a more distinguished representation is the key to improving the performance of this method. Moreover, Transformer-based models have been widely used in various fields, including natural language processing, due to their powerful contextual information processing capabilities. Based on these motivations, we propose a Transformer-based Distinguishing Strong Representation Deep Hashing (TDSRDH). For text modality, since the sequential relations between words imply semantic relations that are not independent relations, we thoughtfully encode them using a transformer-based encoder to obtain a strong representation. In addition, we propose a triple-supervised loss based on the commonly used pairwise loss and quantization loss. The latter two ensure the learned features and hash-codes can preserve the similarity of the original data during the learning process. The former ensures that the distance between similar instances is closer and the distance between dissimilar instances is farther. So that TDSRDH can generate more discriminative representations while preserving the similarity between modalities. Finally, experiments on the three datasets MIRFLICKR-25K, IAPR TC-12, and NUS-WIDE demonstrated the superiority of TDSRDH over the other baselines. Moreover, the effectiveness of the proposed idea was demonstrated by ablation experiments.

Keyword:

Transformers Feature extraction Deep learning Cross-modal retrieval Correlation Hash functions Codes transformer Representation learning discriminative representation Semantics strong representation deep hashing

Author Community:

  • [ 1 ] [Zhou, Suqing]Fujian Polytech Informat Technol, Internet Things & Artificial Intelligence Coll, Fuzhou 350003, Peoples R China
  • [ 2 ] [Han, Yu]China Mobile Grp Shandong Co Ltd, Jinan 250001, Peoples R China
  • [ 3 ] [Chen, Ning]China Univ Petr East China, Qingdao Inst Software, Coll Comp Sci & Technol, Qingdao 266580, Peoples R China
  • [ 4 ] [Huang, Siyu]Chinese Acad Sci, Xiongan Inst Innovat, Baoding 071702, Peoples R China
  • [ 5 ] [Igorevich, Kostromitin Konstantin]South Ural State Univ, Dept Phys Nanoscale Syst, Chelyabinsk 454080, Russia
  • [ 6 ] [Luo, Jia]Beijing Univ Technol, Coll Econ & Management, Beijing 100124, Peoples R China
  • [ 7 ] [Zhang, Peiying]Qilu Univ Technol, Shandong Acad Sci, Shandong Comp Sci Ctr,Minist Educ, Natl Supercomp Ctr Jinan,Key Lab Comp Power Networ, Jinan 250353, Peoples R China
  • [ 8 ] [Zhang, Peiying]Shandong Fundamental Res Ctr Comp Sci, Shandong Prov Key Lab Comp Networks, Jinan 250353, Peoples R China

Reprint Author's Address:

Show more details

Related Keywords:

Source :

IEEE ACCESS

ISSN: 2169-3536

Year: 2023

Volume: 11

Page: 140041-140055

3 . 9 0 0

JCR@2022

Cited Count:

WoS CC Cited Count: 0

SCOPUS Cited Count:

ESI Highly Cited Papers on the List: 0 Unfold All

WanFang Cited Count:

Chinese Cited Count:

30 Days PV: 16

Affiliated Colleges:

Online/Total:228/10507564
Address:BJUT Library(100 Pingleyuan,Chaoyang District,Beijing 100124, China Post Code:100124) Contact Us:010-67392185
Copyright:BJUT Library Technical Support:Beijing Aegean Software Co., Ltd.