Next Article in Journal
Cone-Beam Computed Tomography Analysis of the Root Canal Morphology of Lower Second Molars in the Croatian Subpopulation
Previous Article in Journal
A Comparative Study of Oleuropein Extraction from Wild Olive Leaves (Olea europea subsp. oleaster, Hoffmanns. & Link), Its Gastrointestinal Stability, and Biological Potential
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Grained Similarity Preserving and Updating for Unsupervised Cross-Modal Hashing

College of Information and Intelligence, Hunan Agricultural University, Changsha 410128, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(2), 870; https://doi.org/10.3390/app14020870
Submission received: 18 December 2023 / Revised: 15 January 2024 / Accepted: 17 January 2024 / Published: 19 January 2024
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Unsupervised cross-modal hashing is a topic of considerable interest due to its advantages in terms of low storage costs and fast retrieval speed. Despite the impressive achievements of existing solutions, two challenges remain unaddressed: (1) Semantic similarity obtained without supervision is not accurate enough, and (2) the preservation of similarity structures lacks effectiveness due to the neglect of both global and local similarity. This paper introduces a new method, Multi-Grained Similarity Preserving and Updating (MGSPU), to tackle these challenges. To overcome the first challenge, MGSPU employs a newly designed strategy to update the semantic similarity matrix, effectively generating a high-confidence similarity matrix by eliminating noise in the original cross-modal features. For the second challenge, a novel multi-grained similarity preserving method is proposed, aiming to enhance cross-modal hash code learning by learning consistency in multi-grained similarity structures. Comprehensive experiments on two widely used datasets with nine state-of-the-art competitors validate the superior performance of our method in cross-modal hashing.

1. Introduction

As an efficient information retrieval paradigm for big multimedia data [1,2,3], cross-modal retrieval [4,5,6,7,8,9] utilizes one modality as a query to search another modal data. Among the existing techniques [10,11,12,13,14,15,16,17,18], cross-modal hashing [16,17,18,19,20] is popular for its fast retrieval speed and low storage cost. The core of cross-modal hashing is to map high-dimensional data to a low-dimensional common Hamming space, in which multi-modal instances (text, image, audio and video) with similar semantics are closer, and dissimilar instances are far apart. Notwithstanding the strong practical significance, this task is extremely challenging due to the heterogeneity gap between different modalities.
A number of studies are striving to promote the progress of cross-modal hashing. In general, they can be simply divided into two groups: supervised learning methods [21,22,23,24,25] and unsupervised learning methods [26,27,28,29,30]. The supervised learning methods use human-annotated label to guide model learning, which preserves semantic discrimination well to obtain great performance. However, directly obtaining the category information is really difficult in real-world scenarios, while the cost of manual labeling is so huge as to limit the wide application. In contrast, without relying on category annotations, unsupervised learning methods are great at capturing the underlying correlation patterns between different modalities to bridge the heterogeneity gap. Thus, they have gained considerable attention currently.
Motivation. Although existing researches have shown remarkable results on large-scale datasets, unsupervised cross-modal hashing is yet a Herculean task that await further exploration due to the following challenges. The first challenge is how to accurately measure inter-modal and intra-modal similarity without category information. In the early days, several unsupervised learning methods [31,32] used discrete models to calculate binary similarity, namely, defining the similarity between two instances with small distances as “1”, otherwise “−1”. Unfortunately, the naive similarity measurement is far from accurate. To break this limitation, researchers [12,15,33,34] began to use original feature distance to measure the continuous similarity. However, such similarity is still not accurate enough to represent the real relationship between instances due to the noise existing in their original features. The second challenge is how to preserve similarity structure consistency from both local and global view during hash learning process. Undisputedly, using multi-grained similarity relationships across modalities is crucial to narrow cross-modal heterogeneity. However, current methods [35,36,37] either overlook multi-grained similarity learning across various modalities or struggle to capture a consistent distribution of relations within the content details. Additionally, some state-of-the-art methods [38,39,40,41] do not account for the semantic correlation between feature vectors from one modality and the corresponding hash codes from another modality. To conquer these challenges, this paper has made the following two efforts: (1) designing a novel and effective similarity updating strategy to rise the accuracy of underlying similarity measurement, and (2) combining multi-grained similarity information from local and global view to preserve similarity consistency.
Our Method. We propose a novel cross-modal hashing framework, named Multi-Grained Similarity Preserving and Updating (MGSPU). As shown in Figure 1, this framework consists four module, i.e., feature extraction, semantic similarity matrix updating, hash code learning and hash function learning. To overcome the first challenge, a novel semantic similarity matrix update strategy is developed, which removes noises from the similarity matrix so as to obtain high-confidence supervisory signal. To face the second challenge, we capture the multi-grained similarity structure information from global and local view to preserve similarity consistency better. For global view, we use graph convolutional network (GCN for short) to aggregate the similarity structure information from the neighbors of each instance to enrich the coarse-grained similarity information. For local view, cross-modal attention mechanisms are used to perform cross-modal interaction to enhance similarity learning between instance pairs. Furthermore, we use a similarity consistency reconstruction method to ensure the similarity consistency of hash codes.
Contributions. In summary, the contributions of this paper can be summarized as follows:
  • We propose an effective unsupervised learning framework, called Multi-Grained Similarity Preserving and Updating, which learns high-quality hash codes by comprehensively improving cross-modal similarity learning.
  • We propose a novel semantic similarity matrix updating strategy to effectively remove noises in the original similarity matrix, which produces high-confidence supervisory signal for cross-modal hashing learning.
  • We propose a novel multi-grained similarity preserving method to enhance similarity consistency preserving of the cross-modal hash codes.
  • We conducted extensive experiments on the widely used datasets MIRFLICKR-25K and NUS-WIDE to validate the superiority of the proposed approach and evaluate the effectiveness of each component.
Roadmap. The paper is structured as follows: Section 2 provides a summary of related works. Section 3 outlines the problem definition and details the MGSPU approach. Section 4 presents the experimental results and analysis. Lastly, Section 5 concludes the paper.

2. Related Work

According to literatures, cross-modal hashing techniques are generally categorized into two groups: supervised and unsupervised cross-modal hashing. This section reviews the prevailing solutions related to this paper, which are summarized in Table 1 to ease reading.

2.1. Supervised Cross-Modal Hashing

To obtain common binary representations, supervised cross-modal hashing methods use category information to maintain semantic discrimination. For example, semantic correlation maximization (SCM [42]) learns hash functions by constructing and maintaining semantic similarity matrices. Semantics-preserving hashing (SePH [43]) converts the semantic matrix into a probability distribution and minimizes the Kullback-Leibler divergence so that the learned hash codes are approximately distributed in Hamming space. Thanks to the prosperity of deep learning, researchers began to realize cross-modal retrieval by deep models that explore more discriminative features. Deep cross-modal hashing (DCMH [17]) is a pioneering achievement in this field, which perfectly combines hash function learning with deep feature extraction, and cleverly exploits the labelling information to construct similarity matrices, thus preserving the subtle similarity relationships in cross-modal data. Meanwhile, cross-modal hamming hashing (CMHH [44]) introduces a pairwise focusing loss based on exponential distribution. It penalises instances with similar semantic content but the Hamming distance exceeds a predefined threshold. Semi-supervised adversarial deep hashing (SSAH [45]) is a groundbreaking innovation that shifts the focus to self-supervised methods and integrates adversarial learning into cross-modal hashing, thus making significant progress in the field. Ranking-based deep cross-modal hashing (RDCMH [46]) uses maximum marginal loss to learn uniform Hamming representations. In addition, deep normalized cross-modal hashing with bi-direction relation reasoning (Bi-NCMH [47]) achieved excellent retrieval performance by constructing high-quality similarity matrices to capture similarity relations between instances with multiple labels. Despite the impressive performance of these techniques, their inherent limitations cannot be ignored: they rely heavily on manual annotation to obtain supervision.

2.2. Unsupervised Cross-Modal Hashing

Unsupervised cross-modal hashing methods do not rely on labels so as to apply in real-world scenarios easily. In the existing solutions, constructing a semantic similarity matrix is the key to guide cross-modal relationship learning.Among them, an ingenious method is Deep Joint Semantic Reconstruction Hash (DJSRH [12]), which proposes a joint semantic similarity matrix to simultaneously integrate multi-modal similarity information of cross-modal instances. However, the similarity matrix in DJSRH introduces redundant information from intra-modal fusion items. As an improvement, joint-modal distribution-based similarity hashing (JDSH [36]) involves distribution-based similarity decision and weighting to learn more discriminative hash codes. Deep adaptively-enhanced hashing (DAEH [27]) utilizes distance distributions and similarity ratio information to estimate comparable similarity relationships as complementarity of simple feature distance-based metrics. Although competitive performance is achieved, lack of accurate similarity measurement are these methods since the similarity matrix they rely on are constructed from the original features with noises.
Other unsupervised learning studies make efforts on narrowing the heterogeneity gap by preserving similarity structure consistency. For example, deep graph-neighbor coherence preserving network (DGCPN [48]) explores the relationship information between data and their neighbors to capture neighborhood structure coherence. Aggregation-based graph convolutional hashing (AGCH [26]) employs a GCN to deeply explore the underlying neighborhood structure. Deep relative neighbor relationship preserving hashing (DRNPH [49]) method excavates deep relative neighbor relationships in common Hamming space via binary feature vector based intra- and inter-modal neighbor matrix reconstruction. Impressive progress had been made by these works, however, a common weakness they suffer from is lack of both global and local similarity learning from multi-modal contents.
To address the above shortcomings, we attempt to boost cross-modal hashing learning from two aspects: (1) trying to rise the quality of similarity matrix by reducing noises from original features, and (2) inviting multi-grained similarity learning strategy to capture both global and local similarity relationships. As a results, a novel unsupervised cross-modal hashing technique called multi-grained similarity preserving and updating is developed. The detailed details of this approach will be explored in depth in the next section.

3. Methodology

This section details the proposed MGSPU method across six aspects. Initially, we introduce the notations and problem definition in the Preliminary. Next is an overview of MGSPU, feature extraction, semantic similarity matrix updating, hashing learning (including hash code learning and hashing function learning), and finally a discussion of the optimization algorithm. All abbreviations involved are summarized in Table 2 for easier reading.

3.1. Preliminary

Notations. For clarity and simplicity, calligraphy uppercase letters, such as O , denote sets. Bold uppercase letters, such as W , represent matrices. Bold lowercase letters, such as w , indicate vectors. In addition, the ij-th element of W is represented as W i j , the i-th row of W is represented as W i , the j-th column of W is represented as W j , W T is the transpose of W , I represents identity matrix. . represents the Frobenius norm of a matrix. s g n ( . ) represents a sign function, shown as below:
s g n ( x ) = 1 , x 0 1 , x < 0 .
For ease of comprehension, commonly utilized mathematical notations have been compiled in Table 3 to assist in reference.
Problem Definition. This paper concentrates on cross-modal retrieval between two prevalent modalities, i.e., image v and text t. Let O = ( v i , t i } i = 1 n be a cross-modal dataset, v i R d v and t i R d t refer to the image and text in the pairwise instance respectively, d v and d t represent the dimension of the corresponding features, n denotes the number of instances. Our study aims to learn two hash functions f v = ( · ; θ v ) and f t = ( · ; θ t ) to map images and texts to a common Hamming space: B v = s g n ( f v = ( { v i } i = 1 n ; θ v ) ) , B t = s g n ( f t = ( { t i } i = 1 n ; θ t ) ) , where B v , B t 1 , 1 n × C , C represents the length of the binary code, θ v and θ t are the learnable parameters. In addition, the similarity between two binary codes B i v and B j t is measured by Hamming distance D B i v , B j t = 1 2 C < B i v , B j t > , which is the base for implementing cross-modal hashing retrieval.

3.2. Overview of MGSPU

Figure 1 shows the overview of the MGSPU framework. Specifically, it mainly consists of four modules: (1) feature extraction module, (2) semantic similarity matrix updating module, (3) hash code learning module, and (4) hash function learning module. The feature extraction module maps images and texts into feature subspaces by corresponding encoders: deep CNN for image features and Bag-of-Word (BoW) model for text features. The semantic similarity matrix updating module realizes a novel similarity updating strategy to construct a high-confidence similarity matrix, which effectively enhance the hash code learning and hash function learning. To support hash function learning, the hash code learning module aims at generating real-value hash representations with a novel multi-grained semantic preserving learning strategy. It utilizes a pair of GCNs (each for one modality) to capture global semantic correlation, following by a cross-attention mechanism to enhance cross-modal local feature learning. This technique allows for sufficient semantic correlations preserving of both intra- and inter-modality. The hash function learning module is to learn two hash functions that are implemented by two fully-connected neural networks for image and text, respectively.
In a nutshell, the goal of this method is to optimize the hash function module with other three parts to generate high-quality binary hash codes for cross-modal retrieval task. Thereinafter, we discuss the technical details module by module.

3.3. Feature Extraction

The first step of cross-modal hash code learning is feature extraction from both images and texts. For images, we follow previous works [12,37,48] to extract deep features from CNN model (pre-trained on ImageNet). Specifically, we extract the the 4096-dimensional features from the first fc7 layer (after ReLU) of AlexNet [50] as the original image features F v R m × 4096 for the batch-input images v i i = 1 m . While for texts t i i = 1 m we directly adopt the BoW [51] vectors as their original features F t R m × d t . It is noted that the proposed MGSPU framework can also be compatible with other deep feature extraction models, such as ResNet [52] for images or LSTM [53] for texts. To better validate the superiority of the proposed technique, we use relatively naive but effective feature extraction models (AlexNet and BoW) when implementing MGSPU. The implementation details are presented in Section 4.3.

3.4. Semantic Similarity Matrix Updating

Unlike supervised hashing learning that do not easily suffer from noises in features due to accurate category annotation, the unsupervised method cannot ensure the high accuracy of latent semantic correlations between instances without category information. To address this issue, we focus on constructing a high-confidence unified similarity matrix S as a supervisory signal to filter out noises as much as possible. Accordingly, a novel semantic similarity matrix updating strategy is designed.

3.4.1. Similarity Matrix Construction

To comprehensively consider cross-modal similarity relationships, we first fuse the similarity matrices from image modality and text modality:
S f = α 1 S v + ( 1 α 1 ) S t ,
where S f is the fused similarity matrix, S v and S t represent the image and text similarity matrices, respectively. α 1 [ 0 , 1 ] is a hyperparameter to balance them. Each element in S v and S t is computed by Cosine similarity. Taking image modality as an example:
S i j v = c o s ( v i , v j ) = v i T v j v i F 2 v j F 2 ,
where v i and v j denote the feature of the i-th and j-th image. To rise the accuracy of similarity measurement without supervision, we design a filter-and-augment strategy to refine the similarity matrix, shown as follows:
S r = s g n ( S f ) , S i j f > η 1 2 × s i g m o i d ( 2 × S f ) 1 + I , o t h e r w i s e ,
where S r denotes the refined similarity matrix, η 1 is similarity threshold. the sign function s g n ( S f ) indicates if S i j f > η 1 , S i j f = 1 ; if S i j f < η 1 , S i j f = 1 . Meanwhile, inspired by reference [28], we use the non-linear function s i g m o i d ( · ) to compress the remaining values to further expand the distance between similar and non-similar instances. The coefficient 2 of S f is to control the linear compression, then 2 × s i g m o i d ( · ) 1 maps the compressed value to [ 1 , 1 ] . I is an identity matrix to maintain the similarity between instance pairs.

3.4.2. Dual Instruction Fusion Updating

In order to preserve semantic similarity consistency in cross-modal hash codes, a reliable similarity matrix for unsupervised hash function learning is indispensable. During the training process, unfortunately, we found that there is a evident gap between the similarity relationships of cross-modal hash codes and the real relationships of original instances due to the noises existing in the flawed similarity matrix S r . In other words, the value of | S i j r S i j h | is not small enough to correctly guide cross-modal hash function learning. Intuitively, if we rectify these incorrect similarity values, it will effectively embed the correct similarity relationship information into cross-modal hash codes. For this reason, we attempt to construct dual instruction to further denoise the similarity matrix S r : the one is the difference of similarity value of S i j r and S i j h , the other is difference of the sign of them. Beyond all doubt, these two instructions actually measure the difference between S r and S h from two different perspectives, which could be fused to narrow the gap between S r and S h .
From the analysis mentioned, we develop a novel semantic similarity matrix updating strategy, named dual instruction fusion updating, to gradually eliminate the noises. Firstly, we construct the hashing similarity matrix S h via the real-valued hash code H v and H t :
S h = c o s ( H v , H v ) + c o s ( H t , H t ) + c o s ( H v , H t ) .
Thereafter, we utilize S h to update S r according to the similarity relationship among them. For the convenience of formal description and implementation, we introduce two indicator matrices Λ and Γ at first. The former is defined as Λ i j = S i j r S i j h , therein ⊕ is XOR operation, which means if S i j r × S i j h > 0 , Λ i j = 0 ; otherwise Λ i j = 1 . The latter indicates the numerical difference between S i j h and S i j r , therein the elements | S i j r S i j h | > η 2 are set to 1 and the others are set to 0. η 2 [ 0 , 1 ] be a threshold to measure the difference between two elements in similar matrices. Accordingly, two cases are considered: (1) Not updating: if Λ i j = 0 and Γ i j = 0 , then the updating do not be conducted since S h and S r are similar enough. (2) Updating: according to the value of Λ i j , we introduce two updating rules: if Λ i j = 0 and Γ i j = 1 , we update S i j r by S i j h ; if Λ i j = 1 , we update S i j r by “0”, which is a way with maximizing entropy to enhance generalization capability. The reason behind this rule is intuitive: if the difference between S i j r and S i j h is too large, a wise way is not to biased towards either side. To clearly show the update rules, we list a truth table for Λ and Γ in Table 4. Thereby, the updated semantic similarity matrix S is generated by the following updating process:
S i j = S i j r , Λ i j = 0 Γ i j = 0 α 2 S i j r + ( 1 α 2 ) S i j h , Λ i j = 0 Γ i j = 1 0 , o t h e r w i s e ,
where α 2 [ 0 , 1 ] is a parameter. Finally, we use S to guide the cross-modal hash function learning avoiding the interference from similarity noise. Figure 2 illustrates the updating process.

3.5. MGSP for Hash Code Learning

To enhance the cross-modal similarity consistency preserving for hash code learning, a novel technique, called Multi-Grained Similarity Preserving (MGSP) is developed, as shown in Figure 3a. It consists two key sub-modules, i.e., global similarity aggregation (GSA) and local similarity interaction (LSI), which explore multi-grained similarity information: global and local similarity structure information. Furthermore, another sub-module, named Similarity Consistency Reconstruction (SCR) is involved to narrow heterogeneity gap between original features and reconstructed features from hash codes.

3.5.1. Global Similarity Aggregation

The similarity relationship between instances of any modality (image or text) is essential for similarity preserving, which reflects the latent feature distributions of each kind of data. As we known, the intra-modal similarity relationships can be represented as a graph therein each node is an instance and the similarity relationships are represented by edges. In such a graph, each node may be related to others through complex linked structure that can be captured to learn latent global similarity relationships within each modality. To this end, we developed a global similarity aggregation (GSA) sub-module involving the following process: two graphs firstly are constructed, each per modality. Then, we aggregate global similarity information of each node from its neighbors via a graph convolutional network (GCN). Specifically, let { v , t } , for a batch of samples F * , we feed them into an encoder to generate the latent representations L * . To further focus on the crucial features, the key-value self-attention mechanism is used to obtain global similarity aggregation representations  X g * . The query  Q g * , key  K g * and value  V g * are calculated as:
Q g * = L * W g , Q * , K g * = L * W g , K * , V g * = L * W g , V * ,
where W g , Q * , W g , K * , W g , V * are the learnable weight matrices. The attention map M g * is calculated as:
M g * = s o f t m a x Q g * ( K g * ) T d .
Thus, the global similarity enhanced representations is X g * = M g * V g * . To avoid the issue of smaller feature value caused by attention weighting, we add the latent representation L * with X g * , then normalize it as the input of the GCN:
G * ( 0 ) = σ n o r m ( L * + X g * ) .
where the superscript ( 0 ) denotes the 0-th layer of convolution, namely the input of GCN, n o r m ( · ) is normalization function, σ ( · ) represents a nonlinear activation function, usually the R e L U ( · ) function. The layer-by-layer propagation rule of GCN is formulated as:
G ( l ) = σ D ˜ 1 2 A ˜ D ˜ 1 2 G ( l 1 ) W ( l ) ,
where D ˜ i i = j A ˜ i j , A ˜ is an adjacency matrix constructed from the refined similarity matrix S r via kNN algorithm, k represents the number of neighbors, G ( l 1 ) and G ( l ) represent the input and output of the l-th layer, W ( l ) denotes the learnable parameters of the l-th layer, σ ( · ) is R e L U ( · ) function. The final outputs of the GCN are denoted by G . Finally, to further mitigate the cross-modal heterogeneity, a two-layer fully-connected network with weigh-sharing is integrated on the top of GCN, and its output are denoted as Z .

3.5.2. Local Similarity Interaction

Capturing rich similarity structure information among each image-text pair ( v i , t i ) is another desiderata to further enhance cross-modal similarity preserving. This kind of information is mainly reflected in the latent local feature relationships between modalities. To this end, we develop a local similarity interaction (LSI) sub-module to improve similarity consistency preserving within each image-text pair. Specifically, a key-value cross-attention mechanism is used on the top of GSA sub-module to learn fine-grained cross-modal similarity associations. We obtain query  Q l v and Q l t , key  K l v and K l t , as well as value  V l v and V l t through linear transformation to obtain the visual and textual attention map M l v and M l t :
Q l v = Z v W l , Q v , K l v = Z v W l , K v , V l v = Z v W l , V v , Q l t = Z t W l , Q t , K l t = Z t W l , K t , V l t = Z t W l , V t .
M l v = s o f t m a x Q l v ( K l t ) T d , M l t = s o f t m a x Q l t ( K l v ) T d .
Thus, the attention-weighted representations are X l v = M l v V l v and X l t = M l t V l t . Similar to the above self-attention, the local are calculated as follows:
P v = σ n o r m ( X l v + Z v ) , P t = σ n o r m ( X l t + Z t ) ,
where σ ( · ) is still R e L U ( · ) .

3.5.3. Similarity Consistency Reconstruction

The similarity consistency preserving should not only be reflected in cross-modal features, but more importantly, in hash codes. Deep feature reconstruction, namely reducing the heterogeneity between original features and reconstructed features from continuous hash codes, has been verified to be an effective technique. Current literatures indicate that, two different reconstruction strategies, i.e., intra-modal [28] and inter-modal reconstruction [37], are developed for similarity consistency preserving. In this work, we integrate inter-modal reconstruction into our method to narrow cross-modal gap. Specifically, the continuous hash codes H v and H t are generated by a couple of two-layer fully-connected networks from P v and P t , which then are fed into two re-constructors to output reconstructed features. Accordingly, inspired by [37], a similarity consistency reconstruction loss function L r e c is introduced to preserve the similarity structure information into hash codes:
L r e c = F v ^ F v F 2 + F t ^ F t F 2 ,
where F v ^ and F t ^ denote the reconstructed feature representations.

3.5.4. Hash Code Learning Loss

Furthermore, to preserve similarity consistency of cross-modal hash codes from both intra-modal and inter-modal perspectives, we introduce hash code similarity consistency loss:
L m o d = c o s ( H v , H v ) c o s ( H t , H t ) F 2 + c o s ( H v , H t ) c o s ( H t , H t ) F 2 + c o s ( H v , H t ) c o s ( H v , H v ) F 2 .
Meanwhile, to ensure the accuracy of the similarity matrix, we designed a new loss function L s i m as follows:
L s i m = S c o s ( H v , H v ) F 2 + S c o s ( H t , H t ) F 2 + S c o s ( H v , H t ) F 2 + Γ S S r F 2 ,
where Γ is an indicator function, when S i j r S i j h > η 2 , Γ i j = 1 , otherwise Γ i j = 0 .

3.6. Hashing Function Learning

Using the semantic similarity matrix S and the hash codes H v , H t , we learn two modality-specific hash functions s g n ( f v ( . ; θ v ) ) and s g n ( f t ( . ; θ t ) ) to project deep features F v and F t to Hamming space. Specifically, to preserving the similarity structure between hash codes and cross-modal features, we introduce similarity loss function L f 1 as follows:
L f 1 = S c o s ( f v ( F v ; θ v ) , f v ( F v ; θ v ) ) F 2 + S c o s ( f t ( F t ; θ t ) , f t ( F t ; θ t ) ) F 2 + S c o s ( f v ( F v ; θ v ) , f t ( F t ; θ t ) ) F 2 ,
where f v ( F v ; θ v ) and f t ( F t ; θ t ) denote the relaxed hash codes generated by the hash functions, θ v and θ t denote the parameters. In addition, in order to numerically align the hash codes generated by the hash function with the hash codes H v and H t obtained during the training process, we also introduced loss function L f 2 :
L f 2 = H v f v ( F v ; θ v ) F 2 + H t f t ( F t ; θ t ) F 2 + f v ( F v ; θ v ) f t ( F t ; θ t ) F 2 .
Finally, the quantization loss L f 3 transfers semantic information from relaxed hash codes to binary hash codes:
L f 3 = H v s i g n H v F 2 + H t s i g n H t F 2 .

3.7. Optimization

The learning process of MGSPU can be divided into two stages: (1) hash code learning stage, and (2) hash function learning stage. During the hash code learning stage, the optimization of the objective function L c o d e is performed by minimizing Equations (14)–(16):
m i n θ c L c o d e = λ 1 L r e c + λ 2 L m o d + λ 3 L s i m .
During the hash function learning stage, the optimization of the objective function L h a s h entails minimizing Equations (17)–(19):
m i n θ v , θ t L h a s h = β 1 L f 1 + β 2 L f 2 + β 3 L f 3 .
The entire algorithm process of MGSPU is outlined in Algorithm 1, which is implemented by Adam adaptive algorithm [54].
Algorithm 1 Algorithm of MGSPU.
  • Input: Training set size n; number of epochs e; hash code length C; batch size m; learning rate of the network l r c o d e and l r h a s h ; hyperparameter α 1 , α 2 , η 1 , η 2 , k , λ 1 , λ 2 , λ 3 , β 1 , β 2 , β 3 ;
  • Output: Network parameters for the optimal hash function learning part;
1:
Initialize network parameters θ c , θ v , θ t ;
2:
Construct a unified semantic matrix S r using image features F v and text features F t ;
3:
repeat
4:
   Iterative training e times;
5:
   for  i = 1 , 2 , . . . , n m  do
6:
  Arbitrarily select m training data;
7:
  Generate continuous hash codes H v , H t by forward propagation algorithm;
8:
  Generate the semantic matrix S h through the hash code H v , H t , and update the original semantic matrix S r to generate a unified semantic matrix S ;
9:
  Calculate the loss function Equations (14)–(16) and update the parameters through backpropagation θ c ;
10:
   end for
11:
   for  i = 1 , 2 , . . . , n m  do
12:
  Select m training image-text pairs and hash code H v , H t ;
13:
  Map the original feature F v , F t into a hash code through f v and f t ;
14:
  Calculate the loss Equations (17)–(19), and update the parameter θ v , θ t with backpropagation;
15:
   end for
16:
until Convergence.

4. Experiments

This section presents experiments and analysis to assess the retrieval performance of the proposed method. We begin by introducing the experimental settings, including datasets, evaluation metrics, baselines, and implementation details. Subsequently, we provide a performance comparison between our method and several baselines, along with an ablation analysis to validate the impact of each component.

4.1. Datasets

We conduct thorough experiments on two prominent multimedia benchmark datasets, namely MIRFLICKR-25K [55] and NUS-WIDE [56], widely employed for cross-modal retrieval evaluation. A concise introduction to these datasets is provided below.
MIRFLICKR-25K. The MIRFLICKR-25K dataset includes 25,000 image-text pairs from the popular photo-sharing platform Flickr. Each image is accompanied by multiple text labels. In our experiment, we specifically selected instances with at least 20 text tags. With AlexNet [50], we converted each image into a depth feature of 4096 dimensions, while the text labels were converted into a BoW [51] vector of 1386 dimensions. In addition, each instance is manually annotated with at least one of 24 unique tags. We experimented with 20,015 examples selected from the dataset.
NUS-WIDE. The NUS-WIDE dataset is a substantial real-world web image collection, featuring more than 269,000 images accompanied by over 5000 user-provided tags and 81 concepts across the entire dataset. Using AlexNet [50], each image instance is represented as a 4096-dimensional deep feature, while the textual content is condensed into a 1000-dimensional BoW [51] vector. For our experiment, we excluded instances lacking labels and focused on those associated with the 10 most frequent categories, resulting in a curated set of 186,577 image-text pairs.
Table 5 presents the statistics of the above two datasets, and some samples of these two datasets are shown in Figure 4.

4.2. Evaluation Metrics

In our experiments, we conducted two types of cross-modal retrieval tasks: retrieving texts using image queries (denoted as “I2T”) and retrieving images using text queries (denoted as “T2I”). Next, we utilized two standard hashing performance protocols, Hamming ranking and hash lookup [57], to assess the effectiveness of our method and its competitors. For the Hamming ranking protocol, we utilized mean average precision (mAP) to measure accuracy, while precision-recall curves (P-R curves) were employed for the Hash lookup protocol. For mAP and P-R curves, we considered images and texts to be similar if they shared at least one label; otherwise, they were considered dissimilar. Specifically, given a query q i , the average precision (AP) of the top-N results is defined as:
A P ( q i ) = 1 N r = 1 R p ( r ) d ( r ) ,
where N is the number of relevant instances in the result set, R represents the total amount of data. p ( r ) denotes the precision of the top-r results. If the r-th retrieved result is relevant to the query instances, d ( r ) = 1 ; otherwise, d ( r ) = 0 . The mAP value is defined as the average AP across all queries q i :
m A P = 1 M i = 1 M A P ( q i ) ,
where M represents the number of queries.

4.3. Baselines and Implementation Details

Baselines. we compare the proposed MGSPU method with nine baselines, including CMFH [31], DBRC [58], UDCMH [35], DJSRH [12], JDSH [36], DSAH [37], AGCH [26], DAEH [27], DRNPH [49], which are briefly described as follows:
  • CMFH: This method learns uniform binary feature vectors for different modalities through collective matrix factorization of latent factor models.
  • DBRC: This approach proposes a deep binary reconstruction model to preserve inter-modal correlation.
  • UDCMH: This method utilizes deep learning and matrix factorization with binary latent factor models for multi-modal data search.
  • DJSRH: This approach integrates original neighborhood information from different modalities into a joint-semantics affinity matrix to extract latent intrinsic semantic relations.
  • JDSH: This method introduces a distribution-based similarity decision and weighting scheme for generating a more discriminative hash code.
  • DSAH: This approach explores similarity information across modalities and incorporates a semantic-alignment loss function to align features’ similarities with those between hash codes.
  • AGCH: This method utilizes GCNs to uncover semantic structures, coupled with a fusion module for correlating different modalities.
  • DAEH: This approach attempts to train hash functions with discriminative similarity guidance and an adaptively-enhanced optimization strategy.
  • DRNPH: This method implements unsupervised deep relative neighbor relationship preserving cross-modal hashing for achieving cross-modal retrieval in a common Hamming space.
Except for CMFH, All other approaches use deep features to generate cross-modal hash codes.
Implementation Details. As discussed above, the learning process is divided into two stages. In hash code learning stage, three hyperparameters λ 1 , λ 2 , λ 3 are used to weight L r e c , L m o d , L s i m , respectively. In hash function learning stage, three other hyperparameter β 1 , β 2 , β 3 are used to adjust the ratio between L f 1 , L f 2 , L f 3 . On the MIRFLICKR-25K dataset, we set λ 1 = 0.1 , λ 2 = 1 , λ 3 = 10 , β 1 = 1 , β 2 = 0.01 , β 3 = 1 . On NUS-WIDE dataset, we set λ 1 = 1 , λ 2 = 1 , λ 3 = 10 , β 1 = 10 , β 2 = 0.01 , β 3 = 10 . In the process of semantic similarity matrix construction, we set α 1 = 0.6 , η 1 = 0.8 , α 2 = 0.4 , η 2 = 0.7 on the MIRFLICKR-25K, and set α 1 = 0.4 , η 1 = 0.5 , α 2 = 0.7 , η 2 = 0.6 on the NUS-WIDE. In GSA sub-module, kNN algorithm is used to aggregate nodes in a certain neighborhood for each modality. We set k = 40 and k = 60 for MIRFLICKR-25K and NUS-WIDE, respectively. The optimization algorithm used is the Adam optimization algorithm [54]. For MIRFLICKR-25K, we set the learning rates for hash code learning and hash function learning to 0.001 and 0.0001, respectively, and for NUS-WIDE, both are set to 0.0001. The batch size is consistently set at 512. The number of iterations is defined as 60 for MIRFLICKR-25K and 100 for NUS-WIDE. It’s worth noting that, under the same experimental setup, we directly utilize mAP@50 results provided in the original papers of the baseline methods.
Experimental Environment. All the experiments are performed on a workstation with Intel(R) Core i9-12900K 3.9 GHz CPU, 128 GB RAM, 1 TB SSD storage, 2TB HDD storage, and 1 NVIDIA GeForce RTX 3090Ti GPU with ubuntu-22.04.1 operating system. All the techniques are implemented by Python 3.9 on PyTorch 2.0.1.

4.4. Performance Evaluation

We compare the proposed method with nine baselines on MIRFLICKR-25K and NUS-WIDE datasets. The performance of all these methods are evaluated by Hamming Ranking protocol and hash lookup protocol. Table 6 and Table 7 illustrates the mAP@50 results of our method and the competitors varying hash code lengths (16, 32, 64, 128 bits) on MIRFLICKR-25K and NUS-WIDE. Figure 5 and Figure 6 show the P-R curves on these two datasets in various code length. The detailed analysis and observation are presented as follows.
Hamming Ranking. It is clearly from the Table 6 and Table 7 that the proposed method performs better than the baselines. Specifically, on MIRFLICKR-25K, we found that our method achieved the highest mAP@50 score for both retrieval tasks (I2T: mAP@50 = 0.898 (16 bits), mAP@50 = 0.915 (32 bits), mAP@50 = 0.927 (64 bits), mAP@50 = 0.936 (128 bits); T2I: mAP@50 = 0.876 (16 bits), mAP@50 = 0.883 (32 bits), mAP@50 = 0.889 (64 bits), mAP@50 = 0.900 (128 bits)). For example, our method beats out the strongest competitor, DRNPH, by a significant margin on both two tasks, especially in shorter hash code length: 0.022 (16 bits), 0.013 (32 bits), 0.013 (64 bits) on I2T task, and 0.016 (16 bits), 0.011 (32 bits) on T2I task. The reason behind these results is obvious: with the engagement of the proposed similarity matrix updating strategy, our method can gradually eliminate the noise of the original features used in the similarity relationship construction so as to improve similarity consistency preserving, which is unfortunately ignored by DRNPH. In all but a few case (32 bits code length on I2T and I2T task, MGSPU was defeated by AGCH and DSAH marginally), our method won the competition again on NUS-WIDE by stand-out performance: mAP@50 = 0.811 (16 bits), 0.826 (32 bits), 0.844 (64 bits), 0.858 (128 bits) on I2T task, and mAP@50 = 0.780 (16 bits), 0.786 (32 bits), 0.806 (64 bits), 0.813 (128 bits) on T2I task. Compared with mainstream solutions, complex similarity correlations can be greatly mined through the semantic similarity matrix update strategy, and the MGSP module further retains the potential similarity structure between data.
Hash Lookup. To comprehensively showcase the comprehensive performance comparison of MGSPU with baselines, we draw P-R curves in Figure 5 and Figure 6 with different code lengths on both two datasets. As expected, in addition to dramatically defeating hand-crafted feature based method CMFH, MGSPU outperforms state-of-the-art competitors DRNPH, DSAH and AGCH on various hash code length. This observation is mainly due to the search performance boosting from the interplay of similarity updating and multi-grained similarity preserving of hash codes.
Discussion. It is no secret that, the main reason of poor performance of early models (such as CMFH and DBRC) is their shallow feature extraction techniques that cannot obtain feature representations with rich semantic information. With the help of powerful deep learning techniques, deep neural networks based methods such as DJSRH, JDSH, AGCH and DAEH achieved good results. Among them, DJSRH is equiped with a reconstruction framework for training, which is more competitive than batch training. AGCH uses GCN to aggregate neighborhood information and enhance feature expression. DAEH leverages teacher networks to enhance weaker hashing networks. However, all of them build similarity matrices based on original features, which inevitably bring noises into semantic relationships so as to introduce biases. Furthermore, these methods either maintain local or global similarities to preserve the semantic relationships. For example, the similarity matrix in DJSRH contains redundant information from intra-modal fusion items, while DAEH ignores the semantic relationships of intra-modal details. Comparing with these solutions, therefore, we argue that stepwise denoising through a similarity matrix update strategy can greatly mine complex similarity correlations, thereby generating high-confidence supervision signals. In addition, the MGSP method can effectively improve the hash code quality due to further preserving the potential similar structures within and between modalities. Both the mAP@50 score on the hash ranking protocol and the area under the P-R curves on the hash lookup procotol strongly support our view.

4.5. Ablation Study

To verify the validity of each design in MGSPU, we conducted ablation experiments on the MIRFLICKR-25K and NUS-WIDE datasets, serveral variations were considered for this purpose:
  • MGSPU-1: it removes semantic similarity matrix updating from MGSPU.
  • MGSPU-2: it removes the similarity consistency reconstruction from MGSPU.
  • MGSPU-3: it modifies similarity consistency reconstruction by replacing inter-modal reconstruction with intra-modal reconstruction.
  • MGSPU-4: it removes the GCN module from MGSPU.
From Table 8 and Table 9, the following observations can be obtained: firstly, the comparison of MGSPU-1 with our the full MGSPU method verifies that the proposed dual instruction fusion updating strategy can improve the quality of instance similarity matrix to enhance the retrieval performance. Specifically, the retrieval accuracy of MGSPU-1 for both I2T and T2I task show decrease in some extent: on MIRFLICKR-25K, mAP@50 results of I2T task drop from 0.898 (16 bits), 0.915 (32 bits), 0.927 (64 bits), 0.936 (128 bits) to 0.894 (16 bits), 0.912 (32 bits), 0.924 (64 bits), 0.933 (128 bits), respectively; mAP@50 results of T2I task drop from 0.876 (16 bits), 0.883 (32 bits), 0.889 (64 bits), 0.900 (128 bits) to 0.872 (16 bits), 0.876 (32 bits), 0.883 (64 bits), 0.892 (128 bits), respectively. It indicates that without the semantic similarity matrix updating, complex similarity relationship learning suffers from disturbance by noise. Secondly, we can clearly observe that MGSPU-2 has a remarkably performance degradation compared with the full version of MGSPU. This phenomenon confirms that similarity consistency reconstruction is beneficial to preserve semantic information into hash code. Thirdly, with intra-modal reconstruction, MGSPU-3 performs better than MGSPU-2 especially on long hash codes (e.g., 64 or 128 bits). However, compared with the inter-modal reconstruction used in our method, the performance of MGSPU-3 is slightly weaker, which indicates that the inter-modal reconstruction is more helpful to reduce the heterogeneity between the original feature and the hash code. Lastly, after removing GCN module, MGSPU-4 achieves lower retrieval accuracy than ours. these results show that the structural similarity aggregated from neighborhoods by the GCN module is essential to enrich the similarity relationship information of each instance.

4.6. Sensitivity to Hyperparameters

This section, we conduct an analysis of the sensitivity of all hyperparameters used in the model: α 1 , η 1 ; α 2 , η 2 ; k ; λ 1 , λ 2 , λ 3 ; β 1 , β 2 , β 3 . To explore the comprehensive impact of them, the accuracy of I2T and T2I task is used to visualize the trend of cross-modal hash performance. All these analysis are carried out in 16 bit hash code length on NUS-WIDE dataset.
Hyperparameter  α 1 , η 1 , α 2 , η 2 . In semantic similarity matrix updating module, α 1 , η 1 are used to construct the refined similarity matrix S r , α 2 , η 2 are used to execute the updating strategy to generate semantic similarity matrix S . We observe the performance change of MGSPU by varying α 1 , η 1 and α 2 , η 2 . According to the experimental results in Figure 7, it is clearly that the retrieval accuracy is more susceptible to changes of α 1 when the value of η 1 is small. We speculate that this phenomenon is caused by the noises that are injected in refined similarity matrix when the value of η 1 is small. On the other hand, when we change the α 2 and η 2 , the fluctuations in model performance are relatively less severe, which still shows that our model performs a bit better if η 2 is set to a large value (r.g., η 2 = 0.70 ). The reason behind this results is understandable: if a larger threshold η 2 is taken, the discrimination of whether S i j r and S i j h are dissimilar will be more rigorous. Under this circumstance, the semantic similarity matrix updating will be executed more cautiously to preserve robust.
Hyperparameter k. We recorded the performance change by varying the value of k on NUS-WIDE dataset to evaluate the effect by the the number of neighbors in kNN algorithm. As demonstrated in Figure 8, when k [ 40 , 80 ] , the curve changes sharply, while the curve changes more modestly in the rest of the interval. We conjecture that when too many or too few neighbors we selected, noise will be introduced into the intra-modal similarity relationship representation, thereby affecting the latent similarity relationship learning within modality. Particularly, if k is set to 60, MGSPU achieves the highest mAP@50 score for both I2T and T2I task. It indicates that by selecting an appropriate number of neighbors, high-quality intra-modal similarity structure information can be aggregated by GCN to improve intra-modal similarity consistency preserving.
Hyperparameter λ 1 , λ 2 , λ 3 . As presented in Equation (20), λ 1 , λ 2 , λ 3 are used to balance three components, i.e., L r e , L m o d and L s i m of hash code learning loss function. To analyze the effect by these three losses, we recorded the performance change of our method in Figure 8 by varying λ 1 , λ 2 , λ 3 from 0.001 to 100 with a 10-fold increase. It is noteworthy that our method obtains the best performance if we set λ 1 = 1 , λ 2 = 1 , λ 3 = 10 . Among these three losses, we found that the new designed loss L s i m has a relatively greater effect on the learning of hash code. We argue that this is mainly due to the indispensability of a reliable similarity matrix for unsupervised hash learning.
Hyperparameter  β 1 , β 2 , β 3 . As depicted in Equation (21), β 1 , β 2 , β 3 serves as weight factor to balance L f 1 , L f 2 , and L f 3 . The observations in Figure 8 indicates that L f 1 and L f 3 contributes more than L f 2 . We infer that although the duty of L f 2 is to ensure numerical consistency between the generated hash codes H * and trained hash codes and s g n ( f * ( F * ; θ * ) ) , the main goal of cross-modal hash learning is still to eliminate cross-modal heterogeneity, which is only achieved by L f 1 . Besides, the quantization error is cannot reduced by other losses but L f 3 . In addition, by setting β 1 = 10 , β 2 = 0.01 , β 3 = 10 , our method achieves the best performance.

5. Conclusions

In this paper, we propose a novel unsupervised hashing learning framework, called multi-grained similarity preserving and updating to improve cross-modal hashing performance. To obtain a high-confidence similarity matrix, we develop an update strategy that corrects the similarity values after the original feature fusion using a matrix constructed in Hamming space, and a loss function is designed to guide the similarity update. Also to learn high-quality hash codes, we co-model from multiple granularity to preserve semantic correlations within and between modalities. Specifically, GCNs is used to capture the global similarity relationship within each modality, and a cross-attention mechanism is used to perform interactions between modalities to bridge the heterogeneous gap. In addition, deep feature reconstruction further enhances inter-modal correlations and reduces modal gaps.
In our experiments, we use the hamming ranking protocol and the hash lookup protocol to compare with other benchmark models, and the experimental results on both datasets show that our approach achieves impressive performance. In addition, we set up four different ablation experiments to verify the performance of the designed module, and the results also validate the effectiveness of the designed module.
Although our experimental results achieved excellent results, for text data, we only used a simple bag-of-words model, and the gap between the results of image-retrieval text and text-retrieval image is relatively large. As an exploratory work, we plan to further improve the feature extraction stage in the future to achieve a more accurate alignment of the semantic information of images and text, which will lead to a more balanced cross-modal retrieval.

Author Contributions

R.W. and L.Z. designed the methodology, wrote the original draft, designed and prepared all figures. X.Z. contributed to conceptualization, project administration. X.Z. and L.Z. acquired funding, reviewed and edited the manuscript. R.W. and L.Z. conceived the experiments. R.W., Z.Y. and Y.L. conducted the experiments and acquired experimental results. R.W., L.Z. and Z.Z. analyzed the experimental results. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (62202163, 62072166), the Natural Science Foundation of Hunan Province (2022JJ40190), the Scientific Research Project of Hunan Provincial Department of Education (22A0145), the Key Research and Development Program of Hunan Province (2020NK2033), the Hunan Provincial Department of Education Scientific Research Outstanding Youth Project (21B0200), and the Hunan Provincial Natural Science Foundation Youth Fund Project (2023JJ40333).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

This paper uses the publicly available dataset MIRFLICKR-25K, NUS-WIDE.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, Y. Survey on deep multi-modal data analytics: Collaboration, rivalry, and fusion. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM) 2021, 17, 1–25. [Google Scholar]
  2. Zhu, L.; Zhang, C.; Song, J.; Liu, L.; Zhang, S.; Li, Y. Multi-graph based hierarchical semantic fusion for cross-modal representation. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021; pp. 1–6. [Google Scholar]
  3. Zhang, B.; Hu, H.; Sha, F. Cross-modal and hierarchical modeling of video and text. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 374–390. [Google Scholar]
  4. Xie, L.; Shen, J.; Zhu, L. Online cross-modal hashing for web image retrieval. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; Volume 30. Number 1. [Google Scholar]
  5. Tian, Y.; Zhou, L.; Zhang, Y.; Zhang, T.; Fan, W. Deep cross-modal face naming for people news retrieval. IEEE Trans. Knowl. Data Eng. 2019, 33, 1891–1905. [Google Scholar] [CrossRef]
  6. Zhen, L.; Hu, P.; Wang, X.; Peng, D. Deep supervised cross-modal retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 10394–10403. [Google Scholar]
  7. Wang, K.; Yin, Q.; Wang, W.; Wu, S.; Wang, L. A comprehensive survey on cross-modal retrieval. arXiv 2016, arXiv:1607.06215. [Google Scholar]
  8. Huang, X.; Peng, Y.; Yuan, M. MHTN: Modal-adversarial hybrid transfer network for cross-modal retrieval. IEEE Trans. Cybern. 2018, 50, 1047–1059. [Google Scholar] [CrossRef] [PubMed]
  9. Yu, T.; Yang, Y.; Li, Y.; Liu, L.; Fei, H.; Li, P. Heterogeneous attention network for effective and efficient cross-modal retrieval. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Montreal, QC, Canada, 11–15 July 2021; pp. 1146–1156. [Google Scholar]
  10. Chun, S.; Oh, S.J.; De Rezende, R.S.; Kalantidis, Y.; Larlus, D. Probabilistic embeddings for cross-modal retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 8415–8424. [Google Scholar]
  11. Gao, D.; Jin, L.; Chen, B.; Qiu, M.; Li, P.; Wei, Y.; Hu, Y.; Wang, H. Fashionbert: Text and image matching with adaptive loss for cross-modal retrieval. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Xi’an, China, 25–30 July 2020; pp. 2251–2260. [Google Scholar]
  12. Su, S.; Zhong, Z.; Zhang, C. Deep joint-semantics reconstructing hashing for large-scale unsupervised cross-modal retrieval. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Long Beach, CA, USA, 16–20 June 2019; pp. 3027–3035. [Google Scholar]
  13. Gu, J.; Cai, J.; Joty, S.R.; Niu, L.; Wang, G. Look, imagine and match: Improving textual-visual cross-modal retrieval with generative models. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7181–7189. [Google Scholar]
  14. Cheng, M.; Jing, L.; Ng, M.K. Robust unsupervised cross-modal hashing for multimedia retrieval. ACM Trans. Inf. Syst. (TOIS) 2020, 38, 1–25. [Google Scholar]
  15. Yao, H.L.; Zhan, Y.W.; Chen, Z.D.; Luo, X.; Xu, X.S. Teach: Attention-aware deep cross-modal hashing. In Proceedings of the z, Taipei, Taiwan, 21–24 August 2021; pp. 376–384. [Google Scholar]
  16. Zhang, C.; Song, J.; Zhu, X.; Zhu, L.; Zhang, S. Hcmsl: Hybrid cross-modal similarity learning for cross-modal retrieval. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM) 2021, 17, 1–22. [Google Scholar] [CrossRef]
  17. Jiang, Q.Y.; Li, W.J. Deep cross-modal hashing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3232–3240. [Google Scholar]
  18. Yang, E.; Deng, C.; Liu, W.; Liu, X.; Tao, D.; Gao, X. Pairwise relationship guided deep hashing for cross-modal retrieval. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. Number 1. [Google Scholar]
  19. Ma, L.; Li, H.; Meng, F.; Wu, Q.; Ngan, K.N. Global and local semantics-preserving based deep hashing for cross-modal retrieval. Neurocomputing 2018, 312, 49–62. [Google Scholar] [CrossRef]
  20. Shen, X.; Zhang, H.; Li, L.; Yang, W.; Liu, L. Semi-supervised cross-modal hashing with multi-view graph representation. Inf. Sci. 2022, 604, 45–60. [Google Scholar]
  21. Li, C.; Deng, C.; Li, N.; Liu, W.; Gao, X.; Tao, D. Self-supervised adversarial hashing networks for cross-modal retrieval. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–20 July 2018; pp. 4242–4251. [Google Scholar]
  22. Zhang, D.; Wu, X.J.; Yu, J. Label consistent flexible matrix factorization hashing for efficient cross-modal retrieval. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM) 2021, 17, 1–18. [Google Scholar] [CrossRef]
  23. Chen, Z.D.; Li, C.X.; Luo, X.; Nie, L.; Zhang, W.; Xu, X.S. SCRATCH: A scalable discrete matrix factorization hashing framework for cross-modal retrieval. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 2262–2275. [Google Scholar] [CrossRef]
  24. Hu, P.; Zhen, L.; Peng, D.; Liu, P. Scalable deep multimodal learning for cross-modal retrieval. In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, Paris, France, 21–25 July 2019; pp. 635–644. [Google Scholar]
  25. Dong, X.; Liu, L.; Zhu, L.; Nie, L.; Zhang, H. Adversarial graph convolutional network for cross-modal retrieval. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 1634–1645. [Google Scholar] [CrossRef]
  26. Zhang, P.F.; Li, Y.; Huang, Z.; Xu, X.S. Aggregation-based graph convolutional hashing for unsupervised cross-modal retrieval. IEEE Trans. Multimed. 2021, 24, 466–479. [Google Scholar]
  27. Shi, Y.; Zhao, Y.; Liu, X.; Zheng, F.; Ou, W.; You, X.; Peng, Q. Deep adaptively-enhanced hashing with discriminative similarity guidance for unsupervised cross-modal retrieval. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 7255–7268. [Google Scholar]
  28. Zhu, L.; Wu, X.; Li, J.; Zhang, Z.; Guan, W.; Shen, H.T. Work together: Correlation-identity reconstruction hashing for unsupervised cross-modal retrieval. IEEE Trans. Knowl. Data Eng. 2022, 35, 8838–8851. [Google Scholar] [CrossRef]
  29. Li, C.; Deng, C.; Wang, L.; Xie, D.; Liu, X. Coupled cyclegan: Unsupervised hashing network for cross-modal retrieval. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, January 27–1 February 2019; Volume 33, Number 01. pp. 176–183. [Google Scholar]
  30. Wang, W.; Shen, Y.; Zhang, H.; Yao, Y.; Liu, L. Set and rebase: Determining the semantic graph connectivity for unsupervised cross-modal hashing. In Proceedings of the Twenty-Ninth International Conference on International Joint Conferences on Artificial Intelligence, YoKohama, Japan, 7–15 January 2021; pp. 853–859. [Google Scholar]
  31. Ding, G.; Guo, Y.; Zhou, J. Collective matrix factorization hashing for multimodal data. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2075–2082. [Google Scholar]
  32. Song, J.; Yang, Y.; Yang, Y.; Huang, Z.; Shen, H.T. Inter-media hashing for large-scale retrieval from heterogeneous data sources. In Proceedings of the 2013 ACM SIGMOD International Conference on Management of Data, New York, NY, USA, 22–27 June 2013; pp. 785–796. [Google Scholar]
  33. Tu, R.C.; Mao, X.L.; Lin, Q.; Ji, W.; Qin, W.; Wei, W.; Huang, H. Unsupervised Cross-modal Hashing via Semantic Text Mining. IEEE Trans. Multimed. 2023, 25, 8946–8957. [Google Scholar] [CrossRef]
  34. Zhao, Y.; Zhu, Y.; Liao, S.; Ye, Q.; Zhang, H. Class concentration with twin variational autoencoders for unsupervised cross-modal hashing. In Proceedings of the Asian Conference on Computer Vision, Macau, China, 4–8 December 2022; pp. 349–365. [Google Scholar]
  35. Wu, G.; Lin, Z.; Han, J.; Liu, L.; Ding, G.; Zhang, B.; Shen, J. Unsupervised Deep Hashing via Binary Latent Factor Models for Large-scale Cross-modal Retrieval. In Proceedings of the IJCAI, Stockholm, Sweden, 13–19 July 2018; Volume 1, Number 3. p. 5. [Google Scholar]
  36. Liu, S.; Qian, S.; Guan, Y.; Zhan, J.; Ying, L. Joint-modal distribution-based similarity hashing for large-scale unsupervised deep cross-modal retrieval. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Xi’an, China, 25–30 July 2020; pp. 1379–1388. [Google Scholar]
  37. Yang, D.; Wu, D.; Zhang, W.; Zhang, H.; Li, B.; Wang, W. Deep semantic-alignment hashing for unsupervised cross-modal retrieval. In Proceedings of the 2020 International Conference on Multimedia Retrieval, Dublin, Ireland, 26–29 October 2020; pp. 44–52. [Google Scholar]
  38. Zhang, Z.; Lin, Z.; Zhao, Z.; Xiao, Z. Cross-modal interaction networks for query-based moment retrieval in videos. In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, Paris, France, 21–25 July 2019; pp. 655–664. [Google Scholar]
  39. Wang, Z.; Liu, X.; Li, H.; Sheng, L.; Yan, J.; Wang, X.; Shao, J. Camp: Cross-modal adaptive message passing for text-image retrieval. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Long Beach, CA, USA, 16–20 June 2019; pp. 5764–5773. [Google Scholar]
  40. Yu, Y.; Xiong, Y.; Huang, W.; Scott, M.R. Deformable siamese attention networks for visual object tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, DC, USA, 14–19 June 2020; pp. 6728–6737. [Google Scholar]
  41. Gu, W.; Gu, X.; Gu, J.; Li, B.; Xiong, Z.; Wang, W. Adversary guided asymmetric hashing for cross-modal retrieval. In Proceedings of the 2019 on International Conference on Multimedia Retrieval, Ottawa, ON, Canada, 10–13 June 2019; pp. 159–167. [Google Scholar]
  42. Zhang, D.; Li, W.J. Large-scale supervised multimodal hashing with semantic correlation maximization. In Proceedings of the AAAI Conference on Artificial Intelligence, Québec, QC, Canada, 27–31 July 2014; Volume 28. Number 1. [Google Scholar]
  43. Lin, Z.; Ding, G.; Hu, M.; Wang, J. Semantics-preserving hashing for cross-view retrieval. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3864–3872. [Google Scholar]
  44. Cao, Y.; Liu, B.; Long, M.; Wang, J. Cross-modal hamming hashing. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 202–218. [Google Scholar]
  45. Jin, S.; Zhou, S.; Liu, Y.; Chen, C.; Sun, X.; Yao, H.; Hua, X.S. SSAH: Semi-supervised adversarial deep hashing with self-paced hard sample generation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, Number 07. pp. 11157–11164. [Google Scholar]
  46. Liu, X.; Yu, G.; Domeniconi, C.; Wang, J.; Ren, Y.; Guo, M. Ranking-based deep cross-modal hashing. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, Number 01. pp. 4400–4407. [Google Scholar]
  47. Sun, C.; Latapie, H.; Liu, G.; Yan, Y. Deep normalized cross-modal hashing with bi-direction relation reasoning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 4941–4949. [Google Scholar]
  48. Yu, J.; Zhou, H.; Zhan, Y.; Tao, D. Deep graph-neighbor coherence preserving network for unsupervised cross-modal hashing. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, Number 5. pp. 4626–4634. [Google Scholar]
  49. Yang, X.; Wang, Z.; Wu, N.; Li, G.; Feng, C.; Liu, P. Unsupervised Deep Relative Neighbor Relationship Preserving Cross-Modal Hashing. Mathematics 2022, 10, 2644. [Google Scholar] [CrossRef]
  50. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  51. Ko, Y. A study of term weighting schemes using class information for text classification. In Proceedings of the 35th International ACM SIGIR Conference on Research and Development in Information Retrieval, Portland, OR, USA, 12–16 August 2012; pp. 1029–1030. [Google Scholar]
  52. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, June 26–1 July 2016; pp. 770–778. [Google Scholar]
  53. Memory, L.S.T. Long short-term memory. Neural Comput. 2010, 9, 1735–1780. [Google Scholar]
  54. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  55. Huiskes, M.J.; Lew, M.S. The mir flickr retrieval evaluation. In Proceedings of the 1st ACM International Conference on Multimedia Information Retrieval, Vancouver, BC, Canada, 30–31 October 2008; pp. 39–43. [Google Scholar]
  56. Chua, T.S.; Tang, J.; Hong, R.; Li, H.; Luo, Z.; Zheng, Y. Nus-wide: A real-world web image database from national university of singapore. In Proceedings of the ACM International Conference on Image and Video Retrieval, Santorini Island, Greece, 8–10 July 2009; pp. 1–9. [Google Scholar]
  57. Liu, W.; Mu, C.; Kumar, S.; Chang, S.F. Discrete graph hashing. Adv. Neural Inf. Process. Syst. 2014, 27, 3419–3427. [Google Scholar]
  58. Li, X.; Hu, D.; Nie, F. Deep binary reconstruction for cross-modal hashing. In Proceedings of the 25th ACM International Conference on Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 1398–1406. [Google Scholar]
Figure 1. The framework of MGSPU. It comprises four main modules: feature extraction, semantic similarity matrix updating, hash code learning and hash function learning module.
Figure 1. The framework of MGSPU. It comprises four main modules: feature extraction, semantic similarity matrix updating, hash code learning and hash function learning module.
Applsci 14 00870 g001
Figure 2. The process of dual instruction fusion updating. Γ is an indicator matrix that tells us where values need to be updated, when | S r S h | > η 2 , Γ i j = 1 ; otherwise Γ i j = 0 . Λ i j = S i j r S i j h , ⊕ is XOR operation. If Γ i j = 1 and Λ i j = 0 , S i j is weighted by S i j r and S i j h ; if Γ i j = 0 and Λ i j = 0 , then S i j = S i j r ; the rest of S is consistent with 0.
Figure 2. The process of dual instruction fusion updating. Γ is an indicator matrix that tells us where values need to be updated, when | S r S h | > η 2 , Γ i j = 1 ; otherwise Γ i j = 0 . Λ i j = S i j r S i j h , ⊕ is XOR operation. If Γ i j = 1 and Λ i j = 0 , S i j is weighted by S i j r and S i j h ; if Γ i j = 0 and Λ i j = 0 , then S i j = S i j r ; the rest of S is consistent with 0.
Applsci 14 00870 g002
Figure 3. The details of MGSP. (ac) illustrate the pipeline of global similarity aggregation, local similarity interaction, and similarity consistency reconstruction, respectively. Best view in color.
Figure 3. The details of MGSP. (ac) illustrate the pipeline of global similarity aggregation, local similarity interaction, and similarity consistency reconstruction, respectively. Best view in color.
Applsci 14 00870 g003
Figure 4. Some examples from the MIRFLICKR-25K and NUS-WIDE datasets. (a) MIRFLICKR-25K. (b) NUS-WIDE.
Figure 4. Some examples from the MIRFLICKR-25K and NUS-WIDE datasets. (a) MIRFLICKR-25K. (b) NUS-WIDE.
Applsci 14 00870 g004
Figure 5. P-R curves of different models at different lengths on MIRFLICKR-25K dataset. (a) I2T:16bit. (b) I2T:32bit. (c) I2T:64bit. (d) I2T:128bit. (e) T2I:16bit. (f) T2I:32bit. (g) T2I:64bit. (h) T2I:64bit.
Figure 5. P-R curves of different models at different lengths on MIRFLICKR-25K dataset. (a) I2T:16bit. (b) I2T:32bit. (c) I2T:64bit. (d) I2T:128bit. (e) T2I:16bit. (f) T2I:32bit. (g) T2I:64bit. (h) T2I:64bit.
Applsci 14 00870 g005
Figure 6. P-R curves of different models at different lengths on MIRFLICKR-25K dataset. (a) I2T:16bit. (b) I2T:32bit. (c) I2T:64bit. (d) I2T:128bit. (e) T2I:16bit. (f) T2I:32bit. (g) T2I:64bit. (h) T2I:128bit.
Figure 6. P-R curves of different models at different lengths on MIRFLICKR-25K dataset. (a) I2T:16bit. (b) I2T:32bit. (c) I2T:64bit. (d) I2T:128bit. (e) T2I:16bit. (f) T2I:32bit. (g) T2I:64bit. (h) T2I:128bit.
Applsci 14 00870 g006
Figure 7. Sensitivity analysis of α 1 , η 1 and α 2 , η 2 on NUS-WIDE dataset. (a) I2T: α 1 and η 1 . (b) T2I: α 1 and η 1 . (c) I2T: α 2 and η 2 . (d) T2I: α 2 and η 2 .
Figure 7. Sensitivity analysis of α 1 , η 1 and α 2 , η 2 on NUS-WIDE dataset. (a) I2T: α 1 and η 1 . (b) T2I: α 1 and η 1 . (c) I2T: α 2 and η 2 . (d) T2I: α 2 and η 2 .
Applsci 14 00870 g007
Figure 8. Sensitivity analysis of k; λ 1 , λ 2 , λ 3 ; β 1 , β 2 , β 3 on NUS-WIDE dataset. (a) I2T: k. (b) I2T: λ . (c) I2T: β . (d) T2I: k. (e) T2I: λ . (f) T2I: β .
Figure 8. Sensitivity analysis of k; λ 1 , λ 2 , λ 3 ; β 1 , β 2 , β 3 on NUS-WIDE dataset. (a) I2T: k. (b) I2T: λ . (c) I2T: β . (d) T2I: k. (e) T2I: λ . (f) T2I: β .
Applsci 14 00870 g008
Table 1. Related works summary.
Table 1. Related works summary.
TypeReferenceSourceAuthorsDescription
SupervisedSCM [42]AAAI (2014)Zhang, D., et al.Integrate semantic tags into the data modeling process.
SePH [43]CVPR (2015)Lin, Z., et al.The semantic similarity given by the training data is converted into a probability distribution, and then the hash code is learned by minimizing the KL divergence.
DCMH [17]CVPR (2017)Jiang, Q.Y., et al.Feature learning and hash learning are integrated for the first time.
CMHH [44]ECCV (2018)Cao, Y., et al.Hamming distance replaces the inner product and uses pairwise loss based on exponential distribution to solve the problem of uneven positive and negative samples.
SSAH [45]AAAI (2020)Jin, S., et al.Derive semantic ranking information using data features and tags and integrate it into cross-modal hashes.
RDCMH [46]AAAI (2019)Liu, X., et al.A multi-level semantic similarity matrix is constructed by considering the bidirectional relation, the relation between the whole instances.
Bi_NCMH [47]CVPR (2022)Sun, C., et al.The potential semantic relationships of data are captured by constructing a joint semantic matrix, but redundant information is introduced.
UnsupervisedDJSRH [12]ICCV (2019)Su, S., et al.The potential semantic relationships of data are captured by constructing a joint semantic matrix, but redundant information is introduced.
JDSH [36]ACM (2020)Liu, S., et al.The proposed sampling weighting scheme can generate more efficient hash codes.
DAEH [27]TCSVT (2022)Shi, Y., et al.Design information mixed similarity estimation integrates distance distribution and similarity rate information.
DGCPN [48]AAAI (2021)Yu, J., et al.Using a graph model to integrate neighborhood information, the overall loss is maintained by designing three types of data similarity.
AGCH [26]TMM (2021)Zhang, P., et al.The semantic structure of data is mined by GCN, and the structural information is explored from multiple perspectives by using a variety of similarity measures.
DRNPH [49]Mathematics (2022)Yang, X., et al.Cross-modal triples are used to maintain data similarity, and the distance between similar samples is required to be smaller than that between dissimilar pairs.
Table 2. Abbreviation summary.
Table 2. Abbreviation summary.
AbbreviationMeaning
MGSPUMulti-Grained Similarity Preserving and Updating
CNNConvolutional Neural Network
BoWBag of words
GCNsGraph Convolutional Networks
ResNetResidual Networks
LSTMLong Short-Term Memory Network
MGSPMulti-Grained Similarity Preserving
GSAGlobal Similarity Aggregation
LSILocal Similarity Interaction
SCRSimilarity Consistency Reconstruction
Table 3. Notation summary.
Table 3. Notation summary.
NotationDefinition
v i The i-th image instance
t i The i-th text instance
F v The original image features after feature extraction
F t The original text features after feature extraction
f v Image hash function
f t Text hash function
S v Image similarity matrix
S t Text similarity matrix
S f Fused similarity matrix
S r Refined similarity matrix
S h Hashing similarity matrix
Γ The indicator matrix
Λ The indicator matrix
S Unified similarity matrix
L * The corresponding latent representations generated by the image or text encoder
X g * Image or text global similarity aggregation representations by self-attention mechanism
G * Final outputs of image or text GCN
Z * Image or text share the outputs of a fully connected network through a two-layer of weights
X l v Image attention-weighted representations by cross-attention mechanism
X l t Text attention-weighted representations by cross-attention mechanism
P v Image features output by the local similarity interaction sub-module
P t Text features output by the local similarity interaction sub-module
F v ^ Reconstructed image feature representations
F t ^ Reconstructed text feature representations
H v Real-valued hash codes for image instances
H t Real-valued hash codes for text instances
B v Binary code for image instances
B t Binary code for text instances
Table 4. Indicator matrices truth table for dual instruction fusion updating strategy.
Table 4. Indicator matrices truth table for dual instruction fusion updating strategy.
Λ i j Γ i j UpdateRules
00 S i j r
01 α 2 S i j r + ( 1 α 2 ) S i j h
100
110
Table 5. Details about the datasets used in our experiments.
Table 5. Details about the datasets used in our experiments.
DatasetTrainQueryRetrievalText Feature
MIRFLICKR-25K5000200018,0151386 d
NUS-WIDE50002000184,5771000 d
Table 6. mAP@50 score of our method and the baselines at various code lengths (bits) on MIRFLICKR-25K.
Table 6. mAP@50 score of our method and the baselines at various code lengths (bits) on MIRFLICKR-25K.
MethodsMIRFLICKR-25K
I2TT2I
16 bits 32 bits 64 bits 128 bits 16 bits 32 bits 64 bits 128 bits
CMFH0.6210.6240.6250.6270.4620.6620.6760.685
DBRC0.6170.6190.6200.6210.6180.6220.6260.628
UDCMH0.6890.6980.7140.7170.6920.7040.7180.733
DJSRH0.8100.8430.8620.8760.7860.8220.8350.847
JDSH0.8320.8530.8820.8920.8250.8640.8780.880
DSAH0.8630.8770.8950.9030.8460.8600.8810.882
AGCH0.8650.8870.8920.9120.8290.8450.8520.880
DAEH0.8120.8350.8470.8450.7780.8190.8250.831
DRNPH0.8760.9020.9140.9330.8600.8720.8850.897
Ours0.8980.9150.9270.9360.8760.8830.8890.900
Table 7. mAP@50 score of our method and the baselines at various code lengths (bits) on NUS-WIDE.
Table 7. mAP@50 score of our method and the baselines at various code lengths (bits) on NUS-WIDE.
MethodsNUS-WIDE
I2TT2I
16 bits 32 bits 64 bits 128 bits 16 bits 32 bits 64 bits 128 bits
CMFH0.4550.4590.4650.4670.5290.5770.6140.652
DBRC0.4240.4590.4470.4470.4550.4590.4680.473
UDCMH0.5110.5190.5240.5580.6370.6530.6950.716
DJSRH0.7240.7730.7980.8170.7120.7440.7710.789
JDSH0.7360.7930.8320.8350.7210.7850.7940.804
DSAH0.7750.8050.8180.8270.7700.7900.8040.815
AGCH0.8090.8300.8310.8520.7690.7800.7980.802
DAEH0.7660.7890.8090.8220.7180.7510.7660.767
DRNPH0.7900.8110.8260.8370.7800.7950.8040.811
Ours0.8110.8260.8440.8580.7800.7860.8060.813
Table 8. mAP@50 score for ablation study on MIRFLICKR-25K.
Table 8. mAP@50 score for ablation study on MIRFLICKR-25K.
MethodsMIRFLICKR-25K
I2TT2I
16 bits 32 bits 64 bits 128 bits 16 bits 32 bits 64 bits 128 bits
MGSPU-10.8940.9120.9240.9330.8720.8760.8830.892
MGSPU-20.8860.9020.9090.9130.8490.8570.8660.876
MGSPU-30.8970.9130.9260.9340.8720.8850.8890.894
MGSPU-40.8950.9040.9240.9260.8680.8730.8860.887
Ours0.8980.9150.9270.9360.8760.8830.8890.900
Table 9. mAP@50 score for ablation study on NUS-WIDE.
Table 9. mAP@50 score for ablation study on NUS-WIDE.
MethodsNUS-WIDE
I2TT2I
16 bits 32 bits 64 bits 128 bits 16 bits 32 bits 64 bits 128 bits
MGSPU-10.8030.8210.8370.8490.7630.7830.8040.805
MGSPU-20.7960.8090.8190.8210.7460.7750.7770.776
MGSPU-30.8010.8210.8270.8510.7680.7910.7990.812
MGSPU-40.7940.8260.8430.8470.7730.7860.7980.797
Ours0.8110.8260.8440.8580.7800.7860.8060.813
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, R.; Zhu, X.; Yi, Z.; Zou, Z.; Liu, Y.; Zhu, L. Multi-Grained Similarity Preserving and Updating for Unsupervised Cross-Modal Hashing. Appl. Sci. 2024, 14, 870. https://doi.org/10.3390/app14020870

AMA Style

Wu R, Zhu X, Yi Z, Zou Z, Liu Y, Zhu L. Multi-Grained Similarity Preserving and Updating for Unsupervised Cross-Modal Hashing. Applied Sciences. 2024; 14(2):870. https://doi.org/10.3390/app14020870

Chicago/Turabian Style

Wu, Runbing, Xinghui Zhu, Zeqian Yi, Zhuoyang Zou, Yi Liu, and Lei Zhu. 2024. "Multi-Grained Similarity Preserving and Updating for Unsupervised Cross-Modal Hashing" Applied Sciences 14, no. 2: 870. https://doi.org/10.3390/app14020870

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop