A Novel Adaptive Feature Fusion Strategy for Image Retrieval

In the era of big data, it is challenging to efficiently retrieve the required images from the vast amount of data. Therefore, a content-based image retrieval system is an important research direction to address this problem. Furthermore, a multi-feature-based image retrieval system can compensate for the shortage of a single feature to a certain extent, which is essential for improving retrieval system performance. Feature selection and feature fusion strategies are critical in the study of multi-feature fusion image retrieval. This paper proposes a multi-feature fusion image retrieval strategy with adaptive features based on information entropy theory. Firstly, we extract the image features, construct the distance function to calculate the similarity using the information entropy proposed in this paper, and obtain the initial retrieval results. Then, we obtain the precision of single feature retrieval based on the correlation feedback as the retrieval trust and use the retrieval trust to select the effective features automatically. After that, we initialize the weights of selected features using the average weights, construct the probability transfer matrix, and use the PageRank algorithm to update the initialized feature weights to obtain the final weights. Finally, we calculate the comprehensive similarity based on the final weights and output the detection results. This has two advantages: (1) the proposed strategy uses multiple features for image retrieval, which has better performance and more substantial generalization than the retrieval strategy based on a single feature; (2) compared with the fixed-feature retrieval strategy, our method selects the best features for fusion in each query, which takes full advantages of each feature. The experimental results show that our proposed method outperforms other methods. In the datasets of Corel1k, UC Merced Land-Use, and RSSCN7, the top10 retrieval precision is 99.55%, 88.02%, and 88.28%, respectively. In the Holidays dataset, the mean average precision (mAP) was 92.46%.


Introduction
Image retrieval aims to find images related to images of interest to users from the image dataset. Over the past 20 years, research in this area has focused on improving the accuracy and efficiency of retrieval [1]. There are two common types of image retrieval systems: text-based image retrieval systems and content-based image retrieval systems. Text-based image retrieval systems require experienced experts to tag images, which is expensive and time-consuming [2]. In contrast, content-based image retrieval uses computers to replace the manual processing of many repetitive tasks, reducing the cost of human, financial, and material resources, so content-based image retrieval is gradually becoming the mainstream method. Content-based image retrieval systems fall into two categories [3]: global features indexed with hashing strategies; local scale-invariant features indexed by a vocabulary tree or a k-d tree. Both approaches have advantages and disadvantages, and their performance complements each other [1,3].
More advanced feature extraction techniques and more appropriate feature selection and fusion strategies are crucial to improve image retrieval performance. In recent years, with the development of deep learning and convolutional neural networks, feature extraction techniques for images have improved substantially compared to the previous ones.
Compared with traditional features, CNN features have more powerful representation capabilities. The authors in [4] proposed a fast multi-label feature selection method based on information theory feature ranking, which sorts features according to the importance of features, and then selects the top-ranking features. In feature fusion, fusion at the metric level is widely used, but selecting appropriate features and their corresponding weights for fusion to improve retrieval performance is still an important issue. The authors in [5] used average global weights to fuse Color and Texture features for image retrieval. The authors in [6] traverse the weight domain to select optimal weights for retrieval performance as the final weights.
Unlike previous feature selection and weight processing methods, this paper first selects valid features automatically based on the correlation feedback of single features and gives them the same initial weight; then updates the initial weights using the PageRank algorithm to obtain the final weight of each feature; finally calculates a comprehensive metric based on the final weights to output the retrieved results. For feature selection, we first extract some features and then use their retrieval accuracy as trust, then rank the trust and select some of them with higher trust for fusion. In feature fusion, we first give the same initial weights to the selected features; then construct a probability transfer matrix based on their trust; and finally update the feature weights using the PageRank algorithm, normalizing the final result obtained as the weights of each feature.
The rest of the paper is organized as follows. The second section describes some of the related work involved in this paper. The third section describes the feature extraction and processing process. The fourth section describes the feature selection and fusion strategy of this paper in detail. The fifth section presents the experimental results. The sixth section summarizes the work in this paper and gives suggestions for subsequent work.

Related Work
Image retrieval starts with creating an image database, inputting the retrieved image, calculating the similarity to the image in the image database, and finally outputting the retrieval result based on the image similarity.
Feature extraction technology is an essential factor affecting the retrieval effect. Before deep learning became popular, feature extraction methods were mainly based on features of color and texture, such as color histograms, color moments, local binary pattern (LBP) features, or some manual feature extractors such as scale invariant feature transform (SIFT), Gabor filters, generalized search trees (GIST) and histogram of oriented gradients (HOG). The development of deep learning technology has brought a significant breakthrough in image retrieval tasks, especially in feature extraction. CNN can be used to extract more representative depth features in images and fusing multiple depth features can help image retrieval tasks achieve better performance.
There has been a lot of work on fusing multiple features to improve retrieval system performance [7,8]. The relevance feedback algorithm [9] has addressed the semantic gap problem that arises in content-based image retrieval. The results obtained through relevance feedback are very similar to human understanding [10,11]. The main steps of relevance feedback are as follows: first, the retrieval system provides primary retrieval results based on the retrieval keywords provided by the user; then the user makes a judgment on the current retrieval results, which retrieval results are relevant to their retrieval needs and which retrieval results are irrelevant; and then the retrieval system will provide new retrieval results based on the feedback of users. This process is repeated until the search results meet the user's needs.
The PageRank algorithm is a link analysis algorithm proposed by Google founders Larry Page and Sergey Brin in 1997 and is often used in the field of web search [12,13]. It creates a probability transfer matrix based on the links between web pages and continuously updates the initial probability vector based on the probability transfer matrix to obtain a stable probability vector to represent the probability distribution of users visiting each web page. There have been a lot of improved applications based on the traditional PageRank algorithm. The authors in [14] used the PageRank algorithm to enhance the Peer-Led Team Learning (PLTL) teaching method. The authors in [15] proposed a new algorithm for ranking online shops based on the PageRank algorithm. The authors in [16] proposed a new method based on PageRank to find the most influential individuals in online social networks.
In this paper, we propose a feature selection and fusion strategy based on information entropy and correlation feedback. We first extract features and preprocess the features, then construct similarity using information entropy and obtain the retrieval accuracy of each feature based on relevant feedback and use the accuracy as a trust to select features. Finally, we initialize the feature weights with the mean value, update the weights with the PageRank algorithm to obtain the comprehensive similarity, and obtain the retrieval results based on the comprehensive similarity.

Feature Extraction and Preprocess
In this paper, we extract multiple CNN features and traditional color and texture features simultaneously. The specific extraction process is as follows: for CNN features, we use the pre-trained model trained in the ImageNet dataset to extract the fc2 layer features of the original image and grayscale map, respectively. Here, fc2 refers to the features of the second fully connected layer of the network. Of course, in specific experiments, we can also extract features from other appropriate layers. Then, we stitch the extracted two parts of features to get a one-dimensional feature vector. For color and texture features, we extract color histogram features, denoted as color, LBP features, and GIST features for the original image only.
The dimensionality of tandem features is high, and noise points in image feature descriptors may reduce retrieval accuracy. In addition, using the high-dimensional image feature descriptors as a benchmark to calculate the similarity between the query image and the image in the database is time-costly and reduces retrieval efficiency. Based on the above two points, we need to reduce the dimensionality of the data features. Principal component analysis (PCA) is one of the common methods for data dimensionality reduction, which maps high-dimensional data to low-dimensional space with large variance by matrix mapping. It can reduce the dimensionality of data, reduce the computational effort, and avoid the loss of information as much as possible. In the experiments of this paper, we use the classical PCA algorithm to reduce the image features to different dimensions and compare the retrieval results of the features under different dimensions, which are described in Part 5 of the paper.
Since it takes a lot of time to calculate the similarity measure of multiple features, hash coding is needed to reduce the subsequent computation. The authors use binary to encode image features [17] as follows: where i ∈ {1, 2, . . . , k} ave(F i ) is the mean of the feature F i , m is the dimension of the feature F i , F(i, j) and is the j-th component of the feature F i .

Method Introduction
In image retrieval methods based on multi-feature fusion, feature selection and weight determination are very important to improve retrieval accuracy. The traditional method first extracts the color, texture, or other effective features of the image and then manually selects several features for fusion. The disadvantage of this approach is that the same feature contributes differently to image retrieval in different datasets. For a given dataset, color information may be more prominent, texture information may be more prominent, or other information may be more prominent. It is difficult to determine the features suitable for the dataset and determine the optimal combination of features in advance in image retrieval work. If we try different combinations for all datasets, we can get better results, but the time consumption is vast, the efficiency is low, and the algorithm is not fixed. In response, this paper proposes a feature selection strategy using relevant feedback based on information entropy. The basic process of the retrieval system is shown in Figure 1.

Method Introduction
In image retrieval methods based on multi-feature fusion, feature selection and weight determination are very important to improve retrieval accuracy. The traditional method first extracts the color, texture, or other effective features of the image and then manually selects several features for fusion. The disadvantage of this approach is that the same feature contributes differently to image retrieval in different datasets. For a given dataset, color information may be more prominent, texture information may be more prominent, or other information may be more prominent. It is difficult to determine the features suitable for the dataset and determine the optimal combination of features in advance in image retrieval work. If we try different combinations for all datasets, we can get better results, but the time consumption is vast, the efficiency is low, and the algorithm is not fixed. In response, this paper proposes a feature selection strategy using relevant feedback based on information entropy. The basic process of the retrieval system is shown in Figure 1 Figure 1. The proposed retrieval system framework.
As shown by Figure 1, for each retrieved image, we first extract the features of the image and process the features using PCA algorithm and hash coding; then combine information entropy and correlation feedback to select the features. Next, we initialize the feature weights using the mean value, update the weights using the PageRank algorithm. Finally, we calculate the combined similarity using selected features and final weights and output the retrieval result based on similarity.

Feature Selection Strategy
Information entropy is often used as a quantitative indicator of the information content of a system. In statistics, the greater the entropy, the smaller the distribution of discrimination; the smaller the entropy, the greater the discrimination of distribution. In [18], the authors used information entropy to calculate the weights of clusters, which is calculated as follows: Influenced by this idea, this paper proposes a new form of information entropy to measure the differentiation in each dimension of a single feature and assign different weights to each dimension. The smaller the entropy, the greater discrimination of that As shown by Figure 1, for each retrieved image, we first extract the features of the image and process the features using PCA algorithm and hash coding; then combine information entropy and correlation feedback to select the features. Next, we initialize the feature weights using the mean value, update the weights using the PageRank algorithm. Finally, we calculate the combined similarity using selected features and final weights and output the retrieval result based on similarity.

Feature Selection Strategy
Information entropy is often used as a quantitative indicator of the information content of a system. In statistics, the greater the entropy, the smaller the distribution of discrimination; the smaller the entropy, the greater the discrimination of distribution. In [18], the authors used information entropy to calculate the weights of clusters, which is calculated as follows: Influenced by this idea, this paper proposes a new form of information entropy to measure the differentiation in each dimension of a single feature and assign different weights to each dimension. The smaller the entropy, the greater discrimination of that feature dimension, the greater contribution to the distance calculation and the greater the weight, and vice versa. Then, we construct a weighted distance function to obtain the similarity between images.
In this paper, we use the trust obtained based on entropy and relevant feedback to select these functions. Suppose there are k extracted features, denoted as F i (p), F i (q)(i ∈ {1, 2, . . . , k}), q is the image to be retrieved, p is the image in the database. Firstly, we calculate the retrieval accuracy of a single feature based on users' feedback on the retrieval results. Then, the retrieval accuracy is regarded as the trust of a single feature, and the trust is used to select features.
Firstly, under feature F i , calculate the distance between the retrieved image q and all images in the database, and get the similarity S i (q) between them based on the distance. The calculation process of similarity S i (q) is: where S i (q) is the similarity of the retrieved image q with other images in the database calculated based on the feature F i , n is the total number of images in the image database.
When calculating distances, we use Manhattan distances with weights. These weights are constructed using information entropy. w j is the weight of the j-th component of feature F i , is calculated as Equations (6) and (7): here, i ∈ {1, 2, . . . , n}, j ∈ {1, 2, . . . , m}, n is the total number of images in the image library, m is the dimensionality of the feature, f ij is the j-th dimensional feature of the i-th image, and h j is the entropy of the j-th component of F i . When the values of each dimension of the feature are equal, the entropy and weights of each dimension are also equal, and the resulting entropy h j is 1. Then, the similarity S i (q) is sorted, and t samples from the image database with high similarity with the retrieved image q are returned as the retrieval results, denoted as L i = L i 1 , L i 2 , . . . , L i t , where t is the predefined number of returned images. The query precision is calculated according to the retrieval results and relevant feedback. The query precision is the trust of a single feature, denoted as P(i), i ∈ {1, 2, . . . , k}. In the Holidays dataset, the accuracy is measured by mAP, and in the other datasets, the accuracy is measured by precision. After the single feature trust is obtained, they are ranked from highest to lowest, and the top 5 features are selected for fusion based on trust.

Post-Processing of Weights by the PageRank Algorithm
Influenced by the PageRank algorithm, this paper takes the probability vector as a feature weight to measure the importance of each feature in feature fusion and uses the PageRank algorithm to obtain the final weights. Firstly, we construct the probability transfer matrix according to the retrieval trust of all selected features as the preference and then update the initial weights so that the features with higher trust can obtain higher fusion weights. Using the PageRank algorithm to update weights has the following two advantages: (1) the weight of the adaptive means that we do not need to design the weight function manually; (2) the time complexity is greatly reduced. Compared with other weight updating algorithms, the PageRank algorithm does not need the help of query results or loss function and only needs to iteratively use the probability transfer matrix, which greatly reduces the calculation amount of weight updating.
Firstly, we calculate the transfer matrix of the degree of retrieval performance preference between every single feature, denoted as TM kk = {TM(x, y)}. TM(x, y) is the preference of feature x, x ∈ {1, 2, . . . , k} to feature y, y ∈ {1, 2, . . . , k}, which is constructed as follows: Then, each column TM kk = {TM(x, y)} is normalized by dividing each value in the column by the sum of the values in that column. Finally, using TM kk to obtain the weights of single features. The process is as following Algorithm 1: Algorithm 1. Post-Processing of Weights by the PageRank Algorithm [13] Input: the probability transfer matrix TM, hyper-parameter of γ ∈ [0, 1] and ε, the initial weight w 0 . Output: the final weight w d . 1: Initialize the weight w 0 = 1/k. 2: d = 1 3: repeat 4: where w 0 is the initial weight, d refers to the number of iterations, ε, γ is the parameter, in this paper ε = 0.005, γ = 0.4. After post-processing by the PageRank algorithm, the final weights are used to calculate the integrated similarity measure: Finally, the retrieval results are output according to the integrated similarity metric.

Datasets
To test the effectiveness of the method proposed in this paper, we conducted experiments in four image retrieval datasets: Holidays, Corel-1k, UC Merced Land-Use, and RSSCN7. The details of which are as follows.
Holidays dataset, containing 1491 holiday photos, in 500 categories, where the first image in each category is used as the retrieved image and the remaining images are the corresponding related images. The evaluation index for its retrieval performance is mAP.
Wang (Corel 1K) dataset, which contains 1000 images, is divided into ten categories: flowers, horses, buildings, beaches, etc. Each category contains 100 images. The precision of Top-N is used as the evaluation index for the retrieval system.
The UC Merced Land-Use dataset, a remote sensing dataset describing land use, was released by the University of California and includes 2100 images divided into ten categories. The image size is 256 × 256, and the pixel resolution is 30 cm. Again, we used precision as the evaluation index for the retrieval system.
The RSSCN7 dataset, the remote sensing scene dataset, contains seven categories; each category includes 400 images, each with a pixel size of 400 × 400. These images are sampled at four different scales in different seasons and weather variations. We also used precision as the evaluation index for the retrieval system.

Evaluation Index
The precision of the first N images (top-N) is calculated as follows: Nr refers to the number of the first N returned images that belong to the same category as the retrieved image and N refers to the total number of returned images. The mAP is the mean value of average precision (AP) obtained from multiple queries, which is presented by: where AP k refers to the AP k value obtained by the query of the k-th image, Nr the number of the first N returned images that belong to the same category as the retrieved image LOC i refers to the location of the i-th image in the returned image that belongs to the same category as the retrieved image. For example, if five images belonging to the same category as the query image appear at positions 1, 3, 6, 9, and 10, the AP is represented as follows: AP = (1/1 + 2/3 + 3/6 + 4/9 + 5/10)/5 = 0.62 (13)

Experiment Introduction
In order to prove the effectiveness of the proposed method, we conducted experiments in four datasets: Holiday, Corel-1k, UC Merced Land Use, and RSSCN7, and analyzed the experimental results. First, we obtained and compared the following three kinds of accuracy: (1) the retrieval accuracy using a single feature; (2) the accuracy obtained by selecting five features using the method in this paper and then fusing them using the average value as the weight (AVG); (3) the accuracy obtained using the feature selection and fusion strategy proposed in this paper (Ours). Then, to further prove the effectiveness of the feature selection strategy, we analyze the relationship between the retrieval accuracy of features and the number of times they are selected. After that, to further prove the effectiveness of the PageRank algorithm, we analyzed the relationship between the retrieval accuracy of features and the final weights. Finally, we give a visualization map of the retrieval results for each dataset. All the code used in this article is shared at https://github.com/JNZYBOBO/AZYBOBO.

Holidays Dataset
In the Holidays dataset, we selected the first image of each category as the retrieved images, 500 images in total, and the others as image database images. Firstly, we obtained the retrieval accuracy of each feature, AVG, and ours under different dimensionality reduction results, and the results are shown in Table 1. From Table 1, we can see the impact of different dimensions on the retrieval results. The retrieval results of each dimension are basically the same. If only the retrieval accuracy is considered, we can choose to reduce the dimension to about 300 dimensions. If the retrieval efficiency is considered, we can choose to reduce the dimension to a lower level.
The results based on single-feature retrieval show that CNN features achieve better results in single-feature image retrieval, while Color, LBP, Gist, and other color or texture features are less effective. This is because color features cannot describe the local distribution of colors and the spatial location of each color in the image; texture features are easily affected by factors such as illumination and reflection. Meanwhile, CNN features can fully consider the local and global information of the image since they are not easily affected by factors such as location and illumination and are highly resistant to noise.Different features have different effects on image retrieval results and features with stronger expressive power can be selected using the feature selection strategy proposed in this paper. Compared with any single-feature image retrieval, using our feature selection strategy can obtain more than 10 percentage points of mAP improvement. Further, using our PageRank algorithm to update the initial weights also improves the mAP by 1-2 percentage points in different dimensions.
To further prove the rationality of the feature selection strategy proposed in this paper, we counted the ratio of each feature that was selected in 500 retrieved images and compared the relationship between the retrieval accuracy of each feature and the number of times it was selected; the results are shown in Figure 2. In the figure, "Mean mAP" refers to the average value of mAPs for single feature retrieval in different dimensions in Table 1, and "Ratio" refers to the ratio of each feature selected, i.e., the number of times it is selected divided by the number of retrieved images (500). According to Figure 2, The trend of the two broken lines is approximately the same, and the higher the retrieval accuracy of the features, the higher the probability of being selected. Thus, it can be shown that our feature selection method can select features with good results. 23,1670 noise.Different features have different effects on image retrieval results and stronger expressive power can be selected using the feature selection stra in this paper. Compared with any single-feature image retrieval, using our tion strategy can obtain more than 10 percentage points of mAP improve using our PageRank algorithm to update the initial weights also improves t 2 percentage points in different dimensions.

21,
To further prove the rationality of the feature selection strategy pro paper, we counted the ratio of each feature that was selected in 500 retriev compared the relationship between the retrieval accuracy of each feature an of times it was selected; the results are shown in Figure 2. In the figure, refers to the average value of mAPs for single feature retrieval in different Table 1, and "Ratio" refers to the ratio of each feature selected, i.e., the num is selected divided by the number of retrieved images (500). According to trend of the two broken lines is approximately the same, and the higher th curacy of the features, the higher the probability of being selected. Thus, it that our feature selection method can select features with good results. In order to further prove the effectiveness of the PageRank algorithm this paper, we use 20 retrieval images to obtain the mAP of the selected fiv the final weight after updating, a total of 100 mAP values and 100 weig relationship between the two is shown in Figure 3. From the figure, w things: (1) the retrieval accuracy of the selected features is usually high, illustrates the rationality of our feature selection strategy; (2) in general, fo the greater the retrieval accuracy, the greater the final weight. Thus, in feat PageRank algorithm can give higher weight to the features with higher re In order to further prove the effectiveness of the PageRank algorithm proposed in this paper, we use 20 retrieval images to obtain the mAP of the selected five features and the final weight after updating, a total of 100 mAP values and 100 weight values. The relationship between the two is shown in Figure 3. From the figure, we can see two things: (1) the retrieval accuracy of the selected features is usually high, which further illustrates the rationality of our feature selection strategy; (2) in general, for each feature, the greater the retrieval accuracy, the greater the final weight. Thus, in feature fusion, the PageRank algorithm can give higher weight to the features with higher retrieval accuracy, so that the fused features have a better retrieval effect.
illustrates the rationality of our feature selection strategy; (2) in general, f the greater the retrieval accuracy, the greater the final weight. Thus, in fea PageRank algorithm can give higher weight to the features with higher r cy, so that the fused features have a better retrieval effect.  In the multi-feature fusion-based image retrieval system, compared with manual feature selection and design of weight functions, the feature selection strategy proposed in this paper can automatically select the most superior features for different images, while using the PageRank algorithm can automatically update the feature weights and adjust the weight combinations according to the feature trust. Therefore, it has higher retrieval accuracy and robustness. Table 2 shows the comparison between the experimental results in the Holidays dataset and other methods for manual selection of features and fusion weights. The optimal precision of image retrieval based on this method is 0.9246. Compared with the literature [19], retrieval accuracy increased by about 0.0126, compared with the literature [17], retrieval accuracy increased by about 0.0401, compared with the literature [20], retrieval accuracy increased by about 0.0446, compared with the literature [21], retrieval accuracy increased by about 0.0526, compared with the literature [2], retrieval accuracy increased by about 0.0556, compared with the literature [3], and retrieval accuracy increased by about 0.0694. For the Corel-1K dataset, we randomly selected 20% of the images in each category as the images to be retrieved, for a total of 200 images, and the rest as the database images. Firstly, we reduce the dimensionality of each extracted feature by the PCA dimensionality reduction algorithm and then obtain the precision of top10, top20, top30, top40, and top50 in different dimensions. The results are shown in Figure 4. As we can see, image retrieval accuracy is highest when the features drop to 10 dimensions among any number of returned images. The reduction in dimensionality not only improves precision but also reduces computational effort and improves retrieval efficiency.
From the comparison of Tables 1 and 3, we can see that the same features have different retrieval effects in different datasets, like AlexNet, ResNet50 works well in the Holidays dataset but poorly in the Corel-1k dataset, and NASNetLarge on the contrary. Furthermore, the features required for different images in the same dataset are different. According to the traditional method, we cannot determine the suitable features for different datasets or different images in the same dataset in advance. In contrast, using our feature selection strategy ensures that for each retrieved image, as long as there is a feature to be selected with better effect, it will definitely be selected, and the fused effect will definitely be better than the best retrieval effect of all features. Table 3 compares three kinds of accuracy on the different number of returned images when the feature dimension is 10. It can be seen that in different return numbers, the feature selection strategy in this paper has achieved better detection results. The detection accuracy is nearly 10 percentage points better than that of any single feature. At the same time, the use of the PageRank algorithm also makes the detection accuracy in different top indicators achieve a certain degree of improvement.
as the images to be retrieved, for a total of 200 images, and the res ages. Firstly, we reduce the dimensionality of each extracted featur sionality reduction algorithm and then obtain the precision of top10 and top50 in different dimensions. The results are shown in Figu image retrieval accuracy is highest when the features drop to 10 dim number of returned images. The reduction in dimensionality not onl but also reduces computational effort and improves retrieval efficien From the comparison of Tables 1 and 3, we can see that the sa ferent retrieval effects in different datasets, like AlexNet, ResNet5 Holidays dataset but poorly in the Corel-1k dataset, and NASNetL Furthermore, the features required for different images in the same According to the traditional method, we cannot determine the suit ferent datasets or different images in the same dataset in advance. I feature selection strategy ensures that for each retrieved image, as l ture to be selected with better effect, it will definitely be selected, and definitely be better than the best retrieval effect of all features. Ta kinds of accuracy on the different number of returned images whe  Similarly, we also counted the ratio of each feature being selected on 200 retrieved images, and the relationship between these ratios and retrieval accuracy is shown in Figure 5. In the figure, "Mean Precision" refers to the average value of precision for single feature retrieval in different top indicators in Table 3, and "Ratio" still refers to the rate at which each feature is selected. We can see that compared with Figure 2, the trend between the two is more consistent in Figure 5, indicating that the feature selection strategy is more effective in the Corel-1K dataset, which can also be seen from the final experimental results in Tables 1 and 3. This is because the Holidays dataset has more categories and fewer samples in each category, which are easily affected by randomness and lead to unstable results. In contrast, the Corel-1K dataset has fewer categories and a larger number of images in each category, which is more robust. and fewer samples in each category, which are easily affected by r unstable results. In contrast, the Corel-1K dataset has fewer ca number of images in each category, which is more robust.  Figure 6 shows the optimization effect of the PageRank algori According to Figure 6, it can be seen that the PageRank algorithm weights according to the retrieval accuracy, giving higher weigh  Figure 6 shows the optimization effect of the PageRank algorithm on fusion weights. According to Figure 6, it can be seen that the PageRank algorithm optimizes the fusion weights according to the retrieval accuracy, giving higher weights to the excellent features to make the best final retrieval results. Compared with the Holidays dataset, the effect is more obvious in the Corel-1K dataset. 23,1670 tures to make the best final retrieval results. Compared with the Holid effect is more obvious in the Corel-1K dataset. At the same time, we compared the results with the results of oth Corel-1K dataset, as shown in Table 4. The accuracy of our method is 0.9 with the methods proposed in other papers, our method has obvious adv Similarly, we conducted experiments in the UC Merced Land Use tasets. As with the Corel-1K dataset, we also used 20% of each category At the same time, we compared the results with the results of other papers in the Corel-1K dataset, as shown in Table 4. The accuracy of our method is 0.9955. Compared with the methods proposed in other papers, our method has obvious advantages. Similarly, we conducted experiments in the UC Merced Land Use and RSCN7 datasets. As with the Corel-1K dataset, we also used 20% of each category of images as the retrieved images and the rest as the database images. According to the results after tuning the parameters, we choose to reduce the features extracted from the RSCN7 dataset to 40 dimensions and the features extracted from the UC Merced Land Use dataset to 20 dimensions, after which we obtain the retrieval accuracy under different top metrics, and the results are shown in Tables 5 and 6. Table 5 shows the precision on different tops in the UC Merced Land Use dataset, and Table 6 shows the precision on different tops in the RSSCN7 dataset. The detection results in both datasets also demonstrate the rationality of the proposed method in this paper. With different evaluation metrics, the detection results are greatly improved by using our method in both cases. We also demonstrate the effectiveness of the feature selection strategy and weight update strategy in both datasets. In the UC Merced Land Use and RSSCN7 datasets, we selected 10 images in each category to obtain the relationship between the selected ratio and the mean precision of each feature. In addition, we selected 21 images to obtain the relationship between the retrieval accuracy and the final weights, and all the relationships are shown in Figure 7. From Figure 7, we can see that the proposed method still has good results in these two datasets. selected 10 images in each category to obtain the relationship between the selected ra and the mean precision of each feature. In addition, we selected 21 images to obtain relationship between the retrieval accuracy and the final weights, and all the relati ships are shown in Figure 7. From Figure 7, we can see that the proposed method still good results in these two datasets.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Visualization Result
Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the "Results and similarity" column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval re-0.99974986

Conclusions
The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature's weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval results. Experimental results in the Holidays, Corel-1k, UC Merced Land-Use, and RSSCN7 show that the proposed method has better retrieval performance than single feature image retrieval systems and image retrieval systems without the PageRank algorithm. The experiments in the Holidays and Corel-1k datasets show that the proposed method is better than the existing methods. The next step is to investigate how to select the appropriate parameters and improve the retrieval performance after the features have been selected.