You are currently viewing a new version of our website. To view the old version click .
Journal of Imaging
  • Article
  • Open Access

17 January 2020

Multiple Query Content-Based Image Retrieval Using Relevance Feature Weight Learning

,
and
1
Computer Science Department, College of Computer and Information Sciences, King Saud University, Riyadh 11362, Saudi Arabia
2
Computer Science Department, Community College, Taibah University, P.O.B.344, Madinah 41477, Saudi Arabia
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Advances in Image Feature Extraction and Selection

Abstract

We propose a novel multiple query retrieval approach, named weight-learner, which relies on visual feature discrimination to estimate the distances between the query images and images in the database. For each query image, this discrimination consists of learning, in an unsupervised manner, the optimal relevance weight for each visual feature/descriptor. These feature relevance weights are designed to reduce the semantic gap between the extracted visual features and the user’s high-level semantics. We mathematically formulate the proposed solution through the minimization of some objective functions. This optimization aims to produce optimal feature relevance weights with respect to the user query. The proposed approach is assessed using an image collection from the Corel database.

1. Introduction

The proliferation of social networks along with the wide dispersion of smart devices has yielded an exponential growth of digital image databases. This massive increase has precipitated the challenge of mining specific images among huge collections. Thus, image retrieval has become an active field of research.
Text-based image retrieval systems require the annotation of images in a database. This is a tedious and expensive task. Content-based image retrieval (CBIR) represents an alternative that overcomes this drawback. Hereby, the user selects a query image that conveys the information he or she is looking for. The query content is then used to mine the database. Several CBIR approaches have been reported in the literature [1,2,3,4] during the last decade, and other CBIR applications have been proposed recently [5,6,7]. For these CBIR systems, the visual properties of an image are described using low-level feature descriptors [8]. More specifically, these descriptors can include the color, texture, and shape properties of the image. These low-level features translate the visual content of the image into numerical vectors that allow for a quantitative estimation of the similarity between two images [9,10,11,12]. In the literature, several visual descriptors have been proposed. Some of these descriptors are generic, while others are specific to some applications [13,14]. However, there is a gap between the semantic interest of the user and an extracted visual feature descriptor. For instance, if the user can provide an image containing a red apple as a query, the retrieved images may contain a red rose, red balloon, or green apple depending on the visual feature adopted by the CBIR system. Recently, multiple query retrieval systems have emerged as a way to overcome this disadvantage and to bridge the semantic gap with low-level image features [15,16]. For these systems, the user provides a set of query images. This conveys a better understanding of the user’s high-level interest to the retrieval system. The main point of multiple query CBIR systems is how to compute the distances between the set of query images and each image in the database.
Based on the investigation of existing multiple query retrieval systems, in this paper, we have designed a new approach for a multiple-query-based CBIR system, named weight-learner, which overcomes the limitations of existing approaches. The proposed approach retrieves a set of images relevant to the semantic learnt from the set of query images without learning the content of the image or its annotation. The proposed query formulation is practical for the user in such a way that 2–3 images are sufficient to express retrieval interest. Moreover, the user does not have to provide ambiguous information, such as image query scoring or feature weighting. Weight-learner expresses the semantics of the user query using feature relevance weights that are automatically learned from the set of queries. The optimal feature weights are derived through the minimization of a new formulated objective function. We should mention here that any visual feature can benefit from this approach, i.e., it is not restricted to any feature. The considered features in this work are used as examples.
In order to assess the performance of the proposed multiple query approach, we compare it to the performance of a simple query image. Moreover, we study the effects of the number of query images on weight-learner. Finally, we implement existing multiple query approaches and compare their retrieval performance to the proposed solution. The rest of the paper is organized as follows. In Section 2, we review the work related to multiple query retrieval systems. The proposed approach, weight-learner, is described in Section 3. In Section 4, we assess the performance of the weight-learner approach. Finally, in Section 5, we conclude and outline future works.

3. Multiple Query Content-Based Image Retrieval Based on Relevance Feature Weight Learning

We propose a new approach for a multiple-query-based CBIR system, weight-learner, that overcomes the limitations of existing approaches. We make the query formulation practical for the user in such a way that 2–3 images is sufficient to express user retrieval interest. Moreover, the user does not have to provide extra information, such as image query scoring or feature weighting. We aim to express the semantics of the user query using feature relevance weights that are automatically learned from the set of query images. The weight-learner is based on a linear weighted combination of the distances between each query image and a given image in the dataset. The weight with respect to each considered feature descriptor will be learned in an unsupervised manner. These weights would allow one to discriminate between the features with respect to each query image in such a way that a large weight will be assigned to the most relevant feature. This optimization will be achieved through the minimization of the following objective function:
J = i = 1 M t = 1 S w t 2 ( x it c t ) 2
subject to
t w t [ 0 . . 1 ]
and
t = 1 S w t = 1
where M is the number of query images, S is the number of feature entries, and x it represents the t th feature value of the S-dimensional data point. On the other hand, w t represents the discrimination weight of feature t, and c t is the t th component of the center vector. The term t = 1 S w t 2 ( x it c t ) 2 is the weighted Euclidean distance from query image I Q i to the center of the query image set, which is defined as in Equation (4):
c t = 1 M i = 1 M x it
To optimize J with respect to w t , we use the Lagrange multiplier technique [28] and obtain
J L = i = 1 M t = 1 S w t 2 ( x it c t ) 2 λ ( t = 1 S w t 1 ) .
The derivative with respect to w t is
J L w t = 2 w t i = 1 M ( x it c t ) 2 λ .
Setting the derivative to zero yields
w t = λ 2 i = 1 M ( x it c t ) 2 .
The derivative with respect to λ is
J L λ = t = 1 S w t 1.
Setting J L λ to 0 gives
t = 1 S w t = 1.
Substituting Equation (7) in (9) gives
t = 1 S λ 2 i = 1 M ( x i t c t ) 2 = 1
which yields
λ = 2 t = 1 S 1 i = 1 M ( x i t c t ) 2
from which,
w t = 1 i = 1 M ( x it c t ) 2 q = 1 S 1 i = 1 M ( x iq c q ) 2 .
Simplifying Equation (13), we obtain
w t = D t q = 1 S D q
where D t = 1 / i = 1 M ( x it c t ) 2 , which reflects the inverse of the dispersion of the set of query images along the t th dimension, and q = 1 S D q can be perceived as the overall inverse dispersion of the set of query images. Therefore, w t is inversely proportional to the ratio of the dispersion along the t th dimension to the total image query dispersion. Thus, the more compact the set of queries is along the t th dimension, the higher the discriminating weight, w t , will be for the t th dimension. Once the set of weights, w t , is computed using Equations (4) and (12), the distance between the set of query images, I Q = [ x it ] 1 i M 1 t S , and the images in the dataset, I D j = [ x jt ] 1 j N 1 t S , is defined as
d i s t ( I Q , I D j ) = i = 1 M   t = 1 S w t 2 ( x it x jt ) 2 .
Algorithm 1 describes the steps of the proposed approach.
Algorithm 1 Weight-learner
Input:
I Q = { I Q i ( i = 1 , , M ) } : set of M query images,
I D = { I D j ( j = 1 , , N ) } : image dataset of size N
Output: K retrieved images
  • Extract the set of feature descriptors from each image I D = { I D j ( j = 1 , , N ) }
  • Extract the set of feature descriptors from each query I Q = { I Q i ( i = 1 , , M ) }
  • Compute the weights using (4) and (12),
  • For each image I D j in the database ( j = 1 , , N )
      Compute the distance D ( I Q 1 , , I Q M , I D j ) between the set of query images I Q and an image I D j using (14)
  • Sort D ( I Q 1 , , I Q M , I D j ) , ( j = 1 , , N ) in increasing order
  • Return the K first images with the smallest distances

4. Experiments

In this chapter, we assess the performance of the proposed weight-learner approach using image collection from the Corel dataset [29]. In the first experiment, we compare the performance of the proposed multiple query approach to the performance of a simple query image. In the second experiment, we study the effect of the number of query images on the proposed weight-learner approach. In the third experiment, we implement existing multiple query approaches and compare their retrieval performances to the proposed solution. In the following section, a description of the considered dataset, an outline of the retrieval performance measure used, and details about the performed experiments are provided.

4.1. Dataset Description

In order to assess the performance of the proposed approach, we use a subset of 1300 images from the Corel dataset [29]. This dataset includes 12 categories: air shows, arabian horse, antelope, beach, bears, birds, butterfly, bulling, buses, car performance, car race, and dogs. Each class contains 100 images, except for the last category, which contains 200 images.

4.2. Experimental Setting

Since any visual feature can benefit from this approach and it not restricted to any feature, from the 1300 images, we extract the color histogram [30] and the edge histogram features [31] as examples. After concatenating the two features, entries between 1 and 150 correspond to the edge histogram, and entries between 151 and 341 correspond to the color histogram. We set this number of query images to five, as we consider more than five query images to be impractical for the user. Next, we build 260 groups of five images randomly selected from each category to be used as query sets for every analysis.
For multiple query approaches that require feature weights, we considered three cases. Table 1 summarizes the three considered cases for the two feature weights, WC and WT.
Table 1. The three considered cases for manually set feature weights.
For the other multiple query approaches requiring a goodness score of v, we attribute a score of 1 to 5 to the five query images. We consider five cases obtained by permuting the scores between the query images. Table 2 reports the five considered cases.
Table 2. The five considered cases for manually set query image scores.
We use the average normalized modified rank retrieval (ANMRR) score [32] to assess the performance of the multiple query approaches. The ANMRR score is between zero and one. This value is small for good retrieval results and large for bad retrieval results.

4.3. Experiment 1

In order to show that the weight-learner reduces the semantic gap between the extracted visual features and the user’s high-level semantics, we will compare the results of the proposed approach to a classical single query retrieval result. For each of the two systems, the proposed approach and the single query image, we will convey the edge histogram feature, the color histogram feature, and the concatenation of both. Table 3 reports a comparison of the performance of the proposed multiple query approach vs. the performance of a single query image using an edge histogram, color histogram, and the concatenated features based on ANMRR. For the three considered cases, the proposed multiple query approach gives a lower ANMRR score. This means that weight-learner outperforms the single query approach. Moreover, we notice from Table 3 that, in the case of a single query approach, the performance decreases when using both features. This decrease in performance is due to the problem of dimensionality. This effect will be worse if more features are used. Since the feature descriptor choice for a specific query is challenging, a consideration of all features will lower the performance of the system. However, the proposed multiple query approach has a low ANMRR (0.2906) when using both features, reflecting a very good retrieval result. Indeed, weight-learner is based on a linear weighted combination of the distances between each query image and a given image in the dataset. These weights allow one to discriminate between the features with respect to each query image in such a way that a large weight will be assigned to the most relevant feature. These feature relevance weights that are learned from the set of query images express the semantics of the user query.
Table 3. Comparison of the performance of the proposed multiple query approach vs. the performance of a simple query image using a texture histogram, color histogram, and concatenated features based on average normalized modified rank retrieval (ANMRR).
In order to show examples of how the proposed approach outperforms the single query approach, we display in Figure 1 the retrieval results of the sample query for the category Bears using concatenated features. Figure 1a shows the five query images provided as input for the proposed approach. Figure 1b reports the retrieval result using a single query approach, where only the first query image is used. Figure 1c displays the retrieval results using the new proposed approach. The single query approach does not return any correct images among the first retrieved images. However, the five retrieved images using weight-learner are all correct.
Figure 1. Retrieval results of the sample query for the category “bears” using concatenated features. (a) The 5 query images, (b) the corresponding retrieval result using a single query approach (only the first query image is used), and (c) the corresponding retrieval result using the proposed approach.

4.4. Experiment 2

In this experiment, we aim to study the effect of the number of query images on the proposed approach. We will convey both features to the proposed multiple query approach and tune the number of query images from two to seven and record the ANMRR score for each considered image query number. Figure 2 reports the proposed approach performance with respect to the number of query images. Ultimately, the AVNMRR decreases when the number of query images increases. This is expected since more query images would provide more information and thus guide the proposed retrieval system towards a better retrieval result. Moreover, the decrease rate becomes slower starting at four query images. Table 4 shows the performance of the proposed multiple query approach when the number of query images is equal to two compared to the performance of a simple query image using the concatenated features based on ANMRR. Even when using only two query images, the performance of the retrieval system increases drastically. We can thus conclude that the proposed approach has a query formulation that is practical for the user in such a way that 2–3 images would be sufficient to express retrieval interest.
Figure 2. Proposed approach performance with respect to the number of query images.
Table 4. The performance of the proposed multiple query approach when the number of query images is equal to 2 compared to the performance of a simple query image using the concatenated features based on ANMRR.

4.5. Experiment 3

In this experiment, we compare the proposed multiple query image to the existing multiple query images described in Section 2. Namely, we compare the proposed approach to the weighted color and texture histograms approach [17], the Image Grouper approach [19], the Minimum weighted distance combination approach [21], the standard-deviation-based weights approach [22], the MindReader approach [23], the Logic AND-based distance approach [25], the Multi-feature query approach [26], and the distance combination [27] approach. In this experiment, we consider five query images for all approaches since we assume that five query images are the maximum number of query images that can be considered for the query to remain practical for the user.
Table 5 compares the performance of the proposed multiple query approach to the existing multiple query approaches and to the single query approach based on ANMRR. The proposed approach noticeably outperforms all existing approaches. Indeed, our approach has the lowest ANMRR when using the concatenated feature (0.2906). Moreover, even when using color features or textures feature only, the proposed approach still outperforms the other approaches. Moreover, the weighted color and texture histograms approach [17] has a lower performance than the single query approach even when considering the best results over the five considered feature weight scenarios. This performance discrepancy is due to the sensitivity to the choice of the feature weights. The Multi-feature query approach [26] with an ANMRR of 0.4266 is the next best multiple query retrieval system after the proposed one. Although this approach does not formally use feature weights (refer to Section 2), by taking the minimum over the feature values, it offers a type of crisp weighting, where the feature with the minimum value is assigned a value of 1, and all the others have a value of 0. This explains why this model outperforms the other existing approaches. However, this other approach takes into account only one feature and discards the others, so the information provided by the other features is lost. Since our approach uses a weighted combination of all features, all the information provided by the features is taken into account with different levels of importance according to the weight value. These weights allow for discrimination between the different feature values. This explains the outstanding performance of the proposed approach. Moreover, the proposed approach does not require any feature weight tuning since the feature weights are automatically learnt from the user’s set of queries.
Table 5. The performance of the proposed multiple query approach compared to the existing multiple query approaches and to the single query approach based on ANMRR.
To further investigate how the proposed approach functions better than the other approaches, we provide a sample query set corresponding to the category of buses (Figure 3). Figure 3a shows the corresponding five query images. Figure 3b–f,i reports the first five retrieved images returned by the existing approaches. For each existing approach, the parameters giving the best results are considered. The last set of images (Figure 3j) is related to the proposed approach. The images highlighted by the black box show that the image was correctly retrieved.
Figure 3. Sample query 2 and its corresponding retrieval result with respect to all considered approaches. (a) The query images. (b) Results obtained using the method in [17] with color weight = 0.1 and texture weight = 0.9. (c) Results obtained using the Image Grouper [19]. (d) Results obtained using the method in [22] with color weight = 0.5 and texture weight = 0.5. (e) Results obtained using the method in [22]. (f) Results obtained using the method in [23]. (g) Results obtained using the method in [26]. (h) Results obtained using the method in [27]. (i) Results obtained using the method in [29]. (j) Results obtained using the proposed method.
From Figure 3, we can see that, for some existing multiple query approaches, the “bus” category does not appear in the set of retrieved images while images related to other categories are retrieved. This means that these approaches failed to understand the semantic “bus” of the query images. On the other hand, our approach noticeably outperforms the other approaches. This means that the proposed approach is able to successfully capture user semantics through the conveyed set of query images. This is done by learning the appropriate feature relevance weights that discriminate between features. Figure 4 plots the weights learned by the proposed approach for the considered sample query set. Entries between 1 and 150 correspond to the edge histogram, and entries between 151 and 341 correspond to the color histogram. We can see from Figure 4 that the edge histogram entries are assigned automatically larger weights than the color histogram ones. This means that the various colors of the buses are not taken into account, but the shapes of the buses are considered. This yields a successful retrieval result. The proposed approach looks at the similarities between the different feature entries of the same query set and learns the appropriate relevance feature weights to yield outperforming retrieval results.
Figure 4. Learned weights using the new proposed approach on Sample query 2. Entries between 1 and 150 correspond to the edge histogram, and entries between 151 and 341 correspond to the color histogram.

5. Conclusions

Despite the availability of advanced visual image low-level descriptors, understanding the semantic concept of an image remains a challenging task. It is even harder to express the interest of the user using one single image. Multiple query images help to better express user interest and narrow the gap between an image’s semantics and its visual descriptors. However, a multiple-query-based CBIR system should be practical and able to effectively aggregate the information provided by different query images to enhance retrieval performance.
In this work, we proposed a novel linear weighted combination of distances between each query image and the images in the database. We learned the relevance weight for each query image with respect to each visual descriptor entry. These weights allowed us to discriminate between the features in such a way that large weights are assigned to the most relevant feature entry. These weights are learned through the minimization of an appropriate objective function. The experimental results show that weight-learner outperforms single query retrieval results. Moreover, it makes query formulation practical for the user in such a way that two to three images would be sufficient to express a retrieval interest.
For future work, we intend to investigate the image query goodness score. We plan to learn a weight wit for each image query i with respect to the feature entry t. This way, the learned weights will reflect both the relevance feature weight and the image query goodness score. This work will yield a new definition for the distance between a set of query images and the images of the database. We also intend to use the new distance of the proposed approach as a way of including the feedback of the user in the retrieval. Thus, instead of considering a set of multiple query images, we use a set of feedback images reported as correctly retrieved by the user. The learned weight would then improve the next query.

Author Contributions

Conceptualization, O.B. and M.M.B.I.; methodology O.B. and M.M.B.I.; software, A.A.-M.; validation, M.M.B.I.; formal analysis, A.A.-M. and O.B.; investigation, A.A.-M.; writing—original draft preparation, A.A.-M.; writing—review and editing, M.M.B.I. and O.B.; visualization, A.A.-M.; supervision, O.B.; project administration, M.M.B.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

The authors are grateful for the support by the Research Center of the College of Computer and Information Sciences, King Saud University. This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Faloutsos, C.; Barber, R.; Flickner, M.; Hafner, J.; Niblack, W.; Petkovic, D.; Equitz, W. Efficient and effective querying by image content. J. Intell. Inf. Syst. 1994, 3, 231–262. [Google Scholar] [CrossRef]
  2. Flickner, M.; Sawhney, H.; Niblack, W.; Ashley, J.; Huang, Q.; Dom, B.; Gorkani, M.; Hafner, J.; Lee, D.; Petkovic, D.; et al. Query by image and video content: The QBIC system. Computer 1994, 28, 23–32. [Google Scholar] [CrossRef]
  3. Nepal, S.; Ramakrishna, M. A generalized test bed for. In Proceedings of the 10th lnternational Conference of the Information Resources Management Association, Hershey, PA, USA, 16–19 May 1999. [Google Scholar]
  4. Ma, W.Y.; Manjunath, B.S. NETRA: A toolbox for navigating. In Proceedings of the IEEE International Conference on Image Processing, Santa Barbara, CA, USA, 26–29 October 1997. [Google Scholar]
  5. Jiang, M.; Zhang, S.; Li, H.; Metaxas, D.N. Computer-Aided Diagnosis of Mammographic Masses Using Scalable Image Retrieval. IEEE Trans. Biomed. Eng. 2014, 62, 783–792. [Google Scholar] [CrossRef] [PubMed]
  6. Zhang, X.; Liu, W.; Dundar, M.; Badve, S.; Zhang, S. Towards Large-Scale Histopathological Image Analysis: Hashing-Based Image Retrieval. IEEE Trans. Med. Imaging 2014, 34, 496–506. [Google Scholar] [CrossRef] [PubMed]
  7. Alraqibah, H.; Bchir, O.; Ismail, M.M.B. X-ray Image Retrieval System Based on Visual Feature Discrimination. In Proceedings of the International Conference on Digital Image Processing (ICDIP), Athens, Greece, 5–6 April 2014. [Google Scholar]
  8. Komal, J.; Akhilesh, V.; Savita, G.; Swati, G. A Survey on Recent Image Indexing and Retrieval Techniques for Low-Level Feature Extraction in CBIR Systems. In Proceedings of the IEEE International Conference on Computational Intelligence & Communication Technology (CICT), Ghaziabad, India, 13–14 February 2015. [Google Scholar]
  9. Guo, J.M.; Prasetyo, H.; Chen, J.H. Content-Based Image Retrieval Using Error Diffusion Block Truncation Coding Features. IEEE Trans. Circ. Syst. Video Technol. 2014, 25, 466–481. [Google Scholar]
  10. Thepade, S.D.; Shinde, Y.D. Robust CBIR using sectorisation of hybrid wavelet transforms with Cosine-Walsh, Cosine-Kekre, Cosine-Hartley combinations. In Proceedings of the International Conference on Pervasive Computing (ICPC), Pune, India, 8–10 January 2015. [Google Scholar]
  11. Gupta, N.; Das, S.; Chakraborti, S. Extracting information from a query image, for content based image retrieval. In Proceedings of the Eighth International Conference on Advances in Pattern Recognition (ICAPR), Kolkata, India, 4–7 January 2015. [Google Scholar]
  12. Hassekar, P.P.; Sawant, R.R. Experimental analysis of perceptual based texture features for image retrieval. In Proceedings of the International Conference on Communication, Information & Computing Technology (ICCICT), Mumbai, India, 15–17 January 2015. [Google Scholar]
  13. Alraqibah, H.; Ismail, M.M.B.; Bchir, O. Empirical Comparison of Visual Descriptors for Content based X-ray Image Retrieval. In Proceedings of the International Conference on Image and Signal Processing, Cherbourg, France, 20 June–2 July 2014. [Google Scholar]
  14. Otaiba, S.; Qassim, S.; Bchir, O.; Ismail, M.M.B. Empirical comparison of visual descriptors for multiple bleeding spots recognition in wireless capsule endoscopy video. In Proceedings of the International Conference on Computer Analysis of Images and Patterns (CAIP), York, UK, 27–29 August 2013. [Google Scholar]
  15. Hsiao, K.J.; Calder, J.; Hero, A.O. Pareto-Depth for Multiple-Query Image Retrieval. IEEE Trans. Image Process. 2014, 24, 583–594. [Google Scholar] [CrossRef] [PubMed]
  16. Fernando, B.; Tuytelaars, T. Mining Multiple Queries for Image Retrieval: On-the-fly learning of an Object-specific Mid-level Representation. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013. [Google Scholar]
  17. Tang, J.; Acton, S. An Image Retrieval Algorithm Using multiple query Images. In Proceedings of the Seventh International Symposium on Signal Processing and its Applications, Paris, France, 4 July 2003. [Google Scholar]
  18. Swain, M.J.; Ballard, D.H. Color indexing. Int. J. Comput. Vis. 1991, 7, 11–32. [Google Scholar] [CrossRef]
  19. Nakazato, M.; Huang, T.S. Extending Image Retrieval with Group-Oriented Interface. In Proceedings of the IEEE International Conference on Multimedia and Expo, Lausanne, Switzerland, 26–29 August 2002. [Google Scholar]
  20. Mika, S.; Ratsch, G.; Weston, J.; Scholkopf, B.; Mullers, K. Fisher Discriminant Analysis with Kernels. In Proceedings of the IEEE Conference on Neural Networks for Signal Processing, Madison, WI, USA, 25 August 1999. [Google Scholar]
  21. Iqbal, Q.; Aggarwa, J.K. Feature Integration, Multi-image Queries and Relevance Feedback in Image Retrieval. In Proceedings of the International Conference on Visual Information Systems, Miami, FL, USA, 24–26 September 2003. [Google Scholar]
  22. Brunelli, R.; Mic, O. Image Retrieval by Examples. IEEE Trans. Multimed. 2000, 2, 164–170. [Google Scholar] [CrossRef]
  23. Ishikawa, Y.; Faloutsos, R.S.C. MindReader: Querying databases through multiple examples. In Proceedings of the VLDB Conference, New York, NY, USA, 24–27 August 1998. [Google Scholar]
  24. Hazewinkel, M. Mahalanobis Distance, Encyclopedia of Mathematics; Springer: Berlin/Heidelberg, Germany, 2001. [Google Scholar]
  25. Joseph, S.; Balakrishnan, K. Multi Query Image Retrieval System with Application to Mammogram Images. Int. J. Adv. Res. Comput. Sci. 2012, 3, 469–474. [Google Scholar]
  26. Nepal, S.; Ramakrishna, M.V. MultiFeature Query by Multiple Examples in Image Databases. In Advances in Data Management; McGrawHill Publish Co. Ltd.: New York, NY, USA, 2000. [Google Scholar]
  27. Zhu, L.; Zhang, A. Supporting multi-example image queries in image databases. In Proceedings of the IEEE International Conference on on Multimedia and Expo, New York, NY, USA, 30 July–2 August 2000. [Google Scholar]
  28. Hazewinkel, M. Lagrange Multipliers, Encyclopedia of Mathematics; Springer: Berlin/Heidelberg, Germany, 1998. [Google Scholar]
  29. Wang, J.Z.; Li, J.; Wiederhold, G. SIMPLIcity: Semantics-Sensitive Integrated Matching for Picture Libraries. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 23, 947–963. [Google Scholar] [CrossRef]
  30. Mangijao, S.S.; Hemachandran, K. Image Retrieval based on the combination of Color Histogram and Color Moment. Int. J. Comput. Appl. 2012, 58, 27–34. [Google Scholar]
  31. Sun, W.C.; Park, D.K.; Park, S.J. Efficient use of MPEG-7 edge histogram descriptor. ETRI J. 2002, 24, 23–30. [Google Scholar]
  32. Park, G.; Baek, Y.; Lee, H.K. Re-ranking algorithm using post-retrieval clustering for content-based image retrieval. Inf. Process. Manag. 2005, 41, 177–194. [Google Scholar] [CrossRef]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.