Open Access
This article is

- freely available
- re-usable

*Algorithms*
**2018**,
*11*(8),
115;
doi:10.3390/a11080115

Article

Color-Based Image Retrieval Using Proximity Space Theory

^{1}

College of Science, Dalian Maritime University, Dalian 116026, China

^{2}

School of Automation, Shenyang Aerospace University, Shenyang 110136, China

^{*}

Authors to whom correspondence should be addressed.

Received: 30 June 2018 / Accepted: 26 July 2018 / Published: 28 July 2018

## Abstract

**:**

The goal of object retrieval is to rank a set of images by their similarity compared with a query image. Nowadays, content-based image retrieval is a hot research topic, and color features play an important role in this procedure. However, it is important to establish a measure of image similarity in advance. The innovation point of this paper lies in the following. Firstly, the idea of the proximity space theory is utilized to retrieve the relevant images between the query image and images of database, and we use the color histogram of an image to obtain the Top-ranked colors, which can be regard as the object set. Secondly, the similarity is calculated based on an improved dominance granule structure similarity method. Thus, we propose a color-based image retrieval method by using proximity space theory. To detect the feasibility of this method, we conducted an experiment on COIL-20 image database and Corel-1000 database. Experimental results demonstrate the effectiveness of the proposed framework and its applications.

Keywords:

color features; proximity space; image retrieval; similarity measurement## 1. Introduction

Image retrieval is the procedure of retrieving relevant images from a big database of images, which usually occurs in one of two ways: text-based image retrieval or content-based image retrieval. Text-based image retrieval describes an image by using one or more keywords, but manual annotation has a huge cost burden. To address this defect, content-based image retrieval is introduced in [1]. Content-based image retrieval differs from traditional retrieval methods, as it uses visual contents such as key points, colors and textures to retrieve similar images from large scale image databases. Content-based image retrieval has been widely employed in several research fields, such as machine vision, artificial intelligence, signal processing, etc. Any content-based image retrieval system has two steps that can be explained as follows. The first step is feature extraction; the key point in this step is to find a method of feature extraction to precisely describe the content of image. Color features, which are the basic characteristic of the image content, are widely used in image retrieval [2,3,4,5]. For example, in [6], the color features was extracted in HSV color space by using color moment method. Gopal et.al. considered that only using grayscale information is imprecise in the image matching [7], because some images have same garyscale but have different color information, therefore they put color information into SURF descriptor making the precision better than only used SURF [8]. The second step is similarity measurement; in this step, the query image is selected, and the similarities between it and images of database are computed. In recent years, many different similarity computation methods have been proposed. Similarity is obtained by calculating the distance between two feature vectors, such as Euclidean, Manhattan and Chi square distance [9]. Chen et al. calculated similarities between the query image and images in database by comparing the coding residual [10]. Kang et al. proposed a similarity measure that is a sparse representation issue via matching the feature dictionaries between the query image and images of database [11]

With the rapid development of modern computer technology and soft computing theory, many modern algorithms and theories are applied to the image segmentation and image retrieval, which include fuzzy sets [12], rough set theory [13,14,15,16,17,18,19,20], near sets [21], mathematical morphology, neural network, etc. Computational Proximity (CP) [21] is a novel method to deal with image processing, which can identify nonempty sets of points that are either close to each other or far apart by comparing descriptions of pixels points or region in digital images or videos. Pták and Kropatsch [22] were the first to combine nearness in digital images with proximity spaces. In CP, pixel is an object, and the various features of pixel, such as grayscale, color, texture and geometry, are regarded as its attributes. For a point x, $\mathsf{\Phi}(x)$ denotes the feature vector of x, then

$$\mathsf{\Phi}(x)=({\varphi}_{1}(x),\cdots ,{\varphi}_{i}(x),\cdots ,{\varphi}_{n}(x)),i\in \{1,\cdots ,n\}.$$

Similarity between either points or regions in CP is defined in terms of the feature vectors that describe either the points or the regions. Feature vectors are expressed by introducing probe function. A probe $\varphi $ maps a geometric object to a feature value that is a real number [21]. The similarity of a pair of nonempty sets is determined by the number of the same element. Put another way, nonempty sets are strongly near, provided sets have at least one point in common (introduced in [23]).

**Example**

**1.**

[21] Let $A,B,C$ be a nonempty sets of points in a picture, $\mathsf{\Phi}=\{r,g,b\}$, where $r(A)$ equals the intensity of the redness of A, $g(A)$ equals the intensity of the greenness of A and $b(A)$ equals the intensity of the blueness of A. Then,

$$\begin{array}{c}\hfill \begin{array}{c}\mathsf{\Phi}(A)=(1,\text{}1,\text{}1)(description\text{}of\text{}A)\hfill \\ \mathsf{\Phi}(B)=(0,\text{}0,\text{}1)(description\text{}of\text{}B)\hfill \\ \mathsf{\Phi}(C)=(0,\text{}1,\text{}1)(description\text{}of\text{}C)\hfill \\ \mathsf{\Phi}(D)=(0.4,\text{}0.4,\text{}0.4)(description\text{}of\text{}D)\hfill \\ \mathsf{\Phi}(E)=(0.4,\text{}0.2,\text{}0.3)(description\text{}of\text{}E).\hfill \end{array}\end{array}$$

From this, Region A resembles Region B, since both regions have equal blue intensity. Region A strongly resembles (is strongly near) Region C, since both regions have equal green and blue intensities. By contrast, the description of Region A is far from the descriptions of Region D or E, since Regions A and D or E do not have equal red or green or blue intensities.

From the above mentioned example, the description of an image is defined by a feature vector, which contains feature values that are real numbers (values are obtained by using probes, e.g., color brightness, and gradient orientation). It can be found that there are different feature values for different values. However, only several features play a distinguishing role for a object (set), so it is necessary to consider the order of features in matching description of another object. Specifically, for any nonempty set of points in an image, first we transform RGB color space into HSV (Hue, Saturation, Value) color space, and then 72 kinds of color features will be obtained by the color quantization. The HSV color space, which is closer to the human eye, is selected as the color feature representation of the image. Because human eye is limited to distinguish colors and if the range of each color component is too large, feature extraction will become very difficult. Thus, it is necessary to reduce the amount of computation and improve efficiency by quantifying the HSV color space component [24], and several dominant colors [25] are chosen, which can effectively match the image and reduce the computation complexity.

## 2. Proposed Method

This part presents the proposed method from two aspects, first is the transformation from RGB color space into HSV color space, and 72 kinds of color features will be obtained by the color quantization in HSV color space. Then, the similarity of images is identified by using proximity space theory. Second, we put forward an improved dominance granule structure formula as a similarity calculation method of proximity sets.

#### 2.1. Feature Extraction

RGB color space and HSV color space are often used in digital image processing, and RGB color space is the basic color space that can be represented by a cube model. Red, green and blue are called the three primary colors and any point corresponds to a color in the RGB color model. HSV color space is established based on human visual perception characteristics and it can be represented by cone model. “Hue” means different color, “saturation” means the intensity of the color, whose value is in the unit interval [0,1], and “value” means the degree of shade of color. Circumferential color of the top of cone is pure color, and the top of the cone corresponds to $V=1$, which contains three surfaces (R = 1, G = 1, and B = 1) in RGB color space. The conversion formula from RGB color space to HSV color space is given by:
where $\sigma \to 0$, $v=v/255$ and r, g, and b are the red, green and blue coordinates of a color, respectively, in RGB color space.
where $h\in [0,\cdots ,360]$ and $s,v\in [0,\cdots ,1]$ are the hue, saturation and value coordinates of a color, respectively, in HSV color space. Then, according to the property of the HSV color space, we can obtain 72 kinds of colors by the non-uniform quantization method, which can reduce the dimension of the color feature vectors and the amount of calculation. The quantization formula is described as follows:

$$v=max(r,g,b),\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}s=\frac{v-min(r,g,b)}{v},$$

$${r}^{\prime}=\frac{v-r+\sigma}{v-min(r,g,b)+\sigma},\phantom{\rule{0.166667em}{0ex}}{g}^{\prime}=\frac{v-g+\sigma}{v-min(r,g,b)+\sigma},\phantom{\rule{0.166667em}{0ex}}{b}^{\prime}=\frac{v-b+\sigma}{v-min(r,g,b)+\sigma},$$

$$\begin{array}{c}\hfill {h}^{\prime}=\left\{\begin{array}{cccc}(5+{b}^{\prime}),& r=max(r,g,b)& and& g=min(r,g,b)\\ (1-{g}^{\prime}),& r=max(r,g,b)& and& g\ne min(r,g,b)\\ (1+{r}^{\prime}),& g=max(r,g,b)& and& b=min(r,g,b)\\ (3-{b}^{\prime}),& g=max(r,g,b)& and& b\ne min(r,g,b)\\ (3+{g}^{\prime}),& b=max(r,g,b)& and& r=min(r,g,b)\\ (5-{r}^{\prime}),& b=max(r,g,b)& and& r\ne min(r,g,b),\end{array}\right.\end{array}$$

$$\begin{array}{c}\hfill h=60\times {h}^{\prime},\end{array}$$

$$\begin{array}{c}\hfill H=\left\{\begin{array}{cc}0& \phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}316\le h\le 360\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}or\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}0\le h\le 20\\ 1& 21\le h\le 40\\ 2& 41\le h\le 75\\ 3& 76\le h\le 155\\ 4& 156\le h\le 190\\ 5& 191\le h\le 270\\ 6& 271\le h\le 295\\ 7& 296\le h\le 315,\end{array}\right.\end{array}$$

$$\begin{array}{c}\hfill S=\left\{\begin{array}{cc}0,& 0\le s\le 0.2\\ 1,& 0.2<s\le 0.7\\ 2,& 0.7<s\le 1,\end{array}\right.\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}V=\left\{\begin{array}{cc}0,& 0\le v\le 0.2\\ 1,& 0.2<v\le 0.7\\ 2,& 0.7<v\le 1,\end{array}\right.\end{array}$$

Based on the above quantization method, the three-dimensional feature vector is mapped into a one-dimensional feature vector by using the equation: $l=h{Q}_{s}{Q}_{v}+s{Q}_{v}+v$, where ${Q}_{s}$ and ${Q}_{v}$ are set to 3 because saturation and value are divided into three levels, respectively. Thus, the above equation can be rewritten as

$$l=9h+3s+v.$$

According to the value range of h, s, and v, we can obtain the range of $l=[0,1,\cdots ,71]$, and then we can get 72 kinds of color features.

#### 2.2. Similarity Measure Based on Dominance Perceptual Information

Feature extraction and similarity measurement are two important steps in image retrieval. In this paper, color feature is selected as the retrieval feature, and we quantify color features into 72 numbers, $\mathsf{\Phi}=\{0,1,\cdots ,71\}$. If an image X has T pixels, and there are Q pixels with the color feature c in an image, ${\varphi}_{c}(X)=\frac{Q}{T}$ and $c\in \mathsf{\Phi}$. For example, $c=0$, ${\varphi}_{0}(X)$ represents the probability that the color feature equals 0 in the image X, ${\varphi}_{1}(X)$ represents the probability that the color feature equals 1 in the image X, and ${\varphi}_{71}(X)$ represents the probability that the color feature equals 71 in the image X; then, we can obtain that ${\mathsf{\Phi}}^{\prime}(X)=\{{\varphi}_{0}(X),{\varphi}_{1}(X),\cdots ,{\varphi}_{71}(X)\}$, using which we can get a color histogram (see Figure 1) of a query image. The Top-m color features ${\mathsf{\Phi}}^{m}(X)$ is obtained by descending order of ${\mathsf{\Phi}}^{\prime}(X)$, ${\mathsf{\Phi}}^{m}(X)=\{{l}_{1}^{x},{l}_{2}^{x},\cdots ,{l}_{m}^{x}\},{l}_{i}\in \mathsf{\Phi}=\{0,1,2,\cdots ,71\}$.

In this paper, we take Top-m dominant color features in the HSV space to match images to improve the efficiency of retrieval; therefore, we use color features vector ${\mathsf{\Phi}}^{m}(X)$ as the distinguishing description of image X. In addition, image similarity measurement is also important, such as Euclidean distance formula or cosine formula. However, if we use Euclidean distance formula or cosine formula to compute the similarity between the two images, in essence, it will lose significance. For example, the similarity between matrix $[1,2,3]$ and matrix $[2,4,6]$ is equal to 1 by cosine formula. Inspired by dominance rough set theory, we propose a novel dominance granule structure similarity method as similarity measurements, defined as follows:

**Definition**

**1.**

Let ${\mathsf{\Phi}}^{m}(X)=\{{l}_{1}^{x},{l}_{2}^{x},\cdots ,{l}_{m}^{x}\},{\mathsf{\Phi}}^{m}(Y)=\{{l}_{1}^{y},{l}_{2}^{y},\cdots ,{l}_{m}^{y}\}$ and ${\mathsf{\Phi}}^{m}(X),{\mathsf{\Phi}}^{m}(Y)\subseteq \mathsf{\Phi}$, where X and Y represent the query image and the image of image database, respectively; and ${a}_{x}$ and ${a}_{y}$ represent the color attribute of X and Y, respectively. A is a set of all pre-order relationships on ${\mathsf{\Phi}}^{m}(X)$ and ${\mathsf{\Phi}}^{m}(Y)$. $\forall \phantom{\rule{4pt}{0ex}}{\u2ab0}_{{a}_{x}},{\u2ab0}_{{a}_{y}}\in A$ [26], ${\mathsf{\Phi}}^{m}(X)/{\u2ab0}_{{a}_{x}}=\{{\left[{l}_{1}^{x}\right]}_{{\u2ab0}_{{a}_{x}}},{\left[{l}_{2}^{x}\right]}_{{\u2ab0}_{{a}_{x}}},\cdots ,{\left[{l}_{m}^{x}\right]}_{{\u2ab0}_{{a}_{x}}}\}$ and ${\mathsf{\Phi}}^{m}(Y)/{\u2ab0}_{{a}_{y}}=\{{\left[{l}_{1}^{y}\right]}_{{\u2ab0}_{{a}_{y}}},{\left[{l}_{2}^{y}\right]}_{{\u2ab0}_{{a}_{y}}},\cdots ,{\left[{l}_{m}^{y}\right]}_{{\u2ab0}_{{a}_{y}}}\}$ are the corresponding dominance granule structure, then the similarity between X and Y can be denoted by
because Equation (8) needs to calculate the global color feature similarity, the efficiency is reduced due to the large amount of calculation. Therefore, the Top-m color features of the image are selected for image matching, and Equation (8) is improved as follows:
where $\left|\mathsf{\Phi}\right|$ is the number of all the elements in set Φ, $\left|{\left[{l}_{i}^{x}\right]}_{{\u2ab0}_{{a}_{x}}}\mathsf{\Delta}{\left[{l}_{j}^{y}\right]}_{{\u2ab0}_{{a}_{y}}}\right|=|{\left[{l}_{i}^{x}\right]}_{{\u2ab0}_{{a}_{x}}}\cup {\left[{l}_{j}^{y}\right]}_{{\u2ab0}_{{a}_{y}}}|-|{\left[{l}_{i}^{x}\right]}_{{\u2ab0}_{{a}_{x}}}\cap {\left[{l}_{j}^{y}\right]}_{{\u2ab0}_{{a}_{y}}}|$, and ${\left[{l}_{i}\right]}_{{\u2ab0}_{{a}_{x}}}=\{{l}_{j}|{l}_{j}{\u2ab0}_{{a}_{x}}{l}_{i},1\le j\le m\}$,${\left[{l}_{i}\right]}_{{\u2ab0}_{{a}_{y}}}=\{{l}_{j}|{l}_{j}{\u2ab0}_{{a}_{y}}{l}_{i},1\le j\le m\}$. The ordinal relation between objects in terms of the attribute ${a}_{x}$ or ${a}_{y}$ is denoted by ⪰, and ${l}_{1}{\u2ab0}_{{a}_{x}}{l}_{2}$ or ${l}_{1}{\u2ab0}_{{a}_{y}}{l}_{2}$ means that ${l}_{1}$ is at least as good as ${l}_{2}$ with respect to the attribute ${a}_{x}$ or ${a}_{y}$.

$$\begin{array}{c}\hfill S(\mathsf{\Phi}/{\u2ab0}_{{a}_{x}},\mathsf{\Phi}/{\u2ab0}_{{a}_{y}})=1-\frac{1}{\left|\mathsf{\Phi}\right|-1}\sum _{i=1}^{\left|\mathsf{\Phi}\right|}\frac{\left|{\left[{l}_{i}\right]}_{{\u2ab0}_{{a}_{x}}}\mathsf{\Delta}{\left[{l}_{i}\right]}_{{\u2ab0}_{{a}_{y}}}\right|}{\left|\mathsf{\Phi}\right|}.\end{array}$$

$$\begin{array}{c}\hfill S({\mathsf{\Phi}}^{m}(X)/{\u2ab0}_{{a}_{x}},{\mathsf{\Phi}}^{m}(Y)/{\u2ab0}_{{a}_{y}})=1-\frac{1}{m}\sum _{i=1}^{m}\frac{\underset{j}{min}\left|{\left[{l}_{i}^{x}\right]}_{{\u2ab0}_{{a}_{x}}}\mathsf{\Delta}{\left[{l}_{j}^{y}\right]}_{{\u2ab0}_{{a}_{y}}}\right|}{m},j\in \{1,2,\cdots ,m\},\end{array}$$

**Example**

**2.**

To prove the feasibility of this method, we give an applied instance. Suppose $m=3$, ${\mathsf{\Phi}}^{3}(I)=(7,9,1)$ (description of query image I), where 7, 9 and 1 refer to the color features by quantization and ${\varphi}_{7}(I)>{\varphi}_{9}(I)>{\varphi}_{1}(I)$. Let ${\mathsf{\Phi}}^{3}({I}_{1})$ (description of ${I}_{1}$), ${\mathsf{\Phi}}^{3}({I}_{2})$ (description of ${I}_{2}$), ${\mathsf{\Phi}}^{3}({I}_{3})$ (description of ${I}_{3}$), etc represent the set of Top-3 color features of each image in image database, respectively. Thus, the similarity can be obtained by Equation (9) and results of the similarity are shown by Equation(10). ${I}_{1}$, ${I}_{2}$, ⋯, and ${I}_{13}$ represent images in the image database.

According to Equation (9), if $\phantom{\rule{4pt}{0ex}}7{\u2ab0}_{{a}_{x}}9{\u2ab0}_{{a}_{x}}1$, $\phantom{\rule{4pt}{0ex}}7{\u2ab0}_{{a}_{y}}1{\u2ab0}_{{a}_{y}}9$,

${\left[7\right]}_{{\u2ab0}_{{a}_{x}}}=\{7\}$ | $\text{}{\left[7\right]}_{{\u2ab0}_{{a}_{y}}}=\{7\}$ |

${\left[9\right]}_{{\u2ab0}_{{a}_{x}}}=\{7,9\}$ | $\text{}{\left[1\right]}_{{\u2ab0}_{{a}_{y}}}=\{7,1\}$ |

${\left[1\right]}_{{\u2ab0}_{{a}_{x}}}=\{7,9,1\}$, | $\text{}{\left[9\right]}_{{\u2ab0}_{{a}_{y}}}=\{7,1,9\}$, |

${\mathsf{\Phi}}^{3}(I)/{\u2ab0}_{{a}_{x}}=\left\{\{7\},\{7,9\},\{7,9,1\}\right\}$, ${\mathsf{\Phi}}^{3}({I}_{2})/{\u2ab0}_{{a}_{y}}=\left\{\{7\},\{7,1\},\{7,1,9\}\right\}$,

$$\begin{array}{ccc}\hfill S({\mathsf{\Phi}}^{m}(I)/{\u2ab0}_{{a}_{x}},{\mathsf{\Phi}}^{m}({I}_{2})/{\u2ab0}_{{a}_{y}})& =& 1-\frac{1}{3}\sum _{i=1}^{3}\frac{\underset{j}{min}\left|{\left[{l}_{i}^{x}\right]}_{{\u2ab0}_{{a}_{x}}}\mathsf{\Delta}{\left[{l}_{j}^{y}\right]}_{{\u2ab0}_{{a}_{y}}}\right|}{3}\hfill \\ & =& 1-\frac{1}{3}\times (\frac{0}{3}+\frac{1}{3}+\frac{0}{3})\hfill \\ & =& 1-\frac{1}{3}\times \frac{1}{3}\hfill \\ & =& \frac{8}{9}\approx 0.8,\hfill \end{array}$$

$$\begin{array}{c}\hfill Sim=\left\{\begin{array}{ccc}1,& if& {\mathsf{\Phi}}^{3}({I}_{1})=(7,9,1)\\ 0.8,& if& {\mathsf{\Phi}}^{3}({I}_{2})=(7,1,9)\\ 0.8,& if& {\mathsf{\Phi}}^{3}({I}_{3})=(9,7,1)\\ 0.8,& if& {\mathsf{\Phi}}^{3}({I}_{4})=(1,7,9)\\ 0.8,& if& {\mathsf{\Phi}}^{3}({I}_{5})=\{(7,9,{l}_{3})|{l}_{3}\notin {\mathsf{\Phi}}^{3}(I)\}\\ 0.7,& if& {\mathsf{\Phi}}^{3}({I}_{6})=\{(7,1,{l}_{3})|{l}_{3}\notin {\mathsf{\Phi}}^{3}(I)\}\\ 0.6,& if& {\mathsf{\Phi}}^{3}({I}_{7})=(9,1,7)\\ 0.6,& if& {\mathsf{\Phi}}^{3}({I}_{8})=(1,9,7)\\ 0.6,& if& {\mathsf{\Phi}}^{3}({I}_{9})=\{(7,{l}_{2},{l}_{3})|{l}_{2},{l}_{3}\notin {\mathsf{\Phi}}^{3}(I)\}\\ 0.5,& if& {\mathsf{\Phi}}^{3}({I}_{10})=\{(9,1,{l}_{3})|{l}_{3}\notin {\mathsf{\Phi}}^{3}(I)\}\\ 0.4,& if& {\mathsf{\Phi}}^{3}({I}_{11})=\{(9,{l}_{2},{l}_{3})|{l}_{2},{l}_{3}\notin {\mathsf{\Phi}}^{3}(I)\}\\ 0.2,& if& {\mathsf{\Phi}}^{3}({I}_{12})=\{(1,{l}_{2},{l}_{3})|{l}_{2},{l}_{3}\notin {\mathsf{\Phi}}^{3}(I)\}\\ \vdots \\ 0,& if& {\mathsf{\Phi}}^{3}({I}_{13})=\{({l}_{1},{l}_{2},{l}_{3})|{l}_{1},{l}_{2},{l}_{3}\notin {\mathsf{\Phi}}^{3}(I)\}.\end{array}\right.\end{array}$$

As shown in Equation (10), we can see that the color feature’s order of ${I}_{1}$ is the same as the color feature’s order of I, so the similarity between ${I}_{1}$ and I is 1. Image ${I}_{5}$ strongly resembles (is strongly near) image I better than image ${I}_{6}$, since the color feature’s order of ${I}_{5}$ is near to the color feature’s order of I. The image ${I}_{13}$ does not resemble image I, since both color features are different, so the similarity between ${I}_{13}$ and I is 0. Therefore, it makes it easier to match images if there exist more identical elements between color features. Simultaneously, it is also critical whether corresponding elements between features vector are equal, such as ${I}_{6}$ and image ${I}_{7}$. The similarity measure proposed in this paper can improve the subjective and objective consistency of retrieval results.

#### 2.3. Algorithm Flow Chart

As discussed above, the framework of our proposed method can be shown in Figure 2. The main procedures of the proposed framework are summarized as follows:

**Step 1**: A query image in the RGB value (r,g,b) by nonlinear transformation converts to the HSV space value (h,s,v).

**Step 2**: We get 72 kinds of colors by the quantization.

**Step 3**: We obtain the color histogram of a query image and take the Top-m dominant color features forming the set of ${\mathsf{\Phi}}^{m}(X)$. We do the same with each image in the image database and get the set of ${\mathsf{\Phi}}^{m}(Y)$.

**Step 4**: We use our proposed method for similarity measurements to calculate the similarity between the query image and the image of image database.

Experimental results with Top-20 images are shown in Figure 3 by using proposed similarity measurement, and the query image is the first one in each category.

## 3. Experiment and Results

In this section, experimental results of the proposed method on COIL-20 [7] image database and Corel-1000 [10] database are presented. The COIL-20 database involves 1440 images, in which the images are separated into twenty categories (see Figure 4), and every category includes 72 images. The size of each image is $128\times 128$. Table 1 shows the results of enhanced-surf method and our method, respectively, on COIL-20 database. ${O}_{R}$ is the number of relevant images retrieved, while ${O}_{T}$ is the total number of images retrieved. Thus, the precision and the recall of this retrieval system are given by Equation(11).

$$Precision=\frac{{O}_{R}}{{O}_{T}},\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}Recall=\frac{{O}_{R}}{72}.$$

According to the results in Table 1b, we can get the average precision is 0.4818 and the average recall is 0.8387 by calculating the mean value of the third and fourth line, respectively. Compared with the enhanced SURF, these results demonstrate that the average precision and the average recall of our method increase by $16.77\%$ and $28.61\%$, respectively.

The Corel-1000 database includes 1000 images from ten categories (African tribes, Beaches, Buildings, Buses, Dinosaurs, Elephants, Flowers, Horses, Mountains, and Foods) (see Figure 5).

Each category contains 100 images with size of $384\times 256$ or $256\times 384$. Each image in this database is taken to query. Table 2 shows the precision and recall results of our proposed method compared with CLD [27], Color Moment [28] and CSD [29].

Precision is computed from the Top-10 retrieved images and recall is calculated from the Top-100 retrieved images. CLD takes spatial information into account, but fails to find rotated or translated images that are similar to query images. Color moment has nine color components, with three low-order moments for each color channel, but low-order moments are less discriminating, so it is often used in combination with other features. CSD is mainly used for static image retrieval, and only considers whether the structure contains a certain color, and not about its frequency.

Table 2 shows that the average precision of our experimental results are $17.4\%$, $16.1\%$ and $5.6\%$ higher than CLD, color moment and CSD, respectively; the average recall of our experimental results are $9.4\%$, $5.5\%$ and $2.5\%$ higher than CLD, color moment and CSD, respectively; and our method retrieves buildings category images, the elephants category images and foods category images better than the other algorithms. For example, buildings images can be more precisely retrieved by using our method, and it can be seen that the average precision of our experimental results are $32.3\%$, $28.2\%$ and $11.1\%$ higher than CLD, color moment and CSD, respectively; and the average recall of our experimental results are $24\%$, $18.6\%$ and $5.6\%$ higher than CLD, color moment and CSD, respectively. The precision of dinosaurs category and flower category are highly retrieved in our proposed algorithm. Table 3 shows the precision rate of retrieving the Top-20 database images for each query image, while some retrieved images in the dinosaurs category, horses category and foods category are shown in Figure 6.

## 4. Conclusions

This paper provides a novel image similarity measurement method, which uses the improved dominance granule structure formula as a similarity calculation method of proximity sets. By simulating public datasets and comparing with other methods (CLD, Color Moment and CSD), it was found that the image retrieval based on this method produces a better retrieval result. The proposed method is meaningful to improve the retrieval accuracy, and provides a new attempt in the field of image retrieval.

## Author Contributions

Conceptualization, J.W. and L.W.; Methodology, J.W.; Software, J.W.; Validation, J.W. and L.W.; Formal Analysis, J.W.; Writing-Original Draft Preparation, J.W.; Writing-Review & Editing, J.W., L.W., X.L. and Y.R.; Visualization, J.W.; Supervision, L.W., X.L., Y.R. and Y.Y.

## Funding

This work was partially supported by Natural Science Foundations of China under Grant Nos. 61602321, 61773352, Aviation Science Foundation with No. 2017ZC54007, Science Fund Project of Liaoning Province Education Department with No. L201614, Natural Science Fund Project of Liaoning Province with No. 20170540694, the Fundamental Research Funds for the Central Universities (Nos. 3132018225, 3132018228) and the Doctor Startup Foundation of Shenyang Aerospace University13YB11.

## Conflicts of Interest

The authors declare no conflict of interest.

## References

- Amores, J.; Sebe, N.; Radeva, P.; Gevers, T.; Smeulders, A. Boosting Contextual Information in Content-Based Image Retrieval. In Proceedings of the 6th ACM SIGMM International Workshop on Multimedia Information Retrieval, New York, NY, USA, 15–16 October 2004; pp. 31–38. [Google Scholar]
- Zhang, Y.P.; Zhang, S.B.; Yan, Y.T. A Multi-view Fusion Method for Image Retrieval. In Proceedings of the International Congress on Image and Signal Processing, BioMedical Engineering and Information, Shanghai, China, 14–16 October 2017; pp. 379–383. [Google Scholar]
- Chen, Y.Y. The Image Retrieval Algorithm Based on Color Feature. In Proceedings of the IEEE International Conference on Software Engineering and Service Science, Beijing, China, 26–28 August 2016; pp. 647–650. [Google Scholar]
- Liang, C.W.; Chung, W.Y. Color Feature Extraction and Selection for Image Retrieval. In Proceedings of the IEEE International Conference on Advanced Materials for Science and Engineering, Tainan, Taiwan, 12–13 November 2016; pp. 589–592. [Google Scholar]
- Li, M.Z.; Jiang, L.H. An Improved Algorithm Based on Color Feature Extraction for Image Retrieval. In Proceedings of the International Conference on Intelligent Human-Machine Systems and Cybernetics, Hangzhou, China, 25–26 August 2016. [Google Scholar]
- Alhassan, A.K.; Alfaki, A.A. Color and Texture Fusion-Based Method for Content-Based Image Retrieval. In Proceedings of the International Conference on Communication, Control, Computing and Electronics, Khartoum, Sudan, 16–18 Janary 2017; pp. 1–6. [Google Scholar]
- Gopal, N.; Bhooshan, R.S. Content Based Image Reterieval using Enhanced SURF. In Proceedings of the Fifth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics, Patna, India, 16–19 December 2015; pp. 1–4. [Google Scholar]
- Bay, H.; Ess, A.; Tuytelaars, T.; Gool, L.V. Speeded-Up Robust Features (SURF). Comput. Vis. Image Underst.
**2008**, 110, 346–359. [Google Scholar] [CrossRef] - Kaur, M.; Sohi, N. A Novel Techinique for Content Based Image Retrieval Using Color, Texture and Edge Features. In Proceedings of the International Conference on Communication and Electronics System, Coimbatore, India, 21–22 October 2016; pp. 1–7. [Google Scholar]
- Chen, Q.S.; Ding, Y.Y.; Li, H.; Wang, X.; Wang, J.; Deng, X. A Novel Multi-Feature Fusion and Sparse Coding Based Framework for Image Retrieval. In Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics Octobor, San Diego, CA, USA, 5–8 October 2014; pp. 2391–2396. [Google Scholar]
- Kang, L.W.; Hsu, C.Y.; Chen, H.W.; Lu, C.S.; Lin, C.Y. Feature-Based Sparse Representation for Image Similarity Assessment. IEEE Trans. Multimed.
**2011**, 13, 1019–1030. [Google Scholar] [CrossRef] - Bakhshali, M.A. Segmentation and Enhancement of Brain MR Images using Fuzzy Clustering Based on Information Theory. Soft Comput.
**2016**, 21, 6633–6640. [Google Scholar] [CrossRef] - Pawlak, Z. Rough sets, Proceedings of the 6th ACMSIGMM. Int. J. Comput. Inform. Sci.
**1995**, 256–341. [Google Scholar] - Lim, K.Y.; Rajeswari, M. Segmenting Object with Ambiguous Boundary using Information Theoretic Rough Sets. Int. J. Electron. Commun.
**2017**, 77, 50–56. [Google Scholar] [CrossRef] - Senthilkumaran, N.; Rajesh, R. Brain Image Segmentation using Granular Rough Sets. Int. J. Arts Sci.
**2009**, 3, 69–78. [Google Scholar] - Roy, P.; Goswami, S.; Chakraborty, S.; Azar, A.T.; Dey, N. Image Segmentation using Rough Set Theory: A Review. Int. J. Rough Sets Data Analy.
**2014**, 1, 62–74. [Google Scholar] [CrossRef] - Khodaskar, A.A.; Ladhake, S.A. Emphasis on Rough Set Theory for Image Retrieval. In Proceedings of the 6th ACM SIGMM International Conference on Computer Communication and Informatics, Coimbatore, India, 8–10 Janary 2015; pp. 1–6. [Google Scholar]
- Thilagavathy, C.; Rajesh, R. Rough Set Theory and its Applications Image Processing-A Short Review. In Proceedings of the 3rd World Conference on Applied Sciences, Engineering, Technology, Kathmandu, Nepal, 27–29 September 2014; pp. 27–29. [Google Scholar]
- Rocio, L.M.; Raul, S.Y.; Victor, A.R.; Alberto, P.R. Improving a Rough Set Theory-Based Segmentation Approach using Adaptable Threshold Selection and Perceptual Color Spaces. J. Electron. Imag.
**2014**, 23, 013024. [Google Scholar] - Wasinphongwanit, P.; Phokharatkul, P. Image Retrieval using Contour Feature with Rough Set Method. In Proceedings of the International Conference on Computer, Mechatronics, Control and Electronic Engineering, Changchun, China, 24–26 August 2010; Volume 6, pp. 349–352. [Google Scholar]
- Peters, J.F. Computational Proximity—Excursions in the Topology of Digital Images; Springe: Berlin, Germany, 2016. [Google Scholar]
- Pták, P.; Kropatsch, W.G. Nearness in Digital Images and Proximity Spaces. In Proceedings of the 9th International Conference on Discrete Geometry, Nantes, France, 18–20 Apirl 2016; pp. 69–77. [Google Scholar]
- Peters, J.F. Visibility in Proximal Delaunay Meshes and Strongerly Near Wallman Proximity. Adv. Math.
**2015**, 4, 41–47. [Google Scholar] - Ma, J. Content-Based Image Retrieval with HSV Color Space and Texture Features. In Proceedings of the International Conference on IEEE Web Information Systems and Mining, Shanghai, China, 7–8 November 2009; pp. 61–63. [Google Scholar]
- Yan, Y.; Ren, J.; Li, Y.; Windmill, J.; Ijomah, W. Fusion of Dominant Colour and Spatial Layout Features for Effective Image Retrieval of Coloured Logos and Trademarks. In Proceedings of the IEEE International Conference on Multimedia Big Data, Beijing, China, 20–22 April 2015; Volume 106, pp. 306–311. [Google Scholar]
- Schroder, B.S.W. Ordered Sets: An Introduction; Springer: Berlin, Germany, 2002. [Google Scholar]
- Messing, D.S.; Beek, P.V.; Errico, J.H. The MPEG-7 Colour Struture Descriptor: Image Description using Colour and Local Spatial Information. In Proceedings of the 2001 International Conference on Image Processing, Thessaloniki, Greece, 7–10 October 2001; pp. 670–673. [Google Scholar]
- Shih, J.L.; Chen, L.H. Colour Image Retrieval Based on Primitives of Colour Moments. IEE Proc. Vis. Image Signal Process.
**2002**, 149, 370–376. [Google Scholar] - Kasutani, E.; Yamada, A. The MPEG-7 Color Layout Descriptor: A Compact Image Feature Description for High-Speed ImageNideo Segment Retrieval. In Proceedings of the International Conference on Image Processing, Thessaloniki, Greece, 7–10 October 2001; pp. 674–677. [Google Scholar]

**Figure 3.**The left image shows car category results, while the right image shows cup category results.

**Figure 6.**Retrieved images for the Top 20: the dinosaurs category (

**a**); (

**b**) the horses category; and (

**c**) the foods category. The query image is the first image in each class.

**Table 1.**Comparisons of precision and recall results among enhanced SURF method and our Proposed Method.

(a) Note: The Results of Enhanced SURF Method. | ||||||||||||||||||||

Object category | a | b | c | d | e | f | g | h | i | j | k | l | m | n | o | p | q | r | s | t |

${O}_{R}$ | 65 | 45 | 24 | 43 | 25 | 65 | 13 | 30 | 54 | 17 | 42 | 49 | 60 | 34 | 21 | 46 | 62 | 35 | 38 | 28 |

${O}_{T}$ | 238 | 215 | 56 | 92 | 268 | 232 | 31 | 31 | 1087 | 168 | 91 | 762 | 1105 | 57 | 96 | 417 | 184 | 97 | 89 | 77 |

Precision | 0.273 | 0.209 | 0.428 | 0.467 | 0.093 | 0.280 | 0.419 | 0.968 | 0.05 | 0.101 | 0.462 | 0.064 | 0.054 | 0.596 | 0.219 | 0.110 | 0.337 | 0.361 | 0.427 | 0.364 |

Recall | 0.903 | 0.625 | 0.333 | 0.597 | 0.347 | 0.902 | 0.180 | 0.417 | 0.75 | 0.236 | 0.583 | 0.680 | 0.833 | 0.472 | 0.292 | 0.639 | 0.861 | 0.486 | 0.528 | 0.389 |

(b) Note: The Results of Our Proposed Method. | ||||||||||||||||||||

Object category | a | b | c | d | e | f | g | h | i | j | k | l | m | n | o | p | q | r | s | t |

${O}_{R}$ | 72 | 53 | 55 | 47 | 59 | 72 | 33 | 24 | 72 | 68 | 70 | 72 | 72 | 51 | 59 | 72 | 72 | 65 | 68 | 50 |

${O}_{T}$ | 247 | 219 | 58 | 95 | 272 | 242 | 33 | 31 | 1116 | 176 | 96 | 788 | 1197 | 59 | 109 | 453 | 194 | 100 | 89 | 77 |

Precision | 0.291 | 0.256 | 0.948 | 0.494 | 0.202 | 0.298 | 1 | 0.774 | 0.065 | 0.375 | 0.729 | 0.091 | 0.060 | 0.881 | 0.541 | 0.159 | 0.371 | 0.680 | 0.772 | 0.649 |

Recall | 1 | 0.778 | 0.763 | 0.652 | 0.764 | 1 | 0.458 | 0.333 | 1 | 0.917 | 0.972 | 1 | 1 | 0.722 | 0.819 | 1 | 1 | 0.944 | 0.958 | 0.694 |

**Table 2.**Comparisons of precision and recall results among CLD, Color Moment, CSD and our Proposed Method.

Category Name | CLD [27] | Color Moment [28] | CSD [29] | Proposed Method | ||||
---|---|---|---|---|---|---|---|---|

Precision | Recall | Precision | Recall | Precision | Recall | Precision | Recall | |

African Tribes | 0.213 | 0.144 | 0.389 | 0.260 | 0.562 | 0.347 | 0.505 | 0.312 |

Beaches | 0.529 | 0.283 | 0.295 | 0.169 | 0.288 | 0.153 | 0.442 | 0.209 |

Buildings | 0.341 | 0.157 | 0.382 | 0.211 | 0.553 | 0.341 | 0.664 | 0.397 |

Buses | 0.255 | 0.113 | 0.580 | 0.319 | 0.447 | 0.258 | 0.703 | 0.435 |

Dinosaurs | 0.907 | 0.510 | 0.938 | 0.785 | 0.794 | 0.614 | 0.962 | 0.539 |

Elephants | 0.477 | 0.235 | 0.491 | 0.302 | 0.630 | 0.260 | 0.706 | 0.339 |

Flowers | 0.700 | 0.288 | 0.679 | 0.360 | 0.537 | 0.253 | 0.733 | 0.300 |

Horses | 0.980 | 0.703 | 0.747 | 0.401 | 0.641 | 0.338 | 0.885 | 0.549 |

Mountains | 0.566 | 0.357 | 0.260 | 0.157 | 0.613 | 0.294 | 0.441 | 0.229 |

Foods | 0.127 | 0.057 | 0.463 | 0.277 | 0.320 | 0.138 | 0.792 | 0.482 |

Average Precision | 0.509 | 0.285 | 0.522 | 0.324 | 0.627 | 0.354 | 0.683 | 0.379 |

Query Image | Precision | Query Image | Precision |
---|---|---|---|

African Tribes | 46% | Elephants | 59% |

Beaches | 34% | Flowers | 60% |

Buildings | 60% | Horses | 83% |

Buses | 63% | Mountains | 36% |

Dinosaurs | 92% | Foods | 72% |

© 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).