Next Article in Journal
Single Image Super-Resolution Based on Global Dense Feature Fusion Convolutional Network
Next Article in Special Issue
Pose Estimation for Straight Wing Aircraft Based on Consistent Line Clustering and Planes Intersection
Previous Article in Journal
A Handy Flexible Micro-Thermocouple Using Low-Melting-Point Metal Alloys
Previous Article in Special Issue
Camera Calibration Using Gray Code
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Local Parallel Cross Pattern: A Color Texture Descriptor for Image Retrieval

1
College of Information Science and Engineering, Northeastern University, Shenyang 110004, China
2
College of Intelligence and Computing, Tianjin University, Tianjin 300350, China
3
Institute of Informatics and Applications, University of Girona, 17017 Girona, Spain
4
School of Software, Jiangxi Normal University, Nanchang 330022, China
5
School of Computer Engineering, Weifang University, Weifang 261061, China
*
Authors to whom correspondence should be addressed.
Sensors 2019, 19(2), 315; https://doi.org/10.3390/s19020315
Submission received: 2 December 2018 / Revised: 11 January 2019 / Accepted: 11 January 2019 / Published: 14 January 2019
(This article belongs to the Special Issue Visual Sensors)

Abstract

:
Riding the wave of visual sensor equipment (e.g., personal smartphones, home security cameras, vehicle cameras, and camcorders), image retrieval (IR) technology has received increasing attention due to its potential applications in e-commerce, visual surveillance, and intelligent traffic. However, determining how to design an effective feature descriptor has been proven to be the main bottleneck for retrieving a set of images of interest. In this paper, we first construct a six-layer color quantizer to extract a color map. Then, motivated by the human visual system, we design a local parallel cross pattern (LPCP) in which the local binary pattern (LBP) map is amalgamated with the color map in “parallel” and “cross” manners. Finally, to reduce the computational complexity and improve the robustness to image rotation, the LPCP is extended to the uniform local parallel cross pattern (ULPCP) and the rotation-invariant local parallel cross pattern (RILPCP), respectively. Extensive experiments are performed on eight benchmark datasets. The experimental results validate the effectiveness, efficiency, robustness, and computational complexity of the proposed descriptors against eight state-of-the-art color texture descriptors to produce an in-depth comparison. Additionally, compared with a series of Convolutional Neural Network (CNN)-based models, the proposed descriptors still achieve competitive results.

1. Introduction

Since a huge number of image corpora have been produced by visual sensor equipment, an increasing demand for efficient encoding and indexing of these image corpora has attracted the attention of a considerable number of researchers [1,2,3,4,5,6,7]. Thanks to the investigators’ breakthroughs, a myriad of methods [8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28] have continuously been developed for encoding and indexing.
In early work, the local binary pattern (LBP) [8], a grayscale texture descriptor, was first proposed for encoding the center pixel and its neighborhood pixels. Afterwards, owing to the disadvantage of losing global information, the LBP was extended to the LBP variance (LBPV) [9] which was amalgamated with global rotation-invariant matching for texture classification. Further, in order to completely detail the local differences among the central pixel and its neighborhood pixels, the completed local binary pattern (CLBP) operator [10] was designed for rotation-invariant feature representation. The local derivative pattern (LDP) [11] was then produced by refining the magnitude difference in local neighborhoods. Along another line, taking into account the situation of non-uniform lighting conditions, the local ternary pattern (LTP) [12] was introduced, and it was combined with kernel principal component analysis (KPCA) to improve its robustness to illumination. Later, the LTP was further modified into the local tetra pattern (LTrP) [13] by using the first-order derivatives in the vertical and horizontal directions. After that, the LTP was again extended to the local maximum edge binary patterns (LMEBP) [14], and the LMEBP were combined with the Gabor transform used for image retrieval and object tracking. Besides these, to achieve robustness to uniform and Gaussian noises, the noise-resistant LBP (NRLBP) [15] was constructed to preserve local structure information. Inspired by the fusion strategy, the local neighborhood difference pattern (LNDP) [16] was concatenated with the LBP map to integrate the local intensity difference information and the local binary information in a parallel manner. However, all the above methods are confined within grayscale image processing, so the major drawback of these construction processes is the inevitable loss of color information.
In recent years, a series of color texture descriptors [17,18,19,20,21,22,23,24,25,26,27,28,29] have been sequentially developed for color image processing. Among them, the local oppugnant color texture pattern (LOCTP) [17], a variant of the multi component LTrP, was combined with the colored pattern appearance model (CPAM) in the YCbCr, HSV, and RGB color spaces. In reference [18], according to the adder and decoder concepts, the multi-channel adder local binary pattern (maLBP) and the multi-channel decoder local binary pattern (mdLBP) were designed to combine the LBP maps in the R, G, and B components. After that, a class of pairwise-based local binary patterns [19,20] and a series of color-edge approaches [21,22,23] were introduced to classify and retrieve natural color images. Recently, on the basis of intra- and inter-channel encoding concepts, the opponent color local binary patterns (OCLBP) were proposed by Mäenpää et al. [24]. Further, in reference [25], Bianconi et al. extended the OCLBP to the improved opponent color local binary patterns (IOCLBP), in which the point-to-point thresholding was replaced by point-to-average thresholding. Considering the graph-based fusion framework, the bag-of-words of local features and the color local Haar binary pattern were integrated by Li et al. [26]. Quite recently, in order to systematically analyze the robustness to illumination changes, a bag of color texture descriptors [27] was studied under 46 lighting conditions. At the same time, with the help of a non-linear support vector machine, the orthogonal combination of local binary patterns (OC-LBP) was concatenated with the color histogram (CH) [28].
In this paper, we present the main following contributions:
  • We design a six-layer color quantizer that is applied to quantize the a* and b* components for color map extraction.
  • We construct a local parallel cross pattern (LPCP) in which the LBP map and the color map are integrated into a whole framework.
  • We further extend the LPCP to the uniform local parallel cross pattern (ULPCP) and the rotation-invariant local parallel cross pattern (RILPCP) to reduce the computational complexity and achieve robustness to image rotation.
  • We benchmark the comparative experiments with eight state-of-the-art color texture descriptors on eight benchmark datasets to illustrate the effectiveness, efficiency, robustness, and computational complexity of the proposed descriptors.
  • We additionally develop a weight-based optimization scheme that shows better improvement.
The rest of this paper is organized as follows. Section 2 briefly introduces the local binary pattern and the color distribution prior in the L*a*b* color space. Section 3 details the feature representation. The experiments and discussion are presented in Section 4. Section 5 concludes this paper and indicates future directions.

2. Related Work

2.1. Local Binary Pattern

In reference [8], Ojala et al. first designed the local binary pattern (LBP) for texture feature representation. Given a gray pixel G(i, j), the computational results among G(i, j) and its neighbors Gx(i, j) are encoded as the LBP value. The formula for the LBP value in pixel G(i, j) is as follows:
L B P n , r ( i , j ) = x = 0 n 1 ϑ ( G ( i , j ) G x ( i , j ) ) × 2 x ,
ϑ ( t ) = { 1 , t 0 0 , t < 0 ,
where Gx(i, j) is a pixel G(i, j)’s x-th neighbor, and n and r represent the number of neighbors and the radius of the neighborhood, respectively.
To reduce the computational complexity, Ojala et al. [8] defined the uniform LBP, in which each LBP pattern has, at most, two bitwise changes among its neighbors. The measure operator “U” of an LBP pattern is defined as the number of bitwise changes. Mathematically, the uniform LBP is defined as follows:
U ( L B P n , r ( i , j ) ) = | ϑ ( G ( i , j ) G 0 ( i , j ) ) ϑ ( G ( i , j ) G n 1 ( i , j ) ) | + x = 1 n 1 | ϑ ( G ( i , j ) G x ( i , j ) ) ϑ ( G ( i , j ) G x 1 ( i , j ) ) | .
If U(LBPn,r(i,j)) ≤ 2, then LBPn,r(i,j) is classified as the uniform LBP pattern; otherwise, LBPn,r(i,j) belongs to the non-uniform LBP pattern.
To achieve robustness to image rotation, Pietikäinen et al. [29] designed the rotation-invariant LBP, in which all types of the same transition were considered as one pattern. The measure operator ROR (LBPn,r(i, j), x) is defined as a circular bitwise right shift for x times on the n-bit number LBPn,r(i, j). Mathematically, the rotation-invariant LBP is expressed as follows:
L B P n , r r i ( i , j ) = min { ROR ( L B P n , r ( i ,   j ) ,   x ) | x 0 , 1 , , n 1 } .
For details, please refer to references [8,29,30]. For simplicity, referring to reference [19], n and r were set to 8 and 1, respectively. In the rest of this paper, we refer to the LBP map as LBP(i, j), the uniform LBP map as ULBP(i, j), and the rotation-invariant LBP map as RILBP(i, j).

2.2. The Selection of the Color Space

The selection of the color space is acknowledged as an important preprocessing stage [1]. Currently, RGB, HSV, HIS, CMYK, YUV, and L*a*b* are widely adopted in feature representation. Among these, the most commonly used color space is RGB. However, the inferiority of the RGB color space can be summarized as follows: (1) the yellow is lost; (2) there is a plethora from green to blue; and (3) it is not suited for the visual perception mechanism. Different from RGB, the superiority of the L*a*b* color space can be summarized as follows: (1) the L*a*b* remedies the missing yellow in RGB; (2) there is no plethora from green to blue; and (3) it is suited for the visual perception mechanism. Besides this, L*a*b* provides excellent decoupling between color (represented by the a* and b* components) and intensity (represented by the L* component) [31]. Therefore, all images are transformed from RGB to the L*a*b* color space in the preprocessing stage. Referring to reference [32], the standard RGB to L*a*b* transformation is carried out as follows:
{ L * = 116 ( Y Y n ) 1 / 3 16 for   Y Y n > 0.08856 L * = 903.3 ( Y Y n ) 1 / 3 for   Y Y n 0.08856 ,
a * = 500 ( f ( X X n ) f ( Y Y n ) ) ,
b * = 500 ( f ( X X n ) f ( Y Z n ) ) ,
with
{ f ( u ) = u 1 / 3 for   u > 0.08856 f ( u ) = 7.78 u + Y Y n for   u 0.08856 ,
where
[ X Y Z ] = [ 0.412453 0.357580 0.180423 0.212671 0.715160 0.072169 0.019334 0.119193 0.950227 ] [ R G B ] ,
where Xn, Yn, and Zn are set to 0.950450, 1.000000, and 1.088754.

2.3. Color Distribution Prior Knowledge in the L*a*b* Color Space

In reference [1], the color distribution prior knowledge in the L*a*b* color space was analyzed and summarized for different color image sets. In Figure 1a,b, an example of the Stex database [33] is illustrated. It can be seen that the frequency of pixels is mainly concentrated in the middle range of the a* and b* components. To validate the consistency of this prior knowledge, extensive experiments were performed on different color image sets, and these results show that the prior knowledge is consistent.
Further, the stability of this prior knowledge was also studied when the image database was changed. Examples of 50% and 10% of the Stex image datasets are presented in Figure 1c–f. From those figures, except for the frequency of pixels, we can easily see that the pixels are still distributed in the middle range of the a* and b* components. These phenomena illustrate that the prior knowledge is stable.
The reason for this was studied more deeply in reference [20]. As depicted in Figure 1a–f, the frequency of pixels in the a* and b* components gradually declines from the middle to both sides, but the saturation in the a* and b* components inversely goes up from the middle to both sides. Through extensive experiments, we propose that the color probability distribution has a negative correlation with the saturation of the a* and b* components because higher saturation occurs with a lower frequency.

3. Feature Representation

3.1. Six-Layer Color Quantizer

Inspired by the color distribution prior knowledge in the L*a*b* color space, a novel six-layer color quantizer was designed, as shown in Figure 2, in which each layer includes a set of bins and its corresponding indices. In the proposed quantizer, the original range [−128, +127] is first divided into two equal bins, 28/3, on both sides and two refined bins, 27/3, in the middle. Sequentially, the indices are named 0, 1, 2, and 3 at layer 1. Second, to further refine the two middle bins, 27/3, they are uniformly divided into four equal bins, 26/3, from layers 1 to 2. Meanwhile, the remaining bins are copied from the layer 1 to 2. Third, the operators “Copy” and “Divide” are continuously repeated until the two middle bins at the layer 6. Finally, combining the layer 1 to 6, we construct a six-layer color quantizer. The quantization layers in the a* and b* components are denoted Wa* and Wb*, where Wa*, Wb* {1, 2, …, 6}, and the indices are denoted Ŵa* and Ŵb*, Ŵa* {0, 1, …, a*} and Ŵb* {0, 1, …, b*}, where a* = 2(Wa* + 1) – 1 and b* = 2(Wb* + 1) – 1 respectively.
Referring to reference [34], we discuss the quantization error under different quantization layers, Wa* and Wb*. Figure 3 shows the quantization errors on Stex, 50% of Stex, and 10% of Stex. From the figures, it can be seen that along with the refinement of layers Wa* and Wb*, where Wa*, Wb* {1, 2, …, 6}, the quantization error decreases obviously. This phenomenon illustrates the effectiveness of the proposed quantizer. Moreover, we also note that the values of the quantization errors on Stex, 50% of Stex, and 10% of Stex are extremely close to each other. This phenomenon confirms the stability and consistency of the six-layer color quantizer. In particular, the quantization error from layers 5 to 6 decreases only slightly on the three datasets. These results demonstrate that stopping the quantization layers Wa* and Wb* at the 6th layer is appropriate.
Furthermore, referring to the visual perception mechanism in reference [35], the original range [0, +100] in the L* component is divided into three bins, namely, [0, +25], [+26, +75], and [+76, +100]. Herein, the quantization layer of the L* component is denoted WL*, where WL* = 1, and the indices of the three bins are denoted ŴL*, ŴL* {0, 1, …, L*}, where L* = 2WL*. For a pixel (i, j) in image I, we combine the indices of WL*, Wa*, and Wb* to construct the color map C(i, j), and the index of C(i, j) is denoted Ĉ, Ĉ {0, 1, …, Č}, where Č = 3 × 2(Wa* + 1) × 2(Wb* + 1) − 1.

3.2. Local Parallel Cross Pattern

As elucidated in Gray’s Anatomy [36], the human visual system is an important pathway that codes low-layer visual cues to construct the high-layer semantics perception in parallel and cross manners. On the basis of the human visual system, we propose a novel local parallel cross pattern (LPCP) to integrate the color map and the LBP map as a unified framework in “parallel“ and “cross“ manners.
Given an original map I(i, j), the central point and its eight neighbors are denoted I0(i, j) and Ik(i, j), where k {1, 2, …, 8}. Firstly, we extract the LBP map LBP(i, j) (see Section 2.1) and the color map C(i, j) ( see Section 3.1). Secondly, all eight neighbors of the LBP map and the color map are mutually crossed to construct the LBP and color cross maps. Thirdly, we calculate the frequency of each neighborhood in the LBP and color cross maps to construct the LBP and color frequency maps, respectively. Finally, the values of the maximum frequency in the LBP and color frequency maps are considered the feature vectors, and the central values in the LBP and color frequency maps are flagged as the indices. For clarity, Figure 4 presents a detailed schematic diagram of the local parallel cross pattern, in which LPCP is encoded as LPCPLBP(3) = 4 and LPCPcolor(7) = 5. Mathematically, LPCP is defined as follows:
L P C P L B P ( L B P 0 ( i , j ) ) = arg   max   Fr { C k ( i , j ) | k = 1 ,   2 ,   ,   8 } ,
L P C P c o l o r ( C 0 ( i , j ) ) = arg   max   Fr { L B P k ( i , j ) | k = 1 ,   2 ,   ,   8 } ,
where Fr{·} represents the frequency of each neighbor. In particular, when the value of LBP0(i, j) is equal to C0(i, j), LBP0(i, j) and C0(i, j) can still be separated and encoded as LPCPLBP(LBP0(i, j)) and LPCPcolor(C0(i, j)), respectively. Herein, the feature dimensions of LPCPLBP and LPCPcolor are 256 and 3 × 2(Wa* + 1) × 2(Wb* + 1). Given a color image dataset T, the optimal quantization layers Wa* and Wb* are calculated based upon the retrieval accuracy score. This process is defined as the maximization problem as follows:
max W a * , W b * Acc ( T | W a * , W b * ) , W a * ,   W b * { 1 , 2 , , 6 } ,
where Acc(T|Wa*, Wb*) represents the retrieval accuracy score. We provide the optimal color quantization layers Wa* and Wb* in Section 4.4.
To reduce the computational complexity, we define the uniform local parallel cross pattern (ULPCP) in which ULBP(i, j) replaces LBP(i, j). To achieve robustness to image rotation, we design the rotation-invariant local parallel cross pattern (RILPCP) in which RILBP(i, j) replaces LBP(i, j).
As presented in Figure 5, RILPCP achieves exactly the same feature encoding in spite of the image rotation. For Figure 5a, we first extract the RILBP map and the color map. Then, according to Equations (10) and (11), RILPCP is encoded as [RILPCPRILBP(3) = 4; RILPCPcolor(7) = 4]. When Figure 5a is rotated 90° to Figure 5b, we can also extract the same RILBP and color map because both color and RILBP are rotation invariant. Further, RILPCP is still calculated as [RILPCPRILBP(3) = 4; RILPCPcolor(7) = 4]. In fact, an image can be rotated by an arbitrary degree.

4. Experiments and Discussion

4.1. Distance Metric

The distance metric is considered as the measure of similarity between a query image and a database image. In our retrieval framework, we first convert the query image and database images into feature vectors, and we then calculate the distance measure between the query image and database images. Referring to references [1,18,20,21,37], the Extended Canberra Distance is exploited, and it is given as
R ( F q , F d ) = ν = 1 k | F q ( ν ) F d ( ν ) | | F q ( ν ) + μ q | + | F d ( ν ) + μ d | ,
where Fd, Fq, and k respectively denote the image in the database, the query image by the user, and the feature vector length, and R is the measurement result between the query image Fq and the image Fd in the database. Further, μ q = ν = 1 k F q ( ν ) / k and μ d = ν = 1 k F d ( ν ) / k .

4.2. Evaluation Criteria

In this section, we introduce the most popular evaluation criteria, such as the precision rate, the recall rate, the average precision rate (APR) value, the average recall rate (ARR) value, and the precision-recall (PR) curve to validate the proposed descriptors.
First, the precision and recall rates are formulated as follows:
P r e c i s i o n = 1 N α d = 1 N α ξ ( η ( F q ) , η ( F d ) )   ,
R e c a l l = 1 N β d = 1 N β ξ ( η ( F q ) , η ( F d ) )   ,
ξ ( η ( F q ) , η ( F d ) ) = { 1 , η ( F q ) = η ( F d ) 0 , η ( F q ) η ( F d ) ,
where N β , N α , and η ( · ) denote the total number of images in the same category, the total number of returned images, and the category information. ξ ( · ) is defined as the binarized function. If ξ(η(Fq), η(Fd))) = 1, then η(Fq) and η(Fd) are determined to be in the same category; if ξ(η(Fq), η(Fd))) = 0, then η(Fq) and η(Fd) do not belong to the same category.
Then, the average precision rate (APR) and average recall rate (ARR) values are given by the following equations:
A P R = n ^ = 1 N ^ P r e c i s i o n ( n ^ ) N ^   ,
A R R = n ^ = 1 N ^ R e c a l l ( n ^ ) N ^   ,
where n ^ and N ^ are the total number of query images and the n ^ th query image, respectively.
Finally, the precision-recall (PR) curve can be considered an ancillary criterion that evaluates the dynamic precision by computing the threshold recall. Mathematically, the PR curve is defined as follows:
P R ( τ ) = N β N τ τ ×   100 % ,
where Nβ and Nτ are the total number of images in the same category and the number of retrieval images at the recall of τ , where τ {1, 2, …, Nα − 1}.

4.3. Image Databases

In our experiments, the eight benchmark databases reported in Table 1, comprising one natural object image database (Coil-100 [38]), two facial image databases (Face95 [39] and Face96 [40]), and five color textural image databases (Outex-00031 [41], Outex-00032 [41], Outex-00033 [41], Outex-00034 [41], and MIT-VisTex [42]), were used to provide a comprehensive evaluation.
The Coil-100 (No. 1) database was produced by a charge-coupled device (CCD)-camera (Sony XC-77P). It has 7200 images in 100 objects. Each object contains 72 images, each with a size of 128 × 128 in JPG format. Because each image was collected by rotating an object at 5 degrees, the Coil-100 not only evaluates the effectiveness, but also investigates the robustness to rotation. Some samples are depicted in Figure 6a in which each row represents the same semantic category.
The Face 95 (No. 2) and Face 96 (No. 3) databases were collected by an S-VHS camcorder. Among them, the Face 95 consists of 1440 images in 72 male and female subjects. For each subject, there are 20 images, each with a size of 180 × 200 in JPG format. The Face 96 has 91 male and female subjects. For each subject, there are 1814 images with a size of 196 × 196 in JPG format. Specifically, all images are variations of head turns, head scales, face expressions, and illumination changes. Some samples are presented in Figure 6b,c.
The Outex-00031 (No. 4), Outex-00032 (No. 5), Outex-00033 (No. 6), and Outex-00034 (No. 7) databases were produced by a charge-coupled device (CCD)-camera (Sony DXC-775P), and the MIT-VisTex (No. 8) was collected from real-world photographs and videos. Some samples of these databases are presented in Figure 6d–h. As documented in Table 1, the Outex-00031, Outex-00032, and Outex-00033 consist of 2720 images in 68 categories. Each category includes 40 images, each with a size of 128 × 128 in BMP format. Differently, the Outex-00034 has 4048 images in 204 categories. Each category includes 20 images, each with a size of 128 × 128 in BMP format. Next, the MIT-VisTex consists of 640 images in 40 categories. Each category includes 16 images, each with a size of 128 × 128 in PPM format. There are resolution differences on Outex-00031, noise differences on Outex-00032, blur differences on Outex-00033, and illumination differences on Outex-00034. Thus, these four databases were used to evaluate the robustness to resolution, noise, blur, and illumination.
In addition, all above databases can be freely downloaded from the corresponding websites. To guarantee accuracy and reproducibility, we chose all images as the query images in the dataset. Referring to references [1,18,20,37], if not specified, the total number of returned images was set to 10 in all of the following experiments.

4.4. Evaluation of Color Quantization Layers

Table 2 shows the highest APR values of LPCP, RILPCP, and ULPCP with the optimal color quantization layers (Wa*, Wb*) on the eight databases. The best values are shown in bold. On Coil-100, three facts are noted: (1) LPCP achieves the highest APR value of 99.47% when (Wa* = 6, Wb* = 5); (2) RILPCP achieves the highest APR value of 99.44% when (Wa* = 5, Wb* = 5); and (3) ULPCP yields the highest APR value of 99.52% when (Wa* = 4, Wb* = 6). Next, on Face 95 and Face 96, LPCP, RILPCP, and ULPCP all achieve their top APR values when (Wa* = 6, Wb* = 6). Similarly, on the remaining color textural databases, the corresponding color quantization layers Wa* and Wb* result in the best accuracy scores. As noted above, it can be easily summarized that when LPCP, RILPCP, and ULPCP are used on the same image database, the coefficients Wa* and Wb* are extremely close to each other. These phenomena again confirm the stability and consistency of the color distribution prior knowledge. Moreover, it is noteworthy that a single color quantization layer cannot be suitable for all image databases. In the following experiments, the optimal color quantization layers, Wa* and Wb*, were adaptively selected according to the different databases.

4.5. Comparison with LBP-Based Descriptors

Table 3 reports comparisons among the proposed descriptors and the LBP-based descriptors in terms of the APR and ARR. The best {APR, ARR} values are shown in bold. As documented in this table, the {APR, ARR} values of LPCP are far better than those of LBP by {9.64%, 1.34%} on Coil-100, {28.88%, 14.43%} on Face95, {32.10%, 16.12%} on Face96, {11.63%, 2.90%} on Outex-00031, {15.44%, 3.86%} on Outex-00032, {12.26%, 3.10%} on Outex-00033, {38.07%, 19.04%} on Outex-00034, and {4.96%, 3.10%} on MIT-VisTex. Analogous to the proposed LPCP, the {APR, ARR} values of ULPCP and RILPCP are also definitely higher than those of ULBP and RILBP on all eight databases. Meanwhile, three points can be observed below: (1) ULPCP has the maximum {APR, ARR} value of {99.52%, 13.82%} on Coil-100; (2) RILPCP produces the best results of {97.12%, 48.56%} on Face95, and {97.77%, 49.04%} on Face96; and (3) LPCP has the highest performance of {89.62%, 22.40%} on Outex-00031, {84.86%, 21.21%} on Outex-00032, {87.80%, 21.99%} on Outex-00033, {84.36%, 42.18%} on Outex-00034, and {98.33%, 61.46%} on MIT-VisTex. According to the above results, it can be asserted that the improvements given by the proposed descriptors are very considerable. The main reason for this is that the LBP information is amalgamated with the color information.

4.6. Comparison with Other Color Texture Descriptors

To evaluate the effectiveness, efficiency, robustness, and computational complexity, the proposed descriptors were compared with eight state-of-the-art color texture descriptors in terms of the average precision rate (APR) value, the average recall rate (ARR) value, the precision-recall (PR) curve, the feature vector length, and the memory consumption. All experiments were performed under the leave-one-out cross-validation principle. For clarity, all comparative color texture methods are summarized as follows:
  • Multi-channel adder local binary patterns (mdLBP) [18]: The 2048-dimensional color texture descriptor in the RGB color space.
  • Multi-channel decoded local binary patterns (maLBP) [18]: The 1024-dimensional color texture descriptor in the RGB color space.
  • Color difference histogram (CDH) [21]: The 90-dimensional color histogram and the 18-dimensional edge orientation histogram in the L*a*b* color space.
  • Multi-texton histogram (MTH) [22]: The 64-dimensional color histogram and the 18-dimensional edge orientation histogram in the HSV color space.
  • Micro-structure descriptor (MSD) [23]: The 72-dimensional color histogram and the 6-dimensional edge orientation histogram in the HSV color space.
  • Opponent color local binary patterns (OCLBP) [24]: The 1536-dimensional color texture descriptor in the RGB color space.
  • Improved opponent color local binary patterns (IOCLBP) [25]: The 3072-dimensional color-texture descriptor in the RGB color space.
  • Orthogonal combination of local binary patterns and color histogram (OC-LBP + CH): The 12-dimensional color histogram in the L*a*b* color space and the 96-dimensional LBP variation in the gray-scale space [28].
  • Local parallel cross pattern (LPCP).
  • Rotation-invariant local parallel cross pattern (RILPCP).
  • Uniform local parallel cross pattern (ULPCP).
Table 4 lists the APR and ARR values for the proposed descriptors and the existing descriptors on the eight databases. The best values are shown in bold. On Coil-100, the {APR, ARR} value of RILPCP is significantly superior to those of mdLBP, maLBP, CDH, MTH, MSD, OCLBP, IOCLBP, and OC-LBP+CH by {7.10%, 0.98%}, {11.11%, 1.54%}, {1.26%, 0.17%}, {0.99%, 0.14%}, {1.43%, 0.20%}, {14.04%, 1.95%}, {10.32%, 1.43%}, and {3.81%, 0.53%}, respectively. Meanwhile, the {APR, ARR} value of RILPCP is slightly improved (by {+0.03%, +0.01%}) by using LPCP. In this scenario, ULPCP acquires a higher {APR, ARR} value than RILPCP: {99.44%, 13.81} compared with {99.52%, 13.82}. Similar to Coil-100, there are more competitive results on Face95, Face96, Outex-00031, Outex-00032, Outex-00033, Outex-00034, and MIT-VisTex. As a consequence of the above results, it can be summarized that the effectiveness of the proposed descriptors is demonstrated in terms of APR and ARR. Remarkably, there are rotation differences on Coil-100, resolution differences on Outex-00031, noise differences on Outex-00032, blur differences on Outex-00033, and illumination differences on Outex-00034. Thus, the robustness is also illustrated to a certain extent.
Figure 7a–h depict the precision-recall (PR) curves of all the comparative descriptors on the eight databases. As we can see from Figure 7a, the curves of LPCP, RILPCP, and ULPCP are higher than those of former color texture descriptors. As shown in Figure 7b, although the curves of ULPCP, MTH, and MSD are interleaved with one another, both LPCP and RILPCP are superior to all other descriptors. One possible reason for this is that the Face95 database emphasizes the importance of the color information. As depicted in Figure 7c, considering the complex background, the curves of the proposed descriptors are better than all remaining descriptors on the Face96 database. Analogous to the results for the Face96 database, LPCP, RILPCP, and ULPCP are much better than other descriptors on the Outex-00031 (see the Figure 7d), Outex-00032 (see the Figure 7e), Outex-00033 (see the Figure 7f), and Outex-00034 (see the Figure 7g) databases, respectively. As depicted in Figure 7h, although mdLBP provides a competitive performance, LPCP achieves the highest curve. The main reason for this is that the MIT-VisTex database contains more textural structure information. Based on all the above observations and analyses, it can be easily noticed that the proposed descriptors are effective in terms of the precision-recall (PR) curve in most of the cases.
Table 5 compares the computational complexity and memory cost in terms of the feature vector length by dimensionality (D) and memory consumption in kilobytes (Kb). All the experiments were performed on a 4.20 GHz four-core CPU with 16 GB of memory. Herein, analogous to RILPCP and ULPCP, 760/844/844/424/400/676/676/616 (D) and 5.94/6.59/6.59/3.31/3.13/5.28/5.28/4.81 (Kb) show that LPCP performs retrieval requiring 760 D and 5.94 Kb on Coil-100, 844 D and 6.59 Kb on Face95, 844 D and 6.59 Kb on Face96, 424 D and 3.31 Kb on Outex-00031, 400 D and 3.13 Kb on Outex-00032, 676 D and 5.28 Kb on Outex-00033, 676 D and 5.28 Kb on Outex-00034, and 616 D and 4.81 Kb on MIT-VisTex. As documented in Table 5, the feature vector length and memory consumption of the proposed descriptors are inferior to those of CDH, MTH, MSD, and OC-LBP+CH, but they are superior to those of mdLBP, mdLBP, OCLBP, and IOCLBP. From the results, although the computational complexity is larger than those of some existing methods, there are several superiorities of LPCP, RILPCP, and ULPCP as follows:
  • The added computational complexity is effective because the retrieval accuracy is enhanced by a large margin.
  • The proposed descriptors can adaptively code the color and texture information from different image databases.
  • The practicability and feasibility of the proposed descriptors, with acceptable feature vector length and competitive memory consumption, are well shown for a realistic system configuration.

4.7. Comparison with CNN-Based Descriptors

On a different note, we also compared the proposed descriptors with emerging CNN-based descriptors including VGGm, VGGm128, VGGm1024, VGGm2048, ALEX, GoogleNet, and Inception-v3 [43]. Referring to references [44,45], the last fully connected layer was first extracted from the pretrained models. Then, L2 normalization was performed on the extracted fully connected layer. Finally, the distance measure was calculated on the normalized feature vector. For fairness, all images in the database were chosen as the query images, and the number of returned images was set to 10.
Figure 8 compares the proposed descriptors and the CNN-based descriptors. On the Coil-100 database, VGGm and ALEX perform slightly better than the proposed LPCP, RILPCP, and ULPCP methods. On the Face95, Face96, Outex-00031, Outex-00032, Outex-00033, and Outex-00034 databases, more significant APR values are obtained by using the proposed LPCP, RILPCP, and ULPCP methods. On the MIT-VisTex database, VGGm achieves a performance that is competitive with those of RILPCP and ULPCP, but LPCP achieves the highest APR value. Although VGGm and ALEX yield relatively competitive APR values, the superior abilities of LPCP, RILPCP, and ULPCP are revealed as follows:
  • The CNN-based descriptors must be pretrained on a large-scale and annotated dataset (e.g., ImageNet), while the proposed LPCP, RILPCP, and ULPCP methods do not need any pretraining process.
  • The pretrained CNN-based descriptors are computationally expensive (e.g., cloud servers and mainframe computers), but LPCP, RILPCP, and ULPCP can be performed on almost all realistic systems (e.g., personal smartphones and home security cameras).
  • LPCP, RILPCP, and ULPCP are more effective than the CNN-based descriptors on six datasets out of the eight examined.

4.8. Additional Experiments on a Weight-Based Optimization Scheme

To further investigate the optimized coefficient scheme, we added weighting-based experiments on the eight databases. Referring to references [35,37], we defined the weighting parameter v, where v {0, 0.01, …, 1.00}, and the weighting local parallel cross pattern (LPCPweight) was formulated as follows:
L P C P w e i g h t = [ ( 1 v ) · L P C P L B P , v · L P C P c o l o r ] ,
Observe that LPCPweight degenerates into LPCPLBP when v = 0.00 and into LPCPcolor when v = 1.00. Further, LPCP is derived from LPCPweight when v is set to 0.50. We also extended LPCPweight to the weighting uniform local parallel cross pattern (ULPCPweight) and the weighting rotation-invariant local parallel cross pattern (URILPCPweight).
Figure 9 shows the curves of the average precision rate (APR) under the weighting parameter v by using LPCPweight, ULPCPweight, and RILPCPweight on the eight databases. On the Coil-100, Face95, and Face96 databases, with an increasing value v, the APR values first rapidly go up and then gradually become stable. These results illustrate that a higher proportion of color information is crucial for object and facial images. On the five color textural databases, with the addition of the color information, the APR values firstly show fast growth for v near 0.00, gradually become stable, and then abruptly decrease for v near 1.00. These phenomena demonstrate that an appropriate weighting optimization scheme can achieve greater enhancements.
In summary, an optimized parameter v is not only beneficial to integrating the merits of LPCPcolor and LPCPLBP, but also yields more notable improvements.

5. Conclusions

In this paper, a color texture method called local parallel cross pattern (LPCP) was proposed for encoding the LBP and color information into a unified framework. Three major contributions were summarized in this paper. First, based on the color prior knowledge in the L*a*b* color space, we designed a six-layer color quantizer that is to be used for color map extraction. Second, inspired by the human visual system, we proposed the local parallel cross pattern (LPCP) for combining the color map and the LBP map in “parallel” and “cross” manners. Third, to improve the computational complexity and provide rotation invariant, LPCP was further extended to the uniform local parallel cross pattern (ULPCP) and the rotation-invariant local parallel cross pattern (RILPCP), respectively. We performed a series of experiments on the eight databases to evaluate the proposed descriptors. Depending on the average precision rate (APR) results, the optimal quantization layers were chosen from the six-layer color quantizer. Compared with the LBP-based descriptors, LPCP, ULPCP, and RILPCP achieved promising APR and ARR results. To evaluate the effectiveness, efficiency, robustness, and computational complexity, we performed comparative experiments among the proposed methods and eight state-of-the-art color texture descriptors in terms of the average precision rate (APR) value, the average recall rate (ARR) value, the precision-recall (PR) curve, the feature vector length, and the memory consumption. Moreover, the proposed approaches were also compared with a series of CNN-based models and achieved competitive results. Additionally, the weight-based optimization scheme yielded more notable improvements.
In the future, Locality Sensitive Hashing [46] and feature selection [47,48,49,50] will be considered to cut down the computation complexity and memory consumption. Meanwhile, Query Expansion (QE) [51] and Graph Fusion (GF) [52] will be integrated into the image retrieval system to retrieve more target images. Moreover, normalization methods [53] also will be considered to achieve illumination invariance.

Supplementary Materials

Supplementary File 1

Author Contributions

Q.F. conceived the research idea. Q.F. performed the experiments. Q.F. wrote the paper. Q.F., Q.H., M.S., Y.Y., Y.W., and J.D. gave many suggestions and helped revise this manuscript.

Funding

This research was funded by the National Nature Science Foundation of China, grant number [61871106], the Fundamental Research Grant Scheme for the Central Universities, grant number [130204003], the Project of Shandong Province Higher Educational Science and Technology Program, grant number [J16LN68], the Shandong Province Natural Science Foundation, grant number [ZR2017QF011] and the National Key Technology Research and Development Programme of the Ministry of Science and Technology of China, grant number [2014BAI17B02].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Feng, Q.; Hao, Q.; Chen, Y.; Yi, Y.; Wei, Y.; Dai, J. Hybrid histogram descriptor: A fusion feature representation for image retrieval. Sensors 2018, 187, 1943. [Google Scholar] [CrossRef] [PubMed]
  2. Yang, M.; Song, W.; Mei, H. Efficient retrieval of massive ocean remote sensing images via a cloud-based mean-shift algorithm. Sensors 2017, 17, 1693. [Google Scholar] [CrossRef] [PubMed]
  3. Liu, S.; Wu, J.; Feng, L.; Qiao, H.; Liu, Y.; Lou, W.; Wang, W. Perceptual uniform descriptor and ranking on manifold for image retrieval. Inf. Sci. 2017, 424, 235–249. [Google Scholar] [CrossRef]
  4. Smeulders, A.W.M.; Worring, M.; Santini, S.; Gupta, A.; Jain, R. Content-based image retrieval at the end of the early years. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 22, 1349–1380. [Google Scholar] [CrossRef]
  5. Piras, L.; Giacinto, G. Information fusion in content based image retrieval: A comprehensive overview. Inf. Fusion 2017, 37, 50–60. [Google Scholar] [CrossRef]
  6. Liu, L.; Chen, J.; Fieguth, P.; Zhao, G.; Chellappa, R.; Pietikäinen, M. From BoW to CNN: Two decades of texture representation for texture classification. Int. J. Comput. Vis. 2018, 1–36. [Google Scholar] [CrossRef]
  7. Fernández, A.; Álvarez, M.X.; Bianconi, F. Texture description through histograms of equivalent patterns. J. Math. Imaging Vis. 2013, 45, 76–102. [Google Scholar] [CrossRef]
  8. Ojala, T.; Pietikäinen, M.; Maenpaa, T. Multi resolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  9. Guo, Z.; Zhang, L.; Zhang, D. Rotation invariant texture classification using LBP variance (LBPV) with global matching. Pattern Recognit. 2010, 43, 706–719. [Google Scholar] [CrossRef]
  10. Guo, Z.; Zhang, L.; Zhang, D. A completed modeling of local binary pattern operator for texture classification. IEEE Trans. Image Process. 2010, 19, 1657–1663. [Google Scholar]
  11. Zhang, B.; Gao, Y.; Zhao, S.; Liu, J. Local derivative pattern versus local binary pattern: Face recognition with high-order local pattern descriptor. IEEE Trans. Image Process. 2010, 19, 533–544. [Google Scholar] [CrossRef] [PubMed]
  12. Tan, X.; Triggs, B. Enhanced local texture feature sets for face recognition under difficult lighting conditions. IEEE Trans. Image Process. 2010, 19, 1635–1650. [Google Scholar] [PubMed]
  13. Murala, S.; Maheshwari, R.P.; Balasubramanian, R. Local tetra patterns: A new feature descriptor for content-based image retrieval. IEEE Trans. Image Process. 2012, 21, 2874–2886. [Google Scholar] [CrossRef] [PubMed]
  14. Subrahmanyam, M.; Maheshwari, R.P.; Balasubramanian, R. Local maximum edge binary patterns: A new descriptor for image retrieval and object tracking. Signal Process. 2012, 92, 1467–1479. [Google Scholar] [CrossRef]
  15. Ren, J.; Jiang, X.; Yuan, J. Noise-resistant local binary pattern with an embedded error-correction mechanism. IEEE Trans. Image Process. 2013, 22, 4049–4060. [Google Scholar] [CrossRef] [PubMed]
  16. Verma, M.; Raman, B. Local neighborhood difference pattern: A new feature descriptor for natural and texture image retrieval. Multimed. Tools Appl. 2018, 77, 11843–11866. [Google Scholar] [CrossRef]
  17. Jeena Jacob, I.; Srinivasagan, K.G.; Jayapriya, K. Local oppugnant color texture pattern for image retrieval system. Pattern Recognit. Lett. 2014, 42, 72–78. [Google Scholar] [CrossRef]
  18. Dubey, S.R.; Singh, S.K.; Singh, R.K. Multichannel decoded local binary patterns for content-based image retrieval. IEEE Trans. Image Process. 2016, 25, 4018–4032. [Google Scholar] [CrossRef]
  19. Qi, X.; Xiao, R.; Li, C.; Qiao, Y.; Guo, J.; Tang, X. Pairwise rotation invariant co-occurrence local binary pattern. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 2199–2213. [Google Scholar] [CrossRef]
  20. Hao, Q.; Feng, Q.; Wei, Y.; Sbert, M.; Lu, W.; Xu, Q. Pairwise cross pattern: A color-LBP descriptor for content-based image retrieval. In Proceedings of the 19th Pacific Rim Conference on Multimedia, Hefei, China, 21–22 September 2018. [Google Scholar]
  21. Liu, G.; Yang, J. Content-based image retrieval using color difference histogram. Pattern Recognit. 2013, 46, 188–198. [Google Scholar] [CrossRef]
  22. Liu, G.; Li, Z.; Zhang, L.; Xu, Y. Image retrieval based on micro-structure descriptor. Pattern Recognit. 2011, 44, 2123–2133. [Google Scholar] [CrossRef] [Green Version]
  23. Liu, G.; Zhang, L.; Hou, Y.; Li, Z.; Yang, J. Image retrieval based on multi-texton histogram. Pattern Recognit. 2010, 43, 2380–2389. [Google Scholar] [CrossRef]
  24. Mäenpää, T.; Pietikäinen, M. Texture analysis with local binary patterns. In Handbook of Pattern Recognition and Computer Vision; World Scientific: Singapore, 2005; pp. 197–216. [Google Scholar]
  25. Bianconi, F.; Bello-Cerezo, R.; Napoletano, P. Improved opponent color local binary patterns: An effective local image descriptor for color texture classification. J. Electron. Imag. 2017, 27. [Google Scholar] [CrossRef]
  26. Li, L.; Feng, L.; Yu, L.; Wu, J.; Liu, S. Fusion framework for color image retrieval based on bag-of-words model and color local Haar binary patterns. J. Electron. Imag. 2016, 25. [Google Scholar] [CrossRef]
  27. Cusano, C.; Napoletano, P.; Schettini, R. Evaluating color texture descriptors under large variations of controlled lighting conditions. J. Opt. Soc. Am. A 2016, 33, 17–30. [Google Scholar] [CrossRef]
  28. Singh, C.; Walia, E.; Kaur, K.P. Enhancing color image retrieval performance with feature fusion and non-linear support vector machine classifier. Optik 2018, 158, 127–141. [Google Scholar] [CrossRef]
  29. Pietikäinen, M.; Ojala, T.; Xu, Z. Rotation-invariant texture classification using feature distributions. Pattern Recognit. 2000, 33, 43–52. [Google Scholar] [CrossRef] [Green Version]
  30. Bianconi, F.; González, E. Counting local n-ary patterns. Pattern Recognit. Lett. 2018, 177, 24–29. [Google Scholar] [CrossRef]
  31. Sarrafzadeh, O.; Dehnavi, A.M. Nucleus and cytoplasm segmentation in microscopic images using k-means clustering and region growing. Adv. Biomed. Res. 2015, 4, 174. [Google Scholar]
  32. Gonzalez, R.C.; Woods, R.E. Digital Image Processing, 3rd ed.; Publishing House of Electronics Industry: Beijing, China, 2010; pp. 455–456. ISBN 9787121102073. [Google Scholar]
  33. Salzburg Texture Image Database. Available online: http://www.wavelab.at/sources/STex/ (accessed on 22 August 2014).
  34. Kolesnikov, A.; Trichina, E.; Kauranne, T. Estimating the number of clusters in a numerical data set via quantization error modeling. Pattern Recognit. 2015, 48, 941–952. [Google Scholar] [CrossRef]
  35. Zhang, M.; Zhang, K.; Feng, Q.; Wang, J.; Kong, J. A novel image retrieval method based on hybrid information descriptors. J. Vis. Commun. Image Represent. 2014, 25, 1574–1587. [Google Scholar] [CrossRef]
  36. Standring, S. Gray’s Anatomy: The Anatomical Basis of Clinical Practice, 41st ed.; Elsevier Limited: New York, NY, USA, 2016; pp. 686–708. ISBN 9780702068515. [Google Scholar]
  37. Guo, J.; Prasetyo, H.; Lee, H.; Yao, C. Image retrieval using indexed histogram of void-and-cluster block truncation coding. Signal Process. 2016, 123, 143–156. [Google Scholar] [CrossRef]
  38. Nene, S.A.; Nayar, S.K.; Murase, H. Columbia Object Image Library (COIL-100); Technical Report CUCS; Department of Computer Science, Columbia University: New York, NY, USA, 1996. [Google Scholar]
  39. Libor Spacek’s Facial Image Databases “Face 95 Image Database”. Available online: https://cswww.essex.ac.uk/mv/allfaces/faces95.html (accessed on 8 August 2014).
  40. Libor Spacek’s Facial Image Databases “Face 96 Image Database”. Available online: https://cswww.essex.ac.uk/mv/allfaces/faces96.html (accessed on 8 August 2014).
  41. Outex Texture Image Database. Available online: http://lagis-vi.univ-lille1.fr/datasets/outex.html (accessed on 5 October 2017).
  42. MIT Vision and Modeling Group. Available online: http://vismod.media.mit.edu/pub/ (accessed on 12 August 2014).
  43. Szegedy, C.; Vanhoucke, V.; Loffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  44. Napoletano, P. Hand-crafted vs. learned descriptors for color texture classification. In Proceedings of the International Workshop on Computational Color Imaging, Milan, Italy, 29–31 March 2017. [Google Scholar]
  45. Zou, Q.; Ni, L.; Zhang, T.; Wang, Q. Deep learning based feature selection for remote sensing scene classification. IEEE Trans. Geosci. Remote Sens. Lett. 2015, 12, 2321–2325. [Google Scholar] [CrossRef]
  46. Indyk, P.; Motwani, R. Approximate nearest neighbors: Towards removing the curse of dimensionality. In Proceedings of the IEEE Conference on Multimedia Information Analysis and Retrieval, Dallas, TX, USA, 24–26 May 1998; pp. 604–613. [Google Scholar]
  47. Yi, Y.; Zhou, W.; Liu, Q.; Luo, G.; Wang, J.; Fang, Y.; Zheng, C. Ordinal preserving matrix factorization for unsupervised feature selection. Signal Process. Image Commun. 2018, 67, 118–131. [Google Scholar] [CrossRef]
  48. Yi, Y.; Chen, Y.; Dai, J.; Gui, X.; Chen, X.; Lei, G.; Wang, W. Semi-supervised ridge regression with adaptive graph-based label propagation. Appl. Sci. 2018, 12, 2636. [Google Scholar] [CrossRef]
  49. Yi, Y.; Qiao, S.; Zhou, W.; Zheng, C.; Liu, Q.; Wang, J. Adaptive multiple graph regularized semi-supervised extreme learning machine. Soft Comput. 2018, 22, 3545–3562. [Google Scholar] [CrossRef]
  50. Qi, M.; Wang, T.; Liu, F.; Zhang, B.; Wang, J.; Yi, Y. Unsupervised feature selection by regularized matrix factorization. Neurocomputing 2018, 273, 593–610. [Google Scholar] [CrossRef]
  51. Chum, O.; Mikulik, M.; Perdoch, M.; Matas, J. Total recall II: Query expansion revisited. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Colorado Springs, CO, USA, 20–25 June 2011; pp. 889–896. [Google Scholar]
  52. Zhang, S.; Yang, M.; Cour, T.; Yu, K.; Metaxas, D. Query specific rank fusion for image retrieval. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 803–815. [Google Scholar] [CrossRef]
  53. Cernadas, E.; Fernández-Delgado, M.; González-Rufino, E. Influence of normalization and color space to color texture classification. Pattern Recognit. 2017, 61, 120–138. [Google Scholar] [CrossRef]
Figure 1. The frequency of pixels on Stex and its subsets: (a,b) Stex, (c,d) 50% of Stex, and (e,f) 10% of Stex.
Figure 1. The frequency of pixels on Stex and its subsets: (a,b) Stex, (c,d) 50% of Stex, and (e,f) 10% of Stex.
Sensors 19 00315 g001aSensors 19 00315 g001b
Figure 2. The details of the six-layer color quantizer.
Figure 2. The details of the six-layer color quantizer.
Sensors 19 00315 g002
Figure 3. The quantization errors under different quantization layers, Wa* and Wb*, on Stex and its subsets: (a) Stex, (b) 50% of Stex, and (c) 10% of Stex.
Figure 3. The quantization errors under different quantization layers, Wa* and Wb*, on Stex and its subsets: (a) Stex, (b) 50% of Stex, and (c) 10% of Stex.
Sensors 19 00315 g003
Figure 4. Schematic diagram of the local parallel cross pattern (LPCP).
Figure 4. Schematic diagram of the local parallel cross pattern (LPCP).
Sensors 19 00315 g004
Figure 5. Illustration of the rotation invariance of the rotation-invariant local parallel cross pattern (RILPCP).
Figure 5. Illustration of the rotation invariance of the rotation-invariant local parallel cross pattern (RILPCP).
Sensors 19 00315 g005
Figure 6. Some sample images from the eight databases: (a) Coil-100; (b) Face95; (c) Face96; (d) Outex-00031; (e) Outex-00032; (f) Outex-00033; (g) Outex-00034; and (h) MIT-VisTex.
Figure 6. Some sample images from the eight databases: (a) Coil-100; (b) Face95; (c) Face96; (d) Outex-00031; (e) Outex-00032; (f) Outex-00033; (g) Outex-00034; and (h) MIT-VisTex.
Sensors 19 00315 g006aSensors 19 00315 g006b
Figure 7. The precision-recall curves of eleven descriptors on the eight databases: (a) Coil-100; (b) Face95; (c) Face96; (d) Outex-00031; (e) Outex-00032; (f) Outex-00033; (g) Outex-00034; and (h) MIT-VisTex.
Figure 7. The precision-recall curves of eleven descriptors on the eight databases: (a) Coil-100; (b) Face95; (c) Face96; (d) Outex-00031; (e) Outex-00032; (f) Outex-00033; (g) Outex-00034; and (h) MIT-VisTex.
Sensors 19 00315 g007aSensors 19 00315 g007b
Figure 8. The average precision rate (APR) comparisons among the proposed descriptors and the CNN-based descriptors on the eight databases.
Figure 8. The average precision rate (APR) comparisons among the proposed descriptors and the CNN-based descriptors on the eight databases.
Sensors 19 00315 g008
Figure 9. The average precision rate (APR) under the weight value v by using LPCP, RILPCP, and ULPCP on the eight databases.
Figure 9. The average precision rate (APR) under the weight value v by using LPCP, RILPCP, and ULPCP on the eight databases.
Sensors 19 00315 g009
Table 1. Summary of image databases.
Table 1. Summary of image databases.
No.NameImage SizeClassImages in Each ClassImages TotalFormatWebsite
1Coil-100 (Rotation)128 × 128100727200JPGhttp://www.cs.columbia.edu/CAVE/software/softlib/coil-100.php
2Face95180 × 20072201440JPGhttps://cswww.essex.ac.uk/mv/allfaces/faces95.html
3Face96196 × 1969119 or 201814JPGhttps://cswww.essex.ac.uk/mv/allfaces/faces96.html
4Outex-00031
(Scale)
128 × 12868402720BMPhttp://lagis-vi.univ-lille1.fr/datasets/outex.html
5Outex-00032
(Noise)
128 × 12868402720BMPhttp://lagis-vi.univ-lille1.fr/datasets/outex.html
6Outex-00033
(Blur)
128 × 12868402720BMPhttp://lagis-vi.univ-lille1.fr/datasets/outex.html
7Outex-00034
(illumination)
128 × 128204204080BMPhttp://lagis-vi.univ-lille1.fr/datasets/outex.html
8MIT-VisTex128 × 1284016640PPMhttp://vismod.media.mit.edu/pub/VisTex/
Table 2. The highest average precision rate (APR) values of LPCP, RILPCP, and uniform local parallel cross pattern (ULPCP) with the optimal color quantization layers (Wa*, Wb*) on the eight databases.
Table 2. The highest average precision rate (APR) values of LPCP, RILPCP, and uniform local parallel cross pattern (ULPCP) with the optimal color quantization layers (Wa*, Wb*) on the eight databases.
MethodPerformanceDataset
Coil-100 Face95Face96Outex-00031Outex-00032Outex-00033Outex-00034MIT-VisTex
LPCP(Wa*, Wb*)(6, 5)(6, 6)(6, 6)(6, 1)(5, 1)(6, 4)(6, 4)(5, 4)
APR (%)99.4792.3397.0389.6284.8687.8084.3698.33
RILPCP(Wa*, Wb*)(5, 5)(6, 6)(6, 6)(5, 1)(5, 1)(6, 3)(4, 4)(5, 2)
APR (%)99.4497.1297.7789.5384.4087.5584.0697.22
ULPCP(Wa*, Wb*)(4, 6)(6, 6)(6, 6)(5, 1)(5, 1)(6, 3)(5, 4)(3, 2)
APR (%)99.5296.6597.4589.4484.0787.5984.2097.39
Table 3. The performance comparisons among the proposed descriptors and the local binary pattern (LBP)-based descriptors on the eight databases.
Table 3. The performance comparisons among the proposed descriptors and the local binary pattern (LBP)-based descriptors on the eight databases.
MethodPerformanceDataset
Coil-100 Face95Face96Outex-00031Outex-00032Outex-00033Outex-00034MIT-VisTex
LBPAPR (%)89.8363.4564.9377.9969.4275.5446.2993.37
ARR (%)12.4831.7332.5519.5017.3518.8923.1458.36
RILBPAPR (%)84.5359.7861.1976.5767.0074.1345.9389.75
ARR (%)11.7429.8930.6819.1416.7518.5322.9656.09
ULBPAPR (%)85.4058.2559.4276.0365.6873.0445.5190.83
ARR (%)11.8629.1229.7919.0116.4218.2622.7556.77
LPCPAPR (%)99.4792.3397.0389.6284.8687.8084.3698.33
ARR (%)13.8246.1648.6722.4021.2121.9942.1861.46
RILPCPAPR (%)99.4497.1297.7789.5384.4087.5584.0697.22
ARR (%)13.8148.5649.0422.3821.1021.8942.0360.76
ULPCPAPR (%)99.5296.6597.4589.4484.0787.5984.2097.39
ARR (%)13.8248.3348.8822.3621.0221.9042.1060.87
Table 4. The performance comparisons between the proposed descriptors and the existing descriptors in terms of APR and ARR on the eight databases.
Table 4. The performance comparisons between the proposed descriptors and the existing descriptors in terms of APR and ARR on the eight databases.
MethodPerformanceDataset
Coil-100Face95Face96Outex-00031Outex-00032Outex-00033Outex-00034MIT-VisTex
mdLBPAPR (%)92.3472.9779.0969.5159.6065.6348.6697.05
ARR (%)12.8336.4939.6617.3814.9016.4124.3360.65
maLBPAPR (%)88.3367.9473.9970.4260.2465.4244.5395.80
ARR (%)12.2733.9737.1017.6015.0616.3622.2759.87
CDHAPR (%)98.1894.6994.9785.4979.6582.9074.0394.03
ARR (%)13.6447.3547.6321.3719.9120.7337.0258.77
MTHAPR (%)98.4589.1592.2880.7474.6379.9865.1091.97
ARR (%)13.6744.5746.2820.1818.6620.0032.5557.48
MSDAPR (%)98.0192.9789.9477.1672.4676.1666.3292.20
ARR (%)13.6146.4845.1119.2918.1219.0433.1657.63
OCLBPAPR (%)85.4064.4064.5580.7669.2577.9956.1392.42
ARR (%)11.8632.2032.3720.1917.3119.5028.0757.76
IOCLBPAPR (%)89.1266.4773.2483.8474.5480.7573.5895.59
ARR (%)12.3833.2436.7320.9618.6320.1936.7959.75
OC-LBP+CHAPR (%)95.6380.5088.6785.0775.9381.3372.1892.20
ARR (%)13.2840.2544.4621.2718.9820.3336.0957.63
LPCPAPR (%)99.4792.3397.0389.6284.8687.8084.3698.33
ARR (%)13.8246.1648.6722.4021.2121.9942.1861.46
RILPCPAPR (%)99.4497.1297.7789.5384.4087.5584.0697.22
ARR (%)13.8148.5649.0422.3821.1021.8942.0360.76
ULPCPAPR (%)99.5296.6597.4589.4484.0787.5984.2097.39
ARR (%)13.8248.3348.8822.3621.0221.9042.1060.87
Table 5. Feature vector length (D) and memory consumption (Kb) among the proposed descriptors and other previous descriptors.
Table 5. Feature vector length (D) and memory consumption (Kb) among the proposed descriptors and other previous descriptors.
MethodFeature Vector Length (D)Memory Consumption (Kb)
mdLBP204816.00
maLBP10248.00
CDH1080.84
MTH820.64
MSD780.61
OCLBP153511.99
IOCLBP307224.00
OC-LBP+CH1080.84
LPCP760/844/844/424/400/676/676/6165.94/6.59/6.59/3.31/3.13/5.28/5.28/4.81
RILPCP468/624/624/180/180/372/336/2523.66/4.88/4.88/1.41/1.41/2.91/2.63/1.97
ULPCP479/647/647/203/203/395/419/2033.74/5.05/5.05/1.59/1.59/3.09/3.27/1.59

Share and Cite

MDPI and ACS Style

Feng, Q.; Hao, Q.; Sbert, M.; Yi, Y.; Wei, Y.; Dai, J. Local Parallel Cross Pattern: A Color Texture Descriptor for Image Retrieval. Sensors 2019, 19, 315. https://doi.org/10.3390/s19020315

AMA Style

Feng Q, Hao Q, Sbert M, Yi Y, Wei Y, Dai J. Local Parallel Cross Pattern: A Color Texture Descriptor for Image Retrieval. Sensors. 2019; 19(2):315. https://doi.org/10.3390/s19020315

Chicago/Turabian Style

Feng, Qinghe, Qiaohong Hao, Mateu Sbert, Yugen Yi, Ying Wei, and Jiangyan Dai. 2019. "Local Parallel Cross Pattern: A Color Texture Descriptor for Image Retrieval" Sensors 19, no. 2: 315. https://doi.org/10.3390/s19020315

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop