Next Article in Journal
Rapid, High Affinity Binding by a Fluorescein Templated Copolymer Combining Covalent, Hydrophobic, and Acid–Base Noncovalent Crosslinks
Next Article in Special Issue
IrisDenseNet: Robust Iris Segmentation Using Densely Connected Fully Convolutional Networks in the Images by Visible Light and Near-Infrared Light Camera Sensors
Previous Article in Journal
On-Board, Real-Time Preprocessing System for Optical Remote-Sensing Imagery
Previous Article in Special Issue
Presentation Attack Detection for Iris Recognition System Using NIR Camera Sensor
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Textile Retrieval Based on Image Content from CDC and Webcam Cameras in Indoor Environments

by
Oscar García-Olalla
1,
Enrique Alegre
1,3,
Laura Fernández-Robles
2,3,
Eduardo Fidalgo
1,3,* and
Surajit Saikia
1,3
1
Department of Electrical, Systems and Automation, Universidad de León, 24071 León, Spain
2
Department of Mechanical, Computer Science and Aerospace Engineering, Universidad de León, 24071 León, Spain
3
Researcher at INCIBE (Spanish National Cybersecurity Institute), 24005 León, Spain
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(5), 1329; https://doi.org/10.3390/s18051329
Submission received: 21 March 2018 / Revised: 18 April 2018 / Accepted: 21 April 2018 / Published: 25 April 2018
(This article belongs to the Special Issue Visual Sensors)

Abstract

:
Textile based image retrieval for indoor environments can be used to retrieve images that contain the same textile, which may indicate that scenes are related. This makes up a useful approach for law enforcement agencies who want to find evidence based on matching between textiles. In this paper, we propose a novel pipeline that allows searching and retrieving textiles that appear in pictures of real scenes. Our approach is based on first obtaining regions containing textiles by using MSER on high pass filtered images of the RGB, HSV and Hue channels of the original photo. To describe the textile regions, we demonstrated that the combination of HOG and HCLOSIB is the best option for our proposal when using the correlation distance to match the query textile patch with the candidate regions. Furthermore, we introduce a new dataset, TextilTube, which comprises a total of 1913 textile regions labelled within 67 classes. We yielded 84.94% of success in the 40 nearest coincidences and 37.44% of precision taking into account just the first coincidence, which outperforms the current deep learning methods evaluated. Experimental results show that this pipeline can be used to set up an effective textile based image retrieval system in indoor environments.

1. Introduction

The process of automatically finding objects, textiles, faces, or other patterns in images and videos is one of the most studied topics in computer vision. Nowadays, with the huge amount of digital images and videos, it becomes even more critical. Visual sensors are able to acquire a large quantity of visual information from the surroundings around them. Content Based Image Retrieval (CBIR) consists of retrieving images using their content properties from a collection that match a user’s query [1] based on a similarity measure [2]. Many research fields, e.g., medical image [3,4,5], human retrieval [6], biological analysis [7,8], agricultural retrieval [9] and biometric security [10], achieved interesting results using CBIR techniques.
Most works related to CBIR aim at finding objects in datasets of images. Research groups face this problem using different approaches such as invariant local features (SIFT [11,12], SURF [13]), color description [14,15], template matching [16,17] or, more recently, deep learning techniques [18,19,20]. Nonetheless, these techniques may fail when the object does not present a rigid shape or it has a plain shape, as it is the case of textiles. The same textile can appear in images with very skewed shapes. Moreover, textile retrieval shares the difficulties of object retrieval such as the variety in illumination conditions, occlusions, lack of texture information, etc.
The need of retrieving textiles from image collections captured under a variety of visual sensors can be motivated by many applications [21]—for example, for marketing studies in textile stores that suggest the products that fit a decorated room to users. The recently published book “Applications of computer vision in fashion and textiles” [22] deals with three aspects related to computer vision techniques applied to textile industry: (i) textile defect detection and quality control, (ii) fashion recognition and 3D modeling, and (iii) 2D and 3D human body modeling for improving clothing fit. One of its chapters [23] reviews the computer vision state-of-the-art techniques for fashion textile modeling, recognition, and retrieval. A completely different approach in which textiles are needed to be retrieved is to connect evidence of different crime scenes.
In our case, this work is framed in the Advisory System Against Sexual Exploitation of Children (ASASEC) project, a European project that fights child pornography using forensic analysis, data mining and computer vision techniques. It was demonstrated that perverts usually use the same bedrooms to take their pictures or videos [24]. A way to link two images, and consequently provide relationships among the many cases of child pornography, is finding the exact same textiles such as carpets, blankets or any other repeated texture. In our specific case, we aim at evaluating textiles in order to retrieve images in huge datasets (thousands of images) of past proven cases of child abuse connected with a query textile of interest.
The rest of the paper is organized as follows. The related work is presented in Section 2. Section 3 describes the pipeline of our proposed method for textile based image retrieval. Section 4 introduces the TextilTube dataset, the evaluation metrics, the evaluation set-up and the decision evaluation of a distance measure. We show the results of all the experiments carried out in Section 5. Finally, Section 6 draws the conclusions of the paper.

2. Related Research

Material retrieval is related to textile retrieval in some aspects and it is more broadly studied in the literature. Zhu and Brilakis [25] presented a system for detecting concrete based on the account of the colour of the regions in the image. After that, they described the regions using color features and trained a machine learning classifier to determine if the region contains concrete or not. This method cannot deal with very heterogeneous regions due to the way of creating the image partitions. Son et al. [26] proposed a method based on ensemble classifiers in order to distinguish between concrete, steel and wood. One of the main disadvantages of this work is the necessity of uniform areas of the same material for the segmentation step, which consists of dividing the original image into sub-regions of a fixed size. If the material region is smaller than the grid division, a lot of information of the background is processed as a material resulting in a non-accurate description. In [27,28], the authors proposed two methods able to identify multiple materials in object surfaces without the need of segmentation. They first recognized the object class and then used correlations of material labels for such object. In this approach, the correct definition of detailed semantic cues of objects and materials is needed. In 2017, Xue et al. [29] focused on material recognition of real-world outdoor surfaces for which they presented a new very useful dataset for autonomous agents. They exploited the idea of extracting characteristics of materials encoded in the angular and spatial gradients of their appearance from images taken with small angular variations. We refer the reader to [30] for an overview of methods and applications for the automatic characterization of the visual appearance of materials. Material retrieval systems are effective when construction materials are involved, but they may fail with other kind of textiles. The main three differences with general textiles are: the classes of construction materials are well defined, the texture of the construction materials is more homogeneous and the image patches of construction materials are usually big and present regular shapes.
Besides material retrieval, there are also few textile retrieval works in the bulk of the literature. Bashar et al. [31] proposed a system based on three wavelet-domain based features called symmetry, regularity and directionality. In this paper, the authors demonstrated outperformance of the combination of the three features versus just the isolated descriptors using two datasets formed by 150 and 300 images of curtain patterns. Similarly, in 2009, Carbunaru et al. [32] proposed a method that applies independent component analysis over wavelet-domain images. In that case, the researchers chose a dataset composed of images of 30 different fabrics, obtaining an average recognition rate of 94.86%. Recently, in Chun et al. [33], a new method which uses composite feature vectors of color from spatial domain and texture from wavelet-transformed domain is proposed. In contrast with the other papers described before, Chun et al. carried out a retrieval system using a large dataset composed of 1343 textile images. In 2014, Huang and Lin [34] proposed a system based on the combination of color, texture and shape features in order to retrieve textiles over more than 4000 images downloaded from Globle-Tex Co., (http://www.globle-tex.com/). The retrieval system was based on a signature process extracted by different k-means clusters achieving an 83% of success rate. Nevertheless, in all cases, the material or textile datasets are already segmented, usually as a plain piece of fabric, and the system is only focused in the retrieval process. On the contrary, in our proposed work, the textiles are located in real environments presenting diverse shapes, under a wide range of capturing conditions and exposed to occlusions. Recently, a bunch of papers deal with query targets such as cloth worn on human bodies [35,36,37,38,39,40]. It is quite challenging to extract robust features from different images presenting different poses. Our paper encompasses a wider definition of the word textiles, and it is used to retrieve not only cloth on human bodies but any other textile that may appear in indoor environments.
The segmentation of regions of interest is thus critical for an efficient method. In 2016, Zheng and Sarem proposed a method called NAMES, which stands for Non-symmetry and Anti-packing Model and Extended Shading and is based on the idea of packing pixels with a very high performance in terms of time [41]. In 2015, Yang et al. [42] presented a method based on color histogram segmentation using HSV color space. In 2004, Matas et al. [43] proposed a method based on the extraction of Maximally Stable Extremal Regions (MSER) taking into account a binary threshold that varies along all the gray scale spectrum. In our work, we segmented our images using the latter method due to its tolerance against regions with little changes of intensity and the possibility of adjusting it using the binary threshold correctly. After the segmentation step, the description of the regions is another key step of our CBIR system. The detected textile regions can be described using texture descriptors, which are widely used for texture analysis. Texture analysis is a challenging and still open problem in computer vision that consists of detecting and describing the gray level spatial variations of the image pixels. Nowadays, there are multiple fields that profit from automatic texture retrieval, as it makes processes faster with no need for many qualified staff. For this purpose, local descriptors are yet extensively employed for texture description due to their high performance in terms of time and accuracy. Histogram of Oriented Gradients (HOG) is a very popular texture descriptor since Dalal and Triggs presented it in 2005 [44]. This method has demonstrated a great performance in multiple fields, such as pedestrian detection [44] or face recognition [45]. Another very popular descriptor is Local Binary Pattern (LBP) proposed by Ojala et al. [46] due to their simplicity and high capability to extract the intrinsic features from the textures. Guo et al. developed several modifications to LBP such as LBP variance (LBPV) [47], complete LBP (CLBP) [48] or adaptive LBP (ALBP) [49]. García-Olalla et al. introduced algorithms to enhance LBP description [50,51,52], developing a new booster method that can be fused with LBP in order to improve accuracy results [53]. We refer the reader to [54,55] for a general framework and a taxonomy of local binary patterns variants. Recent approaches are focusing on deep Convolutional Neural Networks (CNN) such as AlexNet [56], GoogleNet [57] or VGG-Net [58]. The activations generated at the fully connected layers are used as feature descriptors for image understanding [59], scene recognition [60], semantic segmentation [61], among others.
In this work, we propose a new method for textile based image retrieval in indoor scenes under diverse capturing conditions and subjected to different shapes and occlusions. In accordance with that goal, we present in this paper a new labeled dataset that we created and made publicly available (http://pitia.unileon.es/varp/node/483). It is composed of 684 images extracted from videos with 67 different classes of textiles and it is called TexilTube. We used videos recorded with different visual sensors, such as compact digital cameras (CDC) and webcams. This dataset reproduces, at a small scale, a typical scenario of image evidence related to child pornography. We used MSER on high pass filetered images of the RGB, HSV and Hue channels of the original images to extract the regions of textiles. To describe the textile regions, we used local texture features, i.e., LBP, ALBP, HOG and Faster R-CNN (Region based Convolutional Neural Network) [62]. To enhance the description of the texture descriptors, we used Complete Local Oriented Statistical Information Booster (CLOSIB) booster. We evaluated several distance measures, i.e., Spearman rank, Cityblock, Euclidean and Correlations distances and two evaluation metrics, i.e., precision at n and success at n. We consider the following contributions of this work: (i) we propose a method for extracting the regions of interest based on computing MSER on high pass filtered images of the RGB, HSV and Hue channels of the original images; (ii) we evaluated the performance several descriptors; (iii) we assessed several distance measures by means of a voting schema; and (iv) we present a new dataset for textile retrieval.

3. Method

3.1. Overview

In Figure 1, we illustrate the pipeline of our new method for textile based image retrieval. We can divide the method in two main stages: feature extraction and matching. By means of the experimentation carried out, we were able to determine that the regions in a scene containing textiles have to be extracted from three different transformations of the picture: a high pass filtered image of the RGB, HSV and Hue channels of the original image. We took this into consideration for building up the pipeline of the method.
The feature extraction is comprised of four steps. First, we convert the images to RGB and HSV colour spaces and we also extract the Hue channel. Then, we sharpen the image representations to increase the contrast along the edges where different colors meet. We adopt the unsharp masking method in which an image is sharpened by subtracting a blurred (unsharp) version of the image from itself. We use a Gaussian lowpass filter of standard deviation 1.5 for blurring the image. We use these three image representations for the extraction and description of the regions of interest in the images. The MSER [43] of the sharpened image representations define the regions of interest of the images. Finally, we describe the regions of interest by computing texture descriptors on the gray scale patch. We create a database in which we store for each detected region: the image coordinates of the bounding box of the region of interest in the image of reference, the images of reference themselves and the descriptors.
The matching stage allows for retrieving a given number of images that present the most similar regions to a query region (textile) of interest. It is made up of three steps. First, we describe the gray scale query region by means of the same texture descriptors. Second, we compute some distance measures among the descriptors of the query region and the descriptors of the database. Finally, the hit list is ranked by sorting the regions of the database in ascending order in relation with the distance measure.
Below, we briefly describe the methods used to build the novel pipeline.

3.2. Region Extraction: MSER

We use the MSER method [43] to automatically extract the regions (textiles) of interest due to the good results achieved in preliminary tests. Other methods apart from MSER could be evaluated for finding distinguished regions that possess some distinguishing or singular properties and allow for repeatedly detecting them over a range of image conditions, such as “sieve” [63]. However, an intensive evaluation of such methods is out of the scope of this paper.
MSER is a method for blob detection that extracts from an image a number of co-variant regions called MSERs. These high contrast regions are connected areas characterized by almost uniform intensity, surrounded by a contrasting background. MSERs are constructed by binarizing the image at multiple threshold levels and selecting the connected components that maintain their sizes over a large set of thresholds.
Experimentally, we chose to extract a great diversity of sizes for the areas of the regions of interest, specifically, comprehended between 3000 and 540,000 pixels, step size between intensity threshold levels equal to 3, and a maximum area variation between extremal regions of 0.7.

3.3. Region Description

We use the following methods to describe the textiles: LBP [64], ALBP [49] and HOG [44], and early fusion concatenations of the previous descriptors with CLOSIB and Half Complete Local Oriented Statistical Information Booster (HCLOSIB) enhancers [65].

3.3.1. Local Binary Pattern (LBP)

LBP describes the texture of gray scale images by means of the local spatial structure on the image. For every pixel, a pattern code is computed by comparing its gray level value with the value of its neighbors.
In this work, we used uniform rotational invariant LBP [64] with 16 neighbors and a radius of two pixels, LBP 16 , 2 r i u 2 . The dimension of the descriptor is P + 2 , in this case, 16 + 2 = 18 elements. However, for simplicity, we call it LBP henceforth.

3.3.2. Adaptive Local Binary Pattern (ALBP)

Guo et al. [49] presented a variation of LBP that considers the mean and the standard deviation along given orientation of the pixels in the image. This information is used in the matching step and makes it more robust against changes in the local spatial structure of the images.
We consider also the uniform rotation invariant version, ALBP 16 , 2 r i u 2 , and we call it ALBP for simplicity.

3.3.3. Histogram of Oriented Gradients (HOG)

Histograms of Oriented Gradients [44] evaluates local histograms of image gradient orientations over a grid. HOG characterizes the local appearance of objects taking into account the local edge direction distributions. The method is implemented by dividing the image into small uniform regions called cells, often overlapped. Then, for each cell, a histogram of the gradient orientations over the pixels is extracted. The final descriptor is yielded by concatenation of the gradients along all the cells. In this work, we use overlapped cells of size 64 × 64 pixels on images resized to 256 × 256 pixels.

3.3.4. Complete Local Oriented Statistical Information Booster (CLOSIB) Variants

CLOSIB [65] is obtained from the statistical information of the gray scale gradient magnitude of each pixel of the image. The statistical information of the gradient magnitudes is rarely taken into account to describe the image and provides useful information for texture classification. Equation (1) shows how to compute the CLOSIB enhancer of an image:
CLOSIB P , R , θ = p = 1 P / η ( θ 1 ) μ 2 p ( 1 ) θ ( μ 1 p ) θ 1 / θ ,
where ∥ symbolizes the concatenation function, θ { 1 , 2 } is the order of the statistical moment considered, μ 1 p and μ 2 p are the first and second statistical raw moments, respectively, defined in Equation (2) and η is a factor that controls the portion of the considered orientations in the quantized angular space. We set η = 1 for CLOSIB and η = 2 for Half CLOSIB (HCLOSIB):
μ i p = 1 N c = 1 N | g c g p | i ,
where N is the number of pixels in the image, g c the gray value of the center pixel and g p the gray value of the neighbor, located at a distance R with orientation 2 π p / P from the center pixel. In this work, we use the boosters C L O S I B 16 , 2 , 1 | | C L O S I B 16 , 2 , 2 and H C L O S I B 16 , 2 , 1 | | H C L O S I B 16 , 2 , 2 , which we name CLOSIB and HCLOSIB, respectively, henceforth. Furthermore, we also use the early fusion (concatenation) of LBP, ALBP and HOG descriptors with CLOSIB and HCLOSIB enhancers to describe the texture of the textile regions. We denote the concatenation with the symbol +. For instance, LBP + CLOSIB stands for the early fusion of LBP and CLOSIB.

3.3.5. Faster R-CNN

Faster R-CNN [62] is a Region based Convolutional Neural Network (R-CNN) that generates region of interest proposals by a Region Proposal Network (RPN). Faster R-CNN is basically composed of two parts: a RPN for creating a list of region proposals and a Fast R-CNN network [66] for classifying the regions into objects.
For the RPN, we applied a sliding window of size 3 × 3 on the features obtained at the last convolution layer, which yields an intermediate layer of dimension 512. We fed the intermediate layer into a box classification layer and a box regression layer. We fed the region proposals into a Fully Connected (FC) layer and we extracted the neural codes. Similarly to the MSER approach, we saved a database with the coordinates of the region proposals, the image reference and the neural codes. The matching step remains the same as for MSER approach. We used Faster R-CNN algorithm with VGG-16 [58] architecture pre-trained with an MS-COCO [67] dataset.

3.4. Distance Measures

We use five distance measures to compute the distances among the descriptors of the query region and the descriptors of the automatically detected regions of interest of the database. These are: Spearman, Cosine, Cityblock, Euclidean and Correlation distances. Spearman rank correlation coefficient is a nonparametric measure of rank correlation and it measures the strength and direction of association between two ranked variables. This measure uses a variable’s rank which is the average of their positions in the ascending order of the values. Spearman rank correlation coefficient of two vectors A and B is mathematically defined in Equation (3):
d s p e ( A , B ) = 1 ( ( r ( A ) r ( A ) ¯ ) ( r ( B ) r ( B ) ¯ ) T ( ( r ( A ) r ( A ) ¯ ) ( r ( A ) r ( A ) ¯ ) T ( ( r ( B ) r ( B ) ¯ ) ( r ( B ) r ( B ) ¯ ) T ,
where r ( A ) ¯ and r ( B ) ¯ are the mean value of the ranked vector A, r ( A ) , and ranked vector B, r ( B ) , respectively. The superscript T indicates the transpose of the matrix. Hereafter, we keep the same notation. Cosine distance calculates the angular cosine between two vectors following Equation (4):
d c o s ( A , B ) = 1 A B T ( A A T ) ( B B T ) .
Cityblock distance is calculated using Equation (5) and is defined by the sum of the absolute distances of every coordinate between two vectors. n is the dimension of the vectors. This measure distance depends on the rotation of the coordinate system but is invariant to reflection and translation:
d c i t ( A , B ) = j = 1 n | A j B j | .
Euclidean distance (see Equation (6)) is the most commonly used distance measure and calculates the length of the straight segment that connects two vectors:
d e u c ( A , B ) = ( ( A B ) ( A B ) T .
The Correlation distance is obtaining by dividing the distance covariance of two vectors by the product of their distance standard deviations. See Equation (7):
d c o r ( A , B ) = 1 ( A A ¯ ) ( B B ¯ ) T ( A A ¯ ) ( A A ¯ ) T ( B B ¯ ) ( B B ¯ ) T .

4. Experiments

4.1. TextilTube Dataset

Textile retrieval in real environments is a poorly investigated research field besides fashion cloth retrieval. Up to our knowledge, there is no publicly available dataset that focuses on the recognition of rigid and non-rigid textiles presented in different sizes, shapes and capturing conditions. For this reason, we created a new dataset for the retrieval of textiles in bedrooms (http://pitia.unileon.es/varp/node/483).
The dataset is composed of 684 images of sizes that range between 480 × 360 and 1280 × 720 pixels obtained from 15 videos of YouTube. The videos were recorded in bedrooms with different visual sensors, such as CDCs and webcams. The videos contain plenty of textiles, different camera poses, illumination conditions, occlusions, etc., which makes the textile retrieval task very challenging. The dataset contains 67 classes of textiles such as curtains, carpets, sofas, shirts or dresses, among others. In one image, several classes of textiles may appear. Figure 2 shows a mosaic encompassing one region sample of each class and indicates the number of regions in each class. The number of elements of each class varies from 4 to 116. There is a total of 1913 regions. Therefore, the dataset is highly skewed, simulating a real scenario.
We labelled the dataset in order to provide a ground truth that allows the user to automatically evaluate the performance of a method on the dataset. The ground truth includes the bounding box coordinates and the class labels of each textile region in the images of the dataset. We provide the ground truth in the form of an XML file. We show the diversity in terms of type, size, pose, etc. of some textile classes of the dataset in Figure 3.
TextilTube dataset can be very interesting in fields like child sexual abuse or robbery to connect evidence of different investigations and also for marketing studies in textile stores to suggest the products that best fit the decoration of users’ rooms.

4.2. Performance Evaluation Metrics

In retrieval systems, it is important that the retrieved images are ranked according to their relevance to the query region forming a hit list, rather than being returned as a set. The most relevant hits must be within the top images of the hit list returned for a query region. To account for the quality of ranking the hits in the hit list, we used relevance ranking measures, i.e., precision at n and success at n.

4.2.1. Precision at n

Precision at n, p@n, is the rate of the top-n images of the hit list correctly classified in relation to the class of the query region. Likewise, the precision at a cut-off of n elements of the hit list. We define H i t L i s t n as the set that contains the n images with smallest distance to the query region, q. Equation (8) presents the mathematical definition of precision at n:
p @ n = # H ( q ) n ,
where # H ( q ) is the cardinal of H i t L i s t n in which the query class is actually present in the image and the detected region overlaps the bounding box of the ground truth. It is formally defined in Equation (9):
H ( q ) = h i / ( h i H i t L i s t n ) ( c l a s s ( h i ) = c l a s s ( q ) ) i = 1 n ,
where h i is i-th retrieved image in the hit list.

4.2.2. Success at n

There are occasions in which the user does not need to see many relevant images but is disappointed by a completely irrelevant top-n [68]. This is the case of the ASASEC project, in which finding at least one hit in all the hit list would be a satisfactory result. Success at n, s@n, measures if a relevant image was retrieved within the top-n hits of the hit list. Success at n is equal to 1 if the top-n images contain a relevant document and 0 otherwise (see Equation (10)):
s @ n = 1 , if H ( q ) , 0 , otherwise ,
where H ( q ) is the set of images defined in Equation (9).

4.3. Experimental Setup

We applied the method described in Section 3 to the 684 images of TextilTube dataset, extracting a total of 58,031 regions. In order to evaluate the performance of our method, we used the ground truth textile regions as query regions of interest. For each query region, we calculated p@n and s@n metrics for the retrieved hit list when computing a given distance measure among the texture descriptors of the query region and the analogous texture descriptors of the database. Experiments using Faster R-CNN were developed using the Caffe [69] deep learning framework in Nvidia Titan X GPU https://www.nvidia.com/en-us/geforce/products/10series/titan-x-pascal/.

4.4. Distance Measure Evaluation

In order to determine the best distance measure and present uniform results, we carried out the following voting system. For each texture descriptor in Section 3.2, we computed s@n for n N n = 1 , 2 , , 40 with all distance measures described in Section 3.4. We assigned three, two and one points to the distance measures that achieved the highest, second highest and third highest s@n, respectively, for each experiment. Finally, we summed up the points along all combinations. Figure 4 shows a scheme of the procedure. We disregarded a voting system that only relies on the best distance measure of each experiment because the results for the different distance measures were not enough distinctive.
Figure 5 presents the results in parts per unity achieved with each distance measure. Correlation distance achieved the best results with a 32% of votes, followed by Cosine distance (27%) and Spearman rank correlation coefficient (20%). The commonly used Euclidean distance only yielded a 7% of the votes. Therefore, we carried out our experiments using the Correlation distance.

5. Results

In this section, we present the results obtained following the proposed method and experimentation for each evaluated texture descriptor and the neural codes extracted by Faster R-CNN.
Figure 6 shows the precision at n (p@n) achieved for all texture descriptors. We used values of n N n = 1 , 2 , , 40 . For n 11 , HOG + HCLOSIB descriptor outperformed the rest with a precision of 37.17% for n = 1 . The early fusion of CLOSIB and HCLOSIB with HOG outperforms HOG alone. However, the early fusion of CLOSIB and HCLOSIB with LBP obtained the worst results. In the case of ALBP, the descriptor alone outperforms the early fusion for small values of n, whereas the opposite is true for high values of n. It is worth noting the better performance of ALBP (28.83% for n = 1 ) versus LBP (16.60% for n = 1 ). At a cut of 20, the precision at n values starts to stabilize. We present the numerical results for precision at cuts 1 , 2 , , 20 in Table 1. For high cuts of the hit list, Faster-RCNN slightly outperformed the rest. The best performance was not achieved by some LBP variant as we expected, but by HOG combined with HCLOSIB. HOG is oriented to gather the external and internal shape and HCLOSIB represents the statistical distributions of the texture. The combination of both represents both the shape of the textile’s texture (HOG) and how this texture is organized along the evaluated patch.
Figure 7 illustrates the success at n (s@n). As expected, s@1 is the same as p@1 and for higher cuts of the hit list the success metric increases. In Table 2, we show the numerical results for success at cuts 1 , 2 , , 20 . For values of n 5 , HOG + HCLOSIB yielded the best results, whereas for higher values of n, Faster R-CNN outperformed the others with a 84.94% of s@40 (74.86% with HOG + HCLOSIB). ALBP is the second best descriptor for n = 40 reaching 82.00% of success. CLOSIB enhancer improves the performance of HOG and decreases the performance of LBP and ALBP.
In order to get a unique value to evaluate the performance of each descriptor, we computed the arithmetic mean of the success and precision for three intervals of n N n = 1 , 2 , , j , where j = 10 , 20 , 40 . Figure 8 and Table 3 show the arithmetic mean of the success and the precision in these intervals of values of n. Regarding p@n, HOG + HCLOSIB outperformed the rest of descriptors for the intervals with j = 10 and j = 20 , whereas Faster-RCNN obtained the best results for j = 40 . With respect to s@n, HOG + HCLOSIB yielded the best results for the intervals with j = 10 , whereas Faster R-CNN did for j = 20 and j = 40 . In such a difficult dataset, the outlined method using HOG + HCLOSIB descriptor and Correlation distance measure yielded an arithmetic mean of the precision at 10 of 24.80% and an arithmetic mean of the success at 10 of 51.08%.
Figure 9 and Figure 10 show the visual results for the first five retrieved images in the hit list using HOG + HCLOSIB and Faster R-CNN, respectively. The third textile, artificial flower textile, is a difficult case. HOG + HCLOSIB manages to get one correctly retrieved images for n = 5 . The first two retrieved images are similar textiles of a rug. For the same query image, Faster R-CNN does not retrieve any correct images at a cut of hit list of n = 5 . Pre-trained deep neural networks are trained to classify objects instead of textiles. When a textile does not present a patterned texture, a pre-trained Faster R-CNN is not appropriate to retrieve such queries. However, Faster R-CNN manages to retrieve textiles correctly that present distinctive patterns.

6. Conclusions

In this paper, we presented a new application for textile based image retrieval in indoor environments. Textile based image retrieval is barely studied and when doing so, it is usually applied to fashion cloth retrieval. We introduced a new framework of study, the fight of child sexual abuse. Law enforcement agencies are interested in relating evidence by using textile queries in order to retrieve images or videos that contain the same textile in proven cases of child pornography, usually taken from rooms of houses. We proposed a new effective method for textile based image retrieval in rooms based on texture description of the MSER regions of the images. We assessed LBP, ALBP, HOG and their combination with CLOSIB for describing the image patches and several distance metrics for sorting the hit list. We also evaluated the Faster R-CNN algorithm with VGG-16 architecture pre-trained with MS-COCO dataset. Furthermore, we created and introduced a new public dataset, TextilTube, which consists of 684 frames from 15 Youtube videos of rooms recorded with different visual sensors. The dataset contains 1913 regions of interest that highly vary in terms of capturing conditions, occlusions, illuminations, etc. Moreover, textiles appearing in the images are not rigid and present different shapes. Correlation distance proved to be the most discriminant distance measure based on a voting system analysis. Correlation distance achieved 32% of the votes followed by cosine distance with 27%. HOG + HCLOSIB yielded the best results for low cuts of the hit list, whereas Faster R-CNN performed better for high cuts, closely followed by ALBPS. Taking into account just the most similar image retrieved, HOG + HCLOSIB achieved a precision of 37.17%, which is remarkable due to the number of classes in the dataset (67 classes) and their high intra-class variability. Taking into account the success at n metric, Faster R-CNN achieved a 84.94% retrieving 40 images (ALBP obtained a 82%), which means that about 85 out of 100 images have at least one correspondence in the top 40 retrieved images. This is a very interesting result that can be presented as an application for the criminal police in order to let them evaluate a grid of 40 images at a glance to check if there is a real match to the query image within the hit list. For the application at hand, it is interesting to achieve a high precision at low cuts of the hit list in order to reduce the number of images to visually inspect. To measure this fact, we computed the arithmetic mean of precision at n from 1 to 10. HOG + HCLOSIB outperformed the rest yielding a 24.8% hit rate. The main problem in this application is to find regions containing textiles, rather than objects. To the best of our knowledge, all the deep learning region proposal models are oriented and trained to detect objects. Objects are usually non-homogeneous regions as opposed to textiles. The reason is that CBIR systems are oriented to retrieve objects but not textiles or similar surfaces. Similarly, the deep learning models, are trained with datasets such as MS-COCO or ImageNet, among others, that contain objects and different classes of objects and they are oriented for instance retrieval. In future works, we will train a model for proposing regions with a textile dataset to strengthen the use of deep learning for textile retrieval. Besides evaluating other Region Proposal Networks in future works, different alternatives to MSER for finding distinguished regions, such as Sieve, will be also tested.

Author Contributions

Oscar García-Olalla and Enrique Alegre conceived the proposed method and designed the experiments; Surajit Saikia performed the experiments for CNNs and Oscar García-Olalla performed the rest of experiments; Oscar García-Olalla, Enrique Alegre, Laura Fernández-Robles and Eduardo Fidalgo analyzed the data; Laura Fernández-Robles and Oscar García-Olalla wrote the paper; Eduardo Fidalgo, Enrique Alegre and Laura Fernández-Robles substantively revised the paper, Eduardo Fidalgo provided the visual results.

Funding

This research was funded by [the Spanish Government] grant number [DPI2012-36166] and by [INCIBE (Spanish National Cybersecurity Institute)] grant number [INCIBEC-2015-02493, Addendum 22].

Acknowledgments

This work has been supported by grant DPI2012-36166, the pre-doctoral FPU fellowship program from the Spanish Government (AP2010-0947), grant INCIBEC-2015-02493 corresponding to “Ayudas para la Excelencia de los Equipos de Investigación avanzada en ciberseguridad”, the framework agreement between the University of León and INCIBE (Spanish National Cybersecurity Institute) under Addendum 22, and the PIRTU program of the Regional Government of Castilla y León. We gratefully acknowledge the support of Nvidia Corporation for their kind donation of GPUs (GeForce GTX Titan X and K-40) that were used in this work.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LBPLocal Binary Pattern
ALBPAdaptive Local Binary Pattern
HOGHistogram of Oriented Gradients
CLOSIBComplete Local Oriented Statistical Information Booster
HCLOSIBHalf Complete Local Oriented Statistical Information Booster
MSERMaximally Stable Extremal Regions
CBIRContent Based Image Retrieval
ASASECAdvisory System Against Sexual Exploitation of Children
HSVHue, Saturation, Value
CDCCompact Digital Cameras
p@nprecision at n
s@nsuccess at n
CNNConvolutional Neural Networks
R-CNNRegion based Convolutional Neural Network
RPNRegion Proposal Network
FCFully Connected
RGBRed, Green, Blue
SIFTScale-Invariant Feature Transform
SURFSpeeded Up Robust Features
VGGVisual Geometry Group
MS-COCOMicroSoft Common Objects in COntext
XMLeXtensible Markup Language

References

  1. Czúni, L.; Rashad, M. Lightweight Active Object Retrieval with Weak Classifiers. Sensors 2018, 18, 801. [Google Scholar] [CrossRef] [PubMed]
  2. Domínguez, S. Saliency-based similarity measure. Rev. Iberoam. Autom. Inform. Ind. 2012, 9, 359–370. [Google Scholar] [CrossRef] [Green Version]
  3. Faria, A.V.; Oishi, K.; Yoshida, S.; Hillis, A.; Miller, M.I.; Mori, S. Content-based image retrieval for brain MRI: An image-searching engine and population-based analysis to utilize past clinical data for future diagnosis. NeuroImage Clin. 2015, 7, 367–376. [Google Scholar] [CrossRef] [PubMed]
  4. Srinivas, M.; Naidu, R.R.; Sastry, C.; Mohan, C.K. Content based medical image retrieval using dictionary learning. Neurocomputing 2015, 168, 880–895. [Google Scholar] [CrossRef]
  5. Bugatti, P.H.; Kaster, D.S.; Ponciano-Silva, M.; Traina, C., Jr.; Azevedo-Marques, P.M.; Traina, A.J. PRoSPer: Perceptual similarity queries in medical CBIR systems through user profiles. Comput. Biol. Med. 2014, 45, 8–19. [Google Scholar] [CrossRef] [PubMed]
  6. Jung, J.; Yoon, I.; Lee, S.; Paik, J. Normalized Metadata Generation for Human Retrieval Using Multiple Video Surveillance Cameras. Sensors 2016, 16, 963. [Google Scholar] [CrossRef] [PubMed]
  7. Feng, L.; Bhanu, B.; Heraty, J. A software system for automated identification and retrieval of moth images based on wing attributes. Pattern Recognit. 2016, 51, 225–241. [Google Scholar] [CrossRef]
  8. Mallik, J.; Samal, A.; Gardner, S.L. A content based image retrieval system for a biological specimen collection. Comput. Vis. Image Underst. 2010, 114, 745–757. [Google Scholar] [CrossRef]
  9. Liu, B.; Yue, Y.M.; Li, R.; Shen, W.J.; Wang, K.L. Plant Leaf Chlorophyll Content Retrieval Based on a Field Imaging Spectroscopy System. Sensors 2014, 14, 19910–19925. [Google Scholar] [CrossRef] [PubMed]
  10. Iqbal, K.; Odetayo, M.O.; James, A. Content-based image retrieval approach for biometric security using colour, texture and shape features controlled by fuzzy heuristics. J. Comput. Syst. Sci. 2012, 78, 1258–1277. [Google Scholar] [CrossRef]
  11. Chang, L.; Duarte, M.M.; Sucar, L.; Morales, E.F. A Bayesian approach for object classification based on clusters of SIFT local features. Expert Syst. Appl. 2012, 39, 1679–1686. [Google Scholar] [CrossRef]
  12. Fidalgo, E.; Alegre, E.; González-Castro, V.; Fernández-Robles, L. Compass radius estimation for improved image classification using Edge-SIFT. Neurocomputing 2016, 197, 119–135. [Google Scholar] [CrossRef]
  13. Chen, L.C.; Hsieh, J.W.; Yan, Y.; Chen, D.Y. Vehicle make and model recognition using sparse representation and symmetrical SURFs. Pattern Recognit. 2015, 48, 1979–1998. [Google Scholar] [CrossRef]
  14. Li, H.; Liu, Z.; Huang, Y.; Shi, Y. Quaternion generic Fourier descriptor for color object recognition. Pattern Recognit. 2015, 48, 3895–3903. [Google Scholar] [CrossRef]
  15. Zhu, J.; Yu, J.; Wang, C.; Li, F.Z. Object recognition via contextual color attention. J. Vis. Commun. Image Represent. 2015, 27, 44–56. [Google Scholar] [CrossRef]
  16. Shih, H.C.; Yu, K.C. SPiraL Aggregation Map (SPLAM): A new descriptor for robust template matching with fast algorithm. Pattern Recognit. 2015, 48, 1707–1723. [Google Scholar] [CrossRef]
  17. Tan, M.; Hu, Z.; Wang, B.; Zhao, J.; Wang, Y. Robust object recognition via weakly supervised metric and template learning. Neurocomputing 2016, 181, 96–107. [Google Scholar] [CrossRef]
  18. Salvador, A.; Giró i Nieto, X.; Marqués, F.; Satoh, S. Faster R-CNN Features for Instance Search. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 394–401. [Google Scholar]
  19. Han, J.; Zhang, D.; Cheng, G.; Liu, N.; Xu, D. Advanced Deep-Learning Techniques for Salient and Category-Specific Object Detection: A Survey. IEEE Signal Process. Mag. 2018, 35, 84–100. [Google Scholar] [CrossRef]
  20. Saikia, S.; Fidalgo, E.; Alegre, E.; Fernández-Robles, L. Query Based Object Retrieval Using Neural Codes. In Advances in Intelligent Systems and Computing; Springer: Berlin/Heidelberg, Germany, 2018; Volume 649, pp. 513–523. [Google Scholar]
  21. D’Amato, J.P.; Mercado, M.; Heiling, A.; Cifuentes, V. A proximal optimization method to the problem of nesting irregular pieces using parallel architectures. Rev. Iberoam. Autom. Inform. Ind. 2016, 13, 220–227. [Google Scholar] [CrossRef]
  22. Wong, C. Applications of Computer Vision in Fashion and Textiles, 1st ed.; The Textile Institute Book Series; Woodhead Publishing: Cambreidge, UK; Elsevier Science: New York, NY, USA, 2017. [Google Scholar]
  23. Wen, J.; Wong, W. Chapter 2—Fundamentals of common computer vision techniques for fashion textile modeling, recognition, and retrieval. In Applications of Computer Vision in Fashion and Textiles; Wong, W., Ed.; The Textile Institute Book Series; Woodhead Publishing: Cambreidge, UK, 2018; pp. 17–44. [Google Scholar]
  24. Gangwar, A.; Fidalgo, E.; Alegre, E.; González-Castro, V. Pornography and Child Sexual Abuse Detection in Image and Video: A Comparative Evaluation. In Proceedings of the 8th International Conference on Imaging for Crime Imaging for Crime Detection and Prevention, Madrid, Spain, 13–15 December 2017. [Google Scholar]
  25. Zhu, Z.; Brilakis, I. Parameter optimization for automated concrete detection in image data. Autom. Constr. 2010, 19, 944–953. [Google Scholar] [CrossRef]
  26. Son, H.; Kim, C.; Hwang, N.; Kim, C.; Kang, Y. Classification of major construction materials in construction environments using ensemble classifiers. Adv. Eng. Inform. 2014, 28, 1–10. [Google Scholar] [CrossRef]
  27. Xie, X.; Yang, L.; Zheng, W.S. Learning object-specific DAGs for multi-label material recognition. Comput. Vis. Image Underst. 2016, 143, 183–190. [Google Scholar] [CrossRef]
  28. Yang, L.; Xie, X. Exploiting object semantic cues for Multi-label Material Recognition. Neurocomputing 2016, 173, 1646–1654. [Google Scholar] [CrossRef]
  29. Xue, J.; Zhang, H.; Dana, K.; Nishino, K. Differential Angular Imaging for Material Recognition. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6940–6949. [Google Scholar]
  30. González, E.; Bianconi, F.; Álvarez, M.X.; Saetta, S.A. Automatic Characterization of the Visual Appearance of Industrial Materials through Colour and Texture Analysis: An Overview of Methods and Applications. Adv. Opt. Technol. 2013, 2013, 1–11. [Google Scholar] [CrossRef]
  31. Bashar, M.; Ohnishi, N.; Matsumoto, T.; Takeuchi, Y.; Kudo, H.; Agusa, K. Image retrieval by pattern categorization using wavelet domain perceptual features with LVQ neural network. Pattern Recognit. Lett. 2005, 26, 2315–2335. [Google Scholar] [CrossRef]
  32. Carbunaru, A.E.; Coltuc, D.; Jourlin, M.; Frangu, L. A texture descriptor for textile image retrieval. In Proceedings of the 2009 International Symposium on Signals, Circuits and Systems, Iasi, Romania, 9–10 July 2009; pp. 1–4. [Google Scholar]
  33. Chun, J.C.; Kim, W.G. Textile Image Retrieval Using Composite Feature Vectors of Color and Wavelet Transformed Textural Property. Appl. Mech. Mater. 2013, 333, 822–827. [Google Scholar] [CrossRef]
  34. Huang, Y.F.; Lin, S.M. Searching Images in a Textile Image Database; Springer International Publishing: Cham, Switzerland, 2014; pp. 267–274. [Google Scholar]
  35. Yamaguchi, K.; Kiapour, M.H.; Ortiz, L.E.; Berg, T.L. Retrieving Similar Styles to Parse Clothing. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1028–1040. [Google Scholar] [CrossRef] [PubMed]
  36. Liang, X.; Lin, L.; Yang, W.; Luo, P.; Huang, J.; Yan, S. Clothes Co-Parsing Via Joint Image Segmentation and Labeling with Application to Clothing Retrieval. IEEE Trans. Multimed. 2016, 18, 1175–1186. [Google Scholar] [CrossRef]
  37. Sun, G.L.; Wu, X.; Peng, Q. Part-based clothing image annotation by visual neighbor retrieval. Neurocomputing 2016, 213, 115–124. [Google Scholar] [CrossRef]
  38. Chen, Q.; Huang, J.; Feris, R.; Brown, L.M.; Dong, J.; Yan, S. Deep domain adaptation for describing people based on fine-grained clothing attributes. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 5315–5324. [Google Scholar]
  39. Huang, J.; Feris, R.; Chen, Q.; Yan, S. Cross-Domain Image Retrieval with a Dual Attribute-Aware Ranking Network. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1062–1070. [Google Scholar]
  40. Kiapour, M.H.; Han, X.; Lazebnik, S.; Berg, A.C.; Berg, T.L. Where to Buy It: Matching Street Clothing Photos in Online Shops. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 3343–3351. [Google Scholar]
  41. Zheng, Y.; Sarem, M. A fast region segmentation algorithm on compressed gray images using Non-symmetry and Anti-packing Model and Extended Shading representation. J. Vis. Commun. Image Represent. 2016, 34, 153–166. [Google Scholar] [CrossRef]
  42. Yang, B.; Yu, H.; Hu, R. Unsupervised regions based segmentation using object discovery. J. Vis. Commun. Image Represent. 2015, 31, 125–137. [Google Scholar] [CrossRef]
  43. Matas, J.; Chum, O.; Urban, M.; Pajdla, T. Robust wide-baseline stereo from maximally stable extremal regions. Image Vis. Comput. 2004, 22, 761–767. [Google Scholar] [CrossRef]
  44. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference onComputer Vision and Pattern Recognition (CVPR 2005), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  45. Li, B.; Huo, G. Face recognition using locality sensitive histograms of oriented gradients. Opt. Int. J. Light Electron Opt. 2016, 127, 3489–3494. [Google Scholar] [CrossRef]
  46. Ojala, T.; Pietikainen, M.; Harwood, D. Performance evaluation of texture measures with classification based on Kullback discrimination of distributions. In Proceedings of the 12th International Conference on Pattern Recognition, Jerusalem, Israel, 9–13 October 1994; Volume 1, pp. 582–585. [Google Scholar]
  47. Guo, Z.; Zhang, L.; Zhang, D. Rotation invariant texture classification using LBP variance (LBPV) with global matching. Pattern Recognit. 2010, 43, 706–719. [Google Scholar] [CrossRef]
  48. Guo, Z.; Zhang, L.; Zhang, D. A completed modeling of local binary pattern operator for texture classification. IEEE Trans. Image Process. 2010, 19, 1657–1663. [Google Scholar] [PubMed]
  49. Guo, Z.; Zhang, L.; Zhang, D.; Zhang, S. Rotation invariant texture classification using adaptive LBP with directional statistical features. In Proceedings of the 2010 17th IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2010; pp. 285–288. [Google Scholar]
  50. García-Olalla, O.; Alegre, E.; Fernández-Robles, L.; García-Ordás, M.T. Vitality assessment of boar sperm using an adaptive LBP based on oriented deviation. In Proceedings of the Computer Vision—ACCV 2012 Workshops, Daejeon, Korea, 5–9 November 2012; Park, J.I., Kim, J., Eds.; Springer: Berlin/Heidelberg, Germany, 2013; pp. 61–72. [Google Scholar]
  51. Garcia-Olalla, O.; Alegre, E.; Fernandez-Robles, L.; Garcia-Ordas, M.T.; Garcia-Ordas, D. Adaptive local binary pattern with oriented standard deviation (ALBPS) for texture classification. EURASIP J. Image Video Process. 2013, 2013, 31. [Google Scholar] [CrossRef]
  52. García-Olalla, O.; Alegre, E.; García-Ordás, M.T.; Fernández-Robles, L. Evaluation of LBP Variants using several Metrics and kNN Classifiers. In Similarity Search and Applications; Brisaboa, N., Pedreira, O., Zezula, P., Eds.; Springer: Berlin/Heidelberg, Germany, 2013; pp. 151–162. [Google Scholar]
  53. Garcia-Olalla, O.; Alegre, E.; Fernandez-Robles, L.; Gonzalez-Castro, V. Local Oriented Statistics Information Booster (LOSIB) for Texture Classification. In Proceedings of the 2014 22nd International Conference on Pattern Recognition (ICPR), Stockholm, Sweden, 24–28 August 2014; pp. 1114–1119. [Google Scholar]
  54. Fernández, A.; Álvarez, M.X.; Bianconi, F. Texture Description Through Histograms of Equivalent Patterns. J. Math. Imaging Vis. 2013, 45, 76–102. [Google Scholar] [CrossRef]
  55. Liu, L.; Zhao, L.; Long, Y.; Kuang, G.; Fieguth, P. Extended local binary patterns for texture classification. Image Vis. Comput. 2012, 30, 86–99. [Google Scholar] [CrossRef]
  56. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Advances in Neural Information Processing Systems 25; Pereira, F., Burges, C.J.C., Bottou, L., Weinberger, K.Q., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2012; pp. 1097–1105. [Google Scholar]
  57. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  58. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv, 2014; arXiv:1409.1556. [Google Scholar]
  59. Vallet, A.; Sakamoto, H. Convolutional Recurrent Neural Networks for Better Image Understanding. In Proceedings of the 2016 International Conference on Digital Image Computing: Techniques and Applications (DICTA), Gold Coast, Australia, 30 November–2 December 2016; pp. 1–7. [Google Scholar]
  60. Donahue, J.; Jia, Y.; Vinyals, O.; Hoffman, J.; Zhang, N.; Tzeng, E.; Darrell, T. DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition. In Proceedings of the 31st International Conference on Machine Learning, Bejing, China, 21–26 June 2014; Xing, E.P., Jebara, T., Eds.; PMLR: Bejing, China, 2014; Volume 32, pp. 647–655. [Google Scholar]
  61. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR ’14), Columbus, OH, USA, 23–28 June 2014; IEEE Computer Society: Washington, DC, USA, 2014; pp. 580–587. [Google Scholar]
  62. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  63. Bangham, J.A.; Harvey, R.W.; Ling, P.D.; Aldridge, R.V. Morphological scale-space preserving transforms in many dimensions. J. Electron. Imaging 1996, 5, 5–17. [Google Scholar] [CrossRef]
  64. Ojala, T.; Pietikainen, M.; Maenpaa, T. Multiresolution gray-scale and rotation invariant texture classification with Local Binary Patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  65. García-Olalla Olivera, O. Methods for Improving Texture Description by Using Statistical Information Extracted from the Image Gradient. Ph.D. Thesis, Universidad de León, León, Spain, 2017. [Google Scholar]
  66. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  67. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Proceedings of the Computer Vision—ECCV 2014, Zurich, Switzerland, 6–12 September 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer International Publishing: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar]
  68. Liu, L.; Zsu, M.T. Encyclopedia of Database Systems, 1st ed.; Springer: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
  69. Jia, Y.; Shelhamer, E.; Donahue, J.; Karayev, S.; Long, J.; Girshick, R.; Guadarrama, S.; Darrell, T. Caffe: Convolutional Architecture for Fast Feature Embedding. In Proceedings of the 22nd ACM International Conference on Multimedia (MM ’14), Orlando, FL, USA, 3–7 November 2014; ACM: New York, NY, USA, 2014; pp. 675–678. [Google Scholar]
Figure 1. Scheme of the textile based image retrieval system.
Figure 1. Scheme of the textile based image retrieval system.
Sensors 18 01329 g001
Figure 2. A region sample of each of the 67 classes in TexilTube dataset. The number underneath indicates the amount of regions that belong to that class.
Figure 2. A region sample of each of the 67 classes in TexilTube dataset. The number underneath indicates the amount of regions that belong to that class.
Sensors 18 01329 g002
Figure 3. In rows, images that contain the same textile class in TextilTube dataset. The yellow rectangles overlaid in the images indicate the bounding boxes of the textile regions of the ground truth.
Figure 3. In rows, images that contain the same textile class in TextilTube dataset. The yellow rectangles overlaid in the images indicate the bounding boxes of the textile regions of the ground truth.
Sensors 18 01329 g003
Figure 4. Scheme of the voting procedure to determine the best distance measure.
Figure 4. Scheme of the voting procedure to determine the best distance measure.
Sensors 18 01329 g004
Figure 5. Results of the voting process in parts per unity for the different distance measures.
Figure 5. Results of the voting process in parts per unity for the different distance measures.
Sensors 18 01329 g005
Figure 6. Precision at n (p@n) for all texture descriptors using Correlation distance and n N n = 1 , 2 , , 40 .
Figure 6. Precision at n (p@n) for all texture descriptors using Correlation distance and n N n = 1 , 2 , , 40 .
Sensors 18 01329 g006
Figure 7. Success at n (s@n) for all texture descriptors using Correlation distance and n N n = 1 , 2 , , 40 .
Figure 7. Success at n (s@n) for all texture descriptors using Correlation distance and n N n = 1 , 2 , , 40 .
Sensors 18 01329 g007
Figure 8. Arithmetic mean of precision and success at n for intervals of n from 1 to 10, from 1 to 20 and from 1 to 40.
Figure 8. Arithmetic mean of precision and success at n for intervals of n from 1 to 10, from 1 to 20 and from 1 to 40.
Sensors 18 01329 g008
Figure 9. First five retrieved images in the hit list for three query samples using HOG + HCLOSIB.
Figure 9. First five retrieved images in the hit list for three query samples using HOG + HCLOSIB.
Sensors 18 01329 g009
Figure 10. First five retrieved images in the hit list for three query samples using Faster R-CNN.
Figure 10. First five retrieved images in the hit list for three query samples using Faster R-CNN.
Sensors 18 01329 g010
Table 1. Precision at n (p@n) for all texture descriptors using Correlation distance and n N n = 1 , 2 , , 20 . Results highlighted in bold mark the best results per cut of the hit list.
Table 1. Precision at n (p@n) for all texture descriptors using Correlation distance and n N n = 1 , 2 , , 20 . Results highlighted in bold mark the best results per cut of the hit list.
Descriptorn
12345678910
HOG + HCLOSIB37.232.829.326.824.723.221.720.519.418.6
HOG + CLOSIB35.930.827.925.423.822.520.819.518.517.7
HOG35.230.026.724.422.821.520.219.418.717.8
Faster R-CNN30.127.425.223.822.521.620.619.719.018.2
ALBP28.925.223.021.420.119.018.117.516.916.3
ALBP + HCLOSIB25.521.719.618.717.917.016.215.815.415.0
ALBP + CLOSIB24.823.121.820.419.518.818.117.517.116.6
LBP16.614.814.213.713.313.313.012.612.312.0
LBP + CLOSIB12.210.810.49.69.28.98.58.38.17.9
LBP + HCLOSIB11.110.19.38.78.68.58.48.18.07.7
Descriptorn
11121314151617181920
HOG + HCLOSIB17.717.016.415.815.315.014.614.414.013.7
HOG + CLOSIB17.116.415.815.414.914.614.213.913.613.3
HOG17.016.415.915.514.914.514.113.713.413.0
Faster R-CNN17.617.116.816.315.915.515.114.814.614.3
ALBP15.915.615.215.014.714.414.214.013.813.5
ALBP + HCLOSIB14.714.414.114.013.713.513.313.113.012.8
ALBP + CLOSIB16.215.715.415.114.814.614.414.214.013.9
LBP11.711.411.311.111.010.810.710.510.510.4
LBP + CLOSIB7.97.87.77.67.67.57.47.47.37.2
LBP + HCLOSIB7.77.77.67.57.57.37.37.27.27.2
Table 2. Success at n (p@n) for all texture descriptors using Correlation distance and n N n = 1 , 2 , , 20 . Results highlighted in bold mark the best results per cut of the hit list.
Table 2. Success at n (p@n) for all texture descriptors using Correlation distance and n N n = 1 , 2 , , 20 . Results highlighted in bold mark the best results per cut of the hit list.
Descriptorn
12345678910
Faster R-CNN30.138.744.348.452.255.758.660.962.864.3
ALBP28.937.042.346.749.551.954.456.858.260.0
LBP16.622.627.631.735.540.542.845.247.549.2
ALBP + HCLOSIB25.531.436.540.644.047.148.851.053.054.7
HOG + HCLOSIB37.242.847.550.052.254.656.157.158.460.1
HOG + CLOSIB35.940.644.046.649.852.553.254.856.457.9
HOG35.239.542.444.146.148.950.952.954.956.1
ALBP + CLOSIB24.831.736.739.842.749.046.648.349.850.5
LBP + HCLOSIB10.414.717.421.124.427.129.031.133.434.9
LBP + CLOSIB12.216.520.022.024.126.928.729.931.332.6
Descriptorn
11121314151617181920
Faster R-CNN65.967.268.769.971.171.973.174.375.075.9
ALBP61.563.064.365.866.567.868.769.670.271.5
LBP50.953.054.555.857.659.460.261.062.463.5
ALBP + HCLOSIB56.657.959.260.861.962.663.164.264.564.9
HOG + HCLOSIB60.861.762.363.364.565.065.666.566.867.3
HOG + CLOSIB59.560.260.461.262.062.662.963.263.764.3
HOG56.857.557.758.659.159.960.361.061.461.6
ALBP + CLOSIB51.252.553.353.754.255.656.657.458.158.8
LBP + HCLOSIB36.538.539.741.742.943.945.146.648.148.7
LBP + CLOSIB34.236.137.238.940.441.842.142.743.243.7
Table 3. Arithmetic mean of precision and success at n for intervals of n from 1 to 10, from 1 to 20 and from 1 to 40. Results highlighted in bold mark the best results per performance metric.
Table 3. Arithmetic mean of precision and success at n for intervals of n from 1 to 10, from 1 to 20 and from 1 to 40. Results highlighted in bold mark the best results per performance metric.
DescriptorPrecisionSuccess
Mean (1–10)Mean (1–20)Mean (1–40)Mean (1–10)Mean (1–20)Mean (1–40)
HOG + HCLOSIB24.819.515.151.157.363.9
HOG + CLOSIB23.718.814.748.754.961.4
HOG23.118.514.346.652.659.4
Faster R-CNN22.518.815.350.359.969.8
ALBP20.317.214.547.556.466.1
ALBP + HCLOSIB18.115.713.542.250.960.1
ALBP + CLOSIB19.617.014.640.747.355.1
LBP13.512.210.834.144.456.1
LBP + CLOSIB9.38.47.523.430.639.0
LBP + HCLOSIB8.88.17.422.831.342.0

Share and Cite

MDPI and ACS Style

García-Olalla, O.; Alegre, E.; Fernández-Robles, L.; Fidalgo, E.; Saikia, S. Textile Retrieval Based on Image Content from CDC and Webcam Cameras in Indoor Environments. Sensors 2018, 18, 1329. https://doi.org/10.3390/s18051329

AMA Style

García-Olalla O, Alegre E, Fernández-Robles L, Fidalgo E, Saikia S. Textile Retrieval Based on Image Content from CDC and Webcam Cameras in Indoor Environments. Sensors. 2018; 18(5):1329. https://doi.org/10.3390/s18051329

Chicago/Turabian Style

García-Olalla, Oscar, Enrique Alegre, Laura Fernández-Robles, Eduardo Fidalgo, and Surajit Saikia. 2018. "Textile Retrieval Based on Image Content from CDC and Webcam Cameras in Indoor Environments" Sensors 18, no. 5: 1329. https://doi.org/10.3390/s18051329

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop