Next Article in Journal
CaACBIM: A Context-aware Access Control Model for BIM
Previous Article in Journal
Energy Efficiency and Renewable Energy Management with Multi-State Power-Down Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Crime Scene Shoeprint Retrieval Using Hybrid Features and Neighboring Images

1
School of Information Science and Technology, Dalian Maritime University, Dalian 116026, China
2
School of Physics and Electronics Technology, Liaoning Normal University, Dalian 116026, China
*
Author to whom correspondence should be addressed.
Information 2019, 10(2), 45; https://doi.org/10.3390/info10020045
Submission received: 14 December 2018 / Revised: 17 January 2019 / Accepted: 21 January 2019 / Published: 30 January 2019

Abstract

:
Given a query shoeprint image, shoeprint retrieval aims to retrieve the most similar shoeprints available from a large set of shoeprint images. Most of the existing approaches focus on designing single low-level features to highlight the most similar aspects of shoeprints, but their retrieval precision may vary dramatically with the quality and the content of the images. Therefore, in this paper, we proposed a shoeprint retrieval method to enhance the retrieval precision from two perspectives: (i) integrate the strengths of three kinds of low-level features to yield more satisfactory retrieval results; and (ii) enhance the traditional distance-based similarity by leveraging the information embedded in the neighboring shoeprints. The experiments were conducted on a crime scene shoeprint image dataset, that is, the MUES-SR10KS2S dataset. The proposed method achieved a competitive performance, and the cumulative match score for the proposed method exceeded 92.5% in the top 2% of the dataset, which was composed of 10,096 crime scene shoeprints.

1. Introduction

Shoeprint retrieval aims at retrieving the most similar shoeprints that were collected at different crime scenes, to help investigators to reveal clues about a particular case. In past decades, large numbers of crime scene shoeprint images were collected and recorded for analysis. When there was a new case, investigators could manually compare shoeprints derived at the crime scene with those collected from other crime scenes to reveal clues. It is really difficult and tedious to conduct this work for a huge number of degraded shoeprints. Therefore, it is necessary to propose a more efficient automatic shoeprint retrieval method.
In the past few years, many shoeprint image retrieval methods have been proposed, and they have demonstrated a good performance in forensic investigations. Most of the existing approaches focus on designing low-level features to highlight the most similar aspects of shoeprints. From the perspective of methodology, most of these shoeprint retrieval methods fall into two categories: (i) retrieval of shoeprints using holistic features [1,2,3,4,5,6,7,8,9,10,11,12,13]; and (ii) retrieval of shoeprints using local features [14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29]. Although both kinds of methods are able to search for shoeprint images with visually similar shoe patterns, their retrieval precision may vary dramatically among low quality crime scene shoeprint images. The two main reasons are detailed as follows.
The first reason may be that the descriptive capability of the low-level feature has its own deficiencies. Figure 1 shows the illustrative cases of a failure by either approach. Each pair of shoeprints in Figure 1 does not have the same shoe patterns. The local features (e.g., the Gabor feature) cannot distinguish between the visual patterns in Figure 1a, while it accurately handles the visual patterns in Figure 1b. On the other hand, the holistic features (e.g., the Fourier-Mellin feature) fail to make distinctions in Figure 1b, while it successfully handles shoeprints in Figure 1a, because they consider the overall layout of the images. Therefore, the complementary descriptive capability of the local and holistic features naturally inspires us to integrate their strengths to yield more satisfactory retrieval results.
The second reason may be that the shoeprints derived from crime scenes are usually misaligned, incomplete and degraded shoeprints affected by debris, shadows and other artifacts. Thus, it is difficult to retrieve crime scene shoeprints using pair-wise similarity computed from just two images. Figure 2 shows an example where the similarity estimation can potentially benefit from the neighbors. As shown in Figure 2, three samples A, B, and C are represented as filled circles, and the distance between A and B is equal to the distance between A and C. The feature similarity between A and B is equal to that between A and C, that is, S ( A , B ) = S ( A , C ) . The neighbors of the three samples are represented as circles. From the distribution of their neighbors in Figure 2, it is more reasonable to intuitively let S ( A , B ) < S ( A , C ) , because the neighbors around A and C are much more similar than those around A and B. The descriptive capability of the neighbors inspires us to use the neighborhoods to yield more satisfactory retrieval results.
Wang et al. [29] proposed a manifold ranking-based shoeprint retrieval method, which used opinion scores of shoeprint examples labeled by forensic experts to achieve a good performance. However, the method did not enhance the retrieval precision based on the two aspects stated above. Therefore, in this paper, we propose a shoeprint retrieval method to enhance the retrieval precision considering the following two aspects: (i) integrate the strengths of three kinds of low-level features to yield more satisfactory retrieval results; and (ii) utilize the information contained in the neighboring images to improve the performance of the shoeprint retrieval method.
The main contributions of the proposed method are as follows:
(1) We propose a hybrid feature in a coarse to fine manner from holistic, region and local views. The proposed method integrates the strengths of three kinds of low-level features to yield more satisfactory retrieval results;
(2) We propose a neighborhood-based similarity estimation (NSE) method, which utilizes the information contained in neighbors to improve the performance of a shoeprint retrieval method. The greatest difference, compared to the other existing shoeprint retrieval methods, is that it not only considers the relationship between every two shoeprints, but also the relationship between their neighbors;
(3) We propose a generic manifold based reranking framework, which can narrow the well-known gap between high-level semantic concepts and low-level features;
(4) The proposed method can work well for real crime scene shoeprint image retrieval. The cumulative match score is more than 92.5% in the top 2% of the database, which was composed of 10,096 real crime scene shoeprint images. The evaluation shows our method consistently improves the retrieval precision and compares favorably with the state-of-the-art.
The rest of the paper is organized as follows. Section 2 reviews related works on shoeprint retrieval. Section 3 presents the proposed method. Section 4 provides the experimental results and the analysis, followed by the conclusions in Section 5.

2. Related Works

According to the scope of representation, features roughly fall into two categories: holistic features and local features.
Methods in the holistic features category usually take the whole image into consideration when extracting features. Bouridane et al. [1] used a fractal-based feature to retrieve shoeprint images. They can handle high quality shoeprint images; however, this method is sensitive to variations in rotations and translations. The moment invariant features were used for shoeprint retrieval [2,3], and they can work well for complete shoeprints; however, partial shoeprints are not considered. Chazal et al. [4] and Gueham et al. [5,6] used the Fourier transform to analyze the frequency spectra of shoeprint images, but the methods are sensitive to partial shoeprints. Cervelli et al. [7,8,9] utilized the Fourier transform on the cropped shoeprint images to extract features in frequency domain. However, these methods are sensitive to geometry transformations. Alizadeh et al. [10] retrieved shoeprints by using a sparse representation method. They reported good performance, but their method is sensitive to variations in rotation and translation. Richetelli et al. [11] implemented and tested some shoeprint retrieval methods on a scene-like shoeprint database, that is, the phase-only correlation (POC) method, Fourier-Mellin transformation and scale-invariant feature transform (SIFT) method. Results show that the POC method has better performance than the Fourier-Mellin transformation and the SIFT method; however, the performances of these methods may drop considerably when applied to degraded crime scene shoeprints. Kong et al. [12,13] applied a convolutional neural network to extract multi-channel features, and computed the similarity score using the normalized cross-correlation method. They have achieved a good performance. However, their algorithm requires a large amount of computation.
Methods in the local feature category always divide shoeprint into different regions, and then extract features from these regions. Patil et al. [14] convolved shoeprint images with Gabor filters, and then divided the filtered images into non-overlapping blocks to extract local features for shoeprint retrieval. The method shows good performance for partial shoeprints generated from full shoeprints. Tang et al. [15,16] used an attributed relational graph (ARG) to represent the shoeprint. In the graph, nodes represent fundamental shapes in shoes, such as lines, circles, ellipses, and so on. They reported good performance on distortions and partial shoeprints. However, it is a challenge to handle crime scene shoeprints with random breaks and extrusions, which cannot be represented by above fundamental geometry shapes. Pavlouet et al. [17,18] applied the maximally stable extremal regions (MSER) feature to represent shoeprints. However, the performance may drop a lot when dealing with shoeprint images with noises and distortions. Kortylewsk et al. [19] presented a periodic pattern-based shoeprint image retrieval method. The method firstly detects periodic patterns of the shoeprint, and then evaluates the similarity through comparing the Fourier features of the periodic patterns. The algorithm can deal with shoeprints with periodic patterns. However, it is a challenge to handle degraded shoeprint images. Local interest point based methods can work well for clear shoeprints [20,21,22,23,24,25], but their performance may vary dramatically among crime scene shoeprints. The possible reasons may be that the crime scene shoeprints are highly degraded and randomly occluded, and there are many random extrusions, intrusions or breaks on the shoe patterns. Nevertheless, the local interest point based methods cannot work well on distinguishing the useful information from interferences. Kortylewski et al. [26,27] learned a compositional active basis model to each reference shoeprint, which was used to evaluate against other query images at testing time. The model can be learned well on high quality reference shoeprints. However, how to represent degraded crime scene shoeprint images remains a problem. Wang et al. [28] divided a shoeprint into a top region and a bottom region, and then extracted Wavelet-Fourier-Mellin transform-based features of the two regions for shoeprint retrieval. The method performs well for its invariant features and matching score estimation method. Wang et al. [29] proposed a manifold ranking shoeprint retrieval method that considers not only the holistic and region features but also the relationship between every two shoeprints. The method achieves a good performance on crime scene shoeprint images, but it neglects the effect of local features and the contribution of the neighboring shoeprints.

3. Method

3.1. Notations and Formulations

Let D = { d 1 , d 2 , , d N } + m denote a set of N shoeprint images, and U = q D = { u 1 , u 2 , , u N + 1 , u N + 1 } + m , in which q denotes the query shoeprint. We focus on finding a function f : U + that assigns to each shoeprint u i U a ranking score f i + , 0 f i 1 according to their relevance to the query shoeprint image q . Let f = [ f 1 , f 2 , , f K ] , and K = | U | = N + 1 .
Our motivation is to enhance the shoeprint retrieval precision from the following two perspectives: (i) integrate the strengths of three kinds of low-level features to yield more satisfactory retrieval results; and (ii) utilize the information contained in the neighboring images to improve the performance of the shoeprint retrieval method. Therefore, we have two constraints on the ranking score: (i) closer shoeprint images in multiple feature spaces should share similar ranking scores; and (ii) shoeprint images with similar neighboring shoeprints should share similar ranking scores.
We construct the cost function by employing the above two constraints on f . The shoeprint retrieval problem can be defined as an optimal solution of minimizing the following cost function.
f * = arg min f Q ( f ) = β 1 i = 1 K j = 1 K S i j ( 1 A i i f i 1 A j j f j ) 2 + β 2 i = 1 K j = 1 K W i j ( 1 B i i f i 1 B j j f j ) 2 + γ i = 1 K ( f i y i ) 2
where β 1 , β 2 and γ are the regularization parameters.
The first term weighted by β 1 is the neighborhood correlation term. Shoeprint images with similar neighbors should share similar ranking scores. S i j denotes the neighborhood based similarity between u i and u j , and A is a diagonal matrix, in which A i i = j = 1 K S i j . Intuitively, similarity is usually defined as the feature relevance between two images. But it is difficult to use low level features to describe the shoeprints more clearly, because crime scene shoeprints are usually highly degraded and also randomly partial. Moreover, the traditional image-to-image similarity measure is sensitive to noises. One feasible way to deal with this problem is to use the neighborhoods to provide more information. To this end, we propose a neighborhood-based similarity estimation (NSE) method which regards the neighbors of the images as their features, the more similar neighbors the images have, the higher similarity value they should share. Formally, for shoeprint images u i and u j , the neighborhood-based similarity between the two images can be defined as follows:
S i j = a S u m N ( u i ) u n N ( u j ) W m n | N ( u i ) | | N ( u j ) | + a C | N k ( u i ) N k ( u j ) | | N k ( u i ) N k ( u j ) |
where a S and a C are the weighted parameters, and a S + a C = 1 . W m n denotes the hybrid feature similarity between shoeprint image u m and u n . N ( u i ) = N k ( u i ) u i , N k ( u i ) denotes the k nearest neighbors of u i , which is acquired based on the hybrid feature similarity W i j . | | represents the cardinality of a set. N k ( u i ) denotes the k nearest neighbors of u i , and they are acquired based on the region feature similarity S r ( u i , u j ) that can be calculated according to Equations (7)–(19) in [28]. Here we defined S i j = 1 for u i = u j .
The second term weighted by β 2 is the smoothness term. The shoeprint images nearby in the feature space should share similar ranking scores. W i j denotes the hybrid feature similarity between u i and u j , and B is a diagonal matrix, in which B i i = j = 1 K W i j .
The third term weighted by γ is the fitting term. y = [ y 1 , y 2 , , y K ] T is a vector, in which y i = 1 , if u i is the query, and y i = 0 otherwise.
Equation (1) can also be generalized as a multiple similarity measures manifold ranking frame work, which can be formulated as follows:
[ f * , β * ] = arg min f , β Q ( f ) = p = 1 P i = 1 K j = 1 K β p W i j ( p ) ( 1 C i i ( p ) f i 1 C j j ( p ) f j ) 2 + γ i = 1 K ( f i y i ) 2   subject   to   p = 1 P β p = 1 ,   β p 0 .
where W ( p ) denotes the adjacency matrix calculated using the pth similarity measure, P denotes the number of similarity measures, and C ( p ) is a diagonal matrix, and C i i ( p ) = j = 1 K W i j ( p ) . β p represents the pth regularization parameter. Let β = [ β 1 , β 2 , , β P ] .

3.2. Solution

We solve the optimal question in Equation (3) by constructing a Lagrange function. To get an optimal regularization parameter β , we replace β p with β p q , where q > 1 . Therefore, the Lagrange function is defined as follows:
L ( β , α ) = p = 1 P i = 1 K j = 1 K β p q W i j ( p ) ( 1 C i i ( p ) f i 1 C j j ( p ) f j ) 2 + γ i = 1 K ( f i y i ) 2 + α ( p = 1 P β p 1 )
Letting L ( β , α ) β p = 0 , and L ( β , α ) α = 0 , we can get
q β p q 1 i = 1 K j = 1 K W i j ( p ) ( 1 C i i ( p ) f i 1 C j j ( p ) f j ) 2 α = 0
p = 1 P β p 1 = 0
Then, β p can be acquired as follows:
β p = ( 1 / i = 1 K j = 1 K W i j ( p ) ( 1 C i i ( p ) f i 1 C j j ( p ) f j ) 2 ) 1 / ( q 1 ) p = 1 P ( 1 / i = 1 K j = 1 K W i j ( p ) ( 1 C i i ( p ) f i 1 C j j ( p ) f j ) 2 ) 1 / ( q 1 )
Then, we update f by using the new β p . When β p is fixed, we can get
f * = arg min f , β Q ( f ) = p = 1 P i = 1 K j = 1 K β p q W i j ( p ) ( 1 C i i ( p ) f i 1 C j j ( p ) f j ) 2 + γ i = 1 K ( f i y i ) 2  
The matrix-vector formulation of the function is:
Q ( f ) = p = 1 P 1 4 β p q f T L ( p ) f + 1 2 γ ( f y ) T ( f y )
where L ( p ) = I C ( p ) W ( p ) C ( p ) , which represents a symmetric normalized Laplacian matrix.
Differentiation Q ( f ) with respect to f yields
( p = 1 P β p q L ( p ) + γ I ) f γ y = 0
The ranking score can be obtained as follows:
f * = ( p = 1 P β p q L ( p ) + γ I ) 1 γ y
The algorithm is summarized in Algorithm 1.
Algorithm 1. Solution to the retrieval problem
Input: The affinity matrix W ( p ) , the initial ranking score list y . Iteration number T, tuning parameter q .
Output: The final ranking score list f * .
1: Set r = 0 , set β ( 0 ) = [ 1 / P , 1 / P , , 1 / P ] , assign y to f ( 0 ) .
2: Compute the degree matrix C ( p ) and the Laplacian matrix L ( p ) = I C ( p ) W ( p ) C ( p )
3: Update ranking score f ( r ) = ( p = 1 P β p q L ( p ) + γ I ) 1 γ y
4: Update β ( r ) using Equation (7).
5: Let r = r + 1 . If r > T , quit iteration and output final ranking score list f * = f ( T ) , otherwise go to 3.

3.3. The Affinity Matrix Computation Mothod

In [28], a Wavelet-Fourier-Mellin transform and Similarity Estimation (WFSE) based method is proposed to compute the matching score. The WFSE method has been applied successfully in forensic practice when retrieving crime scene shoeprint images, but it does not take into consideration the local patterns of the shoeprint. Generally, our observation of objects usually is a continuously improving process from the whole to the parts and to the details. Inspired by this rule, we propose a hybrid holistic, region and local features to compute the matching score, which follows the rules of our observation to objects. We define the hybrid feature similarity as follows:
W ( i , j ) = b r S r ( u i , u j ) + b h S h ( u i , u j ) + b l S l ( u i , u j )
where b r , b h and b l denote the weighted parameters, and b r + b h + b l = 1 .
For a shoeprint image u i , the extraction process for its hybrid holistic, region and local features has following six main steps.
Step 1: Acquire and normalize the shoeprint image.
Step 1.1: Acquire the binarized shoeprint image. The shoeprint image is firstly split into a grid of cells, and then a thresholding method (e.g. Otsu’s method) is applied to each cell to extract sub shoeprints. Finally, morphological operations are utilized to eliminate small holes and smooth edges.
Step 1.2: Resolution and orientation normalization. The shoeprint images are rescaled to a predefined resolution measured in dots per inch (DPI). And then we normalize the shoeprint image by using the Shoeprint Contour Model (SPCM) proposed in [28].
Step 2: The normalized shoeprint image u i is divided into the top region and the bottom region, and they are denoted as S top ( i ) and S bottom ( i ) , respectively.
Step 3: The shoeprint image u i and its two regions S top ( i ) and S bottom ( i ) are decomposed at a specified number of levels by using the Haar Wavelet. We can acquire one approximation and three details. The coefficients can have the following forms:
F W ( u i ) = { F W ( u i ) ( l , h , v ) | 0 l L , h , v = 0 , 1 } F W ( S top ( i ) ) = { F W ( S top ( i ) ) ( l , h , v ) | 0 l L , h , v = 0 , 1 } F W ( S bottom ( i ) ) = { F W ( S bottom ( i ) ) ( l , h , v ) | 0 l L , h , v = 0 , 1 }
where L is the maximum level. To avoid merging the useful neighbor patterns, L should be able to meet the criterion: 2 L 1 D min , where D min represents the minimum distance between two neighbor patterns, which can be specified interactively.
Step 4: The Fourier-Mellin transform is applied on each wavelet coefficients to extract features.
Step 4.1: Calculate the Fourier magnitude of the pre-processed image by using the fast Fourier transform (FFT);
Step 4.2: Use a band passed filter proposed in [30] to weaken the effect of the noises, such as small holes, intrusions, extrusions and broken patterns;
Step 4.3: Perform the log-polar mapping of the filtered Fourier magnitude acquired in Step 4.2;
Step 4.4: Calculate the Fourier magnitude of the log-polar mapping calculated in Step 4.3 by using the FFT;
Step 4.5: The filtered Fourier-Mellin domain coefficients of F W ( u i ) are used as holistic features and those of F W ( S top ( i ) ) and F W ( S bottom ( i ) ) are used as region features. Here, we use F M W ( u i ) , F M W ( S top ( i ) ) and F M W ( S bottom ( i ) ) to denote the holistic and region features of the shoeprint u i .
Step 5: Construct Gabor filters. The Gabor filter in the spatial domain has the following form:
G ( x , y ) = 1 δ x δ y exp [ π ( x 2 δ x 2 + y 2 δ y 2 ) ] [ exp ( i 2 π f x ) exp ( π 2 2 ) ]
where x = x sin θ + y cos θ , and x = x cos θ y sin θ . We set the values of f , δ x and δ y as 0.458, 3 and 3, respectively. Here, we construct Gabor filters in eight orientations ( 0 , 22.5 , 45 , 67.5 , 90 , 112.5 , 135 , 157.5 ) according to Equation (14).
Step 6: Extraction of local features. A shoeprint image is convolved with the Gabor filters in eight orientations. Then each filtered shoeprint image is divided into non-overlapping blocks, and there are 8 blocks in each row and 16 blocks in each column. The standard variance σ θ ( m , n ) and the mean M θ ( m , n ) of the pixel intensities in each block across all filtered images are used as the local features. The local feature of the shoeprint image u i is defined as follows:
G W ( u i ) = { σ θ ( m , n ) , M θ ( m , n ) }
where θ { θ 1 , θ 2 , , θ 8 } , m = 1 , 2 , , 8 and n = 1 , 2 , , 16 .
For two shoeprint images u i and u j , the holistic feature similarity S h ( u i , u j ) between them is computed as follows:
S h ( u i , u j ) = ( F M W ( u i ) F M W ¯ ( u i ) ) ( F M W ( u j ) F M W ¯ ( u j ) ) | F M W ( u i ) F M W ¯ ( u i ) | | F M W ( u j ) F M W ¯ ( u j ) |
The regional feature similarity S r ( u i , u j ) between them is a weighted sum of correlation coefficients of both F M W ( S top ( i ) ) and F M W ( S bottom ( i ) ) . Please refer to Equations (7)–(19) in [28] for details about how to set the weights adaptively. The local feature similarity S l ( u i , u j ) between two images is computed as follows:
S l ( u i , u j ) = ( G W ( u i ) G W ¯ ( u i ) ) ( G W ( u j ) G W ¯ ( u j ) ) | G W ( u i ) G W ¯ ( u i ) | | G W ( u j ) G W ¯ ( u j ) |

4. Experiments

4.1. Experiment Configuration

4.1.1. Dataset

The experiments were conducted on two shoeprint datasets. One is the MUES-SR10KS2S dataset [28], and shoeprints in this dataset were collected from real crime scenes. The other is a public available dataset, named the FID-300 dataset [26].
The MUES-SR10KS2S dataset contains one probe set and one gallery set. The gallery set consists of 72 probe images, 432 synthetic versions of the probe images and 9592 crime scene shoeprints. Examples of crime scene shoeprint images in the dataset are shown in Figure 3. It can be seen that shoeprint images with same patterns differ greatly, due to the varying imaging conditions.
The FID-300 dataset consists of 300 probe shoeprints and 1175 gallery shoeprints. The probe shoeprint images were collected from crime scenes by investigators. The gallery shoeprints were generated by using a gelatine lifter on the outsole of the reference shoe, and then by scanning the lifters. The gallery shoeprints are of very high quality. Examples of shoeprints in FID-300 dataset are shown in Figure 4. Figure 4a shows one group of probe shoeprints, and Figure 4b shows their corresponding shoeprints in the gallery set.

4.1.2. Evaluation Metrics

The cumulative match score used in [31] is applied to evaluate the performance of the method, and it is defined as follows:
C M S ( n ) = 100 R n | P |
where | P | and R n denote the number of the probe images and the number of gallery images which match the probe images in the top n rank, respectively.

4.2. Performance Evaluation

4.2.1. Performance Evaluation of the Proposed Hybrid Features and the Proposed NSE Method

To test the performance of the hybrid holistic, region and local features, we used four kinds of features to retrieve images in the dataset and evaluated the performance of these features according to their cumulative match score. We also compared the performance of the proposed NSE method with that of the features. The first kind of feature is the holistic feature, and its matching score was computed according to Equation (16). The second kind of feature is the region feature, in our method, the Wavelet-Fourier-Mellin feature proposed in [28] was used as the region feature, and its matching score was computed according to Equation (8) in [28]. The third kind of feature is the local feature, and its matching score was computed according to Equation (17). The fourth kind of feature is the proposed hybrid features, and the matching score was computed according to Equation (12). The results are listed in Table 1. For single features, region features have better performance than both holistic and local features. For the proposed hybrid feature, its cumulative match score is improved 6.6% on average than that of [28] because of the strengths of three kinds of low-level features. The results also illustrate that the cumulative match score of the proposed NSE method is improved 10.4% on average than that of [28]. Figure 5 provides a visual illustration of the top 10 shoeprint images in the ranking lists of our method and the compared method. The results show that the proposed hybrid feature and NSE method outperform the work of [28] on the MUES-SR10KS2S database.

4.2.2. Comparison with the Traditional Manifold Ranking Method

Zhou et al. [32] provided a manifold based ranking (MR) algorithm, and the manifold regularization term of our proposed ranking cost function is based on their ideas. To evaluate the effectiveness of our method, we compared our method with the traditional manifold ranking method [32]. To ensure a fair comparison, the affinity matrixes used in both methods were the same ones calculated according to Equation (2). The cumulative match scores of the algorithms are listed in Table 2. The results show that the performance of our method is approximately 0.7% on average above that of Zhou et al. [32].

4.2.3. Comparison with the Manifold Ranking Based Shoeprint Retrieval Method

We compared the proposed method with the works of Wang et al. [29] which provided a manifold-based shoeprint retrieval algorithm. The method asks forensic experts to assign the shoeprint example an opinion score according to the similarity between the example and the query, in which the example denotes the shoeprint acquired at the same crime scene with the query shoeprint. To ensure a fair comparison, the vector y = [ y 1 , y 2 , , y K ] T used in our proposed method is same as that of [29], where if u i is the query or the shoeprint sample that has the similar shoe pattern with the query, y i = 1 ; else y i = 0 . The cumulative match scores of the algorithms are listed in Table 3. The experimental results show that the performance of our method surpasses that of [29]. The cumulative match score of ours is approximately 3.0% above that of [29] on average.

4.2.4. Comparison with the State-of-the-art Algorithms

We also compared the proposed method with the state-of-the-art shoeprint retrieval methods on the MUES-SR10KS2S dataset. The results are shown in Table 4. For some state-of-the-art methods do not release codes, the results listed in Table 4 are achieved by running the codes implemented by ourselves. In this section, to ensure a fair comparison, the vector y = [ y 1 , y 2 , , y K ] T used in [29] is the one detailed in Section 3.1. The experimental results show that the cumulative match score of top 2% of the proposed algorithm is more than 92.5%. It can also be found that the cumulative match score of the proposed method is improved 5% compared to the work of Wang et al. [28] on top 2% of the ranked list. The results show that some methods cannot have the similar performance as that reported in the literatures on the MUES-SR10KS2S dataset. We think the possible reasons may be that: (i) the quality and quantity of shoeprints in datasets are greatly different; (ii) the codes and experimental settings may not be the optimal ones.
Our hardware configuration consists of a 3.33-GHz central processing unit (CPU) with 8-GB random access memory (RAM). All of the methods are implemented with MATLAB codes. The mean average running time is applied to evaluate the running times.
M A T = i = 1 n q T ( i ) n q K
where i denotes the ith query image, T ( i ) represents the running time of the ith retrieval, n q denotes the number of query images, and K is the number of the gallery images.
To further verify the effectiveness of the proposed method, we compared our proposed method with the works of Kortylewski et al. [26], Wang et al. [29], Kong et al. [12], Kortylewski [27] and Kong et al. [13] on a public available database named FID-300 [26]. The results are listed in Table 5. The results of Kortylewski et al. [26] and Kong et al. [12] are borrowed from Figure 5 in Kong et al. [12]. The results show that our method achieves a good performance on the FID-300 dataset, and the cumulative match score of our method reaches more than 95.3% in the top 20% of the dataset. The results also show that there is an improvement of approximately 1.3% compared with the works of Kong et al. [13] in the top 20% of the ranked list. However, the cumulative match scores of Kong et al. [13] surpass those of ours in the top 1%, 5% and 15% of the ranking list. The main reason may be as follows. Most of the probe shoeprint images in FID-300 database are of small size, and some of them only provide a small periodical pattern patch to retrieve shoeprint images. Kong et al. [13] used a template matching method to search over both translations (with a stride of 2) and rotations, which can work well for the case that the query image is a small patch. However, our method does not consider this case.

4.3. Analysis and Discussion

In this section, we further analyze the effect of different components of the cost function on the ranking result, which includes the roles of the proposed NSE method and the hybrid feature similarity. The influence caused by each kind of low-level feature is also discussed.

4.3.1. Effectiveness of the proposed NSE method and the hybrid feature similarity

In this section, we conducted experiments to verify the effectiveness of the proposed NSE method and the hybrid feature similarity. We conducted two kinds of experiments by using the MR method with two different affinity matrixes. The first affinity matrix consists of the hybrid feature similarities calculated according to Equation (12), and the second affinity matrix is acquired by using the proposed NSE method. The cumulative match scores of the algorithms are listed in Table 6, and the cumulative match characteristic curves of the algorithms are shown in Figure 6. The cumulative match scores show that the performance of our method is approximately 4.1% above that of Zhou et al. [32] with our proposed hybrid feature similarity on average, and approximately 0.7% on average above that of Zhou et al. [32] with the affinity matrix computed by using our proposed NSE method. The results also show that the proposed method using both hybrid feature similarity and NSE method achieves a higher performance than the one using only the proposed NSE method or the hybrid feature similarity.

4.3.2. Effectiveness of Each Kind of Low Level Feature

We conducted experiments to verify the effectiveness of each kind of the low level features. In the experiments, we replaced the proposed hybrid feature with three kinds of features, and evaluated the performance of these features according to their cumulative match score. The first kind of feature was the hybrid feature of holistic and region feature, and we conducted this kind of experiments to verify the effectiveness of the local features. The second kind of feature was the hybrid feature of holistic and local feature, and we conducted this kind of experiments to verify the effectiveness of the region features. The third kind of feature was the hybrid feature of region and local feature, and we conducted this kind of experiments to verify the effectiveness of the holistic features. The results are listed in Table 7. The results show that the method with the proposed hybrid features achieves a higher performance than the one using only two of them. Furthermore, the proposed hybrid features can integrate the strengths of three kinds of low-level features, and each kind of features can help the proposed method to yield more satisfactory retrieval results.

5. Conclusions

In this paper, we proposed an effective shoeprint image retrieval method. In the proposed method, we enhance the retrieval precision from two perspectives: (i) integrate the strengths of three kinds of low-level features to yield more satisfactory retrieval results; and (ii) enhance the traditional distance-based similarity by leveraging the information embedded in the neighboring shoeprints. The greatest difference between the proposed method and the other existing shoeprint retrieval methods is that we not only consider the relationship between every two shoeprints, but also the relationships between their neighbors. Our proposed method can also be generalized as a generic reranking framework which utilizes information contained in the neighbors to improve the effectiveness of manifold based method. Experiments on the real crime scene datasets have shown that the performance of the proposed algorithm outperforms not only the traditional manifold ranking method, but also the state-of-the-art shoeprint retrieval algorithms.

Author Contributions

Y.W. and X.W. conceived the idea and designed the experiments; Y.W. performed the experiments; Y.W. and T.Z. analyzed the data; Y.W. wrote the original draft; X.W. and T.Z. reviewed and edited the manuscript; X.W. supervised this work.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bouridane, A.; Alexander, A.; Nibouche, M.; Crookes, D. Application of fractals to the detection and classification of shoeprints. In Proceedings of the International Conference on Image Processing, Vancouver, Canada, 10–13 September 2000; pp. 474–477. [Google Scholar] [CrossRef]
  2. Algarni, G.; Amiane, M. A novel technique for automatic shoeprint image retrieval. Forensic Sci. Int. 2008, 181, 10–14. [Google Scholar] [CrossRef] [PubMed]
  3. Wei, C.H.; Hsin, C.; Gwo, C.Y. Alignment of Core Point for Shoeprint Analysis and Retrieval. In Proceedings of the International Conference on Information Science, Electronics and Electrical Engineering (ISEEE), Sapporo, Japan, 26–28 April 2014; pp. 1069–1072. [Google Scholar] [CrossRef]
  4. Chazal, P.D.; Flynn, J.; Reilly, R.B. Automated processing of shoeprint images based on the Fourier transform for use in forensic science. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 341–350. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Gueham, M.; Bouridane, A.; Crookes, D. Automatic Classification of Partial Shoeprints Using Advanced Correlation Filters for Use in Forensic Science. In Proceedings of the International Conference on Pattern Recognition, Tampa, FL, USA, 8–11 December 2008; pp. 1–4. [Google Scholar] [CrossRef]
  6. Gueham, M.; Bouridane, A.; Crookes, D. Automatic recognition of partial shoeprints based on phase-only correlation. In Proceedings of the IEEE International Conference on Image Processing, San Antonio, TX, USA, 16 September–19 October 2007; pp. 441–444. [Google Scholar] [CrossRef]
  7. Cervelli, F.; Dardi, F.; Carrato, S. An automatic footwear retrieval system for shoe marks from real crime scenes. In Proceedings of the International Symposium on Image and Signal Processing and Analysis, Salzburg, Austria, 16–18 September 2009; pp. 668–672. [Google Scholar] [CrossRef]
  8. Cervelli, F.; Dardi, F.; Carrato, S. A texture based shoe retrieval system for shoe marks of real crime scenes. In Proceedings of the International Conference on Image Analysis and Processing, Catania, Italy, 8–11 September 2009; pp. 384–393. [Google Scholar] [CrossRef]
  9. Cervelli, F.; Dardi, F.; Carrato, S. A translational and rotational invariant descriptor for automatic footwear retrieval of real cases shoe marks. In Proceedings of the 18th European Signal Processing Conference, Aalborg, Denmark, 23–27 August 2010; pp. 1665–1669. [Google Scholar]
  10. Alizadeh, S.; Kose, C. Automatic retrieval of shoeprint images using blocked sparse representation. Forensic Sci. Int. 2017, 277, 103–114. [Google Scholar] [CrossRef] [PubMed]
  11. Richetelli, N.; Lee, M.C.; Lasky, C.A.; Gump, M.E.; Speir, J.A. Classification of footwear outsole patterns using Fourier transform and local interest points. Forensic Sci. Int. 2017, 275, 102–109. [Google Scholar] [CrossRef] [PubMed]
  12. Kong, B.; Supancic, J.; Ramanan, D.; Fowlkes, C. Cross-Domain Forensic Shoeprint Matching. In Proceedings of the 28th British Machine Vision Conference (BMVC), London, UK, 4–7 September 2017. [Google Scholar]
  13. Kong, B.; Supancic, J.; Ramanan, D.; Fowlkes, C. Cross-Domain Image Matching with Deep Feature Maps. arXiv, 2018; arXiv:1804.02367,2018. [Google Scholar]
  14. Patil, P.M.; Kulkarni, J.V. Rotation and intensity invariant shoeprint matching using Gabor transform with application to forensic science. Pattern Recognit. 2009, 42, 1308–1317. [Google Scholar] [CrossRef]
  15. Tang, Y.; Srihari, S.N.; Kasiviswanathan, H. Similarity and Clustering of Footwear Prints. In Proceedings of the 2010 IEEE International Conference on Granular Computing, San Jose, USA, 14–16 August 2010; pp. 459–464. [Google Scholar] [CrossRef]
  16. Tang, Y.; Srihari, S.N.; Kasiviswanathan, H.; Corso, J. Footwear print retrieval system for real crime scene marks. In Proceedings of the International Conference on Computational Forensics, Tokyo, Japan, 11–12 November 2010; pp. 88–100. [Google Scholar] [CrossRef]
  17. Pavlou, M.; Allinson, N.M. Automated encoding of footwear patterns for fast indexing. Image Vis. Comput. 2009, 27, 402–409. [Google Scholar] [CrossRef]
  18. Pavlou, M.; Allinson, N.M. Automatic extraction and classification of footwear patterns. In Proceedings of the 7th International Conference on Intelligent Data Engineering and Automated Learning, Burgos, Spain, 20–23 September 2006; pp. 721–728. [Google Scholar] [CrossRef]
  19. Kortylewski, A.; Albrecht, T.; Vetter, T. Unsupervised Footwear Impression Analysis and Retrieval from Crime Scene Data. In Proceedings of the Asian Conference on Computer Vision, Singapore, 1–5 November 2014; pp. 644–658. [Google Scholar] [CrossRef]
  20. Nibouche, O.; Bouridane, A.; Crookes, D.; Gueham, M.; Laadjel, M. Rotation invariant matching of partial shoeprints. In Proceedings of the International Machine Vision and Image Processing Conference, Dublin, Ireland, 2–4 September 2009; pp. 94–98. [Google Scholar] [CrossRef]
  21. Crookes, D.; Bouridane, A.; Su, H.; Gueham, M. Following the Footsteps of Others: Techniques for Automatic Shoeprint Classification. In Proceedings of the Second NASA/ESA Conference on Adaptive Hardware and Systems, Edinburgh, UK, 5–8 August 2007; pp. 67–74. [Google Scholar] [CrossRef]
  22. Su, H.; Crookes, D.; Bouridane, A.; Gueham, M. Local image features for shoeprint image retrieval. In Proceedings of the British Machine Vision Conference, University of Warwick, Coventry, UK, 10–13 September 2007; pp. 1–10. [Google Scholar]
  23. Wang, H.X.; Fan, J.H.; Li, Y. Research of shoeprint image matching based on SIFT algorithm. J. Comput. Methods Sci. Eng. 2016, 16, 349–359. [Google Scholar] [CrossRef]
  24. Almaadeed, S.; Bouridane, A.; Crookes, D.; Nibouche, O. Partial shoeprint retrieval using multiple point-of-interest detectors and SIFT descriptors. Integr. Comput. Aided Eng. 2015, 22, 41–58. [Google Scholar] [CrossRef]
  25. Luostarinen, T.; Lehmussola, A. Measuring the accuracy of automatic shoeprint recognition methods. J. Forensic Sci. 2014, 59, 1627–1634. [Google Scholar] [CrossRef] [PubMed]
  26. Kortylewski, A.; Vetter, T. Probabilistic Compositional Active Basis Models for Robust Pattern Recognition. In Proceedings of the 27th British Machine Vision Conference (BMVC), York, UK, 19–22 September 2016. [Google Scholar]
  27. Kortylewski, A. Model-based IMAGE Analysis for Forensic Shoe Print Recognition. Ph.D. Thesis, University of Basel, Basel, Switzerland, 2017. [Google Scholar]
  28. Wang, X.N.; Sun, H.H.; Yu, Q.; Zhang, C. Automatic Shoeprint Retrieval Algorithm for Real Crime Scenes. In Proceedings of the Asian Conference on Computer Vision, Singapore, 1–5 November 2014; pp. 399–413. [Google Scholar] [CrossRef]
  29. Wang, X.N.; Zhan, C.; Wu, Y.J.; Shu, Y.Y. A manifold ranking based method using hybrid features for crime scene shoeprint retrieval. Multimed. Tools Appl. 2016, 76, 21629–21649. [Google Scholar] [CrossRef]
  30. Reddy, B.S.; Chatterji, B.N. An FFT-based technique for translation, rotation, and scale-invariant image registration. IEEE Trans. Image Process. 1996, 5, 1266–1271. [Google Scholar] [CrossRef] [PubMed]
  31. Phillips, P.J.; Grother, P.; Micheals, R. Evaluation Methods in Face Recognition. In Handbook of Face Recognition; Jain Anil, K., Li Stan, Z., Eds.; Springer: New York, NY, USA, 2005; pp. 328–348. [Google Scholar]
  32. Zhou, D.; Weston, J.; Gretton, A.; Bousquet, O.; Scholkopf, B. Ranking on Data Manifolds. In Advances in Neural Information Processing Systems 16; Thrun, S., Saul, L.K., Schölkopf, B., Eds.; MIT Press: Cambridge, MA, USA, 2003; pp. 169–176. [Google Scholar]
Figure 1. Illustrative cases of a failure by either approach. (a) Holistic features may yield more satisfactory results than local features. (b) Local features may yield more satisfactory results than holistic features.
Figure 1. Illustrative cases of a failure by either approach. (a) Holistic features may yield more satisfactory results than local features. (b) Local features may yield more satisfactory results than holistic features.
Information 10 00045 g001
Figure 2. Example shows that the similarity estimation can benefit from neighbors.
Figure 2. Example shows that the similarity estimation can benefit from neighbors.
Information 10 00045 g002
Figure 3. Examples of crime scene shoeprint images in the MUES-SR10KS2S dataset. (a) The probe shoeprint and its counterparts from real crime scenes in the gallery set. (b) Corresponding binarized versions of the probe shoeprint and its counterparts in the gallery set.
Figure 3. Examples of crime scene shoeprint images in the MUES-SR10KS2S dataset. (a) The probe shoeprint and its counterparts from real crime scenes in the gallery set. (b) Corresponding binarized versions of the probe shoeprint and its counterparts in the gallery set.
Information 10 00045 g003
Figure 4. Examples of the probe shoeprints and their corresponding reference shoeprints in FID-300 dataset. (a) The probe shoeprints from crime scenes. (b) Corresponding shoeprints of the probe shoeprints in the gallery set.
Figure 4. Examples of the probe shoeprints and their corresponding reference shoeprints in FID-300 dataset. (a) The probe shoeprints from crime scenes. (b) Corresponding shoeprints of the probe shoeprints in the gallery set.
Information 10 00045 g004
Figure 5. Results of our method and the compared methods. The top 10 in the ranking list of our method and the compared methods are presented. Shoeprints encircled by the solid boxes are the query shoeprints, and those encircled by dotted boxes are their counterparts. (a) Wang et al. 2014 [28]. (b) The proposed hybrid feature. (c) The proposed neighborhood-based similarity estimation (NSE) method.
Figure 5. Results of our method and the compared methods. The top 10 in the ranking list of our method and the compared methods are presented. Shoeprints encircled by the solid boxes are the query shoeprints, and those encircled by dotted boxes are their counterparts. (a) Wang et al. 2014 [28]. (b) The proposed hybrid feature. (c) The proposed neighborhood-based similarity estimation (NSE) method.
Information 10 00045 g005
Figure 6. Cumulative match characteristic curves of both our method and the traditional manifold ranking method.
Figure 6. Cumulative match characteristic curves of both our method and the traditional manifold ranking method.
Information 10 00045 g006
Table 1. Comparison of the performance of our proposed method with the compared methods.
Table 1. Comparison of the performance of our proposed method with the compared methods.
ApproachesThe Cumulative Match Score of Top Percentage
0.1%0.2%0.3%0.4%0.5%1%2%
Holistic feature11.9%20.2%25.2%29.4%31.9%39.4%49.0%
Region feature [28]45.2%64.1%69.4%73.8%75.8%81.8%87.5%
Local feature11.3%17.1%19.0%20.2%21.2%26.8%32.7%
The proposed hybrid feature53.2%71.4%77.6%81.3%83.1%87.3%89.9%
The proposed NSE method53.8%80.0%84.1%84.7%86.1%89.7%92.3%
Table 2. Comparison with the traditional manifold ranking method.
Table 2. Comparison with the traditional manifold ranking method.
ApproachesThe Cumulative Match Score of Top Percentage
0.1%0.2%0.3%0.4%0.5%1%2%
Manifold ranking method [32]52.4%79.6%83.9%85.1%86.3%89.5%91.9%
Our method53.6%81.0%84.7%85.9%86.7%89.5%92.5%
Table 3. Comparison with the manifold ranking based shoeprint retrieval method.
Table 3. Comparison with the manifold ranking based shoeprint retrieval method.
ApproachesThe Cumulative Match Score of Top Percentage
0.1%0.2%0.3%0.4%0.5%1%2%
Wang et al. 2016 [29]52.6%76.0%81.8%84.5%85.5%90.1%93.5%
Our method54.6%81.7%86.5%87.9%88.9%90.9%94.8%
Table 4. Comparisons with the state-of-the-art algorithms on the MUES-SR10KS2S dataset.
Table 4. Comparisons with the state-of-the-art algorithms on the MUES-SR10KS2S dataset.
MethodsPerformance Reported in the LiteraturePerformance on Our Dataset
PerformanceGallery Set DescriptionThe Cumulative Match Score of Top 2%Mean Average Running Time(ms)
Kortylewski et al. 2014 [19]85.7%@20%#R:1,17538.64%0.5696
Almaadeed et al. 2015 [24]68.5%@2.5%#R:40034.7%14.2
Wang et al. 2014 [28]87.5%@2%#S:10,09687.5%17.8
Wang et al. 2016 [29]93.5%@2%#S:10,09691.1%17.8238
Kong et al. 2017 [12]92.5%@20%#R:1,17545.6%186.0
Our method_#S:10,09692.5%18.2
Table 5. Comparisons with the state-of-the-art shoeprint retrieval algorithms on the FID-300 dataset.
Table 5. Comparisons with the state-of-the-art shoeprint retrieval algorithms on the FID-300 dataset.
MethodThe Cumulative Match Score of the Top Percentage
1%5%10%15%20%
Kortylewski et al. 2016 [26]22.0%47.5%58.0%67.0%71.0%
Wang et al. 2016 [29]67.9%81.3%86.3%91.3%94.0%
Kong et al. 2017 [12]73.0%82.5%87.5%91.0%92.5%
Kortylewski 2017 [27]58.0%72.0%79.0%81.5%84.0%
Kong et al. 2018 [13]79.0%86.3%89.0%91.3%94.0%
Our method71.8%81.7%87.3%92.0%95.3%
Table 6. Comparison with the traditional manifold ranking method.
Table 6. Comparison with the traditional manifold ranking method.
ApproachesThe Cumulative Match Score of Top Percentage
0.1%0.2%0.3%0.4%0.5%1%2%
Hybrid feature similarity+MR53.2%71.8%78.0%81.3%82.7%87.5%90.5%
NSE+MR52.4%79.6%83.9%85.1%86.3%89.5%91.9%
Our method (Hybrid feature similarity +NSE+ MR)53.6%81.0%84.7%85.9%86.7%89.5%92.5%
Table 7. Performance of the proposed method with the three kinds of features.
Table 7. Performance of the proposed method with the three kinds of features.
ApproachesThe Cumulative Match Score of Top Percentage
0.1%0.2%0.3%0.4%0.5%1%2%
Hybrid feature of holistic and region48.8%73.0%76.6%80.4%81.3%86.7%89.3%
Hybrid feature of holistic and local33.9%44.4%50.2%54.4%56.7%70.4%77.2%
Hybrid feature of region and local48.6%72.2%75.8%78.8%83.9%88.5%89.5%
The proposed hybrid feature53.6%81.0%84.7%85.9%86.7%89.5%92.5%

Share and Cite

MDPI and ACS Style

Wu, Y.; Wang, X.; Zhang, T. Crime Scene Shoeprint Retrieval Using Hybrid Features and Neighboring Images. Information 2019, 10, 45. https://doi.org/10.3390/info10020045

AMA Style

Wu Y, Wang X, Zhang T. Crime Scene Shoeprint Retrieval Using Hybrid Features and Neighboring Images. Information. 2019; 10(2):45. https://doi.org/10.3390/info10020045

Chicago/Turabian Style

Wu, Yanjun, Xinnian Wang, and Tao Zhang. 2019. "Crime Scene Shoeprint Retrieval Using Hybrid Features and Neighboring Images" Information 10, no. 2: 45. https://doi.org/10.3390/info10020045

APA Style

Wu, Y., Wang, X., & Zhang, T. (2019). Crime Scene Shoeprint Retrieval Using Hybrid Features and Neighboring Images. Information, 10(2), 45. https://doi.org/10.3390/info10020045

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop