Next Article in Journal
ArtiLock: Smartphone User Identification Based on Physiological and Behavioral Features of Monosyllable Articulation
Previous Article in Journal
Spectral Analysis Methods for Improved Resolution and Sensitivity: Enhancing SPR and LSPR Optical Fiber Sensing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Bag-of-Visual-Words and FeatureWiz Selection for Content-Based Visual Information Retrieval

1
Faculty of Computers and Artificial Intelligence, Sohag University, Sohag 82524, Egypt
2
Institute for Information Technology and Communications (IIKT), Otto-von-Guericke-University Magdeburg, 39106 Magdeburg, Germany
*
Authors to whom correspondence should be addressed.
Sensors 2023, 23(3), 1653; https://doi.org/10.3390/s23031653
Submission received: 29 October 2022 / Revised: 30 January 2023 / Accepted: 30 January 2023 / Published: 2 February 2023
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Recently, content-based image retrieval (CBIR) based on bag-of-visual-words (BoVW) model has been one of the most promising and increasingly active research areas. In this paper, we propose a new CBIR framework based on the visual words fusion of multiple feature descriptors to achieve an improved retrieval performance, where interest points are separately extracted from an image using features from accelerated segment test (FAST) and speeded-up robust features (SURF). The extracted keypoints are then fused together in a single keypoint feature vector and the improved RootSIFT algorithm is applied to describe the region surrounding each keypoint. Afterward, the FeatureWiz algorithm is employed to reduce features and select the best features for the BoVW learning model. To create the codebook, K-means clustering is applied to quantize visual features into a smaller set of visual words. Finally, the feature vectors extracted from the BoVW model are fed into a support vector machines (SVMs) classifier for image retrieval. An inverted index technique based on cosine distance metric is applied to sort the retrieved images to the similarity of the query image. Experiments on three benchmark datasets (Corel-1000, Caltech-10 and Oxford Flower-17) show that the presented CBIR technique can deliver comparable results to other state-of-the-art techniques, by achieving average accuracies of 92.94%, 98.40% and 84.94% on these datasets, respectively.

1. Introduction

An image is more than thousands of words and is the cornerstone of visual media representation. With the tremendous development in information technology, computer engineering, and smartphone, the number of shared images over networks has increased rapidly [1,2]. Thus, the world needs a robotic way to access and represent visual data concerning content rather than meta-data.
Computer interpretation of the contents of an image is less trivial. A machine’s vision is an extensive matrix of numbers, and it has no idea of the thoughts, knowledge, or meaning of the image it tries to convey. To understand the contents of an image, apply image classification and retrieval that utilize machine learning algorithms’ tasks to extract meaning from an image. This process could be as simple as assigning a label to what the image contains. Searching for an image based on its content or example is known as CBIR or Query By Image Content (QBIC). Many modern applications in hospitals, factories, geographical information systems, remote sensing, cultural heritage, mining, education, home entertainment, journalism and advertising, fashion and interior design, crime prevention, military, medical diagnosis, web searching, etc., use CBIR models to make their systems more intelligent and robust [3].
The first search by an image system was built by IBM [4]. After that, CBIR gained acceptance in many areas of computer vision, becoming crucial in producing robust techniques. Many research papers overviewed CBIR [5,6,7,8,9]. Recently, the image features have been based on color, texture, and shape to quantify image content [10,11,12]. Then, various algorithms were developed to describe the features of the image. At the highest level, a feature is the interesting regions of an image that are both unique and recognizable. The process of finding and describing the interesting regions of an image is known as feature extraction, and it is broken down into two phases: Keypoint Feature Detection (KFD) and Feature Description (FD). The first phase is to detect points that determine the interesting regions of an image. These regions may be edges, corners, or blobs where the point intensities are approximately uniform. The algorithms that can detect these interesting regions are Keypoint Feature Detectors (KFDs). The second phase is to describe and quantify the region of the image surrounding each interesting keypoint. To describe a region of an image, extract the features vectors of each keypoint called a local feature. Only the local neighborhood surrounding the keypoint is included in the computation of the descriptor. The algorithms that apply this phase are called FDs.
Feature-level fusion is a critical stage in feature preparation. The advantage of feature fusion lies in two aspects: first, it can derive the most discriminatory information from the original multiple feature sets involved in fusion; second, it can eliminate redundant information resulting from feature set correlation, allowing for real-time decision-making. In other words, feature fusion can derive and obtain the most effective and low-dimensional feature vector sets that will help with the final conclusion. Although the merits and importance of feature fusion in the CBIR applications, it has some challenges where it must run some experiments on the dataset to select the most appropriate features to fuse between them. Moreover, it should tune the appropriate percentage of features which will be used in the learning stage to avoid the overfitting of data in the learning stage. The feature fusion between the KFDs introduced in this paper is more effective than the recent fusion techniques between the FDs, where the extracted features at the fusion between KFDs are less than in the case of the fusion between FDs, which decreases the probability of overfitting and presenting more accurate results. Furthermore, the KFDs fusion has less computational complexity than the FDs fusion.
There are many techniques for CBIR. However, the outcomes techniques obtained using a single feature type are unsatisfactory. Multiple combined feature descriptors are frequently used to achieve better outcomes, but quickly and effectively searching a database for a relevant query image is still challenging. Moreover, the computational complexity is higher for the existing CBIR systems. To overcome the challenges mentioned above, besides giving better results; this paper introduced four main contributions to improving the CBIR systems, first, comparing the strongest nine KFDs parallel with the seven robust FDs. The comparison is to select the best two KFDs with the most convenient FDs. This contribution produces effective low-dimensional Hybrid Keypoint Features Vector (HKFV) that reduce the next steps’ complexity, cost, and time. The second contribution is using a new feature selection algorithm (i.e., FeatureWiz selection) to select the powerful features from the extracted features for the next clustering step. This contribution determines the overall powerful features extracted from the dataset that reduces the time complexity and cost for the clustering stage. Furthermore, it decreases the data overfitting probability at the learning stage. The third contribution is applying a feature selection algorithm on the BoVW to select adequate visual words in the learning and sorting stage, called BoVW FeatureWiz Selection (BFS). This contribution improves the learning stage and decreases the data overfitting probability. It also contributes to reducing the time complexity and cost of similarity matching and sorting. The fourth contribution is the new Keypoint Feature Fusion Visual Words (KFFVW) framework. The rest of the paper is structured as follows: Section 2 reviews the literature related to CBIR. Section 3 discusses in detail the pipeline of the proposed adaptive KFFVW framework, whereas the experimental results are illustrated and evaluated in Section 4. Section 5 concludes the research paper.

2. Related Work

CBIR has been an active growing research area for the last three decades [13]. Feature extraction is the baseline that quantifies a dataset in CBIR by extracting features from each image in the collection. The feature extraction step is the foundation of adaptive image classification, and its success influences the classification performance. Feature extraction is heavily used in computer vision applications to rank the similarity between images. The feature extraction technique may be carried out automatically in the CBIR, significantly reducing the challenges of the pure annotation-based approach in Text-Based Image Retrieval (TBIR). Nilsback et al. [14] proposed a visual vocabulary that explicitly represents color, texture, and shape to distinguish one flower from another in 17 flower species. Latha et al. [15] introduced a new color-image-descriptor through feature fusion between HSV color statistical feature and the multichannel LBP-oriented color descriptor for an enhanced hybrid CBIR system. Sadek et al. [16,17] proposed an architecture of the CBIR system using the neural spline model to employ a cubic-spline activation function to reduce the semantic gap between the high-level semantic human vision concepts and low-level visual features. Specific texture approaches, such as the Gabor features, the Curvelet, and the Wavelet packets, are used in CBIR to reduce the semantic gap [18,19,20].
Ghahremani et al. [21] proposed an image feature integration technique based on the Local Binary Pattern (LBP) descriptor and the Scale Invariant Feature Transform descriptor (SIFT) as a local invariant descriptor, and the Histogram of Oriented Gradients (HOG) as a global feature to produce the Bag-of-Features (BoF). The SIFT and LBP detect feature keypoints, while the HOG produces powerful features for the background clutter. The results of [21] showed that the most accurate CBIR was achieved using a combination of SIFT and LBP.
The BoVW is a common effective technique in image retrieval tasks. Yousuf et al. [22] proposed a novel method based on the SIFT features and a new weighting scheme. A fuzzy representation was used to index images with a more stable signature in the used method. Alkhawlani et al. [23] used the SURF and the SIFT as local invariant descriptors and utilized k-means for clustering the features. Support Vector Machines (SVMs) were used as an effective classifier for the learning and classification stage. Andy et al. [24] used the Gaussian Mixture Model (GMM) to construct a Visual Words Codebook (VWC). Moreover, multiclass SVMs were proposed to classify images using Hellinger’s kernel, chi-square kernel, and linear kernel functions for similarity metrics. The color histograms of the query image and images of the same class were used to compute similarity distance. Mehmood et al. [25] proposed a novel Weighted Average of Triangular Histograms (WATH) technique to improve the performance of visual vocabulary in BoVW. If the vocabulary size is too small, it will not represent all image patches, while the vocabulary size is too large, the overfitting issue will arise. The WATH technique decreases this issue on the larger size of the codebook.
Khan et al. [26] proposed a feature fusion vector by merging Local Tetra Angle Patterns and RGB color features. They also optimized SVMs using Genetic Algorithms (GA) for better image classification. They used the Chi-square quadratic distance function to measure the similarity between the query image and the dataset images. Alarcao et al. [27] proposed an Expert, dynamic Late Feature (ExpertosLF) fusion system for CBIR based on online learning via end-user relevance feedback. At each query, ExpertosLF determines the contribution of each feature collection in the ensemble for the subsequent queries based on the advantage of the user’s feedback.

3. The Proposed Methodology

Lastly, CBIR has many techniques, but each has its specific process. This work focused on the recent techniques that depend on BoVW. The latest fusion techniques in BoVW considered the fusion between feature descriptors. The suggested method improves the extracted features by fusing the keypoint features, and it is then described by feature descriptor algorithms rather than the direct fusion between the feature descriptors.
The proposed method depends on extracting the FAST and the SURF keypoints separately from each image. Then, these keypoints are fused in a single keypoint feature vector. The RootSIFT algorithm is applied to describe the region surrounding each keypoint. A large number of feature vectors characterizes each image. Afterwards, a feature selection algorithm is used to select the appropriate feature vectors from all images as a sample for the clustering step. A k-means clustering algorithm is applied to formulate feature vectors into clusters. Each cluster is a group of feature vectors whose center represents a visual world. The combination of the visual words constructs the codebook. Then, vector quantization is applied to this codebook to describe each image in one dimension feature vector. The output of the vector quantization is a separate feature vector for each image in the training dataset.
The query process requires comparing the query features to every feature vector in the dataset. It is a linear operation O(N). As the size of the dataset increases, the time it takes to perform a query will increase. Building an inverted index of the BoVW to optimize the data structures of the indexed features breaks down the query process to sub-linear O(log(N)). Figure 1 shows the block diagram of the proposed method.

3.1. Feature Extraction

Before the feature extraction phase, a preprocessing step is performed for computational efficiency. The images in the collection are first converted to HSV-color format and then down-sampled to a fixed resolution of 320 × 320 pixels. Feature extraction in CBIR is the process of representing a broad set of images by applying an adequate image descriptor to the keypoints of images to extract features from each image in the dataset. One of the most influencing factors of the BoVW is the feature extraction process, which contributes to constructing powerful visual words. Keypoints detectors fusion in this paper contributes to the success of the feature extraction process for a better BoVW. The feature extraction stage is the baseline of the adaptive image classification, and its success determines the success of the overall classification performance [28]. After running sufficient experiments and validations, the FAST and SURF keypoint detectors were selected to detect the interest points in an image. The Root SIFT descriptor was chosen to describe those points.

3.1.1. Speeded-Up Robust Features (SURF)

Bay et al. [29] provided a quick and reliable SURF algorithm for local, similarity invariant representation and comparison of images. The SURF algorithm key is its ability to compute operators quickly using box filters, enabling real-time applications such as CBIR, tracking, and object detection. SURF is composed of a two-step Fast Hessian keypoint detector and feature description. In this paper, a single first step is utilized to identify keypoints (i.e., points of interest) in an image at locations where the determinant of the Hessian matrix has a maximum value. In the second step, feature description is implemented using the RootSIFT algorithm. In the first step, the Hessian matrix performs well in calculation speed and accuracy, which is why SURF utilizes it. The Hessian of a given pixel at point X = (x,y) is defined as follows:
H ( f ( x , y ) ) = 2 f x 2 2 f x y 2 f x y 2 f y 2
For adapting to any scale, the image is filtered by a Gaussian kernel. The Hessian matrix H (x, σ ) in a given image I can be defined at a position x and scale σ as follows:
H ( f ( x , y ) ) = L x x ( x , σ ) L x y ( x , σ ) L x y ( x , σ ) L y y ( x , σ )
where,
L x x ( x , σ ) = I ( x ) 2 x 2 g ( σ ) , L x y ( x , σ ) = I ( x ) 2 x y g ( σ ) , L y y ( x , σ ) = I ( x ) 2 y 2 g ( σ ) where L x x ( x , σ ) is the convolution of the Gaussian second-order derivative with the image I in point x, and similarly for L x y ( x , σ ) and L y y ( x , σ ) , g ( σ ) as scale σ . The discretized second-order Gaussian derivatives in the yy and y directions are approximated by weighted box filters in Figure 2.
To quickly and accurately determine the sum of values in a given image or a rectangular portion of a grid, utilize the integral image. It can be used to determine the average intensity inside a particular image. They make it possible to compute box-type convolution filters quickly. The total pixels in the input image I inside the rectangular region bounded by the origin and x is represented by the entry of an integral image I s u m ( x ) at a location x = ( x , y ) is given by:
I ( x ) = i = 0 i x j = 0 j x I ( i , j )
Convolution with a Gaussian kernel must be used to determine the Hessian matrix’s determinant and the second-order derivative. SURF uses box filters to enhance the approximation (i.e., convolution and second-order derivative). These approximative second-order Gaussian derivatives can be calculated using an integral image regardless of size at a much lower computational cost, which is a part of why SURF is fast. Applying box filters of various sizes in SURF allows for scale-space implementation. Therefore, rather than iteratively shrinking the image size, the scale space is utilized by upscaling the filter size. A 9 × 9 size filter with scale s = 1.2 (i.e., equivalent to Gaussian derivatives with σ = 1.2) serves as the box filter’s initial scale layer. The subsequent layers are created by filtering the image with progressively larger masks while accounting for integral images’ discrete character and their unique filter structure. This yield filters of sizes 9 × 9, 15 × 15, 21 × 21, and 27 × 27.
The Non-Maximum Suppression (NMS) technique in a 3 × 3 × 3 neighbourhood is used to localize interest points in the image. The Brown et al. approach [30] is then used to interpolate the maxima of the determinant of the Hessian matrix in scale and image space. The scale difference between each octave’s first layers is more extensive, making scale space interpolation crucial. So the NMS technique can be used to locate the keypoints over an image scale. For achieving a good approximation, the determinant is weighted to exclude keypoints having low contrast or points lying on edges or being near to edges:
d e t ( H a p p r o x ) = D x x D y y ( ω D x y ) 2
where D y y and D x y are the approximated and discrete kernels for L y y and L x y , respectively. The ω term is theoretically sensitive to scale, but it can be assumed almost constant at 0.9 as Bay [29] suggested.

3.1.2. Features from Accelerated Segment Test (FAST)

The FAST idea is based on the corner detector known as the Smallest Univalue Segment Assimilating Nucleus (SUSAN). The brighter and darker pixels next to each other are determined using the center of a circular area. Only the pixels on the discretized circle that characterize the segment, not the entire circle, are evaluated. Sixteen pixels must be compared to the nucleus value for a full accelerated segment test as shown in Figure 3. Therefore, the value S defines the detected corner’s greatest angle. The corner detector is more repeatable when S is kept as high as possible. When comparing the brightness of the nucleus with the value of a pixel on the circular pattern, the Accelerated Segment Test (AST) uses a minimal difference threshold (t). This setting controls the sensitivity of the corner reaction. A lower t-value produces corners with smoother gradients which are fewer but stronger corners, whereas a big t-value only produces weak corners. When comparing FAST in [31,32] to other corner detectors such as the Harris detector (HARRIS), Difference of Gaussian (DoG), or SUSAN, it is demonstrated that the AST with S = 9 or 16 has good repeatability. A corner detector’s repeatability is a quality criterion that assesses a method’s ability to identify similar corners of a scene from various points.

3.1.3. Hybrid Keypoint Features Vector (HKFV)

As shown in Table 1, Table 2 and Table 3 the feature extraction process requires a longer time and higher hardware than other CBIR processes. Increasing the number of features will increase the time needed for other steps depending on the number of features in this step. Therefore, avoiding expanding the number of feature detectors or feature descriptors in the feature fusion stage. In our work, the output keypoints of each FAST and SURF were concatenated, only including the unique key points to develop the HKFV. It is essential to mention that FAST or SURF alone cannot be better keypoint features for images of similar color, texture, or objects as the HKFV. Each dataset image has a single HKFV that represents the interesting salient regions of an image. The images of a dataset do not have the same length for HKFV, but the output of describing each keypoint in HKFV is a feature vector of the same dimension. The selection of (FAST + SURF) feature detectors incorporated in our work is based on two main reasons, scientific explanation and experimental results.
Firstly, the scientific explanation. In many computer vision tasks, the FAST is a corner-detection technique that might extract feature points and then track and map objects. The FAST corner detector’s high computing efficiency may be its most advantageous feature. Furthermore, using machine learning techniques, faster and more efficient computations may be made by computer vision tasks using FAST detectors. The FAST detector is quicker than the difference of Gaussians (DoG), employed by the SIFT, SUSAN, Harris detectors, and many other popular feature extraction techniques. In addition, FAST extracts more keypoints than other detectors. Additionally, when machine learning techniques are used, the outcomes are better.
The SURF method is a quick and reliable approach for local, similarity-invariant representation and comparison of images. The SURF approach’s primary appeal is its ability to compute operators quickly using box filters, opening the door to real-time applications, including image matching, object identification, image registration, and classification. The SIFT feature detector stage served as some of the inspiration for SURF authors. The authors of SURF say that the standard version is more resistant to various image alterations than SIFT and is many times quicker than SIFT.
On the objective of feature detection, ORB outperforms SIFT and is better than SURF while being nearly two orders of magnitude quicker. The well-known FAST keypoint detector and the BRIEF description are the foundation for ORB. Both methods are appealing due to their high effectiveness and low expense.
Secondly, the experimental results. The experimental comparisons of the most vital nine Keypoint Feature Detectors (KFDs) parallel with the seven robust Feature Descriptors (FDs) are implemented based on the BOVW technique, as shown in Figure 4. These results indicate that the top two keypoint detectors are FAST and SURF. Furthermore, the Fast or SURF outperforms the ORB feature detectors.
Based on the previous scientific explanation and the experimental results, it was very interesting to find results of the best combination between SIFT, FAST, SURF, and ORB to produce the HKFV, but the experimental results shown in Table 3 indicate that the combination FAST+SURF gives better results and time performance than SIFT+ORB or SIFT+ORB+FAST+SURF. That is because the BOVW technique depends not only on the image keypoints and image features but also on the image patches that represent the visual words.
To form the HKFV of FAST+SURF, Let Image I is represented as:
I = ( P ( x , y ) )
where P(x,y) represents a pixel at the position (x,y). The FAST keypoints (F) and The SURF keypoints (S) are detected from the image and are represented as
F = { f 1 , f 2 , , f i }
S = { s 1 , s 2 , , s j }
where f 1 to f i and s 1 to s j represent the keypoints detected by FAST and SURF, respectively. While i and j are the numbers of those keypoints. To form the unique keypoints ( F S k ) of the two sets F k and S k find the union between them
F S = F S
so the set of FS k keypoints could be represented as
F S = { F S 1 , F S 2 , , F S k }
where k represents the number of unique keypoints of FAST and SURF.

3.1.4. Scale Invariant Feature Transform Descriptor (SIFT)

The SIFT feature descriptor algorithm requires a set of input keypoints. Then, for each of the input keypoints, SIFT takes the 16 × 16 pixel region surrounding the center pixel of the keypoint region, then divide the 16 × 16 pixel region into sixteen 4 × 4 pixel windows. For each of the 16 windows, compute the gradient magnitude and orientation. After that, construct an 8-bin histogram for each 4 × 4 pixel window. Then collect all 16 of these 8-bin orientation histograms and concatenate them to produce a feature vector of 16 × 8 = 128 dimension. It was, lastly, normalizing the entire feature vector.

3.1.5. RootSIFT Descriptor

There are two stages to local feature descriptions. In the first stage, intriguing and conspicuous points of an image need to be detected. These points, known as keypoints, might be edges or corners in an image. This stage is determined using HKFV. The region surrounding each keypoint in an image must first be extracted and quantified once the collection of keypoints has been determined. Since only the local neighborhood region around the keypoint is considered in the descriptor computation, the feature vector associated with a keypoint is referred to as a feature or local feature. It must select a robust feature descriptor algorithm to quantify and describe the surrounding region for each keypoint in HKFV. Arandjelovic et al. [33] introduced a simple extension to SIFT called RootSIFT, which can be used to increase object recognition accuracy, quantization, and retrieval accuracy. The following simple steps are to extend SIFT to RootSIFT.
(i)
Compute the SIFT descriptors.
(ii)
L1-normalize each SIFT vector.
(iii)
Take the square root of each element in the SIFT vector.
In our work, we decided on the RootSIFT feature descriptor for this task. This selection was based on experimental reasons. The experimental comparisons of the most vital nine KFDs parallel with the seven robust FDs shown in Figure 4 indicate that the RootSIFT descriptor gives the highest results with each FAST and SURF detector.

3.2. FeatureWiz Selection

When building predictive models, feature variables are crucial. Many features are undesirable since they could produce overfitting, which would cause the model to resemble the training data closely. Additionally, a lot of characteristics will increase the size of the search space for the problem, which is known as the curse of features. A selection features approach could be employed to give a relevant score for each feature variable and determines which features are essential for predicting the target variable. The feature selection model was down-sampled into two levels. In the first one, the algorithm selects the unique features, approximately 25% to 50% of the extracted features. The second level uses the open-source Python package FeatureWiz, a quick and effective technique to identify essential features in a dataset regarding the target variable. The FeatureWiz processes operate in two stages. The first stage is the Searching for the Uncorrelated List of Variables (SULoV), which identifies a pair of variables beyond an externally passed correlation threshold, so they are highly correlated. After determining the pairs, the Mutual Information Score (MIS), which gauges the relationship between two variables, is calculated for each pair. The pair of variables with the lowest correlation and the highest MIS scores are then considered. The variables chosen from SULOV are recursively sent through XGboost in the second stage (Recursive XGBoost), which aids in identifying the best features by skipping the most similar data. One of the optional FeatureWiz parameters is the percentage of the output features from the total input features. The FeatureWiz Percentage Parameters (FPP) will be tuned based on various experiments to select the feature percentage that gives the best retrieval results. The chosen features percentage for the clustering could be calculated as
R R = C R D
where R R represents the selected features, C is the value of the percentage (0 C 1), and R D is the total extracted features.

3.3. Clustering to Visual Words

It is common to group features in an image to give them meaning. K-Means clustering is the most common clustering algorithm for this grouping. Clustering is splitting a large number of data points into a small number of clusters. Data points are assigned to the closest centroid to build a cluster. Therefore, data points in one cluster are easier to compare with data points in other clusters. The center of each cluster is called a visual word. The combination of visual words constructs the vocabulary. The following Figure 5 shows the K-Means clustering model. After K-Means clustering is applied, the output visual words vocabulary could be denoted as follows:
{ R R } K M e a n s V W = { v w 1 , v w 2 , , v w c }
where v w 1 to v w c are visual words feature vectors.

3.4. Features Quantization

To quantize feature vectors of an image to a single BoVW histogram vector, the Cosine distance matrix D between the total extracted features R D for each image and the visual words vector V W must be constructed. The CDM between the vectors F and W is calculated as follows:
d F , W = C o s ( F , W ) = F · W F W
The Cosine distances matrix D for ( f 1 , f 2 , , f c ) in the total extracted features matrix R D and ( w 1 , w 2 , , w z ) in the visual words V W is calculated as follows:
D ( R D , V W ) = d f 1 w 1 d f 1 w 2 d f 1 w c d f 2 w 1 d f 2 w 1 d f 2 w c d f z w 1 d f z w 1 d f z w c
Then apply image mapping by forming the histogram that describes the image using V W . The histogram is constructed by counting how many each visual word gives the minimum distance d in matrix D. It is calculated as:
h ( w i ) = c o u n t ( a r g m i n ( r o w k ) i s i ) f o r k ( 1 , z ) a n d i ( 1 , c )
where argmin returns the index of the column of the minimum value in the row r o w k in matrix D. Therefore, the output histogram of an image is:
H = { h ( w 1 ) , h ( w 2 ) , , h ( w c ) }

3.5. Visual Words Features Normalization

A scaling technique called normalization shifts and rescales values so that they fall between the ranges of 0 and 1. They are additionally called Min-Max scaling. The normalization step reduces the overfitting probability. The normalization process could be calculated by:
H ` = H h m i n h m a x h m i n
where h m i n and h m i n are the minimum and the maximum features of the H visual words features vector respectively.

3.6. BoVW FeatureWiz Selection (BFS)

The Bag-of-Words (BoW) model represents text data at modelling text using machine learning algorithms. The BoW model is a well-known model for feature extraction in natural language processing. The BoW is a representation of text that describes the occurrence of words within a document. Selecting powerful words from a text is a crucial step for the success of the BoW model. In Visual retrieval, there is a similar model, which is the BoVW model, and it operates as same as the BoW model technique. In our work, we added a step for selecting powerful visual words after clustering. This step is built based on the FeatureWiz selection technique. FeatureWiz is used for the feature selection before clustering and to choose the most powerful visual words after the normalization step.

3.7. The Support Vector Machines (SVMs) Prediction Model

The prediction model is applied using the state-of-the-art SVMs on the BoVW. The SVMs are initially proposed by Vapnik [34]. The SVMs are a supervised machine learning algorithms that are used for regression or classification. Although regression can also benefit greatly from it, classification is where it is most often used. The SVMs algorithm seeks to identify a hyperplane in a d-dimensional space that clearly divides the data points into various classes according to the support vectors. Support vectors in both classes are the data points that are closest to the hyperplane. In a d-dimensional space, where d is the number of features in the dataset, each individual piece of data was plotted. The best hyperplane to divide the data should then be determined. SVMs are limited to binary classification (i.e., choose between two classes). For multi-class issues, though. Each class can have a binary classifier created by the SVMs. Each classifier will produce one of two outcomes based on a score value: either the data point belongs to that class or does not. The final output of the SVMs is the class with the highest score. Consider the case where SVM will divide the two classes, A and B, in a two-dimensional space. The classifier’s goal is to determine if the unidentified feature vector F belongs to Class A or Class B. Suppose a linear equation:
g ( F ) = ω T F + θ
where ω denotes the weight vector’s perpendicular position to the hyperplane. It represents the hyperplane’s orientation in d-dimensional space, while θ shows where the hyperplane is located. This linear equation depicts a straight line in two dimensions. In contrast, the equation represents a plane in three dimensions and a hyperplane in more than three dimensions. Figure 6 shows the SVM in two-dimensional space.
  • If g ( f i ) < 1 then, f i lies in class B.
  • If g ( f i ) = 1 then, f i is a positive hyperplane.
  • If g ( f i ) = 0 then, f i lies on the hyperplane.
  • If g ( f i ) = 1 then, f i is a negative hyperplane.
  • If g ( f i ) > 1 then, f i lies in class A.
The mathematical description of SVM has several approaches, such as optimal separating hyperplanes, linearly separable, linearly non-separable, and nonlinear SVM. In the proposed work, it is recommended to use linear SVM duo to many features as it is more likely that the data are linearly separable in high dimensional space. In Linear SVM, we want only to tune one parameter (c), which controls the amount of regularization applied to the data. A smaller value of C allows the SVM to be a soft classifier. As the value grows, the SVM becomes more and more strict, allowing for fewer misclassifications; however, if it grows too large, it could easily overfit the training data.

4. Experimental Results

This section illustrates the outcomes of the experiments and the discussion of the proposed KFFVW framework. The performance and accuracy are evaluated on the three standard datasets (i.e., Corel-1000, Caltech-10, and Oxford Flower-17). The experiments were focused on the Corel-1000 dataset to select the best two keypoints detectors to apply fusion between them to form robust HKFV for the features’ extraction stage. The FAST and SURF keypoint feature detectors gave the best performance. The performances of each standalone FAST and SURF keypoint feature detector based on traditional CBIR were evaluated respectively on various vocabulary sizes compared to the performance of the proposed HKFV with the BFS for the proposed KFFVW framework. Every experiment was implemented five times based on K-fold Cross-Validation (KCV) to achieve the average performance. Furthermore, the performance of the proposed KFFVW was compared to the recent state-of-the-art methods on the specified datasets.

4.1. Experimental Parameters

The CBIR parameters are a primary factor for evaluating any CBIR system’s performance and running time. The parameters of the different experiments for the proposed framework are detailed below.

4.1.1. Features Percentages

The extracted features were used to construct the vocabulary. Using all the extracted features as input to the clustering step is not recommended. Datta et al. [35] reported that, although the increase in the used features gives better results, it increases the processing cost and tends to overfit and vice versa. Every used vocabulary size was tested over various feature percentages (i.e., 10%, 25%, 50%, 75%, and 100%) of the total unique extracted features to find the best percentage features that give the best performance with minimum processing cost. The features were selected based on featureWiz selection.

4.1.2. Vocabulary Size

The vocabulary size is the number of feature clusters. The vocabulary size is a primary key for evaluating any CBIR System. More extensive size vocabulary increases the system’s performance, but the massive vocabulary size tends to overfit and the processing cost and vice versa [36]. Various vocabulary sizes (i.e., 256, 512, 1024, …, 32,780) were examined to select the most suitable vocabulary size from the training features for efficient vector quantization.

4.1.3. Training and Testing Image Percentage

The experimented datasets are designed in two groups; one is for testing, and the other is for training. The training group is selected based on five-fold cross-validation. The dataset is divided into 80% for the training process and 20% for the testing stage.

4.2. Evaluation Metrics

The KCV is a simple approach for selecting the best model and determining an estimator’s accuracy [37]. The primary parameter for measuring the performance of the suggested CBIR framework is KCV. According to [38], the best value for k is between 5 and 10. The performance of the framework is evaluated using five-fold cross-validation. Precision is measured and summarized by each dataset class. The overall precision of the CBIR system is calculated as the average precision in all classes. As we set the value of k to 5, 80% of data will always be utilized to develop the training model, while the remaining 20% will be used to validate the model. The validation process was repeated five times, with the average being computed (each time was for a different fold). Various evaluation measures are used, including accuracy, precision, MAP, recall, and F1-score, which can be stated in the following ways.
  • Accuracy: Accuracy relates to how close a measurement is to the true or recognized value. That means it is the probability that the CBIR framework test retrieves the correct images. It is one of the most intuitive and widely used performance measures.
    A c c u r a c y = T N + T P F N + F P + T N + T P × 100 %
  • Precision (P): Precision relates to the capability to detect the true class among all predicted valid classes accurately. Moreover, it is expressed as a ratio of the correctly predicted valid classes to all anticipated true classes. High precision is always associated with a low false positive rate, as shown by the equation below.
    P = T P F P + T P × 100 %
  • Mean Average Precision (MAP): Often known as the average precision (AP), it is a well indicator for assessing how well models execute tasks such as document/information retrieval and object recognition. The MAP of a set of queries is defined by
    M A P = q = 1 Q A v e P ( q ) Q × 100 %
    where AveP(q) is the average precision (AP) for a specific query, q, and Q is the total number of queries in the collection.
  • Recall (R): Recall, also known as sensitivity, hit rate, or true positive rate, can be broadly considered as the ability of the model to identify all positive cases so that it can be expressed as
    R = T P F N + T P × 100 %
    It should be noted that the above formula shows that high recalls are certainly associated with low false negative rates.
  • F1-score (F): F1-score score is not as intuitive as accuracy, but it is beneficial when measuring the robustness and accuracy of a CBIR system. The F1-score, which serves as a single measure of test performance that describes both recall and precision, is usually calculated as a weighted average of recall and precision.
    F = 2 R · P R + P × 100 %

4.3. Datasets

The Corel-1000 dataset [39] is a subset of the Corel image dataset [40], which is common in the evaluation of the CBIR [41,42,43]. It includes 1000 images in Corel-1000, divided into ten semantic classes having 100 images for each. Figure 7 shows a sample image of each semantic class of the Corel-1000 dataset. The Caltech-10 dataset is a subset of the Caltech-101 dataset [44]. The Caltech-10 includes 1000 images divided into ten classes, each with 100 images. The Oxford Flower-17 dataset [14] contains 1360 images divided into 17 semantic flower classes, where each class consists of 80 images. Figure 8 and Figure 9 show a sample image of each semantic class of the Oxford Flower-17 and Caltech-10 datasets, respectively.

4.3.1. Corel-1000 Dataset Results

For selecting the best keypoint detectors with the best features descriptor, each keypoint detector of the (FAST, HARRIS, Good Features To Track (GFTT), DoG, SURF, CenSurE, Maximally Stable Extremal Regions (MSER), Binary Robust Invariant Scalable Keypoint (BRISK), and Oriented FAST and Rotated Binary robust independent elementary features (ORB)) with each features descriptor of the (RootSIFT, SIFT, SURF as a descriptor, BRISK as a descriptor, ORB as a descriptor, Fast Retina Keypoint (FREAK), and Binary Robust Independent Elementary Features (BRIEF) have experimented.
The experiments were on the Corel-1000 dataset, where 2048 clusters were used to form the vocabulary set. Each experiment was tested five times with a different collection of images for the training (80%) and the rest for testing. Then, the average of all results was calculated. The performance comparison in Figure 4 shows that the two maximum MAP values on two different keypoint detectors are 82.76% at using FAST as a keypoint detector and RootSIFT as a features descriptor and 77.47% at using SURF as a keypoint detector and RootSIFT as a features descriptor. Thus, fusing the FAST and SURF keypoints is suggested by selecting the unique keypoints. Then, the RootSIFT was used to describe those keypoints.
The FAST and SURF keypoints were detected separately from each image in the dataset. Then, they were concatenated uniquely to prevent any point from duplication. Then the RootSIFT describes the detected keypoints. If an inquiry example image is given, its quantized features histogram will be predicted by the trained SVMs model. Moreover, it will be compared with the quantized features histograms of the images in the target dataset based on the Cosine distance similarity metrics. The result of the Top-20 retrieved images in response to the queries belonging to the Elephants class is shown in Figure 10 The query image appears in the first row. In contrast, the other two rows are the retrieved images in response to the query image. The score of each extracted image is displayed above it, and it represents the position of the query image to the retrieved image.
In Figure 11 and Figure 12, the various vocabulary sizes are shown along with the X-axis and the MAP value on the Y-axis. The Blue and Red lines show the CBIR performance of the standalone FAST and SURF keypoint based on the BoVW technique. The third line shows the proposed KFFVW framework’s performance, giving the max MAP value of 92.49% at vocabulary size 8192. Furthermore, the MAP value of 91.19% on a vocabulary size of 2048 using 50% of the features, as shown in Figure 11 at using 75% features with 8192 vocabulary sizes, the MAP value is 91.59%, as shown in Figure 12.
Table 4 shows the performance and statistical analysis of the suggested HKFV+BFS approach based on the fusion FAST and SURF keypoint detectors on various sizes of the vocabulary and BFS percentages using the Corel-1000 image dataset. The bold value indicates the best MAP performance. As shown in Table 1, the overall calculation time required for the whole retrieval process for a CBIR in the proposed work is divided into feature extraction time, prediction time, and similarity measurement and sorting time. The comparison between the proposed framework and the recent CBIR approaches (GMM-EM [45], Hybrid features [46], CH-LDP [47], CH-LDP-SIFT-DPBoF [48], and GA-SVM [49]) in Table 5 proves the adequate performance of the suggested method. Furthermore, according to the same table, the proposed framework outperforms all other methods in the mean average precision. The suggested framework surpasses the other retrieval results in most of the classes.

4.3.2. Caltech-10 Dataset Results

The Caltech-10 dataset consists of 1000 images of 10 semantic classes, each containing 100 images. It is a subgroup of the Caltech-101 dataset, which includes 101 classes. In our work, the classes Faces, Leopards, Motorbikes, Airplanes, Bonsai, Brain, Car side, Grand piano, Ketch, and Watch were selected from the Caltech-101 dataset to construct the Caltech-10 dataset. To investigate the effect of the combination of BFS with the vocabulary size on the performance of the proposed framework, we have designed a combination between BFS and vocabulary size as mentioned in Table 6. The best precision value (i.e., 98.40%) was at the combination 75% for BFS and 2048 vocabularies. Figure 13 shows the combination of the BFS and the vocabulary size on the Caltech-10 dataset. The results in Table 7 show that the proposed KFFVW framework based on HKFV features using BFS gives better results than using a standalone keypoint feature detector such as SURF and FAST.

4.3.3. Flower-17 Dataset Results

We computed the performance via various experiments of the proposed framework on the Oxford Flower-17 dataset on different combinations of BFS and vocabulary sizes. The results in Table 8 show the effectiveness of the HKFV+BFS techniques on various vocabulary sizes and BFS. The best MAP value is 84.94% at the combination of 25% BFS and 8192 vocabularies. Figure 14 shows the relation between the percentages of BFS and the vocabulary size. It also shows that the low and high percentages of BFS may lead to data overfitting or low results, while the appropriate values for BFS and vocabulary size (i.e., 25% and 8192, respectively) give the best results. Furthermore, the traditional BoVW technique experimented on the Oxford Flower-17 dataset using standalone SURF and standalone FAST keypoint features detector with the RootSIFT feature descriptor. Table 9 shows that the proposed KFFVW framework based on HKFV features and BFS visual words gives better results than a standalone keypoint feature detector.
In this experiment, we evaluated how well our framework performed using various numbers of images it returned. The classes of Sunflower, Pansy, Coltsfoot, Windflower, Tulip, Lilyvalley, Iris, Fritillary, and Buttercup yielded the highest results. In contrast, the classes Crocus, Bluebell, Cowslip, Daffodil, Dandelion, Snowdrop, Tigerlily, and Daisy yielded the lowest results because they have similar features and less dissimilarity over other semantic classes. The proposed framework attained a maximum mean average precision of 84.94% on the Oxford Flower-17 dataset. The comparison between the proposed framework and the recent CBIR approaches [50,51,52] on the Oxford Flower-17 dataset in Table 10 proves the adequate performance of the suggested method.
It was interesting to find the best combination between SIFT, FAST, SURF, and ORB to produce the HKFV. The performance and execution time comparison results between HKFV(SIFT+ ORB)+BFS, HKFV(SIFT+ORB+ FAST+SURF)+BFS, and HKFV(FAST+ SURF)+BFS were included as Table 2. The comparison in Table 3 provides execution time and performance between (SURF+BFS), (FAST+BFS), (HKFV), and (HKFV+BFS) validates the effectiveness of each component in the framework and indicates whether each part is necessary for the framework.
The tests and computations were performed on an AMD Ryzen 5 3400G with Radeon Vega 11 graphics (4 CPUs) 3.7 GHz, RAM 8 GB PC running Ubuntu 18.04 LTS 64bit.

5. Conclusions and Future Works

This paper has presented a novel CBIR framework based on visual word fusion of multiple feature descriptors to improve retrieval performance. The interest points have been separately extracted from an image using FAST and SURF. The retrieved keypoints have been combined into one keypoint feature vector, and the enhanced RootSIFT technique has been used to describe the area around each keypoint. The FeatureWiz selection technique has been applied to reduce and select the robust features for the BoVW learning model. K-means clustering has been used to quantify visual features into a more manageable set of visual words to produce the codebook. Finally, The SVM classifier for image retrieval has been fed by the feature vectors generated from the BoVW model. The obtained images have been sorted according to how closely they resemble the query image using an inverted index technique based on the cosine distance metric. By achieving average accuracies of 92.94%, 98.40%, and 84.94% on these benchmark datasets (Corel-1000, Caltech-10, and Oxford Flower-17), experiments on the three benchmark datasets have demonstrated that the presented CBIR technique could deliver results that were comparable to other state-of-the-art methods. Nevertheless, a possible limitation of this study is that the validation process needs wider datasets. The future work will be in two main directions. The first one is applying advanced deep learning techniques for better retrieval similarity and accuracy as an adaptive version of the framework. The second direction will focus on performing more experiments for testing and evaluating our method for addressing content-based image retrieval systems and presenting it as an online API.

Author Contributions

Conceptualization, S.B.; methodology, S.B. and E.S.; software, E.S. and S.B.; validation, S.B., M.H. and A.A.-H.; formal analysis, S.B., A.A.-H. and M.H; project administration, A.A.-H.; funding acquisition, A.A.-H. All authors have read and agreed to the published version of the manuscript.

Funding

This work is funded by the Federal Ministry of Education and Research of Germany (BMBF) (AutoKoWAT-3DMAt Nr. 13N16336) and DFG-Project Nr. GZ: Al 638/15-1.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding authors.

Acknowledgments

We are extremely grateful to the anonymous referees for their insightful comments and valuable suggestions which helped us to considerably improve the content of the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bagwari, A.; Sinha, A.; Singh, N.; Garg, N.; Kanti, J. CBIR-DSS: Business Decision Oriented Content-Based Recommendation Model for E-Commerce. Information 2022, 13, 479. [Google Scholar] [CrossRef]
  2. Subramanian, M.; Lingamuthu, V.; Venkatesan, C.; Perumal, S. Content-based image retrieval using colour, gray, advanced texture, shape features, and random forest classifier with optimized particle swarm optimization. Int. J. Biomed. Imaging 2022, 2022, 1–14. [Google Scholar] [CrossRef] [PubMed]
  3. Jain, N.; Salankar, S.S. Content Based Image Retrieval Using Improved Gabor Wavelet Transform and Linear Discriminant Analysis. In Proceedings of the 2018 3rd International Conference for Convergence in Technology (I2CT), Pune, India, 6–8 April 2018; pp. 1–4. [Google Scholar]
  4. Alzu’bi, A.; Amira, A.; Ramzan, N. Semantic content-based image retrieval: A comprehensive study. J. Vis. Commun. Image Represent. 2015, 32, 20–54. [Google Scholar] [CrossRef]
  5. Dhaygude, M.A.; Kinariwala, S. A literature survey on content-based information retrieval. J. Comput. Technol. (JCT) 2022, 11, 1–6. [Google Scholar]
  6. Bakheet, S.; Mofaddel, M.; Soliman, E.; Heshmat, M. A survey on content-based visual Information retrieval. Int. J. Eng. Res. Technol. 2020, 9, 499–505. [Google Scholar]
  7. Latif, A.; Rasheed, A.; Sajid, U.; Ahmed, J.; Ali, N.; Ratyal, N.I.; Zafar, B.; Dar, S.H.; Sajid, M.; Khalil, T. Content-based image retrieval and feature extraction: A comprehensive review. Math. Probl. Eng. 2019, 2019, 1–22. [Google Scholar] [CrossRef]
  8. Jadia, A.; Chawla, M.P.S. Image classification and detection of insulators using bag of visual words and speeded up robust features insulators. Int. J. Innov. Sci. Mod. Eng. (IJISME) 2020, 6, 7–13. [Google Scholar] [CrossRef]
  9. Bakheet, S.; Mofaddel, M.; Soliman, E.; Heshmat, M. Adaptive Multimodal Feature Fusion for Content-Based Image Classification. Appl. Math. Inf. Sci. (AMIS) 2020, 14, 699–708. [Google Scholar]
  10. Shukran, M.A.; Abdullah, M.N.; Yunus, M.S. New Approach on the Techniques of Content-Based Image Retrieval (CBIR) Using Color, Texture and Shape Features. J. Mater. Sci. Chem. Eng. 2021, 9, 51–57. [Google Scholar] [CrossRef]
  11. Patil, J.K.; Kumar, R. Analysis of content-based image retrieval for plant leaf diseases using color, shape and texture features. Eng. Agric. Environ. Food 2017, 10, 69–78. [Google Scholar] [CrossRef]
  12. Pradhan, J.; Pal, A.K.; Banka, H. A CBIR system based on saliency driven local image features and multi orientation texture features. J. Vis. Commun. Image Represent. 2022, 83, 103396–103414. [Google Scholar] [CrossRef]
  13. Li, X.; Yang, J.; Ma, J. Recent developments of content-based image retrieval (CBIR). Neurocomputing 2021, 452, 675–689. [Google Scholar] [CrossRef]
  14. Nilsback, M.; Zisserman, A. A Visual Vocabulary for Flower Classification. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; pp. 1447–1454. [Google Scholar] [CrossRef]
  15. Latha, D.; Sheela, C.J. Enhanced hybrid CBIR based on multichannel LBP oriented color descriptor and HSV color statistical feature. Multimed. Tools Appl. 2022, 81, 23801–23818. [Google Scholar] [CrossRef]
  16. Sadek, S.; Al-Hamadi, A.; Sayed, U. Image retrieval using cubic splines neural networks. Int. J. Video Mage Process. Netw. Secur. (IJIPNS) 2009, 9, 17–22. [Google Scholar]
  17. Sadek, S.; Al-Hamadi, A.; Michaelis, B.; Sayed, U. Cubic-splines neural network-based system for image retrieval. In Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 273–276. [Google Scholar] [CrossRef]
  18. Al-Obaide, Z.H.; Al-Ani, A.A. Image retrieval based on discrete curvelet transform. Image Retr. Based Discret. Curvelet Transform. 2021, 1, 94–103. [Google Scholar] [CrossRef]
  19. Mishra, V.P. Texture analysis using wavelet transform. ADCAIJ Adv. Distrib. Comput. Artif. Intell. J. 2021, 10, 5–13. [Google Scholar] [CrossRef]
  20. Bu, H.; Kim, N.; Kim, S. Content-based image retrieval using a combination of texture and color features. Hum.-Centric Comput. Inf. Sci. 2021, 11, 1–12. [Google Scholar]
  21. Ghahremani, M.; Ghadiri, H.; Hamghalam, M. Local features integration for content-based image retrieval based on color, texture, and shape. Multimed. Tools Appl. 2021, 80, 28245–28263. [Google Scholar] [CrossRef]
  22. Yousuf, M.; Mehmood, Z.; Habib, H.A.; Mahmood, T.; Saba, T.; Rehman, A.; Rashid, M. A novel technique based on visual words fusion analysis of sparse features for effective content-based image retrieval. Math. Probl. Eng. 2018, 2018, 2134395. [Google Scholar] [CrossRef]
  23. Alkhawlani, M.; Elmogy, M.; Elbakry, H. Content-based image retrieval using local features descriptors and bag-of-visual words. Int. J. Adv. Comput. Sci. Appl. (IJACSA) 2015, 6, 212–219. [Google Scholar] [CrossRef]
  24. Andy, S.; Santika, D.D. Content based image retrieval using bag of visual words and multiclass support vector machine. Innov. Comput. Inf. Control Express Lett. 2017, 11, 1479–1488. [Google Scholar]
  25. Mehmood, Z.; Mahmood, T.; Javid, M.A. Content-based image retrieval and semantic automatic image annotation based on the weighted average of triangular histograms using support vector machine. Appl. Intell. 2018, 48, 166–181. [Google Scholar] [CrossRef]
  26. Khan, U.A.; Javed, A. A hybrid CBIR system using novel local tetra angle patterns and color moment features. J. King Saud Univ.-Comput. Inf. Sci. 2022, 1, 1–18. [Google Scholar] [CrossRef]
  27. Alarcão, S.M.; Mendonça, V.; Maruta, C.; Fonseca, M.J. ExpertosLF: Dynamic late fusion of CBIR systems using online learning with relevance feedback. Multimed. Tools Appl. 2022, 1, 1–43. [Google Scholar] [CrossRef]
  28. Sadek, S.; Al-Hamadi, A. Robust hand gesture recognition using multiple shape-oriented visual cues. EURASIP J. Image Video Process. 2022, 1, 1–18. [Google Scholar] [CrossRef]
  29. Bay, H.; Ess, A.; Tuytelaars, T.; Gool, L.V. Speeded-up robust features (SURF). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
  30. Brown, M.; Lowe, D.G. Invariant features from interest point groups. In Proceedings of the 13th British Machine Vision Conference, Cardiff, UK, 2–5 September 2002; pp. 398–410. [Google Scholar]
  31. Rosten, E.; Porter, R.; Drummond, T. Faster and better: A machine learning approach to corner detection. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 32, 105–119. [Google Scholar] [CrossRef]
  32. Do, M.N.; Vetterli, M. The contourlet transform: An efficient directional multiresolution image representation. IEEE Trans. Image Process. 2005, 12, 2091–2106. [Google Scholar] [CrossRef]
  33. Arandjelović, R.; Zisserman, A. Three things everyone should know to improve object retrieval. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2911–2918. [Google Scholar]
  34. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  35. Datta, R.; Joshi, D.; Li, L.; Wang, J.Z. Image retrieval: Ideas, influences, and trends of the new age. ACM Comput. Surv. 2008, 40, 1–60. [Google Scholar] [CrossRef]
  36. Sharif, U.; Mehmood, Z.; Mahmood, T.; Javid, M.A.; Rehman, A.; Saba, T. Scene analysis and search using local features and support vector machine for effective content-based image retrieval. Artif. Intell. Rev. 2019, 52, 901–925. [Google Scholar] [CrossRef]
  37. Arlot, S.; Celisse, A. A survey of cross-validation procedures for model selection. Stat. Surv. 2010, 4, 40–79. [Google Scholar] [CrossRef]
  38. Hastie, T.; Tibshirani, R.; Friedman, J. Model assessment and selection. In The Elements of Statistical Learning: Data Mining, Inference, and Prediction; Springer: Berlin/Heidelberg, Germany, 2009; pp. 214–215. [Google Scholar]
  39. Wang, J.Z.; Li, J.; Wiederhold, G. SIMPLIcity: Semantics-sensitive integrated matching for picture libraries. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 23, 947–963. [Google Scholar] [CrossRef]
  40. Li, J.; Wang, J.Z. Real-time computerized annotation of pictures. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 985–1002. [Google Scholar] [PubMed]
  41. Ali, N.; Bajwa, K.B.; Sablatnig, R.; Mehmood, Z. Image retrieval by addition of spatial information based on histograms of triangular regions. Comput. Electr. Eng. 2016, 54, 539–550. [Google Scholar] [CrossRef]
  42. Fadaei, S. New dominant color descriptor features based on weighting of more informative pixels using suitable masks for content-based image retrieval. Int. J. Eng. (IJE) 2022, 35, 1457–1467. [Google Scholar] [CrossRef]
  43. Walia, E.; Pal, A. Fusion framework for effective color image retrieval. J. Vis. Commun. Image Represent. 2014, 25, 1335–1348. [Google Scholar] [CrossRef]
  44. Kinnunen, T.; Kamarainen, J.; Lensu, L.; Lankinen, J.; Käviäine, H. Making Visual Object Categorization More Challenging: Randomized Caltech-101 Data Set. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 476–479. [Google Scholar]
  45. Zeng, S.; Huang, R.; Wang, H.; Kang, Z. Image retrieval using spatiograms of colors quantized by Gaussian mixture models. Neurocomputing 2016, 171, 673–684. [Google Scholar] [CrossRef]
  46. Khokher, A.; Talwar, R. A fast and effective image retrieval scheme using color-, texture-, and shape-based histograms. Multimed. Tools Appl. 2017, 76, 21787–21809. [Google Scholar] [CrossRef]
  47. Zhou, J.; Liu, X.; Xu, T.; Gan, J.; Liu, W. A new fusion approach for content-based image retrieval with color histogram and local directional pattern. Int. J. Mach. Learn. Cybern. 2018, 9, 677–689. [Google Scholar] [CrossRef]
  48. Zhou, J.; Liu, X.; Liu, W.; Gan, J. Image retrieval based on effective feature extraction and diffusion process. Multimed. Tools Appl. 2019, 78, 6163–6190. [Google Scholar] [CrossRef]
  49. Khan, U.A.; Javed, A.; Ashraf, R. An effective hybrid framework for content based image retrieval (CBIR). Multimed. Tools Appl. 2021, 80, 26911–26937. [Google Scholar] [CrossRef]
  50. Lin, C.; Chen, R.; Chan, Y. A smart content-based image retrieval system based on color and texture feature. Image Vis. Comput. 2009, 6, 568–665. [Google Scholar] [CrossRef]
  51. ElAlami, M.E. A novel image retrieval model based on the most relevant features. Knowl.-Based Syst. 2011, 24, 23–32. [Google Scholar] [CrossRef]
  52. ElAlami, M.E. A new matching strategy for content-based image retrieval system. Appl. Soft Comput. 2014, 14, 407–418. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the proposed methodology.
Figure 1. Flowchart of the proposed methodology.
Sensors 23 01653 g001
Figure 2. Box filter approximation of second−order Gaussian partial derivatives.
Figure 2. Box filter approximation of second−order Gaussian partial derivatives.
Sensors 23 01653 g002
Figure 3. An interest point and 16 pixels surrounding on it.
Figure 3. An interest point and 16 pixels surrounding on it.
Sensors 23 01653 g003
Figure 4. Performance of each keypoint detector with each descriptor.
Figure 4. Performance of each keypoint detector with each descriptor.
Sensors 23 01653 g004
Figure 5. K-Means clustering model sample.
Figure 5. K-Means clustering model sample.
Sensors 23 01653 g005
Figure 6. SVMs maximize the hyperplane between two different classes, A and B.
Figure 6. SVMs maximize the hyperplane between two different classes, A and B.
Sensors 23 01653 g006
Figure 7. Sample image of each semantic category of the Corel-1000 dataset.
Figure 7. Sample image of each semantic category of the Corel-1000 dataset.
Sensors 23 01653 g007
Figure 8. Sample image of each semantic category of the Caltech-10 dataset.
Figure 8. Sample image of each semantic category of the Caltech-10 dataset.
Sensors 23 01653 g008
Figure 9. Sample image of each semantic category of Oxford Flower-17 dataset.
Figure 9. Sample image of each semantic category of Oxford Flower-17 dataset.
Sensors 23 01653 g009
Figure 10. The retrieved images for an Elephant image from the Corel-1000 dataset.
Figure 10. The retrieved images for an Elephant image from the Corel-1000 dataset.
Sensors 23 01653 g010
Figure 11. HKFV using 50% BFS vs. standalone FAST and SURF on the Corel-1000 dataset.
Figure 11. HKFV using 50% BFS vs. standalone FAST and SURF on the Corel-1000 dataset.
Sensors 23 01653 g011
Figure 12. HKFV using 75% BFS vs. FAST and SURF keypoints detector on the Corel-1000 dataset.
Figure 12. HKFV using 75% BFS vs. FAST and SURF keypoints detector on the Corel-1000 dataset.
Sensors 23 01653 g012
Figure 13. KFFVW on various vocabularies sizes and BFS on the Caltech -10 dataset.
Figure 13. KFFVW on various vocabularies sizes and BFS on the Caltech -10 dataset.
Sensors 23 01653 g013
Figure 14. KFFVW on various vocabularies sizes and BFS on Oxford Flower-17 dataset.
Figure 14. KFFVW on various vocabularies sizes and BFS on Oxford Flower-17 dataset.
Sensors 23 01653 g014
Table 1. Time spent for each phase of the retrieval process in seconds.
Table 1. Time spent for each phase of the retrieval process in seconds.
PhaseElapsed Time (s)
Query Feature Extraction (QFE)0.140
Query Prediction (QP)0.001
Similarity and Sorting of top 20 Images (SaS20)0.007
Total0.148
Table 2. Various features combination performance for the proposed method on Oxford Flower-17 dataset.
Table 2. Various features combination performance for the proposed method on Oxford Flower-17 dataset.
HKFV (SIFT+ORB) +BFSHKFV (SIFT+ORB+FAST+SURF) +BFSHKFV (FAST+SURF) +BFS
MAP81.7282.8684.94
QFE Time0.12810.26180.1401
QP Time0.00100.00140.0011
SaS20 Time0.00700.00780.0072
Total Time0.13610.2710.1484
Table 3. Comparison validate the effectiveness of each component in the framework.
Table 3. Comparison validate the effectiveness of each component in the framework.
SURFSURF+BFSFASTFAST+BFSHKFVHKFV+BFS
MAP74.8878.1379.0681.2382.9584.94
QFE Time0.06520.07060.05310.05860.12940.1401
QP Time0.00090.00100.00070.00080.00100.0011
SaS20 Time0.00530.00580.00620.00650.00700.0072
Total Time0.07140.07740.06000.06590.13740.1484
Table 4. The KFFVW Performance with various vocabulary sizes and BFS on the Corel-1000 dataset.
Table 4. The KFFVW Performance with various vocabulary sizes and BFS on the Corel-1000 dataset.
F%MAP Performance (in %) on Different Vocabulary Size
256512102420484096819216,38432,76865,536
10%60.3659.7475.8882.4188.5587.3587.1883.8077.57
25%82.2481.6385.8887.5088.5591.6090.3489.5488.34
50%85.6585.6388.0891.1989.9292.4989.9586.5085.08
75%88.4788.4988.6090.7290.0091.5990.2987.2886.49
100%86.6887.0687.7389.7489.4090.2689.0184.7783.59
Average80.8880.7185.4388.5189.4890.8589.5586.5884.21
SE5.185.322.381.610.320.900.591.001.84
SD11.5811.895.333.590.712.011.332.244.11
Table 5. Comparison between metrics the recent proposed CBIR approaches on the Corel-1000 and the proposed method.
Table 5. Comparison between metrics the recent proposed CBIR approaches on the Corel-1000 and the proposed method.
ClassesMetricsGMM-EM [45]Hybrid Features [46]CH-LDP [47]CH-LDP-SIFT-DPBoF [48]GA-SVM [49]Proposed HKFV+BFS
P72.5079.9077.982.6085.0083.57
AfricaR14.5015.9815.8016.5217.0018.64
F24.1726.6326.2727.5328.3330.48
P65.2047.2560.1056.8085.0088.22
BeachR13.409.4512.2011.3617.0015.64
F22.2315.7520.2818.9328.3326.57
P70.6065.7069.1077.1085.0094.63
BuildingR14.1213.1413.8215.4217.0019.78
F23.5321.9023.0325.7028.3332.72
P89.2092.9587.6098.9095.00100.0
BusesR17.8418.5917.5219.7819.0020.00
F29.7330.9829.2032.9731.6733.33
P100.099.8599.40100.0100.0100.0
DinosaursR20.0019.9719.8820.0020.0020.00
F33.3333.2833.1333.3333.3333.33
P70.5063.6059.2575.50100.096.31
ElephantsR14.1012.7211.8515.1020.0019.54
F23.5021.2019.7525.1733.3332.49
P94.8093.3595.8097.3095.0093.08
FlowersR18.9618.6719.1619.4619.0018.86
F31.6031.1231.9332.4331.6731.36
P91.8095.5591.8595.9095.0091.85
HorsesR18.3619.1118.3719.1819.0018.20
F30.6031.8530.6231.9731.6730.38
P72.2541.6064.0077.8085.0093.12
MountainsR14.458.3212.8015.5617.0019.62
F24.0813.8721.3325.9328.3332.41
P78.8076.1578.1089.5085.0084.15
FoodsR15.7615.2315.6217.9017.0018.25
F26.2725.3826.0329.8328.3329.99
MAP80.5775.5978.3185.1491.0092.49
Avg-Recall16.1515.1215.7017.0318.2018.85
Avg-F-score26.8525.1926.1628.3730.3331.31
Table 6. The KFFVW Performance with various vocabulary sizes and BFS on the Caltech-10 dataset.
Table 6. The KFFVW Performance with various vocabulary sizes and BFS on the Caltech-10 dataset.
BFS%MAP Caltech-10 Dataset Performance (in %) on Different Vocabulary Size
25651210242048409681921638432,76865,536
10%84.8290.8794.6394.1596.1897.4595.0294.1193.86
25%95.1396.1497.1297.2998.0698.1097.4396.4895.75
50%95.2096.8497.9697.9698.1098.2097.9897.2895.97
75%96.9498.3098.3098.4098.2098.2098.0696.8896.66
100%94.8795.9395.8496.6996.0896.3296.2695.1994.62
Average93.3995.6296.8396.9697.3797.7296.9595.9995.37
SE2.1741.2570.6830.7460.4880.3620.5800.5860.501
SD4.8622.8111.5271.6681.0920.8091.2971.3111.119
Table 7. HKFV with BFS vs. standalone SURF and FAST on Caltech-10 dataset classes.
Table 7. HKFV with BFS vs. standalone SURF and FAST on Caltech-10 dataset classes.
SURFFASTProposed HKFV+BFS
ClassesPRFPRFPRF
Faces98203310020331002033
Leopards9619329820331002033
Motorbikes921831931931971932
Airplanes921831931931971932
Bonsai8818298818291002033
Brain881829901830971932
Car side881829901830961932
Grand piano911830931931971932
Ketch9118309218311002033
Watch9118309218311002033
Average91.5018.3030.5092.9018.5830.9798.4019.6832.80
Table 8. The KFFVW Performance with various vocabulary sizes and BFS on Oxford Flower-17.
Table 8. The KFFVW Performance with various vocabulary sizes and BFS on Oxford Flower-17.
BFS%MAP Flower-17 Dataset Performance (in %) on Different Vocabulary Size
25651210242048409681921638432,76865,536
10%43.6354.5058.1266.9268.3472.5474.2376.2475.32
25%55.5561.6865.5674.1473.8381.2684.9483.3481.51
50%61.8467.9472.6377.9877.1478.5981.6280.8279.25
75%66.2767.8671.8478.3077.5779.2381.2781.1180.37
100%68.1367.9070.8376.6276.8878.6979.1878.4277.98
Average59.0863.9867.8074.7974.7578.0680.2679.9978.89
SE4.432.662.722.101.731.461.761.221.07
SD9.905.946.074.703.883.273.952.722.39
Table 9. HKFV with BFS vs. standalone SURF and FAST on Oxford Flower-17 dataset.
Table 9. HKFV with BFS vs. standalone SURF and FAST on Oxford Flower-17 dataset.
SURFFASTProposed HKFV+BFS
ClassesPRFPRFPRF
Bluebell7615.225741525801627
Buttercup701423821627851728
Coltsfoot7715.426851728931931
Cowslip6412.821721424801627
Crocus7414.825731524781626
Daffodil6412.821761525801627
Daisy7414.825751525841728
Dandelion701423711424801627
Fritillary6913.823841728851728
Iris8216.427841728851728
Lilyvalley701423791626851728
Pansy8717.429861729931931
Snowdrop6713.422721424801627
Sunflower9418.8319419311002033
Tigerlily7214.424751525801627
Tulip7815.626831728881829
Windflower851728791626881829
Average74.8814.9824.9679.0615.8126.3584.9416.9928.31
Table 10. Comparison between recent CBIR approaches and the HKFV+BFS on Oxford Flower-17 dataset.
Table 10. Comparison between recent CBIR approaches and the HKFV+BFS on Oxford Flower-17 dataset.
Method/MetricsPRF
Color and texture [50]76.415.325.49
Elalami [51]79.316.126.77
Elalami [52]82.117.328.58
Proposed Framwork84.9416.9928.31
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bakheet, S.; Al-Hamadi, A.; Soliman, E.; Heshmat, M. Hybrid Bag-of-Visual-Words and FeatureWiz Selection for Content-Based Visual Information Retrieval. Sensors 2023, 23, 1653. https://doi.org/10.3390/s23031653

AMA Style

Bakheet S, Al-Hamadi A, Soliman E, Heshmat M. Hybrid Bag-of-Visual-Words and FeatureWiz Selection for Content-Based Visual Information Retrieval. Sensors. 2023; 23(3):1653. https://doi.org/10.3390/s23031653

Chicago/Turabian Style

Bakheet, Samy, Ayoub Al-Hamadi, Emadeldeen Soliman, and Mohamed Heshmat. 2023. "Hybrid Bag-of-Visual-Words and FeatureWiz Selection for Content-Based Visual Information Retrieval" Sensors 23, no. 3: 1653. https://doi.org/10.3390/s23031653

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop