PCA-Based Advanced Local Octa-Directional Pattern (ALODP-PCA): A Texture Feature Descriptor for Image Retrieval

: This paper presents a novel feature descriptor termed principal component analysis (PCA)- based Advanced Local Octa-Directional Pattern (ALODP-PCA) for content-based image retrieval. The conventional approaches compare each pixel of an image with certain neighboring pixels providing discrete image information. The descriptor proposed in this work utilizes the local intensity of pixels in all eight directions of its neighborhood. The local octa-directional pattern results in two patterns, i.e., magnitude and directional, and each is quantized into a 40-bin histogram. A joint histogram is created by concatenating directional and magnitude histograms. To measure similarities between images, the Manhattan distance is used. Moreover, to maintain the computational cost, PCA is applied, which reduces the dimensionality. The proposed methodology is tested on a subset of a Multi-PIE face dataset. The dataset contains almost 800,000 images of over 300 people. These images carries different poses and have a wide range of facial expressions. Results were compared with state-of-the-art local patterns, namely, the local tri-directional pattern (LTriDP), local tetra directional pattern (LTetDP), and local ternary pattern (LTP). The results of the proposed model supersede the work of previously deﬁned work in terms of precision, accuracy, and recall.


Introduction
Realization of information and communication technologies, machine learning, and artificial intelligence techniques has opened new horizons for new applications. With the immense growth of internet and digital media, the domain of image processing has evolved rapidly in the past decade.
In image processing, shape, color, and texture play key roles in creating a feature vector [1,2]. A variety of descriptors is proposed using different techniques, such as the gray level co-occurrence matrix (GLCM), HOG, LBP, and SIFTS techniques [3][4][5][6][7]. These descriptors are further used in combination with machine learning techniques (Random Forest, SVM) to extract features of images [8][9][10][11]. Focusing on CBIR techniques, the local descriptor uses local pixel information extraction, which is one vital methodology gaining limelight in the engineering industries and research arena. Within local descriptor methodology, local binary pattern (LBP) was the first methodology that used local pixel information to extract the image features. LBP has multiple extensions such as local ternary patterns (LTP), local tri-directional pattern (LTriDP), and local tetra directional pattern (LTetDP).
Some of these techniques have used directions and magnitude to find the local neighborhood intensity patterns in four directions, whereas some approaches used the hybrid approach, i.e., in combination with other different methodologies (GLCM, HOG) for image retrieval by orchestrating a histogram from the co-occurrence matrix. On the other hand, the same local neighborhood intensity patterns were used in combination with GLCM to obtain more information about the co-relation of pixels and their frequency instead of creating a histogram [7]. In some places, local neighborhood intensity patterns are used in combination with color patterns for content-based image retrieval. Nevertheless, local neighborhood pixel information has always been the key point of interest in content-based image retrieval. Most of the patterns proposed so far utilize uniform approaches to obtain information from neighboring pixels, and only a handful of approaches use direction-based pixel information.

Problem Statement and Contribution
The above-described techniques have limitations; for example, methods focus on obtaining color features and ignore the texture features. In other cases, the focus is on creating histograms, ignoring the structural elements and their spatial correlation (SEC). Moreover, SEC does not fully regard all the structural element information. Some approaches are proposed which use information from neighboring pixels. Such techniques, however, are not quite successful due to the increase in dimensionality and computational cost, thus increasing image retrieval time. However, local pixel information is used in a lot of methods. However, some of them are not quite successful in picking up on changes in local pixel information trends. Another disadvantage includes a large dimension of the feature Focusing on CBIR techniques, the local descriptor uses local pixel information extraction, which is one vital methodology gaining limelight in the engineering industries and research arena. Within local descriptor methodology, local binary pattern (LBP) was the first methodology that used local pixel information to extract the image features. LBP has multiple extensions such as local ternary patterns (LTP), local tri-directional pattern (LTriDP), and local tetra directional pattern (LTetDP).
Some of these techniques have used directions and magnitude to find the local neighborhood intensity patterns in four directions, whereas some approaches used the hybrid approach, i.e., in combination with other different methodologies (GLCM, HOG) for image retrieval by orchestrating a histogram from the co-occurrence matrix. On the other hand, the same local neighborhood intensity patterns were used in combination with GLCM to obtain more information about the co-relation of pixels and their frequency instead of creating a histogram [7]. In some places, local neighborhood intensity patterns are used in combination with color patterns for content-based image retrieval. Nevertheless, local neighborhood pixel information has always been the key point of interest in content-based image retrieval. Most of the patterns proposed so far utilize uniform approaches to obtain information from neighboring pixels, and only a handful of approaches use direction-based pixel information.

Problem Statement and Contribution
The above-described techniques have limitations; for example, methods focus on obtaining color features and ignore the texture features. In other cases, the focus is on creating histograms, ignoring the structural elements and their spatial correlation (SEC). Moreover, SEC does not fully regard all the structural element information. Some approaches are proposed which use information from neighboring pixels. Such techniques, however, are not quite successful due to the increase in dimensionality and computational cost, thus increasing image retrieval time. However, local pixel information is used in a lot of methods. However, some of them are not quite successful in picking up on changes in local pixel information trends. Another disadvantage includes a large dimension of the feature vector. To the best of our knowledge, no study has ever explored the impact of utilizing eight directions to compute the features and retrieve the image. Hence, there is a dire need to develop descriptors that utilize all eight directions of the first radius of a pixel so that the problem of poor accuracy and high dimensionality could be considered.

•
We proposed a novel approach which considers all surrounding pixels to obtain the complete pixel features.

•
We proposed a methodology to reduce the computation cost by applying dimension reduction techniques using PCA.

•
We provided a mechanism to complete direction and magnitude patterns on the edges of the image (at Radius 2) where pixel information is not available.
The proposed methodology is twofold. Initially, eight surrounding pixels (nearest to center pixels) are considered to attain the most relevant information, focusing on the first radius only. Afterwards, the magnitude of the vector is computed, resulting in a joint histogram. At this stage, PCA is introduced to keep the data normalized and attain maximum information. This approach generates a small but enriched feature vector, which has all the qualities of the surrounding eight pixels.
The paper is organized as follows: Section 2 covers related work followed by Section 3 carrying the details of the proposed descriptor. Section 4 presents extensive experiments along with comparative analysis with various state-of-the-art techniques. Section 5 concludes the paper.

State of the Art
Early methods for texture classification includes different approaches majority of them are filtered approaches [6]. One of the popular filter Gabor filter, which uses 40 different kinds of kernel to convolve on an image to extract features [3,5], wavelet transform utilized by [4,9], and wavelet frames extracted by [10]. Haralick et al. [8] proposed a novel approach named the gray level co-occurrence matrix (GLCM). GLCM measures the spatial relationship between two different gray tones and is utilized as a classification technique and extract texture feature of images.
Local binary pattern (LBP), on the other hand, uses local pixel information to extract texture features of the image. Several algorithms have been proposed to improve quality, i.e., noise reduction. Zheng et al. [12] and Zhiqin et al. [13] proposed a novel approach to dehaze images obtained during cloudy and moist weather, which helps to extract robust features. Wavelet transform is a powerful tool for efficiently representing signals and images at multiple levels of detail with many advantages, such as compression, level-ofdetail display, progressive transmission, level-of-detail editing, filtering, modeling, fractals and multifractals, etc. [14]. Ahmed et al. [15] proposed a nonparametric denoising methods based on contourlet transform with sharp frequency localization. Ouahabi et al. [16] proposed wavelet-based image denoising methods. The local ternary pattern (LTP) [11] uses only three directions of information on the pixel. LTP requires users to set the initial threshold value. Murala et al. [17] introduced a new pattern which utilizes local pixel information, named the local tetra pattern (LTP). This pattern utilizes four directions of neighboring pixel information (0, 45, 90, and 135 degrees). As per Murala, only four angles are sufficient to cover the whole pixel, as 0 has same information as 180, 45 is same as 225, 90 degrees is equivalent to 270 degrees, and lastly 135 is equivalent to 315 degrees. The reason for choosing only four angles was to reduce the computation cost. A local ternary pattern was introduced by [18] and used in biomedical images.
The BRINT descriptor was proposed by Liu [19] which was rotation invariant, noise sensitive and computation vise very simple. BRINT descriptor used similar technique as used in LBP, because it utilizes surrounding eight pixels, but the difference is only pixels should be multiple of eight. Vimina et al. and Mistry et al. [20,21] used LBP to extract texture features in combination with color features to create multi-channel feature descriptor.
Noise-Resistant LBP (NR-LBP) [22] and Robust LBP (RLBP) [23] uses LBP technique with enhancement of noise reduction and efficiency, respectively. The local tetra directional pattern [17] focuses only on two directions: horizontal and vertical (0 and 90 degree). Two directional LBP showed better results than existing LBP. Another enhanced version of LBP was proposed by [24], named the local tetra oppugnant pattern, which also uses RGB color space in extracting neighbor pixel information.
Walia et al. [25] proposed the color difference histogram (CDH), which was used for color and texture features. Jain et al. [26] used Fuzzy C-Mean algorithm, Co-occurrence Matrix and HSI Hue algorithm to extract shape, texture, and color features, respectively. Feng et al. [27] used color and texture features to propose a global correlation descriptor (GCD) for efficient image retrieval.
Asif et al. [28] proposed novel approach that combines color vector quantization used for CBIR. The proposed approach extracts features based on edge orientation and color features. Dhiman et al. [29] used GLCM and LBP for content-based image retrieval to apply to the CORAL dataset. Suleman et al. [30] and Zeeshan et al. [31] used contextual techniques to find similarities in 1D and 2D signals. Sukhjeet et al. [32] used a hybrid approach which utilizes color space and quaternion moment vector to create this unique feature vector. Sawat et al. [33] used a heuristic approach for real-time content-based images retrieval. Qin et al. [34] used local gradient of images to extract the directional gradient feature vector and tested their approach on six different datasets.
Current state-of-the-art methodologies using most recent databases, 2D and 3D face recognition methods, were analyzed [35]. Adjabi et al. [36] analysed different approaches and proposed a multi-block color binarized statistical approach for face recognition. Khaldi et al. [37] used unsupervised deep learning approach for human ear recognition from face. El et al. [38] used the CNN approach to detect pain from the human face. Mimouna et al. [39] proposed a heterogeneous multimodal dataset approach OLIMP for advanced environment perception.
A local wavelet decomposition method was proposed by Dubey et al. [40] which produces output by comparing values with its center pixel. Giveki et al. [41] used SIFT in combination with a local directional pattern to produce the feature vector. Wu [42] introduced an enhanced version of LTP, which takes an image as input and creates two patterns called upper and lower binary patterns.
Ouahabi et al. proposed new approach based on discrete wavelet transform (DWT) and used for multifractal analysis [43]. Multifractal analysis technique was used on ultrasound images [44]. Gupta et al. [45] used SIFT and SURF descriptor for face recognition and apply to five different public datasets. Juan et al. [46] used an evolutionary technique to obtain the texture features from the facial images dataset. The evolutionary technique finds several interest points from face image. Shekhar et al. [47] used two new texture feature descriptors based on a zigzag approach named the color zigzag binary pattern (CZZBP) and color median block zigzag binary pattern (CMBZZBP). This ZigZag texture feature descriptor uses three color components RGB to extract features. Agarwal et al. [48] introduced a descriptor based on local pixel information named the Haar-like local ternary co-occurrence pattern (HLTCoP) for image retrieval. Al-Barhi et al. [49] used a PCA-based SIFT for the static face recognition system.
The local binary co-occurrence pattern [50] and local extrema co-occurrence pattern [51] were suggested by Verma, which compare each pixel from the image with its center symmetric pixels. Verma considered only four directions 0, 45, 90, and 135 degrees. This approach reduces the computation cost, but due to considering only four pixels, lot of important information, which could be used to create a good descriptor is left behind. Verma proposed another directional pattern named the local tri-directional pattern (LTriDP) [52], which considers only three directions.
Numerous approaches have been developed using neural network (NN). The CNNbased feature transformation (NFT) approach was introduced [53], which helps in reducing the redundancy in features before performing classification. Charles et al. [54] introduced a local mesh color texture pattern (LMCTP), which uses color and spatial features and merges them to create a local descriptor. Amy et al. [55] used the transfer learning approach to find the TM logo similarity used for images retrieval. Christy et al. [56] proposed new Electronics 2022, 11, 202 5 of 17 methodology for CBIR, which extract pixels information from an image based on its shape, attributes, and tag.
Multi-trend structure descriptor (MTSD) [57] compares the target pixel with its four neighbor directions-0, 45, 90, and 135 degrees-and ignores the rest. MTSD uses the left-to-right/bottom-up approach for convolution and only uses the 3 × 3 kernel for it. MTSD uses the fix threshold to compute the binary value. Bala et al. [58] proposed textonbased local texton XOR patterns (LTxXORP) and used them for CBIR. The PM extract's structure features by using texton XOR. Banerjee et al. [59] introduced a local neighborhood intensity pattern (LNIP) to retrieve the content-based images. One of the best features of this approach was illumination resistance. The color component approach was utilized again by Bhunia et al. [60], in which he combined color and texture information, and HSV (Hue and Saturation) was used for the color component. Manickam et al. [61] used local directional technique (four directions only) named the extrema number pattern (LDIRENP) for content-based image retrieval. Li et al. [62] studied current developments in the field of CBIR.
Recently, a local tetra-directional pattern was introduced by Bedi et al. [63], in which they compared a pixel with the four most adjacent pixels (0 • , 45 • , 90 • , and 135 • ).
Among all the above-described descriptors, the most appropriate was the local tridirectional or tetra-directional patterns. These methods are used to fix directions while leaving very important chunk of image information behind, which compromises the accuracy of the feature descriptor. The proposed methodology has tried to overcome this issue by utilizing all the neighboring pixel's information. It has also proposed a solution to reduce the computation cost.
Each of the eight pixels are as important as the three angles used by [51] or the four angles used by [63]. For example, in Figure 2, considering the tri directional pattern for location I1, value 8 is only compared with the above, below, and central pixels, whereas we have higher intensity pixels available in the second radius, which could be used in making the strongest feature vectors which were left behind; on the other hand, in the tetra directional pattern, the second radius is not fully utilized; in I1, value 8 is only compared with angles (0 • , 45 • , 90 • , and 135 • ) which are (2, 8, 9 and 4), and again we have high intensity pixels available at angles 270 • and 315 • , which were left behind. The second drawback is that Bedi [63] utilized the second radius, but he did not mention any methodology pertaining to how to handle image edges where use of the second radius is not possible. The third disadvantage of both approaches is the need to stick with the Euclidian distance calculation approach while the accuracy of Manhattan distance is proven [64]. The recent state of the artwork did not use all the directions because of high computation cost. Therefore, new methods were introduced to truncate important pixel information. ing the redundancy in features before performing classification. Charles et al. [54] introduced a local mesh color texture pattern (LMCTP), which uses color and spatial features and merges them to create a local descriptor. Amy et al. [55] used the transfer learning approach to find the TM logo similarity used for images retrieval. Christy et al. [56] proposed new methodology for CBIR, which extract pixels information from an image based on its shape, attributes, and tag. Multi-trend structure descriptor (MTSD) [57] compares the target pixel with its four neighbor directions-0, 45, 90, and 135 degrees-and ignores the rest. MTSD uses the leftto-right/bottom-up approach for convolution and only uses the 3 × 3 kernel for it. MTSD uses the fix threshold to compute the binary value. Bala et al. [58] proposed texton-based local texton XOR patterns (LTxXORP) and used them for CBIR. The PM extract's structure features by using texton XOR. Banerjee et al. [59] introduced a local neighborhood intensity pattern (LNIP) to retrieve the content-based images. One of the best features of this approach was illumination resistance. The color component approach was utilized again by Bhunia et al. [60], in which he combined color and texture information, and HSV (Hue and Saturation) was used for the color component. Manickam et al. [61] used local directional technique (four directions only) named the extrema number pattern (LDIRENP) for content-based image retrieval. Li et al. [62] studied current developments in the field of CBIR.
Recently, a local tetra-directional pattern was introduced by Bedi et al. [63], in which they compared a pixel with the four most adjacent pixels (0°, 45°, 90°, and 135°).
Among all the above-described descriptors, the most appropriate was the local tridirectional or tetra-directional patterns. These methods are used to fix directions while leaving very important chunk of image information behind, which compromises the accuracy of the feature descriptor. The proposed methodology has tried to overcome this issue by utilizing all the neighboring pixel's information. It has also proposed a solution to reduce the computation cost.
Each of the eight pixels are as important as the three angles used by [51] or the four angles used by [63]. For example, in Figure 2, considering the tri directional pattern for location I1, value 8 is only compared with the above, below, and central pixels, whereas we have higher intensity pixels available in the second radius, which could be used in making the strongest feature vectors which were left behind; on the other hand, in the tetra directional pattern, the second radius is not fully utilized; in I1, value 8 is only compared with angles (0°, 45°, 90°, and 135°) which are (2, 8, 9 and 4), and again we have high intensity pixels available at angles 270° and 315°, which were left behind. The second drawback is that Bedi [63] utilized the second radius, but he did not mention any methodology pertaining to how to handle image edges where use of the second radius is not possible. The third disadvantage of both approaches is the need to stick with the Euclidian distance calculation approach while the accuracy of Manhattan distance is proven [64]. The recent state of the artwork did not use all the directions because of high computation cost. Therefore, new methods were introduced to truncate important pixel information.

Materials and Methods
The proposed model is shown in Figure 3. The algorithm has three parts: in the first section, the directional pattern is created, and in the second part, the magnitude pattern is created. Both pattern's histograms are then joined to create a combined feature vector, which is then passed to the third part to reduce the dimensionality while using PCA. The proposed model is shown in Figure 3. The algorithm has three parts: in the first section, the directional pattern is created, and in the second part, the magnitude pattern is created. Both pattern's histograms are then joined to create a combined feature vector, which is then passed to the third part to reduce the dimensionality while using PCA.

Advance Local Octa Directional Pattern
The advanced local octa-directional pattern (ALODP-PCA) not only utilizes all the surrounding pixel information, but also reduces the computation cost as well. Instead of taking 3 directions [52] or 4 directions [63], all of the directions in the first degree radius are utilized in ALODP-PCA. True information about any pixel is retained by its closest neighboring pixels. With the help of all surrounding pixel information, we can obtain the true value of the center pixel. Therefore, the proposed methodology considers eight neighbor pixels close to the center pixels (from the first radius) to obtain the direction pattern. To handle the edges and corner, where the surrounding 8 pixels are not available, the value is taken as 0 to compute the value of the center pixel. Figure 4 shows how and which pixels of radii 1 and 2 are involved in the octa directional pattern.

1.
Directional pattern calculation: In step 1 of the proposed algorithm, the directional pattern is created. Detail of the calculation is given in Section 4.1.

2.
Magnitude pattern calculation: In step 2 of the proposed algorithm, the magnitude pattern is created. Detail of the calculation is given in Section 4.2.

3.
Quantization histogram creation: For each directional and magnitude pattern a 40-bin quantization histogram is created.

4.
Joint histogram creation: In this step, all directional and magnitude histograms are combined to create a joint histogram.

5.
Dimensions reduction: In this step, principal component analysis (PCA) is used to select the most important features from the image. Detail is given in Section 4.3.

6.
Similarity measure: In the final step, Manhattan distance is utilized to find the similarity measure. Detail is given in Section 4.3.

Advance Local Octa Directional Pattern
The advanced local octa-directional pattern (ALODP-PCA) not only utilizes all the surrounding pixel information, but also reduces the computation cost as well. Instead of taking 3 directions [52] or 4 directions [63], all of the directions in the first degree radius are utilized in ALODP-PCA. True information about any pixel is retained by its closest neighboring pixels. With the help of all surrounding pixel information, we can obtain the true value of the center pixel. Therefore, the proposed methodology considers eight neighbor pixels close to the center pixels (from the first radius) to obtain the direction pattern. To handle the edges and corner, where the surrounding 8 pixels are not available, the value is taken as 0 to compute the value of the center pixel. Figure 4 shows how and which pixels of radii 1 and 2 are involved in the octa directional pattern.

Direction Pattern
The center pixel is named Ic, which has 8 neighborhood pixels, I1, I2, …, I8, in the first radius and 16 pixels in the second radius. All the mathematical calculations are similar to the calculation performed by Verma et al. [52] The only difference is that Verma created 3

Direction Pattern
The center pixel is named I c , which has 8 neighborhood pixels, I 1 , I 2 , . . . , I 8 , in the first radius and 16 pixels in the second radius. All the mathematical calculations are similar to the calculation performed by Verma et al. [52] The only difference is that Verma created 3 directional patterns, whereas we have created eight directional patterns, and all the equations are modified accordingly. In the first step, the proposed methodology calculates the difference of the pixel against its center pixel. Figure 5 shows how the proposed methodology is used to calculate the directional pattern. To elaborate the working of the proposed methodology, a sample image of 5 × 5 is used. For each pixel location I 1 to I 8 , directional differences are calculated by comparing each pixel with its surrounding eight pixels, as given below.
PM produces 8 directional patterns, which are used to generate 8 histograms using Equation (6).

Magnitude Pattern
Murala et al. [17] has proved that, similar to directional patterns, the magnitude of a pixel also plays a very important role in creating a feature vector. Similar to the directional pattern, the magnitude of the pixel will use all eight directions. Figure 6 shows how the proposed methodology is used to calculate the magnitude pattern. To elaborate the working of PM, a sample image of 5 × 5 is used. The magnitude of pixel location I1 to I8 is calculated using Equation (7), where each pixel is compared with its surrounding eight pixels.
where = 1, 2, … , 8 represents pixels at location( , ). Each value of magnitude is compared with central magnitude value, if the magnitude at pixel ( , ) is found to be greater than the central magnitude, the binary value 1 is assigned to it or else it will be 0, as shown in Equation (8).
After obtaining eight differences, D 1 , D 2 , . . . , D 8 , a pattern with each distance is calculated as given below in Equation (1).
In Equation (1), #(D k < 0) is the count of distances less than 0 processed by taking the mod with the number of directions, which is 8 in our case. This will give us the function value. Next, we calculate the pattern value for each neighborhood pixel using Equation (2).

Magnitude Pattern
Murala et al. [17] has proved that, similar to directional patterns, the magnitude of a pixel also plays a very important role in creating a feature vector. Similar to the directional pattern, the magnitude of the pixel will use all eight directions. Figure 6 shows how the proposed methodology is used to calculate the magnitude pattern. To elaborate the working of PM, a sample image of 5 × 5 is used. The magnitude of pixel location I 1 to I 8 is calculated using Equation (7), where each pixel is compared with its surrounding eight pixels.
where i = 1, 2, . . . , 8 represents pixels at location (i, j). Each value of magnitude is compared with central magnitude value, if the magnitude at pixel (i, j) is found to be greater than the central magnitude, the binary value 1 is assigned to it or else it will be 0, as shown in Equation (8).
Electronics 2022, 11, x FOR PEER REVIEW 9 of 17 After obtaining the magnitude pattern, histogram is generated from each magnitude pattern. Then, all nine histograms concatenated into single joint histogram using Equation (9).
To elaborate the working of ALODP, a sample image of 5 × 5 is used, as shown in Figure 6. The magnitude pattern is given below. Here, Magc is greater than Magi; therefore, as per Equation (8a), Magi will be 0.  Figure 7 shows how our proposed methodology is used to calculate the directional pattern for radius 2. For each pixel location, I1 to I8, directional differences are calculated by comparing each pixel with its surrounding eight pixels, as given below, and taking the value as zero for the pixels where no value was available, especially at the edges. The After obtaining the magnitude pattern, histogram is generated from each magnitude pattern. Then, all nine histograms concatenated into single joint histogram using Equation (9).
To elaborate the working of ALODP, a sample image of 5 × 5 is used, as shown in Figure 6. The magnitude pattern is given below.
Here, Mag c is greater than Mag i ; therefore, as per Equation (8a), Mag i will be 0. Hist = [Hist ALODP 1 , Hist ALODP 2 , . . . , Hist ALODP 8 , Hist m (9) Figure 7 shows how our proposed methodology is used to calculate the directional pattern for radius 2. For each pixel location, I 1 to I 8 , directional differences are calculated by comparing each pixel with its surrounding eight pixels, as given below, and taking the value as zero for the pixels where no value was available, especially at the edges. The same methodology is used to calculate the magnitude for edges as depicted in Figure 7. (9).

Dimension Reduction and Similarity Measure
Principal component analysis (PCA) is one of the popular techniques for reducing multidimensional data. PCA uses eigenvalues and eigenvectors to create a covariance matrix, which is used to find the most variance data. The highest rank in PCA means the

Dimension Reduction and Similarity Measure
Principal component analysis (PCA) is one of the popular techniques for reducing multidimensional data. PCA uses eigenvalues and eigenvectors to create a covariance matrix, which is used to find the most variance data. The highest rank in PCA means the most valuable component. That is the reason we find, utilize, and compare the results of the top 5, 10, and 15 principal components. To quantify the resemblance between images, either we can use the high similarity between two images or use less difference between images-both give the same results. It has been proven that the Manhattan similarity distance works far better compared with the other distance matrix [64].

Dataset Description
In this research, a subset of the Multi-PIE face dataset has been used. This dataset contains over 750,000 images of 337 people's faces taken in different angles and with different expressions. Another reason for choosing this dataset is that this dataset contains images under 19 illumination conditions with different resolutions. The overall size of this dataset is 305 GB. Some of the images from the datasets are shown in Figure 8. the top 5, 10, and 15 principal components. To quantify the resemblance between images, either we can use the high similarity between two images or use less difference between images-both give the same results. It has been proven that the Manhattan similarity distance works far better compared with the other distance matrix [64].

Dataset Description
In this research, a subset of the Multi-PIE face dataset has been used. This dataset contains over 750,000 images of 337 people's faces taken in different angles and with different expressions. Another reason for choosing this dataset is that this dataset contains images under 19 illumination conditions with different resolutions. The overall size of this dataset is 305 GB. Some of the images from the datasets are shown in Figure 8.

Results
In order to evaluate the capability of the proposed methodology accuracy, precision and recall along with computational time are used as performance metrics. The proposed method queries an image from the image database, which results in many images. Some are relevant and some are not. Therefore, the precision, recall, and accuracy are computed by Equations (10) The proposed methodology is tested in two ways: one without the dimension reduction technique, and one with dimension reduction. Each set of time results are compared with state-of-the-art methodologies.

Computaton Time: Without Dimension Reduction Techniques
The proposed methodology is tested on the Multi-PIE dataset against state-of-the-art methodologies, such as LTriDP [42], LTetDP [14], and LTP [11] without dimension reduction. The execution time to retrieve 1000 images is 60 s for LTriDP, 68 s for LTetDP, and 55 s for LTP, as shown in Table 1. The results have shown the computation time, which is increased using proposed methodology because it utilizes all eight directions.

Results
In order to evaluate the capability of the proposed methodology accuracy, precision and recall along with computational time are used as performance metrics. The proposed method queries an image from the image database, which results in many images. Some are relevant and some are not. Therefore, the precision, recall, and accuracy are computed by Equations (10)-(12), respectively.
The proposed methodology is tested in two ways: one without the dimension reduction technique, and one with dimension reduction. Each set of time results are compared with state-of-the-art methodologies.

Computaton Time: Without Dimension Reduction Techniques
The proposed methodology is tested on the Multi-PIE dataset against state-of-the-art methodologies, such as LTriDP [42], LTetDP [14], and LTP [11] without dimension reduction. The execution time to retrieve 1000 images is 60 s for LTriDP, 68 s for LTetDP, and 55 s for LTP, as shown in Table 1. The results have shown the computation time, which is increased using proposed methodology because it utilizes all eight directions.

Computation Time: With Dimension Reduction Techniques
The proposed mythology is tested on the Multi-PIE dataset against state-of-the-art methodologies such as LTriDP [42], LTetDP [14], and LTP [11] with dimension reduction, and the results has shown a significant improvement compared with ALODP. Results of the performance measure clearly show that ALODP outperforms the previously developed state-of-the-art methodologies. Table 2 represents a detailed comparison of results obtained above with dimension reduction by taking 5, 10, and 15 principal components and their execution time. Execution time for the full dataset using the top 5, 10, and 15 principal components is 7.5, 8.75, 9.58 h, respectively. Figure 9 shows the computation time comparison of LTriDP, LTetDP, and LTP with ALODP, which clearly shows that the computation time of ALODP is smallest.

Computation Time: With Dimension Reduction Techniques
The proposed mythology is tested on the Multi-PIE dataset against state-of-the-art methodologies such as LTriDP [42], LTetDP [14], and LTP [11] with dimension reduction, and the results has shown a significant improvement compared with ALODP. Results of the performance measure clearly show that ALODP outperforms the previously developed state-of-the-art methodologies. Table 2 represents a detailed comparison of results obtained above with dimension reduction by taking 5, 10, and 15 principal components and their execution time. Execution time for the full dataset using the top 5, 10, and 15 principal components is 7.5, 8.75, 9.58 h, respectively. Figure 9 shows the computation time comparison of LTriDP, LTetDP, and LTP with ALODP, which clearly shows that the computation time of ALODP is smallest.

Comparative Analysis with Respect to Accuracy
Comparative analysis with respect to accuracy between ALODP and state-of-the-art methodologies shows that ALODP scores 76% accuracy in comparison with LTriDP,

Comparative Analysis with Respect to Accuracy
Comparative analysis with respect to accuracy between ALODP and state-of-theart methodologies shows that ALODP scores 76% accuracy in comparison with LTriDP, LTetDP, and LTP, which are 60%, 68%, and 65%, respectively. Results for the accuracy of the Multi-PIE dataset are given below in Table 3. Table 3. Comparative analysis of accuracy.

Method
Multi-PIE Face Database LTriDP [42] 0.60 LtetDP [14] 0.68 LTP [11] 0.65 ALODP 0.76 Figure 10 shows the accuracy comparison of LtriDP, LtetDP, and LTP with ALODP, which clearly shows that the accuracy of ALODP is high.  Figure 11 shows the precision results of ALODP with LTriDP [42], LTetDP [14], and LTP [11] obtained on the Multi-PIE face dataset. Figure 12 shows the comparison of ALODP and ALODPmag in terms of precision, which clearly shows that all eight directions play a bigger role in image retrieval, which was not considered in previous research.   Figure 11 shows the precision results of ALODP with LTriDP [42], LTetDP [14], and LTP [11] obtained on the Multi-PIE face dataset. Figure 12 shows the comparison of ALODP and ALODP mag in terms of precision, which clearly shows that all eight directions play a bigger role in image retrieval, which was not considered in previous research. LTetDP, and LTP, which are 60%, 68%, and 65%, respectively. Results for the accuracy of the Multi-PIE dataset are given below in Table 3. Table 3. Comparative analysis of accuracy.

Method
Multi-PIE Face Database LTriDP [42] 0.60 LtetDP [14] 0.68 LTP [11] 0.65 ALODP 0.76 Figure 10 shows the accuracy comparison of LtriDP, LtetDP, and LTP with ALODP, which clearly shows that the accuracy of ALODP is high.  Figure 11 shows the precision results of ALODP with LTriDP [42], LTetDP [14], and LTP [11] obtained on the Multi-PIE face dataset. Figure 12 shows the comparison of ALODP and ALODPmag in terms of precision, which clearly shows that all eight directions play a bigger role in image retrieval, which was not considered in previous research.

Comparative Analysis with Respect to Recall
This section explains the recall analysis of ALODP with other techniques on the Multi-PIE face dataset. Figure 13 shows the mean recall for the top 10 retrieval is 96.02%, which clearly shows the efficiency of PM on the Multi-PIE dataset. This proves that the PM can be used to retrieve many images perfectly. Figure 14 shows the comparison of ALODP and ALODPmag in terms of recall, which clearly shows all eight directions playing a bigger role in image retrieval, which was not considered in previous research.

Comparative Analysis with Respect to Recall
This section explains the recall analysis of ALODP with other techniques on the Multi-PIE face dataset. Figure 13 shows the mean recall for the top 10 retrieval is 96.02%, which clearly shows the efficiency of PM on the Multi-PIE dataset. This proves that the PM can be used to retrieve many images perfectly. Figure 14 shows the comparison of ALODP and ALODP mag in terms of recall, which clearly shows all eight directions playing a bigger role in image retrieval, which was not considered in previous research.

Comparative Analysis with Respect to Recall
This section explains the recall analysis of ALODP with other techniques on the Multi-PIE face dataset. Figure 13 shows the mean recall for the top 10 retrieval is 96.02%, which clearly shows the efficiency of PM on the Multi-PIE dataset. This proves that the PM can be used to retrieve many images perfectly. Figure 14 shows the comparison of ALODP and ALODPmag in terms of recall, which clearly shows all eight directions playing a bigger role in image retrieval, which was not considered in previous research.  Demonstration results of five query images are also shown in Figure 15. Regardless of which expression they have on their face, the system was able to locate the correct images.
Without PCA implementation, feature vector length will be high, which can cause slower performance. However, because of the use of PCA for reducing the feature vector dimensions from 9 histograms of 40 bin each to just a few ones (5, 10 and 15 as used in this work), yields the most accurate features to be utilized. By the increase in the PCA count, the results started decreasing but at a slow speed. In all the cases, the proposed methodology shows improved results as compared with LTriDP, LTetDP, and LTP (Table 3) Demonstration results of five query images are also shown in Figure 15. Regardless of which expression they have on their face, the system was able to locate the correct images. Without PCA implementation, feature vector length will be high, which can cause slower performance. However, because of the use of PCA for reducing the feature vector dimensions from 9 histograms of 40 bin each to just a few ones (5, 10 and 15 as used in this work), yields the most accurate features to be utilized. By the increase in the PCA count, the results started decreasing but at a slow speed. In all the cases, the proposed methodology shows improved results as compared with LTriDP, LTetDP, and LTP (Table  3).

Conclusions
This research proposes an improved version of the local directional pattern using the advance local octa-directional pattern based on PCA. To attain all the pixels' surrounding information, we compared the target pixel with its eight surrounding pixels to obtain the 8-directional patterns. In a similar way, the magnitude pattern also utilized 8-side pixels to obtain the magnitude vector, because it is a computationally expansive process to check all the pixels. To minimize this computational overhead, PCA is utilized, which gives an optimal performance. Previously designed approaches skipped the pixels' surrounding information; however, the proposed approach collects all the surrounding pixels' information to create the enriched feature vector and eliminates only the information with less pixel information. Moreover, the proposed methodology also describes the way to handle pixels in the second radius or pixels at the edge, which was not previously, described Demonstration results of five query images are also shown in Figure 15. Regardless of which expression they have on their face, the system was able to locate the correct images. Without PCA implementation, feature vector length will be high, which can cause slower performance. However, because of the use of PCA for reducing the feature vector dimensions from 9 histograms of 40 bin each to just a few ones (5, 10 and 15 as used in this work), yields the most accurate features to be utilized. By the increase in the PCA count, the results started decreasing but at a slow speed. In all the cases, the proposed methodology shows improved results as compared with LTriDP, LTetDP, and LTP (Table  3).

Conclusions
This research proposes an improved version of the local directional pattern using the advance local octa-directional pattern based on PCA. To attain all the pixels' surrounding information, we compared the target pixel with its eight surrounding pixels to obtain the 8-directional patterns. In a similar way, the magnitude pattern also utilized 8-side pixels to obtain the magnitude vector, because it is a computationally expansive process to check all the pixels. To minimize this computational overhead, PCA is utilized, which gives an optimal performance. Previously designed approaches skipped the pixels' surrounding information; however, the proposed approach collects all the surrounding pixels' information to create the enriched feature vector and eliminates only the information with less pixel information. Moreover, the proposed methodology also describes the way to handle pixels in the second radius or pixels at the edge, which was not previously, described

Conclusions
This research proposes an improved version of the local directional pattern using the advance local octa-directional pattern based on PCA. To attain all the pixels' surrounding information, we compared the target pixel with its eight surrounding pixels to obtain the 8-directional patterns. In a similar way, the magnitude pattern also utilized 8-side pixels to obtain the magnitude vector, because it is a computationally expansive process to check all the pixels. To minimize this computational overhead, PCA is utilized, which gives an optimal performance. Previously designed approaches skipped the pixels' surrounding information; however, the proposed approach collects all the surrounding pixels' information to create the enriched feature vector and eliminates only the information with less pixel information. Moreover, the proposed methodology also describes the way to handle pixels in the second radius or pixels at the edge, which was not previously, described anywhere. Extensive experiments proved that the ALODP clearly outperforms the existing ones-LTriDP [52], LTetDP [17], and LTP [11]-in terms of accuracy, precision, and recall.