Next Article in Journal
Analytical MPC Algorithm Using Steady-State Process Model
Previous Article in Journal
Seizure Detection in Medical IoT: Hybrid CNN-LSTM-GRU Model with Data Balancing and XAI Integration
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Three-Dimensional Object Recognition Using Orthogonal Polynomials: An Embedded Kernel Approach

by
Aqeel Abdulazeez Mohammed
1,†,
Ahlam Hanoon Al-sudani
2,†,
Alaa M. Abdul-Hadi
2,†,
Almuntadher Alwhelat
3,†,
Basheera M. Mahmmod
2,†,
Sadiq H. Abdulhussain
2,*,†,
Muntadher Alsabah
4,† and
Abir Hussain
5,†
1
Department of Electronic and Communication Engineering, University of Baghdad, Al-Jadriya, Baghdad 10071, Iraq
2
Department of Computer Engineering, University of Baghdad, Al-Jadriya, Baghdad 10071, Iraq
3
Department of Computer Engineering, Al-Farabi University College, Baghdad 10022, Iraq
4
Medical Technical College, Al-Farahidi University, Baghdad 10071, Iraq
5
Department of Electrical Engineering, University of Sharjah, Sharjah 27272, United Arab Emirates
*
Author to whom correspondence should be addressed.
All authors contributed equally to this work.
Algorithms 2025, 18(2), 78; https://doi.org/10.3390/a18020078
Submission received: 3 January 2025 / Revised: 23 January 2025 / Accepted: 28 January 2025 / Published: 1 February 2025
(This article belongs to the Section Algorithms for Multidisciplinary Applications)

Abstract

:
Computer vision seeks to mimic the human visual system and plays an essential role in artificial intelligence. It is based on different signal reprocessing techniques; therefore, developing efficient techniques becomes essential to achieving fast and reliable processing. Various signal preprocessing operations have been used for computer vision, including smoothing techniques, signal analyzing, resizing, sharpening, and enhancement, to reduce reluctant falsifications, segmentation, and image feature improvement. For example, to reduce the noise in a disturbed signal, smoothing kernels can be effectively used. This is achievedby convolving the distributed signal with smoothing kernels. In addition, orthogonal moments (OMs) are a crucial technique in signal preprocessing, serving as key descriptors for signal analysis and recognition. OMs are obtained by the projection of orthogonal polynomials (OPs) onto the signal domain. However, when dealing with 3D signals, the traditional approach of convolving kernels with the signal and computing OMs beforehand significantly increases the computational cost of computer vision algorithms. To address this issue, this paper develops a novel mathematical model to embed the kernel directly into the OPs functions, seamlessly integrating these two processes into a more efficient and accurate approach. The proposed model allows the computation of OMs for smoothed versions of 3D signals directly, thereby reducing computational overhead. Extensive experiments conducted on 3D objects demonstrate that the proposed method outperforms traditional approaches across various metrics. The average recognition accuracy improves to 83.85% when the polynomial order is increased to 10. Experimental results show that the proposed method exhibits higher accuracy and lower computational costs compared to the benchmark methods in various conditions for a wide range of parameter values.

1. Introduction

Three-dimensional object recognition is considered an important research area in computer vision and artificial intelligence. It focuses on the identification and classification of objects in three-dimensional space. In contrast to 2D, 3D object recognition makes use of depth data, spatial relationships, and geometric features, which lead to a better understanding of the object in the real world [1]. This capability is considered essential for applications where both perception accuracy and real-world interaction are required [2].
The importance of 3D object recognition arises from its wide range of applications across various industries. In augmented and virtual reality, 3D recognition improves the user experiences by allowing for simple interactions with virtual objects in the real world [3]. In robotics, 3D object recognition enables machines to navigate and process objects efficiently, facilitating several tasks such as autonomous navigation and object grasping [4,5]. In healthcare, 3D object recognition enhances advanced medical imaging and surgical planning by providing detailed 3D reconstructions of anatomical structures [6]. In addition, in security and surveillance, 3D recognition improves object detection and tracking in complex environments, which enhances the capabilities of safety and monitoring [7]. Moreover, accurate recognition and interpretation of 3D objects is crucial for advancements in autonomous vehicles, where understanding the 3D structure of the environment is important for safe navigation [8].
Three-dimensional object recognition is important for bridging the gap between digital systems and the real world. As this technology continues to develop, 3D object recognition reveals exciting new possibilities, making it significant in the field of modern computer vision and AI research.
After capturing visual data, image processing algorithms are utilized to identify 3D objects, and areas of interest. These algorithms play a crucial role in extracting meaningful information from visual inputs. This is essential for tasks such as 3D object detection, classification, and scene understanding. Utilizing advanced algorithms, systems can interpret and analyze visual data efficiently across various domains [9]. Within this context, orthogonal polynomials (OPs) can be used, which serve as powerful mathematical tools that significantly enhance the overall performance of the system [10,11,12]. The application of OPs in image preprocessing ensures accurate and efficient processing, making them indispensable for achieving reliable and precise navigation in autonomous robotic systems [13]. A great deal of progress in the OPs field has been made recently in relation to significant analysis branches because OPs are connected to different functions, such as hypergeometric, trigonometric, and elliptic functions. In addition, OPs are related to continued fractions theory, as well as to problems of interpolation, mechanical quadrature, and their occurrence in the theory of integral and differential equations. Moreover, OPs provide instructive illustrations of some situations in the theory of orthogonal systems. In addition, some OPs have recently played important roles in mathematical statistics and quantum mechanics [14]. Moments are considered as the projection of OPs on the signal domain [15].
Discrete moments, also known as discrete orthogonal moments, have been utilized in various signal processing fields such as speech, image, and video processing [16,17,18,19]. They are commonly used as a shape descriptor and are applied in different tasks such as face recognition [20], image classification [21], differential equations [22,23,24,25], speech enhancement [26], and edge detection [27]. Moreover, orthogonal moments have been utilized in other areas such as speech enhancement [28,29], watermarking [30], medical image analysis [31,32,33], and shot boundary detection [34].
In general, moments are quantities that are used to represent 1D, 2D, and 3D data without redundancy, making them an effective and efficient data descriptor. This means that moments have the capability to describe data while avoiding duplication [35,36,37]. Signal preprocessing techniques play a crucial role in enhancing the performance of computer vision algorithms by improving the quality of input data [38]. For instance, noise is removed from signals by utilizing smoothing kernels, such as averaging kernels. This is performed to obtain a more clean and reliable data for the remaining processing steps [19,39]. Furthermore, gradient kernels are usually used to calculate the gradients in the x-direction for 1D signals, x- and y-directions for 2D signals, and x-, y- and z-directions for 3D signals to determine intensity changes in the aforementioned directions, respectively. Extracting moments from images following preprocessing with smoothing and/or gradient image kernels is an effective approach to represent signals [40,41]. However, some existing works are required to reduce the spectral data space and, consequently, the computational load. More improvement is required to obtain more robust and accurate methods.
In order to obtain a smoothed and gradient version of a 3D signal f, kernels h s and h g are convolved with the signal. The convolution process is carried out separately for each required version of the 3D signal. Once the smoothed and gradient versions of the 3D signal are obtained, moments can be extracted from them. However, performing these operations for each signal significantly increases the computational cost of a feature extraction algorithm, such as for 3D object recognition. Recently, 3D object recognition has emerged as a transformative technology with significant potential in fields such as virtual reality, autonomous driving, and commercial manufacturing. While advanced feature extraction methods, including deep learning models, have achieved impressive results in this domain, their high computational cost poses a substantial challenge, particularly for resource-constrained mobile applications [42]. Three-dimensional object recognition is a complex process because 3D objects provide a large variety of projected images based on viewing direction, illumination, distance, and other viewing conditions [43]. To deal with viewing direction effects, in [43], a computational work is demonstrated using sets of a small number of object viewpoints. However, this approach leaves many problems unanswered. The main problems include the use of abstractions, class-based recognition, and the classification problem, which requires a large experimental and computational research. Basically, computer vision is an active task and plays a crucial role in making the computer system see and identify the visual world automatically based on simulating the basic biological ability of human visual perception. Therefore, to achieve a high potential, the focus is on having an accurate algorithm with a lower computational cost [44,45].
Therefore, the aim of this paper is to obtain an accurate algorithm with a lower computational cost by addressing the aforementioned issues. To this end, we propose a new mathematical model for 3D object recognition that computes preprocessed signal moments by combining two operations into one, thereby reducing the cost of clustering. This model can be effectively used in mobile robot applications. To achieve this, we have embedded a smoothing kernel operator into OP functions to extract features quickly and efficiently. Our contributions are as follows:
  • We presented a mathematical model that integrates image kernels directly into orthogonal polynomial (OP) functions, eliminating the need for separate convolution and orthogonal moment (OM) computation processes. This innovation significantly reduces computational overhead while maintaining accuracy.
  • The presented mathematical model shows a reduction in computational cost, making it suitable for applications and large-scale 3D signal processing. Moreover, the method not only enhances efficiency but also achieves an improvement in recognition accuracy. In addition, the presented approach can be used with any discrete orthogonal polynomials.
  • We conducted extensive experiments using the well-known McGill benchmark dataset to validate the effectiveness of the presented approach. These experiments were performed under various noisy environments to evaluate the robustness and reliability of the approach.
This paper is organized as follows: Section 2 discusses the preliminaries related to the orthogonal polynomial (OP) used in this paper and the formulation of the squared Krawtchouk–Tchebichef polynomial. In Section 3, the methodology of the proposed approach is given. Section 4 carries out the performance evaluation of the proposed recognition system based on the proposed technique. Finally, the paper concludes with a summary of findings and future directions in Section 5.

2. Preliminaries

This section describes the OP used in this paper, as well as the process of computing moments, which will be used in the vision and control system of mobile robots to ensure safe navigation in environments. Moments have a significant role in the process of feature extraction. Feature selection is the process of choosing a subset of relevant features from a larger set for a particular task, such as classification. The length of the feature vector is important in the performance of machine learning algorithms. Forming a feature vector using too few features may not capture enough information, which affects the recognition accuracy. On the other hand, constructing a feature vector using too many features can lead to overfitting. Therefore, finding the optimal feature vector length is essential.
Features that possess energy compaction and localization properties in the moment domain are able to simplify feature handling and have a significant impact on processing steps [46]. This is because they allow the signal to be expressed using a small number of moments, reducing computational complexity. Moreover, the localization property in space identifies the region of interest, which speeds up the computation and improves feature classification [47]. To achieve these properties, this paper uses the squared Krawtchouk–Tchebichef polynomial (SKTP) as a discrete transform. The SKTP has been shown to have robust performance in different applications, demonstrating its powerful energy compaction and localization properties [46]. Moreover, its localization property in space contributes an additional reduction in computation and feature classification by locating the region of interest (ROI) [48].

2.1. Formulation of the Squared Krawtchouk–Tchebichef Polynomial

The SKTP is a mathematical function that is generated by combining two orthogonal polynomials according to a defined mathematical operation. This method is efficiently applicable because when two orthogonal polynomials are multiplied, the result is also an orthogonal polynomial. This means that the properties of the original polynomials are maintained in the newly generated polynomial, as explained in [28,46,49]. To this end, the SKTP can be expressed as the nth order of the polynomial R n p ( x ; N ) , where p is a parameter that determines the shift of the polynomial and N is the maximum degree of the polynomial so that
R n p ( x ; N ) = i = 0 N 1 A i p ( x ; N ) B i p ( n ; N ) ,
where A n p ( x ; N ) , and B n p ( x ; N ) represent the discrete Krawtchouk–Tchebichef polynomial (DKTP) [28] and discrete Tchebichef–Krawtchouk polynomial (DTKP) [49], respectively. The aforementioned polynomials are also formed from two orthogonal polynomials, which are defined by [28,49] as follows:
A n p ( x ; N ) = i = 0 N 1 K i p ( n ; N ) T i ( x ; N ) ,
B n p ( x ; N ) = i = 0 N 1 K i p ( x ; N ) T i ( n ; N ) ,
where K n p ( x ; N ) and T n ( x ; N ) are the Krawtchouk and Tchebichef polynomials, respectively. Note that p represents the parameter of the Krawtchouk polynomials. From (2) and (3), the SKTP that is defined in (1) can also be written as follows:
R n p ( x ; N ) = i = 0 N 1 j = 0 N 1 l = 0 N 1 K j p ( i ; N ) T j ( x ; N ) K l p ( n ; N ) T l ( i ; N ) .
The matrix of SKTP, R , can be represented by multiplying the Krawtchouk and Tchebichef polynomials using matrix multiplication properties as
R = ( K T ) 2

2.2. Computation of 3D Moments

In the literature, there are different methods proposed for computing 3D moments. Examples of these methods are the direct method, cubes method, and matrix product method, as well as Kronecker tensor and matrix product method [50]. The primary goal of these methods is to minimize computation time while maintaining accuracy. However, the computational complexity of these methods varies. For example, the direct method has a computational complexity of O ( n 6 ) , while the cubes method has a complexity of O ( ( n C u b o s ) 6 ) , the matrix product method has a complexity of O ( ( n 8 ) 6 ) , and finally, the Kronecker tensor and matrix product method has a complexity of O ( n 3.81 )  [50]. As illustrated above, the computational complexity burden could be very high, especially with 3D signals. To overcome this issue, this paper proposes an efficient method of using the Kronecker tensor as well as the matrix multiplication method to compute the moments in an efficient way with lower computational complexity. To this end, for a 3D signal f ( x , y , z ) , the moments M can be computed as follows [50]:
M : , k = z R k , z R n , x F R m , y n , m , k = 0 , 1 , , O r d ,
where O r d represents the order, and M n , m , k R L × L × L M : , k R L 2 × L .

3. Research Design and Methodology

This section describes the methodology of the proposed method. Suppose we have a 3D object, denoted by f ( x , y , z ) , with dimensions N x × N y × N z . Let us also consider a kernel, denoted by h ( u , v , w ) , that we convolve with f ( x , y , z ) resulting in a new 3D object of size N x × N y × N z , denoted by g ( x , y , z ) that can be written as:
g ( x , y , z ) = f ( x , y , z ) h ( u , v , w )
g ( x , y , z ) = u = a a v = b b w = c c f ( x , y , z ) h ( u x , v y , w z ) .
When convolving a kernel of size A × B × C with a function f ( x , y , z ) , it is possible to decompose the convolution operation into three separate operations. This approach is called spatially separable convolution. In this method, the kernel of size A × B × C is split into three separate kernels: an A × 1 × 1 kernel, a 1 × B × 1 kernel, and a 1 × 1 × C kernel.
In the spatially separable convolution process, the first kernel, i.e., the A × 1 × 1 kernel, is convolved with f ( x , y , z ) , followed by convolving the result with the 1 × B × 1 kernel and then convolving the output of the previous step with the 1 × 1 × C kernel. This technique has been explained in detail in [51]. As a result of this decomposition, the original convolution equation, represented by (7), can be rewritten as follows:
g ( x , y , z ) = u = a a v = b b w = c c f ( x , y , z ) h ( w z ) h ( v y ) h ( u x ) .
To calculate the moments of the function g ( x , y , z ) , we can use Equation (6). This equation will yield the moments ( MC n , m , k ) of the original function f ( x , y , z ) after it has been convolved with the kernel and can be expressed, in the form of a matrix, as follows:
MC : , k = z R k , z R n , x G R m , y T n , m , k = 0 , 1 , , O r d ,
where G denotes the matrix form of the function g ( x , y , z ) and ⊗ represents the mathematical kronecker product. In what follows, we provide proof of the above expression.
Proposition 1. 
Suppose we have a function p that is convolved with a kernel h . The outcome of this convolution operation is a new function q , which can be expressed as follows:
q = p h .
Typically, the convolution operation presented above can be reformulated using the properties of matrix multiplication. The resulting matrix can be expressed as follows:
Q = H P ,
In the equation mentioned above, the matrix H represents the Toeplitz matrix of the kernel h . Using this equation, we can compute the moments Φ as a function of Q as follows:
Φ = R Q .
To this end, we can substitute matrix P given above into the moment function Φ. The resulting moment function Φ can be written as follows:
Φ = R H P
Let us assume that R e = R H , which represents the orthogonal polynomials embedded with the kernel H so that we have the following:
Φ = R e P
The concept we discussed above can also be extended to multidimensional signals. In other words, we can use the same approach to compute moments for signals with more than two dimensions. To achieve this, we can substitute Equation (14) into (10), which yields the following formula:
MC : , k = z R e ( k , z ) R e ( n , x ) F R e ( m , y ) T .
From the formula (16), we can observe that the moments of the function g can be calculated using the original function f and the orthogonal polynomials that contain the kernels within them. These orthogonal polynomials are denoted as R e ( n , x ) , R e ( m , y ) , and R e ( k , z ) for the x, y, and z dimensions, respectively. Here, R e is the embedded version of the orthogonal polynomials and is obtained by multiplying R (the standard orthogonal polynomial matrix) with the Toeplitz matrix H of the corresponding kernel ( h x , h y , or h z ). It is important to note that the Toeplitz matrix H represents the convolution kernel.
Remark 1. 
The current formula for computing moments is faster than the formula that first applies the convolution operation and then computes the moments. This is because in the current formula, the kernels are embedded within the polynomials themselves. This embedding operation is performed at the beginning, after checking the database and extracting relevant information. Once the embedding is complete, the moments can be computed without the need to apply the kernel for each object (function).
In other words, the current method saves computational time because the convolution operation is performed only once, and the resulting orthogonal polynomials containing the embedded kernels are then used to calculate moments for multiple objects without the need for repeated convolution.
Note that the matrix H in the previous equation represents a Toeplitz matrix that is truncated to a size of N × N for the operator h . Suppose h = [ h 0 , h 1 , , h m ] , then the Toeplitz matrix can be constructed as follows:
H = h 0 0 0 0 0 h 1 h 0 0 0 0 h 1 h 0 0 0 h m h 1 0 h m h 0 0 0 0 h m h 1 h 0 .
After presenting the mathematical model for moment computation, it can be applied for object recognition to demonstrate its advantages and importance. This process includes the following steps: generating the orthogonal polynomials, followed by constructing the Toeplitz matrix version of the kernel, and finally, embedding the kernel into the orthogonal polynomials. This approach has several advantages, such as efficient and fast moment computation, without the need to repeatedly apply the kernel for each object, which makes it ideal for object recognition systems.

Research Design

In this section, the design of the methodology is provided. The feature extraction process plays a crucial role in object recognition systems. Features extracted using orthogonal polynomials are considered robust. However, to improve the representation of feature, the selection of orthogonal polynomials is substantial. Therefore, in the proposed recognition system, we have utilized the squared Krawtchouk–Tchebichef polynomial (SKTP) due to its superior performance in signal representation. The SKTP is formed of Krawtchouk and Tchebichef polynomials using (5), which also can be written as follows:
R S = ( R K × R T ) 2 ,
where R S , R K , and R T represent the SKTP, Krawtchouk polynomials, and the Tchebichef polynomials, respectively. To improve the robustness of the features, we embedded kernels in the orthogonal polynomials to perform feature extraction. This will speed up the feature extraction process compared with a traditional feature extraction process. The traditional process starts with processing the object using the image kernels, then the process of feature extraction is performed. However, for the proposed technique, the feature extraction process is performed directly on the object, i.e., the proposed technique eliminates extra process while maintaining its effect. To embed the kernel in the orthogonal polynomials, we follow Equation (12). For more illustration, Figure 1 and Figure 2 show the effect of gradient kernels on 3D object samples using conventional convolution and OP-based convolution.
To generate the kernel, first, the length of the kernel ( l k e r n e l ) is set to generate 1D kernel. Moreover, the object information is extracted from the object dataset, where the object information includes the size of the object N x , N y , and N z . According to the extracted object information, the orthogonal polynomials (SKTP) are generated. After generating the orthogonal polynomials, the generated kernel is embedded into the orthogonal polynomials. The resulting polynomials P x , P y , P z are then used to compute the moments. Figure 3 shows the proposed technique described above.
After computing the moments, the feature vector is generated and it is used for classification. The features are set to a specific size, which is performed by setting up the order of the orthogonal polynomials. In other words, the moments size is limited to the order of the orthogonal polynomials. These orders are O r d x , O r d y , and O r d z for the polynomials P x , P y , and P z , respectively. Therefore, the moment size will be O r d x × O r d y × O r d z . After computing the moment, it is flattened to form the feature vector. Thereafter, the feature vector is normalized and used for classification. It is noteworthy that after obtaining the normalized feature vector, each normalized feature vector is associated with a label (ID) for each object. The labels are used for training and testing processes. In this paper, support vector machine (SVM) is employed for its ability to optimize the margin between two hyperplanes that separate the classes [52]. Moreover, SVM is widely used in recognition for its robustness against signal fluctuations compared to other classifiers [53]. Figure 3 shows the presented technique for embedding kernels. Moreover, Figure 4 shows the methodology using the proposed technique for 3D object recognition.

4. Results and Discussion

The performance of the recognition system based on the proposed technique is evaluated in this section. The evaluation is performed using the McGill dataset formed in [54]. Figure 5 shows samples of the object in the dataset. The dataset (McGill) serves as a benchmark dataset and includes nineteen classes expressed as 3D objects. These classes are categorized as planes, chairs, dinosaurs, fishes, cups, birds, spiders, spectacles, snakes, octopus, teddies, pliers, dolphins, fours, ants, humans, tables, and craps. For the SVM, we utilized the LIB-SVM version 3.24 [55]. It is noteworthy that the experiment is performed using Matlab 2019 on an 8-core core i7-4700MQ CPU with a frequency of 2.4 HGz and 16 GB RAM.
In this paper, we focused on 19 different objects and applied a range of translation effects to enrich the original dataset. The augmentation process starts by shifting each object within a range from −10 to 10 along the x, y, and z axes, using a step size of 5. This systematic shifting creates a diverse array of translations, which helps the recognition system generalize better across various object positions. This augmentation approach guarantee that the dataset captures a wide range of spatial arrangements. A total of 5902 samples are generated as a result of the utilized augmentation techniques. To evaluate the performance of the proposed technique, we implemented different evaluation metric, which are discussed in the following sections.

4.1. Analysis of Optimal Parameter for Orthogonal Polynomial

In this section, first, we performed an experimental analysis to evaluate the performance of the proposed system for different values of the orthogonal polynomial parameter (p). The goal of this analysis is to identify the optimal value that enhances the performance of the proposed system. Support vector machine (SVM) is utilized as a classifier for this analysis. SVM is a powerful tool in the field of machine learning which is widely employed for classification and regression tasks. Compared to other classifiers, SVM’s robustness against signal fluctuations makes it particularly effective for recognition applications [53].
In this paper, SKTP is applied to 3D objects that are chosen from the McGill dataset. Various values of the polynomial parameter (p) are tested for the nth order of the polynomial ( R n p ( x ; N ) ), where (p) affects the polynomial’s shift. The maximum order of the polynomial, denoted as ( O r d ), is set to 8, 10, and 12, while the values of the polynomial parameter (p) include 0.2, 0.3, 0.4, 0.5, and 0.6.
The SVM model was trained using 4134 3D objects from the McGill dataset, and the testing phase involved 1768 3D objects. The details of the dataset are given in Table 1 We examined three different cases of the orthogonal parameter to determine which configuration yields the best fitting effect of SKTP, ultimately aiming to achieve the most accurate signal representation based on the information provided by the 3D objects.
First, the SVM model is trained and tested in clean environment with the results reported in Table 2.
The results shows that the highest performance score is 86.765, which is achieved with ( p = 0.4) and ( O r d = 10). This combination indicates that a moderate value of (p) paired with a polynomial order ( O r d ) of 10 yields optimal results. Furthermore, the results demonstrate that the accuracy increases as the polynomial parameter p increases from 0.2 to 0.4, and the recognition accuracy declines after the polynomial parameter p becomes greater than 0.4 (0.5 and 0.6).
Moreover, the analysis of the average recognition accuracy across different polynomial orders ( O r d ) is important to providing the effect of the polynomial order. For ( O r d = 8), the average recognition accuracy is 79.593, indicating a good ability to capture a relevant representation of the data. This recognition accuracy increases to 83.971 for ( O r d = 10), which shows that the model benefits from increasing the number of features by a higher polynomial order, which enhances the fitting capabilities of data distribution. However, as the polynomial order increases to ( O r d = 12), the average performance score decreases to 81.912. The decrease in average performance shows that as the order of the polynomial increases, the complexity is also increases, potentially resulting in overfitting. These findings show the importance of selecting an appropriate polynomial order, as the results indicate that while higher orders can improve performance, there exists an optimal threshold.
To investigate the performance of the proposed approach, an experiment is performed in different noisy environments, namely Gaussian, salt-and-pepper, and speckle noise. This experiment aims to examine the effects of these various noise types on the ability of the presented technique to generalize across different configurations of the orthogonal polynomial parameter (p) and polynomial orders ( O r d ). The obtained results are reported in Table 3, Table 4 and Table 5.
In Table 3, a comprehensive evaluation of 3D object recognition based on the proposed technique for polynomial order ( O r d = 8) is presented across various environments with different noise parameters. The Gaussian noisy environment reveals a decline in recognition accuracy as the noise level increases. At a Gaussian noise level of 1%, the recognition accuracy is 84.842 at p = 0.4, indicating that the performance can be considered good under low noise conditions. However, as the noise level increases to 5%, the recognition accuracy drops to 78.676. The recognition accuracy in the salt and pepper noisy environment follows a comparable trend when compared to the Gaussian noisy environment, with a maximum recognition accuracy of 84.785 at p = 0.4 and a noise density of 1%. As the noise level increases, the recognition accuracy declines, reaching a score of 80.147 at p = 0.4 and a noise density of 5%. The analysis of speckle noise further emphasizes the challenges posed by noise in maintaining recognition accuracy. The results show that the recognition accuracy is 84.615 at p = 0.4 and a noise variance of 0.2, but recognition accuracy decreases as the noise level increases, with a score of 80.882 at p = 0.4 and a noise variance of 1.
In Table 4, we present a detailed assessment of 3D object recognition performance using the proposed technique for polynomial order ( O r d = 10) across various environments with differing noise parameters. In the clean environment, the system exhibits a high recognition accuracy, achieving a peak score of 86.765 at p = 0.4. This high accuracy indicates that the system effectively captures the underlying features of the data without the interference of noise. The performance remains commendable across other parameter settings, with scores of 85.294 at p = 0.5 and 83.824 at p = 0.3. However, the lowest score of 81.618 is observed at p = 0.2. When Gaussian noise is applied to the objects, the recognition accuracy shows a gradual decline as the noise level increases. At a noise level of 1%, the system maintains a recognition accuracy of 85.464 at p = 0.4, indicating that it performs well under low noise conditions. However, as the noise level rises to 5%, the accuracy decreases to 82.127. The recognition accuracy for Gaussian noise at different levels demonstrate a consistent downward trend, with accuracy values of 84.615 at p = 0.4 for a noise level of 2% and 83.993 at the same parameter for a noise level of 3%. The performance in the salt and pepper noisy environment reveals the same trends observed in the Gaussian noise scenario. The system achieves a maximum recognition accuracy of 85.351 at p = 0.4 and a noise density of 0.01. As the noise density increases, the recognition accuracy declines, reaching 80.882 at p = 0.4 and a noise density of 0.05. This decline highlights that the system is sensitive to high levels of salt and pepper noise. The recognition accuracy for salt and pepper noise also reflect a gradual decrease, with values of 84.333 at p = 0.4 for a noise density of 0.02 and 83.258 at the same parameter for a noise density of 0.03. The analysis of speckle noise further illustrates the challenges associated with maintaining recognition accuracy in the presence of noise. The system achieves a recognition accuracy of 85.238 at p = 0.4 and a noise variance of 0.2. However, as the noise variance increases, the recognition accuracy declines, with a score of 83.088 at p = 0.4 and a noise variance of 1.
In Table 5, we present a thorough evaluation of 3D object recognition performance using the proposed technique for polynomial order ( O r d = 12) across various environments with differing noise parameters. In the clean environment, the system demonstrates comparable performance, achieving a peak recognition accuracy of 86.029 at p = 0.4. When Gaussian noise is applied to 3D objects, the recognition accuracy exhibits a decline as the noise level increases. At a noise level of 1%, the system maintains a recognition accuracy of 84.898 at p = 0.4, indicating that it performs well under low noise conditions. However, as the noise level rises to 5%, the accuracy decreases to 80.769. The accuracy for Gaussian noise at different levels shows a consistent downward trend, with accuracy values of 83.880 at p = 0.4 for a noise level of 2% and 83.258 at the same parameter for a noise level of 3%. The performance in the salt and pepper noisy environment follows a similar performance to that observed with Gaussian noise. The model achieves a maximum recognition accuracy of 84.842 at p = 0.4 and a noise density of 0.01. As the noise density increases, the recognition accuracy decreases, reaching 81.957 at p = 0.4 and a noise density of 0.05. The scores for salt and pepper noise also reflect a gradual decrease, with values of 84.219 at p = 0.4 for a noise density of 0.02 and 83.032 at the same parameter for a noise density of 0.03. The results of speckle noise illustrate the challenges associated with maintaining recognition accuracy in the presence of noise. The model achieves a recognition accuracy of 84.729 at p = 0.4 and a noise variance of 0.2. However, as the noise variance increases, the recognition accuracy declines, with a score of 81.618 at p = 0.4 and a noise variance of 1.
To this end, the results indicate that while the proposed technique performs well in clean environments, its slightly affected in the presence of noise, especially for OP parameter p = 0.4. The findings highlight that the recognition system based on the proposed technique is able to tackle the problem of noisy environments.

4.2. Comparison Between the Proposed System and Existing Works

To assess the effectiveness of the presented recognition system based on the proposed technique, we conducted a comparative analysis of its recognition accuracy against several existing methods. The algorithms included in this comparison are direct Krawtchouk moment invariants (DKMI), Tchebichef moment invariants (TMI), Krawtchouk moment invariants (KMI), Hahn moment invariants (HMI), Tchebichef–Tchebichef–Tchebichef moment invariants (TTTMI), Krawtchouk–Krawtchouk–Krawtchouk moment invariants (KKKMI), Tchebichef–Krawtchouk–Krawtchouk moment invariants (TKKMI), Tchebichef–Tchebichef–Krawtchouk moment invariants (TTKMI), geometric moment invariants (GMI), and overlapped block processing (OBP). The average recognition accuracy for both the presented recognition system and the existing methods are shown in Table 6.
The comparison presented in Table 6 highlights the performance of the proposed recognition system against several existing methods using the McGill database. The average recognition accuracy of each method is reported, providing a clear perspective on the effectiveness of the proposed approach.
The existing methods, including DKMI, HMI, KMI, and TMI, exhibit relatively low recognition accuracy, ranging from 60.32% to 62.01%. These results indicate that they are not too robust for the complexities presented in the McGill database. Moreover, the Tchebichef-based methods, which are TTTMI, KKKMI, TKKMI, and TTKMI, show improved performance, with an accuracy between 71.11% and 72.87%. This indicates that the incorporation of Tchebichef moments enhances the recognition capabilities compared to the earlier methods. However, even the best-performing Tchebichef-based method, TTKMI, achieves only 72.87%, which still falls short of optimal performance. In contrast, the proposed system demonstrates significantly higher recognition accuracy across all tested configurations. The system achieves an accuracy of 82.73% at polynomial order 8, which surpasses the best existing methods by a notable margin. When the polynomial order is increased to 10, the accuracy improves further to 83.85%, indicating that the proposed system benefits from higher polynomial orders, likely due to its enhanced ability to capture complex features in the data. In addition, for polynomial order of 12, the proposed system shows robust performance with an accuracy of 83.33% This stability of performance across different polynomials orders demonstrates that the proposed approach is robust and adaptable. thus, the variations in the dataset are effectively handled by the proposed approach.
The performance of the proposed approach is further evaluated by benchmarking the proposed approach with the traditional method in terms of computational cost. In the traditional method, the 3D object is convolved by the kernel and then the moments are computed. It is noteworthy that the kernel utilized in the experiment is the averaging kernel with a length of 7.
For this purpose, the computation cost is evaluated by performing five runs for all 3D objects across different classes. Then, the average computation time is recorded for each 3D object. The comparison is conducted for three different polynomial orders ( O r d = 8, 10, and 12), and the results are summarized in Table 7. The table includes the computation time for both the proposed approach (Time (ours)) and the traditional method (Time (Trad)), as well as the percentage improvement (Imp) achieved by the proposed method over the traditional method. The percentage improvement is calculated using the following formula:
P e r c e n t a g e I m p r o v e m e n t ( I m p ) = T T r a d T O u r s T T r a d × 100 %
The minimum percentage improvement is observed for Class 16, with improvements of 47.0% at O r d = 8 , 47.7% at O r d = 10, and 46.1% at O r d = 12. In contrast, the highest percentage improvement is achieved for Class 19, where the proposed method attains improvements of 64.2% at O r d = 8 , 64.1% at O r d = 10, and 64.0% at O r d = 12.
The results demonstrate that the proposed algorithm significantly outperforms the traditional algorithm across all polynomial orders and classes. For instance, at O r d = 8 , the average percentage improvement across all classes is 54.7%, while at O r d = 10 and O r d = 12 , the average improvements are 54.8% and 53.4%, respectively. This indicates that the proposed method consistently reduces computation time by more than half compared to the traditional approach. This clearly signifies the robustness and efficiency of the proposed algorithm, particularly when dealing with higher polynomial orders and complex 3D objects. The last row of the table provides the average computation time and improvement per 3D object, further confirming the superiority of the proposed method in terms of computational efficiency.

5. Conclusions

Signal preprocessing plays a crucial role in computer vision applications, as it directly enhances the performance of computer vision algorithms. OMs are considered as robust descriptors for signal analysis and recognition. Convolving kernels with the signal and precomputing OMs increases computational demands. Therefore, this paper introduced a novel mathematical model that embeds the kernel directly into the OPs functions by integrating the processes into a single, streamlined framework. By computing the OMs of smoothed 3D signals directly, the model achieved a substantial reduction in computational overhead while maintaining high accuracy. Our experimental work on 3D objects demonstrated that the proposed algorithm outperforms traditional methods. When the polynomial order is equal to 10, the accuracy improves further to 83.85%, indicating that the proposed work benefits from higher polynomial orders, likely due to its enhanced ability to capture complex features. In the future, we plan to further enhance the performance of the proposed method and apply it to various applications, such as medical imaging and robotics.

Author Contributions

Conceptualization, A.A.M. and A.H.A.-s.; methodology, B.M.M., A.A.M. and S.H.A.; validation, A.A., M.A., A.M.A.-H. and A.H.; investigation, A.H., M.A., A.A., A.M.A.-H. and A.H.A.-s.; resources, B.M.M. and A.M.A.-H.; writing—original draft preparation, B.M.M., S.H.A., M.A., A.H.A.-s. and A.A.M.; writing—review and editing, A.H.A.-s., B.M.M. and A.H.; visualization, A.M.A.-H., S.H.A., A.A. and A.A.M.; supervision, S.H.A.; project administration, S.H.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is contained within the article.

Acknowledgments

The authors would like to thank the University of Baghdad for general help and support.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
1DOne-dimensional
2DTwo-dimensional
3DThree-dimensional
DKMIDirect Krawtchouk moment invariants
TMITchebichef moment invariants
KMIKrawtchouk moment invariants
HMIHahn moment invariants
TTTMITchebichef–Tchebichef–Tchebichef moment invariants
KKKMIKrawtchouk–Krawtchouk–Krawtchouk moment invariants
TKKMITchebichef–Krawtchouk–Krawtchouk moment invariants
TTKMITchebichef–Tchebichef–Krawtchouk moment invariants
GMIGeometric moment invariants
OBPOverlapped block processing
OMOrthogonal moments
OPOrthogonal polynomials
OrdPolynomial order
SKTPSquared Krawtchouk–Tchebichef polynomials

References

  1. Guo, Y.; Wang, H.; Hu, Q.; Liu, H.; Liu, L.; Bennamoun, M. Deep learning for 3d point clouds: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 4338–4364. [Google Scholar] [CrossRef] [PubMed]
  2. Abood, Z.I.; Ismail, T.Z. 3-D Object Recognition using Multi-Wavelet and Neural Network. J. Eng. 2012, 18, 78–94. [Google Scholar]
  3. de Souza Cardoso, L.F.; Mariano, F.C.M.Q.; Zorzal, E.R. A survey of industrial augmented reality. Comput. Ind. Eng. 2020, 139, 106159. [Google Scholar] [CrossRef]
  4. Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-view 3d object detection network for autonomous driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1907–1915. [Google Scholar]
  5. Rusu, R.B.; Cousins, S. 3d is here: Point cloud library (pcl). In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 1–4. [Google Scholar]
  6. Fidvi, S.; Holder, J.; Li, H.; Parnes, G.J.; Shamir, S.B.; Wake, N. Advanced 3D visualization and 3D printing in radiology. In Biomedical Visualisation: Volume 15—Visualisation in Teaching of Biomedical and Clinical Subjects: Anatomy, Advanced Microscopy and Radiology; Springer: Berlin/Heidelberg, Germany, 2023; pp. 103–138. [Google Scholar]
  7. Sohail, S.S.; Himeur, Y.; Kheddar, H.; Amira, A.; Fadli, F.; Atalla, S.; Copiaco, A.; Mansoor, W. Advancing 3D point cloud understanding through deep transfer learning: A comprehensive survey. Inf. Fusion 2024, 113, 102601. [Google Scholar] [CrossRef]
  8. Rao, Q.; Chakraborty, S. In-vehicle object-level 3D reconstruction of traffic scenes. IEEE Trans. Intell. Transp. Syst. 2020, 22, 7747–7759. [Google Scholar] [CrossRef]
  9. Hussain, Z.; Umar, M.; Faisal, M.; Bilal, M.A.; Abbas, S. Synergistic Integration of AI and 6-DOF Robotic Arm for Enhanced Bomb Disposal Capabilities. J. Inst. Eng. India Ser. C 2024, 105, 1247–1261. [Google Scholar] [CrossRef]
  10. Flusser, J.; Suk, T.; Zitová, B. 2D and 3D Image Analysis by Moments; John Wiley & Sons: Hoboken, NJ, USA, 2016. [Google Scholar]
  11. Ali, I.; Saleem, M.T. Applications of orthogonal polynomials in simulations of mass transfer diffusion equation arising in food engineering. Symmetry 2023, 15, 527. [Google Scholar] [CrossRef]
  12. Qi, S.; Zhang, Y.; Wang, C.; Zhou, J.; Cao, X. A survey of orthogonal moments for image representation: Theory, implementation, and evaluation. ACM Comput. Surv. CSUR 2021, 55, 1–35. [Google Scholar] [CrossRef]
  13. Adatrao, S.; Mittal, M. An analysis of different image preprocessing techniques for determining the centroids of circular marks using hough transform. In Proceedings of the 2016 2nd International Conference on Frontiers of Signal Processing (ICFSP), Warsaw, Poland, 15–17 October 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 110–115. [Google Scholar]
  14. Szeg, G. Orthogonal Polynomials; American Mathematical Society: Providence, RI, USA, 1939; Volume 23. [Google Scholar]
  15. Marcellán, F.; Huertas, E.J. Orthogonal Polynomials: Current Trends and Applications: Proceedings of the 7th EIBPOA Conference; Springer Nature: Berlin/Heidelberg, Germany, 2021; Volume 22. [Google Scholar]
  16. Mukundan, R. Some Computational Aspects of Discrete Orthonormal Moments. IEEE Trans. Image Process. 2004, 13, 1055–1059. [Google Scholar] [CrossRef]
  17. Mukundan, R.; Ong, S.; Lee, P. Image analysis by Tchebichef moments. IEEE Trans. Image Process. 2001, 10, 1357–1364. [Google Scholar] [CrossRef] [PubMed]
  18. Abdulqader, D.A.; Hathal, M.S.; Mahmmod, B.M.; Abdulhussain, S.H.; Al-Jumeily, D. Plain, Edge, and Texture Detection Based on Orthogonal Moment. IEEE Access 2022, 10, 114455–114468. [Google Scholar] [CrossRef]
  19. Abdulhussain, S.H.; Ramli, A.R.; Hussain, A.J.; Mahmmod, B.M.; Jassim, W.A. Orthogonal polynomial embedded image kernel. In Proceedings of the International Conference on Information and Communication Technology—ICICT ’19, New York, NY, USA, 15–16 April 2019; pp. 215–221. [Google Scholar] [CrossRef]
  20. Hosny, K.M.; Elaziz, M.A. Face Recognition Using Exact Gaussian-Hermit Moments. In Recent Advances in Computer Vision; Springer: Berlin/Heidelberg, Germany, 2019; pp. 169–187. [Google Scholar]
  21. Batioua, I.; Benouini, R.; Zenkouar, K.; Zahi, A. Image classification using separable invariants moments based on Racah polynomials. Procedia Comput. Sci. 2018, 127, 320–327. [Google Scholar] [CrossRef]
  22. Mizel, A.K.E. Orthogonal Functions Solving Linear functional Differential EquationsUsing Chebyshev Polynomial. Baghdad Sci. J. 2008, 5, 143–148. [Google Scholar] [CrossRef]
  23. Shihab, S.N.; Ouda, E.H.; Ibraheem, S.F. Boubaker Wavelets Functions: Properties and Applications. Baghdad Sci. J. 2021, 18, 1226. [Google Scholar] [CrossRef]
  24. Salih, O.M.; AL-Jawary, M.A. Computational methods for solving nonlinear ordinary differential equations arising in engineering and applied sciences. Iraqi J. Sci. 2023, 64, 4070–4091. [Google Scholar] [CrossRef]
  25. AL-Jawary, M.A.; Salih, O.M. Efficient computational methods for solving the one-dimensional parabolic equation with nonlocal initial and boundary conditions. Iraqi J. Sci. 2024, 65, 3345–3377. [Google Scholar] [CrossRef]
  26. Hussein, H.A.; Hameed, S.M.; Mahmmod, B.M.; Abdulhussain, S.H.; Hussain, A.J. Dual Stages of Speech Enhancement Algorithm Based on Super Gaussian Speech Models. J. Eng. 2023, 29, 1–13. [Google Scholar] [CrossRef]
  27. Rivero-Castillo, D.; Pijeira, H.; Assunçao, P. Edge detection based on Krawtchouk polynomials. J. Comput. Appl. Math. 2015, 284, 244–250. [Google Scholar] [CrossRef]
  28. Mahmmod, B.M.; bin Ramli, A.R.; Abdulhussain, S.H.; Al-Haddad, S.A.R.; Jassim, W.A. Signal compression and enhancement using a new orthogonal-polynomial-based discrete transform. IET Signal Process. 2018, 12, 129–142. [Google Scholar] [CrossRef]
  29. Mahmmod, B.M.; Ramli, A.R.; Abdulhussian, S.H.; Al-Haddad, S.A.R.; Jassim, W.A. Low-Distortion MMSE Speech Enhancement Estimator Based on Laplacian Prior. IEEE Access 2017, 5, 9866–9881. [Google Scholar] [CrossRef]
  30. Yap, P.T.; Paramesran, R. Local watermarks based on Krawtchouk moments. In Proceedings of the 2004 IEEE Region 10 Conference TENCON 2004, Chiang Mai, Thailand, 21–24 November 2004; IEEE: Piscataway, NJ, USA, 2004; Volume B, pp. 73–76. [Google Scholar] [CrossRef]
  31. Bharathi, V.S.; Ganesan, L. Orthogonal moments based texture analysis of CT liver images. Pattern Recognit. Lett. 2008, 29, 1868–1872. [Google Scholar] [CrossRef]
  32. Krishnamoorthy, R.; Amudhavalli, N.; Sivakkolunthu, M. An adaptive mammographic image enhancement in orthogonal polynomials domain. Int. J. Biomed. Biol. Eng. 2010, 4, 687–695. [Google Scholar]
  33. Kumar, S.N.; Ahilan, A.; Haridhas, A.K.; Sebastian, J. Gaussian Hermite polynomial based lossless medical image compression. Multimed. Syst. 2021, 27, 15–31. [Google Scholar] [CrossRef]
  34. Idan, Z.N.; Abdulhussain, S.H.; Mahmmod, B.M.; Al-Utaibi, K.A.; Al-Hadad, S.A.R.; Sait, S.M. Fast Shot Boundary Detection Based on Separable Moments and Support Vector Machine. IEEE Access 2021, 9, 106412–106427. [Google Scholar] [CrossRef]
  35. Zhu, H.; Liu, M.; Shu, H.; Zhang, H.; Luo, L. General form for obtaining discrete orthogonal moments. IET Image Process. 2010, 4, 335–352. [Google Scholar] [CrossRef]
  36. Teague, M.R. Image analysis via the general theory of moments. J. Opt. Soc. Am. 1980, 70, 920–930. [Google Scholar] [CrossRef]
  37. Teh, C.H.; Chin, R.T. On image analysis by the methods of moments. IEEE Trans. Pattern Anal. Mach. Intell. 1988, 10, 496–513. [Google Scholar] [CrossRef]
  38. Vashishth, T.K.; Sharma, V.; Sharma, K.K.; Kumar, B.; Chaudhary, S.; Panwar, R. Enhancing biomedical signal processing with machine learning: A comprehensive review. In Digital Transformation in Healthcare 5.0: Volume 1: IoT, AI and Digital Twin; Walter de Gruyter: Berlin, Germany, 2024; p. 277. [Google Scholar]
  39. Abdulhussain, S.H.; Ramli, A.R.; Saripan, M.I.; Mahmmod, B.M.; Al-Haddad, S.; Jassim, W.A. Methods and Challenges in Shot Boundary Detection: A Review. Entropy 2018, 20, 214. [Google Scholar] [CrossRef]
  40. Barca, E.; Castrignanò, A.; Ruggieri, S.; Rinaldi, M. A new supervised classifier exploiting spectral-spatial information in the Bayesian framework. Int. J. Appl. Earth Obs. Geoinf. 2020, 86, 101990. [Google Scholar] [CrossRef]
  41. Zhou, W.; Yu, L.; Qiu, W.; Zhou, Y.; Wu, M. Local gradient patterns (LGP): An effective local-statistical-feature extraction scheme for no-reference image quality assessment. Inf. Sci. 2017, 397, 1–14. [Google Scholar] [CrossRef]
  42. Song, M.; Guo, Q. Efficient 3D object recognition in mobile edge environment. J. Cloud Comput. 2022, 11, 92. [Google Scholar] [CrossRef]
  43. Ullman, S. Three-dimensional object recognition based on the combination of views. Cognition 1998, 67, 21–44. [Google Scholar] [CrossRef]
  44. Tan, F.; Zhai, M.; Zhai, C. Foreign object detection in urban rail transit based on deep differentiation segmentation neural network. Heliyon 2024, 10, e37072. [Google Scholar] [CrossRef] [PubMed]
  45. Muzahid, A.; Han, H.; Zhang, Y.; Li, D.; Zhang, Y.; Jamshid, J.; Sohel, F. Deep learning for 3D object recognition: A survey. Neurocomputing 2024, 608, 128436. [Google Scholar] [CrossRef]
  46. Abdulhussain, S.H.; Ramli, A.R.; Mahmmod, B.M.; Saripan, M.I.; Al-Haddad, S.; Jassim, W.A. A New Hybrid form of Krawtchouk and Tchebichef Polynomials: Design and Application. J. Math. Imaging Vis. 2018, 61, 555–570. [Google Scholar] [CrossRef]
  47. Jassim, W.A.; Raveendran, P. Face Recognition Using Discrete Tchebichef-Krawtchouk Transform. In Proceedings of the 2012 IEEE International Symposium on Multimedia (ISM), Irvine, CA, USA, 10–12 December 2012; pp. 120–127. [Google Scholar] [CrossRef]
  48. Doukas, C.; Maglogiannis, I. Region of interest coding techniques for medical image compression. IEEE Eng. Med. Biol. Mag. 2007, 26, 29–35. [Google Scholar] [CrossRef]
  49. Mukundan, R.; Raveendran, P.; Jassim, W. New orthogonal polynomials for speech signal and image processing. IET Signal Process. 2012, 6, 713–723. [Google Scholar] [CrossRef]
  50. Mahmmod, B.M.; Abdulhussain, S.H.; Naser, M.A.; Alsabah, M.; Hussain, A.; Al-Jumeily, D. 3D Object Recognition Using Fast Overlapped Block Processing Technique. Sensors 2022, 22, 9209. [Google Scholar] [CrossRef]
  51. Wei, T.; Tian, Y.; Wang, Y.; Liang, Y.; Chen, C.W. Optimized separable convolution: Yet another efficient convolution operator. AI Open 2022, 3, 162–171. [Google Scholar] [CrossRef]
  52. Byun, H.; Lee, S.W. A survey on pattern recognition applications of support vector machines. Int. J. Pattern Recognit. Artif. Intell. 2003, 17, 459–486. [Google Scholar] [CrossRef]
  53. Awad, M.; Motai, Y. Dynamic classification for video stream using support vector machine. Appl. Soft Comput. 2008, 8, 1314–1325. [Google Scholar] [CrossRef]
  54. Siddiqi, K.; Zhang, J.; Macrini, D.; Shokoufandeh, A.; Bouix, S.; Dickinson, S. Retrieving articulated 3-D models using medial surfaces. Mach. Vis. Appl. 2008, 19, 261–275. [Google Scholar] [CrossRef]
  55. Chang, C.C.; Lin, C.J. LIBSVM. ACM Trans. Intell. Syst. Technol. 2011, 2, 27. [Google Scholar] [CrossRef]
  56. Benouini, R.; Batioua, I.; Zenkouar, K.; Najah, S.; Qjidaa, H. Efficient 3D object classification by using direct Krawtchouk moment invariants. Multimed. Tools Appl. 2018, 77, 27517–27542. [Google Scholar] [CrossRef]
  57. Batioua, I.; Benouini, R.; Zenkouar, K.; Zahi, A.; Hakim, E.F. 3D image analysis by separable discrete orthogonal moments based on Krawtchouk and Tchebichef polynomials. Pattern Recognit. 2017, 71, 264–277. [Google Scholar] [CrossRef]
Figure 1. Samples showing the effect of kernels on ant 3D objects using conventional convolution and OP-based convolution.
Figure 1. Samples showing the effect of kernels on ant 3D objects using conventional convolution and OP-based convolution.
Algorithms 18 00078 g001
Figure 2. Samples showing the effect of kernels on plane 3D objects using conventional convolution and OP-based convolution.
Figure 2. Samples showing the effect of kernels on plane 3D objects using conventional convolution and OP-based convolution.
Algorithms 18 00078 g002
Figure 3. Flow diagram of the proposed embedded kernel.
Figure 3. Flow diagram of the proposed embedded kernel.
Algorithms 18 00078 g003
Figure 4. Flow diagram of object recognition using the proposed technique.
Figure 4. Flow diagram of object recognition using the proposed technique.
Algorithms 18 00078 g004
Figure 5. Samples of 3D objects extracted from McGill dataset.
Figure 5. Samples of 3D objects extracted from McGill dataset.
Algorithms 18 00078 g005
Table 1. Details of the McGill dataset.
Table 1. Details of the McGill dataset.
Class IDClass NameNumber of SamplesClass IDClass NameNumber of Samples
1airplanes2611hands20
2ants3012humans26
3birds2113octopus25
4chairs2314pliers20
5crabs3015snakes25
6cups2516spectacles25
7dinosaurs1917spiders31
8dolphins1218tables22
9fishes2319teddy20
10four31
Total number of samples454
Table 2. Recognition accuracy for clean environment with different OP parameter p and order.
Table 2. Recognition accuracy for clean environment with different OP parameter p and order.
p0.20.30.40.50.6Average
Ord
869.85383.25885.29483.08876.47179.593
1081.10983.82486.76585.24982.35383.971
1277.94182.35386.02984.55978.67681.912
Table 3. Recognition accuracy for noisy environments with different OP parameter p and order ( O r d = 8).
Table 3. Recognition accuracy for noisy environments with different OP parameter p and order ( O r d = 8).
Environment p
0.20.30.40.50.6
Clean 69.85383.25885.29483.08876.471
Gaussian0.0169.57082.69284.84282.63676.357
0.0267.98680.82683.93781.84475.452
0.0366.51678.28182.69280.99575.170
0.0464.64977.14980.71380.37372.455
0.0563.46277.09378.67678.95969.514
Salt and Pepper0.0169.34482.52384.78582.69275.848
0.0267.36481.50583.76781.67475.226
0.0366.34680.76982.86280.48674.717
0.0465.66779.01681.50579.80873.756
0.0564.59378.28180.14779.41272.794
Speckle0.269.62782.29684.61582.69276.018
0.469.06180.93983.76782.52375.622
0.668.72279.29983.25881.84475.057
0.868.38277.77181.95781.22273.416
167.59075.90580.88280.88272.059
Average of both clean and noisy environments67.42179.85082.73181.32174.371
Table 4. Recognition accuracy for noisy environments with different OP parameter p and order ( O r d = 10).
Table 4. Recognition accuracy for noisy environments with different OP parameter p and order ( O r d = 10).
Environment p
0.20.30.40.50.6
Clean 81.61883.82486.76585.29482.353
Gaussian0.0181.10982.86285.46484.44682.070
0.0279.92181.50584.61584.05079.016
0.0378.05480.60083.99383.48478.281
0.0476.64079.97782.97581.50577.771
0.0575.73577.60282.12780.14777.206
Salt and Pepper0.0181.33583.14585.35184.10682.353
0.0279.80882.01484.33383.48481.505
0.0377.94180.60083.25882.52380.430
0.0476.86779.35581.61881.95779.299
0.0575.96278.45080.88281.61878.054
Speckle0.281.16583.42885.23884.38982.127
0.479.75181.95784.27683.99382.014
0.678.05480.37384.21982.97581.222
0.875.39678.50783.42882.18380.260
174.43476.47183.08881.44879.412
Average of both clean and noisy environments78.36280.66783.85282.97580.211
Table 5. Recognition accuracy for noisy environments with different OP parameter p and order ( O r d = 12).
Table 5. Recognition accuracy for noisy environments with different OP parameter p and order ( O r d = 12).
Environment p
0.20.30.40.50.6
Clean 77.94182.35386.02984.55978.676
Gaussian0.0177.43281.90084.89884.27677.941
0.0275.39680.82683.88083.48475.905
0.0374.09578.50783.25882.12775.622
0.0473.58677.77182.52381.05275.113
0.0572.73876.98080.76978.95974.717
Salt and Pepper0.0177.48982.12784.84284.33378.563
0.0276.13181.16584.21983.42877.602
0.0375.05779.24283.03282.07076.414
0.0474.09578.39482.57981.44874.038
0.0573.19077.20681.95779.41273.529
Speckle0.277.20681.16584.72984.10677.941
0.474.71779.12983.76783.25877.262
0.673.02076.86782.97582.97575.962
0.871.94674.94382.35381.73175.170
170.92874.88781.61880.99574.265
Average of both clean and noisy environments74.68578.96683.33982.38876.170
Table 6. Comparison between the presented recognition system and existing methods using the McGill database.
Table 6. Comparison between the presented recognition system and existing methods using the McGill database.
Method NameAverage Recognition AccuracyAverage of Recognition Accuracy for Noisy Environment
DKMI [56]62.01%57.92%
HMI [56]60.89%56.69%
KMI [56]60.32%56.44%
TMI [56]60.54%56.40%
TTTMI [57]71.57%67.30%
KKKMI [57]71.11%66.97%
TKKMI [57]72.19%67.64%
TTKMI [57]72.87%68.20%
GMI [57]70.26%65.49%
OBP [50]80.21%80.22%
Proposed System ( p = 0.4 and O r d = 8 )82.73%82.56%
Proposed System ( p = 0.4 and O r d = 10 )83.85%83.66%
Proposed System ( p = 0.4 and O r d = 12 )83.33%83.16%
Table 7. Computation time (in seconds) comparison between the proposed approach (Our) and traditional method (Trad).
Table 7. Computation time (in seconds) comparison between the proposed approach (Our) and traditional method (Trad).
Ord = 8 Ord = 10 Ord = 12
Class IDTime (Trad)Time (Ours)ImpTime (Trad)Time (Ours)ImpTime (Trad)Time (Ours)Imp
11.4460.72250.01.5540.75051.71.5720.80548.8
21.7900.82453.91.8550.86653.31.9950.91654.1
31.1980.58051.61.2390.60451.21.2630.64149.3
41.3020.63851.01.3610.66950.91.4900.71751.9
52.2100.83162.42.2580.87061.52.2860.92559.5
62.0680.71565.42.1480.71866.62.1910.77864.5
71.0730.52750.91.1530.54852.51.1730.58550.2
80.6610.33149.90.7340.34652.90.7620.36851.8
91.3180.63152.11.3970.66952.11.4240.70750.3
101.8340.85653.31.9210.90253.12.0300.95752.9
111.2450.54756.11.3000.57855.51.3540.61454.6
121.4450.71450.51.5090.74950.31.5490.80348.1
131.4990.69053.91.5560.72553.41.6050.77152.0
141.0670.55348.21.1420.57749.41.1980.61548.7
151.4810.68353.91.5650.72353.81.5950.76652.0
161.2890.68347.01.3710.71747.71.4250.76846.1
171.8380.84853.91.9290.88854.01.9780.95851.6
181.3930.60656.51.4200.63555.31.4590.69652.3
191.5450.55364.21.6160.57964.11.7140.61764.0
Average27.7012.53554.729.0313.1154.830.0614.0053.4
Avg for 3D object 10.0610.02854.70.0640.02954.80.0660.03153.4
1 Avg per 3D object represents the average computation time for one 3D object. It is computed by summing all the computation time and dividing it by the number of objects (454)—the test is performed on the original dataset without augmentation.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mohammed, A.A.; Al-sudani, A.H.; Abdul-Hadi, A.M.; Alwhelat, A.; Mahmmod, B.M.; Abdulhussain, S.H.; Alsabah, M.; Hussain, A. Three-Dimensional Object Recognition Using Orthogonal Polynomials: An Embedded Kernel Approach. Algorithms 2025, 18, 78. https://doi.org/10.3390/a18020078

AMA Style

Mohammed AA, Al-sudani AH, Abdul-Hadi AM, Alwhelat A, Mahmmod BM, Abdulhussain SH, Alsabah M, Hussain A. Three-Dimensional Object Recognition Using Orthogonal Polynomials: An Embedded Kernel Approach. Algorithms. 2025; 18(2):78. https://doi.org/10.3390/a18020078

Chicago/Turabian Style

Mohammed, Aqeel Abdulazeez, Ahlam Hanoon Al-sudani, Alaa M. Abdul-Hadi, Almuntadher Alwhelat, Basheera M. Mahmmod, Sadiq H. Abdulhussain, Muntadher Alsabah, and Abir Hussain. 2025. "Three-Dimensional Object Recognition Using Orthogonal Polynomials: An Embedded Kernel Approach" Algorithms 18, no. 2: 78. https://doi.org/10.3390/a18020078

APA Style

Mohammed, A. A., Al-sudani, A. H., Abdul-Hadi, A. M., Alwhelat, A., Mahmmod, B. M., Abdulhussain, S. H., Alsabah, M., & Hussain, A. (2025). Three-Dimensional Object Recognition Using Orthogonal Polynomials: An Embedded Kernel Approach. Algorithms, 18(2), 78. https://doi.org/10.3390/a18020078

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop