Next Article in Journal
CT Rendering and Radiomic Analysis in Post-Chemotherapy Retroperitoneal Lymph Node Dissection for Testicular Cancer to Anticipate Difficulties for Young Surgeons
Previous Article in Journal
GANs for Medical Image Synthesis: An Empirical Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On The Potential of Image Moments for Medical Diagnosis

by
Cecilia Di Ruberto
1,†,
Andrea Loddo
1,*,† and
Lorenzo Putzu
2,†
1
Department of Mathematics and Computer Science, University of Cagliari, Via Ospedale 72, 09124 Cagliari, Italy
2
Department of Electrical and Electronic Engineering, University of Cagliari, Piazza d’Armi, 09123 Cagliari, Italy
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Imaging 2023, 9(3), 70; https://doi.org/10.3390/jimaging9030070
Submission received: 3 February 2023 / Revised: 24 February 2023 / Accepted: 11 March 2023 / Published: 17 March 2023
(This article belongs to the Topic Medical Image Analysis)

Abstract

:
Medical imaging is widely used for diagnosis and postoperative or post-therapy monitoring. The ever-increasing number of images produced has encouraged the introduction of automated methods to assist doctors or pathologists. In recent years, especially after the advent of convolutional neural networks, many researchers have focused on this approach, considering it to be the only method for diagnosis since it can perform a direct classification of images. However, many diagnostic systems still rely on handcrafted features to improve interpretability and limit resource consumption. In this work, we focused our efforts on orthogonal moments, first by providing an overview and taxonomy of their macrocategories and then by analysing their classification performance on very different medical tasks represented by four public benchmark data sets. The results confirmed that convolutional neural networks achieved excellent performance on all tasks. Despite being composed of much fewer features than those extracted by the networks, orthogonal moments proved to be competitive with them, showing comparable and, in some cases, better performance. In addition, Cartesian and harmonic categories provided a very low standard deviation, proving their robustness in medical diagnostic tasks. We strongly believe that the integration of the studied orthogonal moments can lead to more robust and reliable diagnostic systems, considering the performance obtained and the low variation of the results. Finally, since they have been shown to be effective on both magnetic resonance and computed tomography images, they can be easily extended to other imaging techniques.

1. Introduction

Medical imaging is widely used to manage various types of pathologies for diagnosis and postoperative or post-therapy monitoring. As a result, the number of biomedical images has increased significantly, meaning that pathologists and doctors spend much time searching for diseases in images that could be used to treat patients. This is even more important considering that about 80% of biomedical images turn out to be benign [1]. Moreover, the manual interpretation of biomedical images is a time-consuming and challenging task that requires experienced operators, but the results are still subjective, influenced by the operator’s competence or level of fatigue. In contrast, accuracy and efficiency are crucial in this field. Therefore, the development of automated tools capable of effectively handling this massive number of images has become essential.
Computer-aided diagnosis (CAD) systems can assist pathologists and physicians in diagnosing and postoperative or post-therapy monitoring. Such systems attempt to mimic manual inspection [2] but can provide accurate results while reducing subjectivity and speeding up the analysis process [3]. The growing need for automated approaches has attracted the attention of many researchers in this field, who have proposed several CAD systems based on image processing and machine learning techniques [4,5].
Existing CAD systems can be more or less complex, depending on the medical task and the type and quality of images to be analysed. In this study, we focused on methods for direct image classification, which consists of two main steps: feature extraction and classification. The first step generally consists of extracting a set of parameters from an image to characterise the shapes, colour, and texture it contains [6]. In the second step, the extracted features are used to build a model of known cases (during the training phase) and provide a diagnosis of unknown cases (during the testing phase).
Since the advent of convolutional neural networks (CNNs) and in particular, with the proposal of the most famous AlexNet [7] and very deep convolutional networks (VGGNet) [8], many authors have focused on this approach as the sole method of diagnosis. In fact, CNNs can perform feature extraction and classification steps simultaneously because of their internal structure.
Specifically, CNNs can be used as feature extractors by extracting activations from a single layer or multiple layers that can globally describe the images [9]. Such features can then be used to train standard machine learning models, such as ensemble techniques [10]. The use of CNNs in CAD systems dramatically simplifies the design and analysis work but also leads to the development of much more complex CAD systems whose results are often challenging to interpret [11]. For these reasons, many CAD systems still rely on handcrafted features, especially global image descriptors such as texture features and orthogonal moments [6,9,10]. Texture feature methods compute simple or more sophisticated statistical properties, e.g., co-occurrence matrices [12,13,14] or local binary patterns (LBP) [15]. Orthogonal moments are statistical measures for describing objects in images regardless of their position, viewing angle, and illumination, due to their ability to represent global features of an image [16]. In addition, they are less sensitive to noise and can represent the properties of an image without redundancy [17,18].
Due to their significant properties, orthogonal moments have been widely used in various tasks ranging from edge detection [19] to image classification [20]. In particular, they have been applied in several CAD systems, from the diagnosis of breast masses in mammographic images [21] to the analysis of liver textures in computed tomography (CT) scans [22] or the study of ultrasound images of the prostate [23].
The first moments to be introduced were Hu’s moments [24] in 1962, and since then, interest in this type of descriptor has grown considerably. Many works have followed to propose different kinds of moments (circular or Cartesian) [16] or to present more efficient and, at the same time, more accurate formulations [25].
In this work, we perform a detailed comparison between all the best-known types of orthogonal moments in order to evaluate their effectiveness in diagnostics when applied to very different medical images. This work is the first in which the performance of different moments is compared on the same tasks. Indeed, most works in the literature exploit just single moments [21,26,27,28,29,30] or a combination of moments [23,31,32,33] to solve a particular problem. For completeness, we compare the performance of the moments with the best-known texture features and deep features (extracted from pretrained CNNs). It is important to note that deep features are used in this work without any explicit training or fine-tuning of new learners.
Our contribution is fourfold: (i) we reviewed the state of the art on orthogonal moments by describing the main features and differences between them; (ii) we summarised the main types of moments, emphasising their differences; (iii) we examined and compared orthogonal moments in several medical diagnostic tasks; and (iv) we compared orthogonal moments with the best-known texture features and the latest CNNs features. Our numerical experiments allow us to identify the most effective category of moments for the tasks under consideration and also show a particularly encouraging behaviour. Indeed, they confirm that CNNs perform excellently in all tasks, but orthogonal moments seem competitive with them. The orthogonal moments have very little resource consumption, so they can be considered a great alternative to well-known CNN-based descriptors. The rest of this paper is organised as follows. Section 2 discusses related work and the role of orthogonal moments in the medical field. In Section 3, we describe orthogonal moments, their categorisations, and their main properties. In Section 4, we describe the materials and methods used in the experimental evaluation, then we present the results in Section 5.2 and discuss them in Section 5.3. Finally, in Section 6, we summarise the content of the article and outline directions for future work.

2. Related Works

The moments are statistical measures used to characterise a function, capture its most important features, and thus obtain relevant information about an object. Hu was the first to use the invariants of geometric moments in pattern recognition [24] and demonstrated the discriminative power of these features in the case of recognition of printed capital letters. Non-orthogonal moments have lost their attractiveness in favour of orthogonal moments constructed in terms of orthogonal polynomials because the non-orthogonal property of these invariant geometric moments causes information redundancy. It has been shown that orthogonal moments are less sensitive to noise and have an efficient ability to represent features with a low degree of information redundancy. Among the best-known are the discrete Chebyshev moments [34], the discrete Krawtchouk moments [35], the Legendre moments [36], and the Zernike moments [37]. For these reasons, orthogonal moments have received a great deal of attention in recent years, and a large number of novel moments have been proposed.
Recent surveys, such as those conducted by Qi et al. [38] and Singh et al. [16], have reviewed various families of orthogonal moments, their theoretical foundations, implementation methods, and evaluation techniques for measuring their performance in various image analysis tasks. In particular, the latter conducted a comprehensive review of methods for achieving rotation invariance in orthogonal moments and transforms used for image representation, including using rotationally invariant functions, the introduction of rotation parameters, and the use of multiple basis functions.
Image moments have proven to be versatile tools with numerous applications. In quality assessment tasks, moments have been successfully used for the evaluation of authentically distorted images [39]. Additionally, they have also been applied to the recognition of symmetric objects [40], which can be beneficial in robotics and object recognition applications in computer vision, image compression [41], or image watermarking [42].
However, moments find their most extensive application in the medical field. Their usage encompasses a wide range of tasks, including the reconstruction of medical images afflicted by noise, such as CT, magnetic resonance (MR), and X-ray images [43], the detection of pathological conditions, such as brain tumours [44], and the classification of medical images. Classification examples include the work of Wu et al. [23], which used moments to describe the texture of CT liver images and to detect tumours in ultrasound prostate images. Similarly, Iscan et al. [45] and Tahmasbi [21] used moments to detect tumours in brain and mammography images, respectively. Further examples include the works using moments to classify parasites in microscopy images by Dogantekin et al. [46] and spermatogonium by Liyun et al. [47].
In recent decades, the use of moments as features in various applications has become widespread. This popularity can be attributed to their ability to represent an image’s salient features effectively. Numerous studies have employed different categories of moments to serve as image descriptors to accomplish specific tasks. For example, Hu, Legendre, and Zernike moments were investigated and compared to automatically classify boar spermatozoa acrosomes as intact or damaged [48]. In that work, images of boar spermatozoa were then classified by k-nearest neighbour (k-NN) and multilayer perceptron classifiers. Instead, in classifying histological images, Di Ruberto et al. [49] realised a comparative evaluation of different moments, such as Hu, Legendre, and Zernike’s, and texture features, such as local binary patterns and Haralick features. The best results are reported when the orthogonal moments are combined with co-occurrence matrices.
As far as Zernike moments are concerned, on the one hand, Tahmasbi et al. [21] used them as descriptors for identifying breast masses based on their shape and edge characteristics. The study involved extracting two sets of Zernike moments from preprocessed images, each consisting of 32 moments with varying orders and iterations, and using them at the feature selection stage. On the other hand, pseudo-Zernike moments were used to classify leukocyte subtypes in Ryabchykov et al. [26], whereas Elaziz et al. [27] applied fractional-order orthogonal descriptors for bacterial species recognition.
Regarding Chebyshev moments, they were evaluated by Mukundan [28]. The authors explored 5 × 5 neighbourhoods of pixels to encode texture information as a Lehmer code, representing the relative strengths of the moments considered. The generated local descriptors were then used to classify tissue images into six classes.
Krawtchouk moments were used alone for blind integrity verification of medical images in Xu et al. [29] and combined in the extensive work of Batioua et al. [30]. The authors propose new sets of separable discrete moments for 3D image analysis called TKKM (Tchebichef–Krawtchouk–Krawtchouk moments) and TTKM (Tchebichef–Tchebichef–Krawtchouk moments). The study indicated that the proposed 3D separable moments significantly improved representation capabilities compared to traditional Krawtchouk and Tchebichef moments.
Although Hu moments have the well-known problems of noise sensitivity and information redundancy caused by the nonorthogonality of the geometric basis [30], they have been used with equal success. To cite some examples, Siti et al. [50] employed them to classify breast cancer, Laine et al.applied Hu moments to classify the structure of large populations of biopharmaceutical viruses at high resolution [51], and, more recently, in [52], Hu moments were extracted from segmented ultrasound arterial wall images for assessing the risk of atherosclerosis and predicting the chance of myocardial infarction.
To overcome the limitations of Hu moments and improve the performance further, several works proposed combining heterogeneous types of moments, even with different categories of descriptors. For example, in [31], Hu moments were combined with VGG16, an off-the-shelf CNN, to automatically screen COVID-19 based on CT scans. Concerning other combinations, Wu et al. [23] combined Legendre, Zernike, Krawtchouk, and Chebyshev moments to classify CT liver images. In contrast, Lao et al. [32] proposed a framework for Alzheimer’s diagnosis based on combining the 3D discrete wavelet transform and 3D moment invariants features from multimodality images. These features were first provided to a deep neural network with stacked autoencoders for dimensionality reduction, and finally, a softmax regression made the final classification.
Furthermore, the combination of moments appeared beneficial also to improve and speed up the medical image reconstruction of MR, CT, and mammography, as shown in [33]. In that work, an effective method for analysing different types of 2D and 3D medical images based on the discrete orthogonal moments of Chebyshev, Krawtchouk, Hahn, Charlier, and Meixner was presented.
Image moments have generally demonstrated versatility and usefulness in various domains and applications. This makes them a valuable tool in the field of computer vision and image processing, particularly with regard to medical tasks.

3. Overview of Moments

Moments can be defined as the projections of an image function onto particular kernel functions and, mathematically, can be defined as follows.
A general moment M p q f of an image f ( x , y ) , where p , q are non-negative integers and ( p + q ) is called the order of the moment, is defined as in Equation (1) [53]:
M p q f = D h p q ( x , y ) f ( x , y ) d x d y
where h p q ( x , y ) is the kernel function (also known as the polynomial basis function) defined on D.
According to the mathematical properties of the polynomial basis used, we can recognise various categories of moments, as shown in Figure 1.
Depending on whether the basis functions satisfy orthogonality, the image moments can be classified into orthogonal moments or non-orthogonal moments. Orthogonality means that two different basis functions of the set of basis functions are unrelated or are “perpendicular” in geometric terms, which leads to no redundancy in the set of moments. Some of the most popular non-orthogonal moments are geometric moments [54], rotational moments [55], and complex moments [56].
Due to the nonorthogonality of the basis functions, a high information redundancy and high noise sensitivity exist in such moments. As a result, image reconstruction from these moments is quite difficult, and the image representation power is poor. In contrast, orthogonal moments, built on a set of orthogonal polynomials, have received much attention in recent years due to their ability to represent images with minimal information redundancy and a high noise robustness [25]. Orthogonal moments can be divided into those defined in either Cartesian coordinates, ( x , y ) , or polar coordinates, ( r , θ ) .
Orthogonal moments defined in polar coordinates are also called radial or circular orthogonal moments. The kernel function of radial orthogonal moments is built on a particular type of radial orthogonal polynomial and a complex exponential angular Fourier factor. In addition to their image reconstruction property, the significant advantage of radial orthogonal moments lies in their ability to achieve rotation invariance. In this regard, there are mainly three types of orthogonal functions used as definitions, including Jacobi polynomials, harmonic functions, and eigenfunctions.
Circular orthogonal moments based on Jacobi polynomials mainly include Zernike moments [37], pseudo-Zernike moments [55], orthogonal Fourier–Mellin moments [57], Chebyshev–Fourier moments [58], Jacobi–Fourier moments [59], and pseudo Jacobi–Fourier moments [60].
Among the circular orthogonal moments derived from harmonic functions, there are the radial harmonic Fourier moments [61], exponent Fourier moments [62], polar harmonic-transform-based moments [63], i.e., polar complex exponential transform, polar cosine transform, and polar sine transform. The most representative moments based on eigenfunctions are the Bessel–Fourier moments [64].
The kernel functions of orthogonal moments defined in Cartesian coordinates are constructed on normal orthogonal polynomials. The orthogonal moments defined in Cartesian coordinates can be further divided into two types, continuous and discrete orthogonal moments, depending on whether the kernel function h p q ( x , y ) is orthogonal on the continuous or discrete domain. The continuous orthogonal moments mainly include Legendre moments [37] and Gaussian–Hermite moments [65], while the discrete orthogonal moments are mainly composed of Chebyshev moments [34], Krawtchouk moments [35], Hahn moments [66], and Racah moments [67].
All of the orthogonal moments mentioned above are restricted to being of integer order since orthogonal polynomials should be of integer order ( p , q ) Z 2 . In recent years, there has been a focus on investigating fractional-order domain [68,69,70]. In such domain, a fractional-order parameter α R is introduced through certain suitable variable substitution, e.g.,  x x α , y y α [ 0 , 1 ] in Cartesian moments, and  r r α [ 0 , 1 ] in circular moments. Thus, the order of the newly defined moments can be extended to the real domain, e.g.,  ( α p , α q , ) R 2 for Cartesian moments and α r R for circular moments, although so far, mostly the circular moments have been studied and extended to the fractional domain [70,71], even in combination with colour information [72]. The fractional-order category includes radial harmonic Fourier moments, the polar complex exponential transform, the polar cosine transform, the polar sine transform and Jacobi–Fourier moments. They can be described by a recurrence relation which guarantees a high noise robustness in invariant image recognition [70,71].
In Table 1, we summarised the orthogonal moments analysed in this work, with the appropriate reference and the year in which they were first proposed.

4. Medical Image Classification

This section describes the materials and methods used to compare the above-mentioned moment features for CT and MR images classification. In the following, we first describe the data sets used in Section 4.1. Then, in Section 4.2, we list the moments and other features used for comparison purposes and finally, the classification models in Section 4.2.

4.1. Data Sets

We now describe the four different reference data sets used in this work. Each one is publicly available. One sample per class is illustrated in Figure 2. Emphysema-CT [74] is a database of CT images created to develop texture-based CT biomarkers for detecting and diagnosing chronic obstructive pulmonary diseases related to emphysema. It includes 115 high-resolution CT sections from which 168 square patches (of size 61 × 61 ) were manually cropped and annotated by experienced clinicians into three main categories: normal tissue, centrilobular emphysema, and paraseptal emphysema, consisting of 59, 50, and 59 samples, respectively.
Brain-Tumor-MRI [75] is a data set of MR images containing different brain sections from 233 patients, created to develop CAD systems to diagnose the presence of brain tumours. Here, we used the axial sections subset containing 870 images of size 512 × 512 , featuring three types of brain tumours: meningioma, glioma, and pituitary tumour. They are present in 143, 436 and 291 samples, respectively.
OASIS-MRI [76] is also an MR image data set containing different brain sections but was collected to study the clinical dementia rating. Again, we focused only on the axial section images, for a total of 436 images of size 176 × 208 , divided into 336 representing cognitively healthy individuals and 100 with subtypes of dementia (70 with very mild, 28 with mild, and 2 with moderate dementia).
COVIDx CT-2A [77] consists of 194,922 CT images of 3745 patients from 15 different countries, aged 0–93 years (median age 51). Specifically, the CT images are provided with size 512 × 512 pixels and are annotated into three different classes: novel coronavirus pneumonia due to SARS-CoV-2 viral infection, common pneumonia, and normal condition. The available CT images were 3253, 873, and 816, respectively, for a total of 4942 images.
The summarising details for each data set are reported in Table 2.

4.2. Methods

We evaluated different kinds of invariant moments in our experiments for a total of 20 descriptors that we grouped into four important classes, following what we reported in Section 3:
  • Cartesian moments: continuous Legendre moments (LM), discrete Chebyshev moments of the first order (CHM), discrete Chebyshev moments of the second order (CH2M).
  • Circular Jacobi polynomials based moments: Zernike moments (ZM), pseudo-Zernike moments (PZM), orthogonal Fourier–Mellin moments (OFMM), Chebyshev–Fourier moments (CHFM), pseudo-Jacobi–Fourier moments (PJFM), Jacobi–Fourier moments (JFM), fractional-order Jacobi–Fourier moments (FrJFM).
  • Circular eigenfunction-based moments: Bessel–Fourier moments (BFM).
  • Circular harmonic-function-based moments: radial harmonic Fourier moments (RHFM), exponent Fourier moments (EFM), polar complex exponential transform (PCET), polar cosine transform (PCT), polar sine transform (PST), fractional-order radial harmonic Fourier moments (FrRHFM), fractional-order polar complex exponential transform (FrPCET), fractional-order polar cosine transform (FrPCT), fractional-order polar sine transform (FrPST).
A subset of the orthogonal moments representative of all the classes presented in the taxonomy was chosen for this analysis. The order was set to 5 for all the moments. A higher-order would have added irrelevant features or noise [25]. In addition, as mentioned earlier, we also compared the performance of orthogonal moments with some texture and deep features for completeness.
The texture features computed were the rotation invariant grey-level co-occurrence matrix (GLCM) features as proposed in [78] and rotation-invariant LBP features [79]. GLCMs were computed with d = 1 and θ = [ 0 , 45 , 90 , 135 ] . We extracted thirteen rotation-invariant features H a r r i  [78] from them. The LBP map was extracted in the neighbourhood identified by r and n equal to 1 and 8, respectively, and then converted to a rotationally invariant feature vector L B P r i  [79].
The deep features were extracted from four off-the-shelf network architectures: AlexNet [7], VGG19 [8], ResNet50 [80], and GoogLeNet [81]. In both AlexNet and VGG19, we extracted features from the penultimate fully connected layer (fc7), which created a feature vector of size h = 4096 , while for ResNet50 and GoogLeNet, we extracted features from the only fully connected layer present, which created a feature vector of size h = 1000 . Since all referenced CNN architectures were pretrained on the vast image data set ImageNet [82], they had already learned sufficient representation and generalisation capability for various visual recognition tasks [83,84]. Therefore, we directly extracted the features without a fine-tuning process on the target data sets [83].
In order to perform a more detailed evaluation, the results of which are not related to the capabilities of the individual classifier, we chose a single classification technique from each of the following categories: numerical function models, instance-based models, symbolic or logical models, probabilistic-based models and ensemble models. Therefore, we selected five different classification models: support vector machine (SVM) [85], k-NN [86], decision trees (DT) [87], naive Bayes (NB) [88] and bagged trees (BT) [88].

5. Experimental Results

This section describes the experimental evaluation that was conducted and the results that were obtained. The purpose of this study was to analyse the performance of moment-based features for the classification of medical images, as previously mentioned. For this purpose, we conducted various experiments that involved the mentioned data sets, from which the features were extracted and then used to build and test different classification models.

5.1. Experimental Setup

To ensure a fair comparison, the experimental evaluation was performed using a model validation procedure consisting of a k-fold cross-validation applied to the entire data set. The training fold was used for data normalisation and hyperparameter optimisation for each cross-validation. Therefore, the best hyperparameters were optimised for all classifiers using a grid search.
Although we were not interested in the absolute best performance in terms of classification, the optimisation process was repeated for each individual descriptor to determine the best setting for each descriptor and its upper bound in terms of performance.
The classification performance was evaluated using only test fold data; the k-fold cross-validation was repeated with k equal to 10. The performance of the learning models was measured in terms of accuracy. The data sets were multiclass, so the accuracy was calculated as a weighted average of the values computed on the individual classes. The accuracy per class was computed according to the following Equation (2):
A c c u r a c y = T P + T N T P + F P + T N + F N
where T P (true positive) and F P (false positive) provide the number of images correctly and incorrectly classified as positive, respectively. On the other hand, T N (true negative) and F N (false negative) indicate the number of images classified as negative, correctly and incorrectly, respectively.
The experiments were performed on a workstation with the following hardware specifications: an Intel(R) Core(TM) i9-8950HK @ 2.90 GHz CPU, 32 GB RAM, and an NVIDIA GTX1050 Ti GPU with 4 GB of memory. All implementations and experimental evaluations were also performed in MATLAB R2021b. The source code with the moment’s implementation is freely available and was partially written by the same authors and available at http://bugs.unica.it/cana/software (accessed on 10 March 2023, see the orthomoms package), and partly obtained from a previous research article [38] https://github.com/ShurenQi/MomentToolbox (accessed on 10 March 2023).

5.2. Results

Here, we present a detailed account of the various experiments’ results.
For a precise evaluation of the individual descriptors, we report numerical tables detailing the results of the single descriptor categories obtained with each classifier. The classifier’s performance is briefly discussed later, as it is outside the scope of this work. To better examine the achieved results, we also report the feature vector dimensions (size column). We do not report the execution time since only some of the source code for the moment computation is optimised, which would lead to timelines that could be misleading. A detailed analysis of the execution time for our moments’ implementations has been reported in our previous work [25].
Emphysema-CT. The Emphysema-CT data set appears to be the most complex, based on the results obtained and shown in Table 3. Although the k-NN classifier performed best with 11 out of 26 descriptors (all moments), the results were much more varied. None of these 11 features was the absolute best, as deep features were in the top four, with the AlexNet features extracted by the SVM classifier being the absolute best (78.9%).
On average, deep features were by far the best category with an accuracy of 72%, followed by moments (Cartesian with 68.4%, Jacobian with 68%, Bessel–Fourier with 66.9%, and harmonic with 67.7%), and the texture features were last with an average of 63.3%.
Despite the excellent average performance achieved by the deep features, the moments were found to perform particularly well. The accuracy of 75% obtained by the BT classifier with the FrRHFM features was only 3.9% lower than the absolute best obtained by the features extracted from AlexNet (78.9%) and classified with the SVM. Specifically, FrRHFM was composed of a total of 66 features, 4,030 less than those of AlexNet. The k-NN classifier trained with FrPCET moments, consisting of 121 features, achieved the same accuracy of 75%.
Brain-Tumor-MRI. Contrary to the results on the previous data set, the absolute best results on Brain-Tumor-MRI were obtained by the k-NN classifier with 20 out of the 26 descriptors used (see Table 4). Apart from some exceptions caused by moments (JFM better by 1.5% with the SVM, HARri better by 1.7% with the BT), three out of four features extracted from the CNN (VGG19, ResNet50, GoogLeNet) gave better results with the SVM. However, the difference was not significant, except for GoogLeNet, which outperformed the k-NN classifier by 3.3% with the SVM.
On average, Cartesian moments were the best category with an accuracy of 93.7%, followed by the remaining moments: in order, Bessel–Fourier (92%), harmonic (91.8%), and Jacobian (91.7%). Thus, on average, all moments performed significantly better than the deep (90.8%) and texture (90.3%) descriptors.
This data set further illustrated how moments could produce particularly high results with few features. To give an example, the absolute best result was achieved by FrPCET and EFM coupled with a k-NN classifier (both with 98.2% accuracy) with 121 features, and the second best was achieved by CHM with a k-NN classifier (97.9% accuracy) with just 21 features.
OASIS-MRI. The first noteworthy observation from the OASIS MRI experiments is that all results were between 75 and 80%, as shown in Table 5, with a few exceptions, all provided by the NB classifier. In fact, it provided performance above 75% only with the FrPCET moments and AlexNet, VGG19, and ResNet50 features. On average, the best classifier was again the k-NN classifier, which performed best with 18 out of 26 descriptors, with both texture features and 15 out of 20 moments.
The best categories were the harmonic moments and the deep features with a 76.9% accuracy. However, unlike the previous data set, the categories were similar on average, as the Cartesian and Bessel–Fourier moments reached an accuracy of 76.5%, the Jacobian 76.3%, and the texture features 75.7%.
In general, the moments produced excellent results on this data set as well. Particularly noteworthy is the result obtained by the 66 features of CHFM with a k-NN classifier, which was the best overall with an accuracy of 80%.
COVIDx CT-2A. Regarding the COVIDx CT-2A data set, all the classifiers achieved an accuracy greater than 90% with any descriptor, except for the naive Bayes classifier, which achieved exceptionally high results (greater than 90%) only with features obtained from CNNs. As can be seen from Table 6, the best-performing classifier was the SVM, since it was able to obtain the best result with 23 of the 26 descriptors used, as well as being the best on average with an accuracy of 98.1%, 0.2% higher than the average accuracy obtained by the BT.
Moreover, among the identified descriptor categories, the one that performed best in this data set was the deep one, with an average accuracy of 96% among the features extracted from the four CNNs, followed by the texture features with 92.6%, and the Cartesian moments with 92%. In this context, it is crucial to note that the average of all the descriptors (except the deep ones) was strongly influenced by the results obtained by the naive Bayes classifier.
Therefore, we think it is particularly relevant to highlight, using the example of CH2M, how, with only 21 features, we were able to obtain an accuracy of 99%, which is very close to the best absolute performance of 99.4% obtained in this data set from the features extracted from VGG19 and classified by an SVM.

5.3. Discussion

To better understand and analyse the results regarding the performance of the identified moment categories, we report a whisker plot with confidence intervals for each data set in Figure 3, Figure 4, Figure 5 and Figure 6.
To produce such plots, we again grouped and reported on the x-axis the descriptors by categories: CM stands for Cartesian moments, JM for Jacobian moments, BM for Bessel–Fourier moments, HM for harmonic moments, DF for deep features, and TF for texture features. Each whisker along the y-axis represents the mean and standard deviation (minimum and maximum) accuracy value for a specific category of descriptors. The moments were generally very stable in each data set without considering the BM category, which consisted of only one descriptor. In fact, all categories of moments had a relatively low standard deviation, except for JM and HM in the Brain-Tumor-MRI (see Figure 4) and OASIS-MRI (see Figure 5) data sets.
If in the first case, the standard deviation was still lower than that produced by the deep features (1.20% and 1.75%, respectively, versus 2.1%), this trend was only partially confirmed in the second case. In this case, the deep features had a standard deviation of only 0.1% against 0.88% and 0.69%, respectively. In any case, these values were overall acceptable, considering the intrinsic complexity of the classes of the different data sets and, above all, the extremely small dimensions of their feature vectors compared to the deep ones. Moreover, with respect to the large COVIDx CT-2A data set (see Figure 6), where the classes were even closer, the deep features obtained unstable results (i.e., the standard deviation was 3%).
Last but not least, the Emphysema-CT data set was the one that showed a more pronounced difference between the deep features and the moments (see Figure 3). The moments performed better with the average best classifier (SVM) (FrPCET and FrRHFM obtained 75% as single descriptors). However, the absolute best result was obtained by the features extracted (with the k-NN classifier and BT, respectively) from VGG19 (which exceeded the absolute best obtained by the moments by 4.2%). In any case, such a difference could be accepted since the goal was to verify how much the moments were able to generalize their performance with respect to a specific task.
Considering the overall classification performance, recall that we optimized each individual classification algorithm since we wanted to study the absolute best performance of each descriptor. In this case, it can be observed that the naive Bayes classifier provided the worst results, being the classifier with the lowest average accuracy for all data sets. This trend was particularly emphasized for the COVIDx CT-2A data set, where the variance between the results obtained with moments was about 20% smaller than that obtained with deep features. However, even when observing the naive Bayes classifier in the other data sets, its performance when trained with moments tended to be between 6% and 9% lower than the best classifier. This behaviour is hypothetically due to the fact that naive Bayes assumes that the features considered are linearly independent. This aspect has been verified with deep features. In contrast, the features extracted with moments are highly correlated with each other, as can be seen in the work of Putzu et al. [89], where it is highlighted that feature selection applied to moments does not bring any benefit, unlike other categories of descriptors. This aspect provides additional robustness to their performance.
Finally, some useful guidelines emerged from the results of this work:
  • The moments offered comparable performance to deep features, despite being represented by order of magnitude fewer features than deep features.
  • The moments provided extremely robust results, with a reduced standard deviation per category in all the experiments performed.
  • Cartesian and harmonic moments were behaved the best, both on average and in terms of the number of absolute best results they offered in the different experiments. They were also the most stable, with low standard deviations.
  • There was no tendency for the moments to improve over the years; as mentioned in the previous point, the Cartesian moments performed best among the older ones. Moreover, the use of the fractional domain, one of the most recent innovations in the field, did not bring any significant advantage, as can be seen by comparing against the equivalent moments in the integer domain.
  • The naive Bayes classifier was not suitable for the task under consideration, especially when trained by any category of moments, since the assumption that the features were linearly independent was too strong for this type of descriptors.
  • Although moments allowed high accuracies for the SVM, k-NN and BT classifiers, the classifier that benefited the most from the moments was the k-NN classifier. It achieved significant absolute best results in all data sets except COVIDx CT-2A, where it had a marginal difference with the absolute top performer, the SVM.

6. Conclusions

This paper analysed the effectiveness of the best-known types of orthogonal moments in diagnostics when applied to diverse medical images. We first reviewed the state-of-the-art orthogonal moments, summarising their main categories and describing their main characteristics and differences. We then presented their main areas of application, with a particular emphasis on medical applications, which was the target of this work. We then evaluated orthogonal moments in different medical diagnostic tasks, highlighting the performance of orthogonal moments by comparing them with the best-known texture features and the latest CNNs features. This evaluation was carried out with five different classification techniques to guarantee that the results obtained were unrelated to the individual classifiers’ capabilities.
After analysing the experimental results presented in Section 5.3, we found that although CNNs showed impressive performance in all tasks, orthogonal moments proved to be a strong competitor, with comparable or even better results in some cases. It is worth noting that orthogonal moments require significantly fewer resources than CNNs, allowing them to be used in time-critical applications. In addition to being computationally expensive, deep learning approaches require large and diverse training sets and have limited robustness to geometric transformations, requiring data augmentation strategies to improve geometric invariance at the cost of time/space complexity. For these reasons, handcrafted features, particularly orthogonal moments, remain highly competitive in these domains. Orthogonal moments also outperform texture features, especially the well-known LBP descriptors.
Comparing the orthogonal moments among themselves, we observed that those defined in the Cartesian domain, although the least recent, performed the best. On the other hand, the more recent ones, of circular type and belonging to the fractional domain, did not seem to lead to further relevant benefits. It is not completely clear whether this trend was due to the different formulations or whether it was due to the order of moments used (see Section 4.2). From an empirical analysis, we observed that by increasing the order of the moments, the added features were irrelevant or with a low predictive power, but a thorough analysis of the moment order, even with fractional type orders, could be useful to confirm this trend.
Therefore, a challenging future investigation could be related to Cartesian moments on the fractional domain to see if they can bring more benefits to this task. Moreover, given their high performance, a further analysis could be relevant to verify the benefits of a combination of Cartesian moments. Another exciting direction of research could be related to the use of colour information, which is crucial for medical image analysis applications, but at the same time has been overlooked in the calculation of orthogonal moments.

Author Contributions

Conceptualization, C.D.R.; methodology, C.D.R., A.L. and L.P.; software, C.D.R., A.L. and L.P.; validation, C.D.R., A.L. and L.P.; formal analysis, C.D.R., A.L. and L.P.; investigation, C.D.R., A.L. and L.P; resources, C.D.R., A.L. and L.P.; data curation, C.D.R., A.L. and L.P.; writing—original draft preparation, C.D.R., A.L. and L.P.; writing—review and editing, C.D.R., A.L. and L.P.; visualization, C.D.R., A.L. and L.P.; supervision, C.D.R., A.L. and L.P.; project administration, C.D.R., A.L. and L.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

All the features extracted and the data produced in this study are available at the following url: https://github.com/andrealoddo/moments-for-medical-diagnosis GitHub repository (accessed on 10 March 2023).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CADComputer-aided diagnostic
CNNConvolutional neural network
VGGNetVery deep convolutional networks
LBPLocal binary pattern
CTComputed tomography
MRMagnetic resonance
LMContinuous Legendre moments
CHMDiscrete Chebyshev moments
CH2MDiscrete Chebyshev moments of the second order
ZMZernike moments
PZMPseudo-Zernike moments
OFMMOrthogonal Fourier–Mellin moments
CHFMChebyshev–Fourier moments
PJFMPseudo-Jacobi–Fourier moments
JFMJacobi–Fourier moments
FrJFMFractional-order Jacobi–Fourier moments
BFMBessel–Fourier moments
RHFMRadial harmonic Fourier moments
EFMExponent Fourier moments
PCETPolar complex exponential transform
PCTPolar cosine transform
PSTPolar sine transform
FrRHFMFractional-order radial harmonic Fourier moments
FrPCETFractional-order polar complex exponential transform
FrPCTFractional-order polar cosine transform
FrPSTFractional-order polar sine transform
GLCMGrey-level co-occurrence matrix
SVMSupport vector machine
k-NNk-nearest neighbour
DTDecision trees
NBNaive Bayes
BTBagged trees
TPTrue positive
FPFalse positive
TNTrue negative
FNFalse negative

References

  1. Bhattacharjee, S.; Mukherjee, J.; Nag, S.; Maitra, I.K.; Bandyopadhyay, S.K. Review on Histopathological Slide Analysis using Digital Microscopy. Int. J. Adv. Sci. Technol. 2014, 62, 65–96. [Google Scholar] [CrossRef]
  2. Kothari, S.; Phan, J.H.; Young, A.N.; Wang, M.D. Histological image classification using biologically interpretable shape-based features. BMC Med. Imaging 2013, 13, 1–17. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Kowal, M.; Filipczuk, P.; Obuchowicz, A.; Korbicz, J.; Monczak, R. Computer-aided diagnosis of breast cancer based on fine needle biopsy microscopic images. Comput. Biol. Med. 2013, 43, 1563–1572. [Google Scholar] [CrossRef]
  4. Rodríguez, J.H.; Javier, F.; Fraile, C.; Conde, M.J.R.; Llorente, P.L.G. Computer aided detection and diagnosis in medical imaging: A review of clinical and educational applications. In Proceedings of the TEEM ’16: Fourth International Conference on Technological Ecosystems for Enhancing Multiculturality, Salamanca, Spain, 2–4 November 2016; pp. 517–524. [Google Scholar]
  5. Yanase, J.; Triantaphyllou, E. A systematic survey of computer-aided diagnosis in medicine: Past and present development. Expert Syst. Appl. 2019, 138, 112821. [Google Scholar] [CrossRef]
  6. Di Ruberto, C.; Fodde, G.; Putzu, L. Comparison of Statistical Features for Medical Colour Image Classification. In International Conference ICVS on Computer Vision Systems; Lecture Notes in Computer Science; Springer International Publishing: Berlin/Heidelberg, Germany, 2015; Volume 9163, pp. 3–13. [Google Scholar]
  7. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems, Lake Tahoe, Nevada, 3–6 December 2012; Volume 1, pp. 1097–1105. [Google Scholar]
  8. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, Conference Track Proceedings, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  9. Yandex, A.B.; Lempitsky, V. Aggregating Local Deep Features for Image Retrieval. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Washington, DC, USA, 7–13 December 2015; pp. 1269–1277. [Google Scholar]
  10. Kim, W.; Goyal, B.; Chawla, K.; Lee, J.; Kwon, K. Attention-Based Ensemble for Deep Metric Learning. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; pp. 760–777. [Google Scholar]
  11. Loddo, A.; Putzu, L. On the Reliability of CNNs in Clinical Practice: A Computer-Aided Diagnosis System Case Study. Appl. Sci. 2022, 12, 3269. [Google Scholar] [CrossRef]
  12. Gelzinis, A.; Verikas, A.; Bacauskiene, M. Increasing the discrimination power of the co-occurrence matrix-based features. Pattern Recognit. 2007, 40, 2367–2372. [Google Scholar] [CrossRef]
  13. Gong, D.; Zhang, Y.; Jiang, J.; Liu, J. Steganalysis for GIF images based on colors-gradient co-occurrence matrix. Opt. Commun. 2012, 285, 4961–4965. [Google Scholar] [CrossRef]
  14. Peckinpaugh, S.H. An improved method for computing gray-level co-occurrence matrix based texture measures. CVGIP Graph. Model. Image Process. 1991, 53, 574–580. [Google Scholar] [CrossRef]
  15. Ojala, T.; Pietikäinen, M.; Harwood, D. A comparative study of texture measures with classification based on featured distributions. Pattern Recognit. 1996, 29, 51–59. [Google Scholar] [CrossRef]
  16. Singh, C.; Singh, J. A survey on rotation invariance of orthogonal moments and transforms. Signal Process. 2021, 185, 108086. [Google Scholar] [CrossRef]
  17. Yang, B.; Kostková, J.; Flusser, J.; Suk, T.; Bujack, R. Rotation invariants of vector fields from orthogonal moments. Pattern Recognit. 2018, 74, 110–121. [Google Scholar] [CrossRef]
  18. Wang, X.; Yang, T.; Guo, F. Image analysis by circularly semi-orthogonal moments. Pattern Recognit. 2016, 49, 226–236. [Google Scholar] [CrossRef]
  19. Li, X.; Song, A. A new edge detection method using Gaussian-Zernike moment operator. In Proceedings of the IEEE 2nd International Asia Conference on Informatics in Control, Automation and Robotics (CAR), Wuhan, China, 6–7 March 2010; Volume 1, pp. 276–279. [Google Scholar]
  20. Obulakonda, R.R.; Eswara, R.B.; Keshava, R.E. Ternary patterns and moment invariants for texture classification. ICTACT J. Image Video Process. 2016, 7, 1295–1298. [Google Scholar]
  21. Tahmasbi, A.; Saki, F.; Shokouhi, S.B. Classification of benign and malignant masses based on Zernike moments. Comput Biol Med. 2011, 41, 726–735. [Google Scholar] [CrossRef] [PubMed]
  22. Vijayalakshmi, B.; Bharathi, V.S. Classification of CT liver images using local binary pattern with Legendre moments. Curr. Sci. 2016, 110, 687–691. [Google Scholar] [CrossRef]
  23. Wu, K.; Garnier, C.; Coatrieux, J.L.; Shu, H. A preliminary study of moment-based texture analysis for medical images. In Proceedings of the 2010 Annual International Conference of the IEEE Engineering in Medicine and Biology, San Francisco, CA, USA, 31 August–4 September 2010; pp. 5581–5584. [Google Scholar]
  24. Hu, M.K. Visual pattern recognition by moment invariants. IRE Trans. Inf. Theory 1962, 8, 179–187. [Google Scholar]
  25. Di Ruberto, C.; Putzu, L.; Rodriguez, G. Fast and accurate computation of orthogonal moments for texture analysis. Pattern Recognit. 2018, 83, 498–510. [Google Scholar] [CrossRef] [Green Version]
  26. Ryabchykov, O.; Ramoji, A.; Bocklitz, T.; Foerster, M.; Hagel, S.; Kroegel, C.; Bauer, M.; Neugebauer, U.; Popp, J. Leukocyte subtypes classification by means of image processing. In Proceedings of the 2016 Federated Conference on Computer Science and Information Systems (FedCSIS), Gdańsk, Poland, 11–14 September 2016; Volume 8, pp. 309–316. [Google Scholar]
  27. Elaziz, M.; Hosny, K.; Hemedan, A.; Darwish, M. Improved recognition of bacterial species using novel fractional-order orthogonal descriptors. Appl. Soft Comput. 2020, 95, 106504. [Google Scholar] [CrossRef]
  28. Mukundan, R. Local Tchebichef Moments for Texture Analysis. In Moments and Moment Invariants—Theory and Applications; Science Gate Publishing: Thrace, Greece, 2014; pp. 127–142. [Google Scholar]
  29. Xu, Z.; Xilin, L.; Yang, C.; Huazhong, S. Medical Image Blind Integrity Verification with Krawtchouk Moments. Int. J. Biomed. Imaging 2018, 2018, 2572431. [Google Scholar]
  30. Batioua, I.; Benouini, R.; Zenkouar, K.; Zahi, A.; Hakim, E.F. 3D image analysis by separable discrete orthogonal moments based on Krawtchouk and Tchebichef polynomials. Pattern Recognit. 2017, 71, 264–277. [Google Scholar] [CrossRef]
  31. Moung, E.; Hou, C.; Sufian, M.; Hijazi, M.; Dargham, J.; Omatu, S. Fusion of Moment Invariant Method and Deep Learning Algorithm for COVID-19 Classification. Big Data Cogn. Comput. 2021, 5, 74. [Google Scholar] [CrossRef]
  32. Lao, H.; Zhang, X. Diagnose Alzheimer’s disease by combining 3D discrete wavelet transform and 3D moment invariants. IET Image Process. 2022, 16, 3948–3964. [Google Scholar] [CrossRef]
  33. El ogri, O.; Daoui, A.; Yamni, M.; Karmouni, H.; Sayyouri, M.; Qjidaa, H. 2D and 3D Medical Image Analysis by Discrete Orthogonal Moments. Procedia Comput. Sci. 2019, 148, 428–437. [Google Scholar] [CrossRef]
  34. Mukundan, R.; Ong, S.H.; Lee, P.A. Image analysis by Tchebichef moments. IEEE Trans. Image Process. 2001, 10, 1357–1364. [Google Scholar] [CrossRef] [PubMed]
  35. Yap, P.T.; Paramesran, R.; Ong, S.H. Image analysis by Krawtchouk moments. IEEE Trans. Image Process. 2003, 12, 1367–1377. [Google Scholar] [PubMed] [Green Version]
  36. Yap, P.T.; Paramesran, R. An Efficient Method for the Computation of Legendre Moments. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1996–2002. [Google Scholar]
  37. Teague, M.R. Image analysis via the general theory of moments. J. Opt. Soc. Am. 1980, 70, 920–930. [Google Scholar] [CrossRef]
  38. Qi, S.; Zhang, Y.; Wang, C.; Zhou, J.; Cao, X. A Survey of Orthogonal Moments for Image Representation: Theory, Implementation, and Evaluation. ACM Comput. Surv. 2021, 55, 1–35. [Google Scholar] [CrossRef]
  39. Varga, D. No-Reference Quality Assessment of Authentically Distorted Images Based on Local and Global Features. J. Imaging 2022, 8, 173. [Google Scholar] [CrossRef]
  40. Flusser, J.; Suk, T. Rotation Moment Invariants for Recognition of Symmetric Objects. IEEE Trans. Image Process. 2006, 15, 3784–3790. [Google Scholar] [CrossRef] [PubMed]
  41. Rahmalan, H.; Abu, N.A.; Wong, S.L. Using tchebichef moment for fast and efficient image compression. Pattern Recognit. Image Anal. 2010, 20, 505–512. [Google Scholar] [CrossRef]
  42. Tsougenis, E.D.; Papakostas, G.A.; Koulouriotis, D.E. Image watermarking via separable moments. Multim. Tools Appl. 2015, 74, 3985–4012. [Google Scholar] [CrossRef]
  43. Hosny, K.; Papakostas, G.; Koulouriotis, D. Accurate reconstruction of noisy medical images using orthogonal moments. In Proceedings of the 2013 18th International Conference on Digital Signal Processing (DSP), Santorini, Greece, 1–3 July 2013; pp. 1–6. [Google Scholar]
  44. Zhang, Y.; Wang, S.; Sun, P.; Phillips, P. Pathological brain detection based on wavelet entropy and Hu moment invariants. Bio-Med. Mater. Eng. 2015, 26 (Suppl. 1), S1283–S1290. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Iscan, Z.; Dokur, Z.; Ölmez, T. Tumor detection by using Zernike moments on segmented magnetic resonance brain images. Expert Syst. Appl. 2010, 37, 2540–2549. [Google Scholar] [CrossRef]
  46. Dogantekin, E.; Yilmaz, M.; Dogantekin, A.; Avci, E.; Sengur, A. A robust technique based on invariant moments—ANFIS for recognition of human parasite eggs in microscopic images. Expert Syst. Appl. 2008, 35, 728–738. [Google Scholar] [CrossRef]
  47. Liyun, W.; Hefei, L.; Fuhao, Z.; Zhengding, L.; Zhendi, W. Spermatogonium image recognition using Zernike moments. Comput. Methods Programs Biomed. 2009, 95, 10–22. [Google Scholar] [CrossRef]
  48. Alegre, E.; González-Castro, V.; Alaiz-Rodríguez, R.; García-Ordás, M. Texture and moments-based classification of the acrosome integrity of boar spermatozoa images. Comput. Methods Programs Biomed. 2012, 108, 873–881. [Google Scholar] [CrossRef] [PubMed]
  49. Di Ruberto, C.; Loddo, A.; Putzu, L. Histological image analysis by invariant descriptors. In Proceedings of the International Conference ICIAP on Image Analysis and Processing, Catania, Italy, 11–15 September 2017; Lecture Notes in Computer Science. Springer International Publishing: Berlin/Heidelberg, Germany, 2017; Volume 10484, pp. 345–356. [Google Scholar]
  50. Siti, S.; Shaharuddin, S.; Rozi, M. Haralick texture and invariant moments features for breast cancer classification. In Proceedings of the 2016 AIP Conference Proceedings, Bandung, Indonesia, 24–26 October 2016; Volume 1750, pp. 020022-1–020022-6. [Google Scholar]
  51. Laine, R.; Goodfellow, G.; Young, L.; Travers, J.; Carroll, D.; Dibben, O.; Bright, H.; Kaminski, C. Structured illumination microscopy combined with machine learning enables the high throughput analysis and classification of virus structure. eLife 2018, 7, e40183. [Google Scholar] [CrossRef] [PubMed]
  52. Balakrishnan, S.; Joseph, P. Stratification of risk of atherosclerotic plaque using Hu’s moment invariants of segmented ultrasonic images. Biomed. Tech. 2022, 67, 391–402. [Google Scholar] [CrossRef]
  53. Mukundan, R.; Ramakrishnan, K.R. Moment Functions in Image Analysis–Theory and Applications; World Scientific: Singapore, 1998. [Google Scholar] [CrossRef]
  54. Flusser, J.; Zitova, B.; Suk, T. Moments and Moment Invariants in Pattern Recognition; John Wiley & Sons: Hoboken, NJ, USA, 2009. [Google Scholar]
  55. Teh, C.H.; Chin, R.T. On image analysis by the methods of moments. IEEE Trans. Pattern Anal. Mach. Intell. 1988, 10, 496–513. [Google Scholar] [CrossRef]
  56. Abu-Mostafa, Y.S.; Psaltis, D. Recognitive aspects of moment invariants. IEEE Trans. Pattern Anal. Mach. Intell. 1984, 6, 698–706. [Google Scholar] [CrossRef] [Green Version]
  57. Sheng, Y.; Shen, L. Orthogonal Fourier-Mellin moments for invariant pattern recognition. J. Opt. Soc. Am. 1994, 11, 1748–1757. [Google Scholar] [CrossRef] [Green Version]
  58. Ping, Z.; Wu, R.; Sheng, Y. Image description with Chebyshev- Fourier moments. J. Opt. Soc. Am. 2002, 19, 1748–1754. [Google Scholar] [CrossRef] [Green Version]
  59. Ping, Z.; Ren, H.; Zou, J.; Sheng, Y.; Bo, W. Generic orthogonal moments: Jacobi-Fourier moments for invariant image description. Pattern Recognit. 2007, 40, 1245–1254. [Google Scholar] [CrossRef]
  60. Amu, G.; Hasi, S.; Yang, X.; Ping, Z. Image analysis by pseudo-Jacobi (p = 4, q = 3)-Fourier moments. Appl. Opt. 2004, 43, 2093–2101. [Google Scholar] [CrossRef] [PubMed]
  61. Ren, H.; Ping, Z.; Bo, W.; Wu, W.; Sheng, Y. Multidistortion-invariant image recognition with radial harmonic Fourier moments. J. Opt. Soc. Am. 2003, 20, 631–637. [Google Scholar] [CrossRef]
  62. Hu, H.; Zhang, Y.; Shao, C.; Ju, Q. Orthogonal moments based on exponent functions: Exponent-Fourier moments. Pattern Recognit. 2014, 47, 2596–2606. [Google Scholar] [CrossRef]
  63. Yap, P.T.; Jiang, X.; Kot, A.C. Two-dimensional polar harmonic transforms for invariant image representation. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1259–12707. [Google Scholar]
  64. Xiao, B.; Ma, J.; Wang, X. Image analysis by Bessel-Fourier moments. Pattern Recognit. 2010, 43, 2620–2629. [Google Scholar] [CrossRef]
  65. Yang, B.; Flusser, J.; Suk, T. 3D rotation invariants of Gaussian–Hermite moments. Pattern Recognit. Lett. 2015, 54, 18–26. [Google Scholar] [CrossRef]
  66. Yap, P.T.; Paramesran, R.; Ong, S.H. Image Analysis Using Hahn Moments. IEEE Trans. Pattern Anal. Mach. Intellig. 2007, 29, 2057–2062. [Google Scholar] [CrossRef] [PubMed]
  67. Zhu, H.; Shu, H.; Liang, J.; Luo, L.; Coatrieux, J.L. Image analysis by discrete orthogonal Racah moments. Signal Process. 2007, 87, 687–708. [Google Scholar] [CrossRef]
  68. Hoang, T.V.; Tabbone, S. Generic polar harmonic transforms for invariant image representation. Image Vis. Comput. 2014, 32, 497–509. [Google Scholar] [CrossRef] [Green Version]
  69. Hoang, T.V.; Tabbone, S. Generic polar harmonic transforms for invariant image description. In Proceedings of the 2011 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011; pp. 829–832. [Google Scholar]
  70. Xiao, B.; Li, L.; Li, Y.; Li, W.; Wang, G. Image analysis by fractional-order orthogonal moments. Inf. Sci. 2017, 382, 135–149. [Google Scholar] [CrossRef]
  71. Hosny, K.M.; Darwish, M.M.; Eltoukhy, M.M. Novel Multi-Channel Fractional-Order Radial Harmonic Fourier Moments for Color Image Analysis. IEEE Access 2020, 8, 40732–40743. [Google Scholar] [CrossRef]
  72. He, B.; Liu, J.; Yang, T.; Xiao, B.; Peng, Y. Quaternion fractional-order color orthogonal moment-based image representation and recognition. EURASIP J. Image Video Process. 2021, 2021, 1–35. [Google Scholar] [CrossRef]
  73. Yang, H.; Qi, S.; Tian, J.; Niu, P.; Wang, X. Robust and discriminative image representation: Fractional-order Jacobi-Fourier moments. Pattern Recognit. 2021, 115, 107898. [Google Scholar] [CrossRef]
  74. Sørensen, L.; Shaker, S.B.; de Bruijne, M. Quantitative Analysis of Pulmonary Emphysema Using Local Binary Patterns. IEEE Trans. Med Imaging 2010, 29, 559–569. [Google Scholar] [CrossRef] [Green Version]
  75. Cheng, J.; Huang, W.; Cao, S.; Yang, R.; Yang, W.; Yun, Z.; Wang, Z.; Feng, Q. Enhanced Performance of Brain Tumor Classification via Tumor Region Augmentation and Partition. PLoS ONE 2015, 10, e0140381. [Google Scholar] [CrossRef]
  76. Marcus, D.S.; Wang, T.H.; Parker, J.; Csernansky, J.G.; Morris, J.C.; Buckner, R.L. Open Access Series of Imaging Studies (OASIS): Cross-sectional MRI Data in Young, Middle Aged, Nondemented, and Demented Older Adults. J. Cogn. Neurosci. 2007, 19, 1498–1507. [Google Scholar] [CrossRef] [Green Version]
  77. Gunraj, H.; Wang, L.; Wong, A. COVIDNet-CT: A Tailored Deep Convolutional Neural Network Design for Detection of COVID-19 Cases From Chest CT Images. Front. Med. 2020, 7, 1025. [Google Scholar] [CrossRef] [PubMed]
  78. Putzu, L.; Di Ruberto, C. Rotation Invariant Co-occurrence Matrix Features. In Proceedings of the 19th International Conference ICIAP on Image Analysis and Processing, Catania, Italy, 11–15 September 2017; Lecture Notes in Computer Science. Springer International Publishing: Berlin/Heidelberg, Germany, 2017; Volume 10484, pp. 391–401. [Google Scholar]
  79. Ojala, T.; Pietikainen, M.; Maenpaa, T. Multiresolution gray-scale and rotation invariant texture classification with local binary pattern. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  80. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  81. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  82. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE conference on computer vision and pattern recognition, New Orleans, LA, USA, 18–24 June 2009; pp. 248–255. [Google Scholar]
  83. Putzu, L.; Piras, L.; Giacinto, G. Convolutional neural networks for relevance feedback in content based image retrieval. Multimed Tools Appl. 2020, 79, 26995–27021. [Google Scholar] [CrossRef]
  84. Donahue, J.; Jia, Y.; Vinyals, O.; Hoffman, J.; Zhang, N.; Tzeng, E.; Darrell, T. DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition. In Proceedings of the 31st International Conference on International Conference on Machine Learning, Beijing, China, 22–24 June 2014; Volume 32, pp. I-647–I-655. [Google Scholar]
  85. Lin, Y.; Lv, F.; Zhu, S.; Yang, M.; Cour, T.; Yu, K.; Cao, L.; Huang, T.S. Large-scale image classification: Fast feature extraction and SVM training. In Proceedings of the 24th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; IEEE Computer Society: Washington, DC, USA, 2011; pp. 1689–1696. [Google Scholar]
  86. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef] [Green Version]
  87. Coppersmith, D.; Hong, S.; Hosking, J. Partitioning nominal attributes in decision trees. Data Min. Knowl. Discov. 1999, 3, 197–217. [Google Scholar] [CrossRef]
  88. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction; Springer Series in Statistics; Springer: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
  89. Putzu, L.; Loddo, A.; Di Ruberto, C. Invariant Moments, Textural and Deep Features for Diagnostic MR and CT Image Retrieval. In Proceedings of the Computer Analysis of Images and Patterns; Springer International Publishing: Cham, Switzerland, 2021; pp. 287–297. [Google Scholar]
Figure 1. A schematic classification of image moments.
Figure 1. A schematic classification of image moments.
Jimaging 09 00070 g001
Figure 2. Illustrations of the data sets classes. From left to right, (a) shows normal tissue, centrilobular emphysema, and paraseptal emphysema; in (b): meningioma, glioma, and pituitary tumours. (c) represents four classes of dementia: absent, very mild, mild, and moderate. Finally, (d) includes COVID, normal, and pneumonia.
Figure 2. Illustrations of the data sets classes. From left to right, (a) shows normal tissue, centrilobular emphysema, and paraseptal emphysema; in (b): meningioma, glioma, and pituitary tumours. (c) represents four classes of dementia: absent, very mild, mild, and moderate. Finally, (d) includes COVID, normal, and pneumonia.
Jimaging 09 00070 g002
Figure 3. Average accuracy and standard deviation calculated on each descriptor category with the best classifier on average, for the Emphysema-CT data set.
Figure 3. Average accuracy and standard deviation calculated on each descriptor category with the best classifier on average, for the Emphysema-CT data set.
Jimaging 09 00070 g003
Figure 4. Average accuracy and standard deviation calculated on each descriptor category with the best classifier on average, for the Brain-Tumor-MRI data set.
Figure 4. Average accuracy and standard deviation calculated on each descriptor category with the best classifier on average, for the Brain-Tumor-MRI data set.
Jimaging 09 00070 g004
Figure 5. Average accuracy and standard deviation calculated on each descriptor category with the best classifier on average, for the Axial OASIS-MRI data set.
Figure 5. Average accuracy and standard deviation calculated on each descriptor category with the best classifier on average, for the Axial OASIS-MRI data set.
Jimaging 09 00070 g005
Figure 6. Average accuracy and standard deviation calculated on each descriptor category with the best classifier on average, for the COVIDx CT-2A data set.
Figure 6. Average accuracy and standard deviation calculated on each descriptor category with the best classifier on average, for the COVIDx CT-2A data set.
Jimaging 09 00070 g006
Table 1. The year when the moments were first proposed.
Table 1. The year when the moments were first proposed.
Ref.DescriptorsYear
[37]Legendre moments1980
[34]Chebyshev moments of the first order2001
[34]Chebyshev moments of the second order2001
[37]Zernike moments1980
[55]Pseudo-Zernike moments1988
[57]Orthogonal Fourier–Mellin moments1994
[58]Chebyshev–Fourier moments2002
[60]Pseudo-Jacobi–Fourier moments2004
[59]Jacobi–Fourier moments2007
[73]Fractional-order Jacobi–Fourier moments2021
[64]Bessel–Fourier moments2010
[61]Radial harmonic Fourier moments2003
[62]Exponent Fourier moments2014
[63]Polar complex exponential transform2010
[63]Polar cosine transform2010
[63]Polar sine transform2010
[71]Fractional-order radial harmonic Fourier moments2020
[68]Fractional-order polar complex exponential transform2014
[68]Fractional-order polar cosine transform2014
[68]Fractional-order polar sine transform2014
Table 2. The table provides information on the four different data sets used in this research. For each data set, the following details are given: number and size of the images, number of classes, and number of images per class.
Table 2. The table provides information on the four different data sets used in this research. For each data set, the following details are given: number and size of the images, number of classes, and number of images per class.
Data SetNo. of ImagesSizeClassesNo. of Images per Class
Emphysema-CT168 61 × 61 359–50–59
Brain-Tumor-MRI870 512 × 512 3143–436–291
OASIS-MRI436 176 × 208 4336–70–28–2
COVIDx CT-2A4942 512 × 512 33253–873–816
Table 3. Results obtained on the Emphysema-CT data set with all the analysed features, for which we report the categorisation (category column, CM—Cartesian moments, JM—Jacobian moments, BM—Bessel–Fourier moments, HM—harmonic moments, DF—deep features, and TF—texture features), the name (descriptor column), the feature size (size column), the accuracy obtained with all the tested classifiers (SVM, k-NN, DT, NB, BT columns), and the average accuracy value (AVG column). On the right (AVG cat.), we also report the average accuracy value of each category of features and on the bottom (AVG row), we report the average accuracy per classifier.
Table 3. Results obtained on the Emphysema-CT data set with all the analysed features, for which we report the categorisation (category column, CM—Cartesian moments, JM—Jacobian moments, BM—Bessel–Fourier moments, HM—harmonic moments, DF—deep features, and TF—texture features), the name (descriptor column), the feature size (size column), the accuracy obtained with all the tested classifiers (SVM, k-NN, DT, NB, BT columns), and the average accuracy value (AVG column). On the right (AVG cat.), we also report the average accuracy value of each category of features and on the bottom (AVG row), we report the average accuracy per classifier.
CategoryDescriptorSizeClassifiersAVGAVG Cat.
SVMk-NNDTNBBT
CMLM2871.468.564.963.764.366.668.4
CHM2169.669.670.263.769.068.4
CH2M2169.671.470.867.372.070.2
JMZM2170.273.868.570.272.070.968.0
PZM3671.470.866.764.967.368.2
OFMM6670.872.068.561.966.768.0
CHFM6664.372.072.061.363.766.7
PJFM6669.072.070.259.569.668.1
JFM6669.670.869.661.967.968.0
FrJFM6666.173.869.661.361.366.4
BMBFM6669.668.567.360.169.066.966.9
HMRHFM6668.572.069.664.373.269.567.7
EFM12167.370.267.363.767.967.3
PCET12165.567.970.858.969.066.4
PCT6665.672.069.060.766.766.8
PST5569.669.667.961.964.366.7
FrRHFM6670.270.873.266.775.071.2
FrPCET12166.775.070.858.969.068.1
FrPCT6664.970.267.961.963.165.6
FrPST5569.670.269.064.366.768.0
DFAlexNet409678.973.863.172.676.873.072.0
VGG19409676.870.863.170.279.272.0
ResNet50100078.670.864.375.077.473.2
GoogLeNet100076.866.761.972.071.469.8
TFLBP183653.054.256.554.256.554.963.3
HARri5276.870.264.369.677.471.7
- AVG-69.670.367.664.369.1-
Table 4. Results obtained on the Brain-Tumor-MRI data set with all the analysed features, for which we report the categorisation (category column, CM—Cartesian moments, JM—Jacobian moments, BM—Bessel–Fourier moments, HM—harmonic moments, DF—deep features, and TF—texture features), the name (descriptor column), the feature size (size column), the accuracy obtained with all the tested classifiers (SVM, k-NN, DT, NB, BT columns) and the average accuracy value (AVG column). On the right (AVG cat.), we also report the average accuracy value of each category of features and on the bottom (AVG row), we report the average accuracy per classifier.
Table 4. Results obtained on the Brain-Tumor-MRI data set with all the analysed features, for which we report the categorisation (category column, CM—Cartesian moments, JM—Jacobian moments, BM—Bessel–Fourier moments, HM—harmonic moments, DF—deep features, and TF—texture features), the name (descriptor column), the feature size (size column), the accuracy obtained with all the tested classifiers (SVM, k-NN, DT, NB, BT columns) and the average accuracy value (AVG column). On the right (AVG cat.), we also report the average accuracy value of each category of features and on the bottom (AVG row), we report the average accuracy per classifier.
CategoryDescriptorSizeClassifiersAVGAVG Cat.
SVMk-NNDTNBBT
CMLM2895.997.790.791.095.294.193.7
CHM2194.697.989.791.094.793.6
CH2M2194.197.489.990.594.993.4
JMZM2193.493.390.289.793.492.091.8
PZM3694.194.990.589.994.892.8
OFMM6693.895.588.688.592.991.9
CHFM6693.796.287.187.793.791.7
PJFM6692.695.390.588.091.691.6
JFM6694.593.089.188.092.291.4
FrJFM6690.895.589.987.892.691.3
BMBFM6694.395.989.186.994.092.092.0
HMRHFM6693.695.488.487.692.991.691.7
EFM12194.998.287.186.792.691.9
PCET12196.797.787.486.792.092.1
PCT6692.294.487.488.391.490.7
PST5593.495.389.787.093.491.8
FrRHFM6694.195.289.288.392.091.8
FrPCET12195.998.289.487.094.693.0
FrPCT6694.094.588.487.691.591.2
FrPST5591.793.689.887.592.491.0
DFAlexNet409694.694.788.683.993.191.090.8
VGG19409695.494.984.487.793.091.1
ResNet50100095.795.988.587.493.892.3
GoogLeNet100094.491.183.783.191.888.8
TFLBP183692.393.689.477.190.388.590.3
HARri5293.492.388.592.393.992.1
- AVG-94.095.388.787.693.0--
Table 5. Results obtained on the OASIS-MRI data set with all the analysed features, for which we report the categorisation (category column, CM—Cartesian moments, JM—Jacobian moments, BM—Bessel–Fourier moments, HM—harmonic moments, DF—deep features, and TF—texture features), the name (descriptor column), the feature size (size column), the accuracy obtained with all the tested classifiers (SVM, k-NN, DT, NB, BT columns) and the average accuracy value (AVG column). On the right (AVG cat.), we also report the average accuracy value of each category of features and on the bottom (AVG row), we report the average accuracy per classifier.
Table 5. Results obtained on the OASIS-MRI data set with all the analysed features, for which we report the categorisation (category column, CM—Cartesian moments, JM—Jacobian moments, BM—Bessel–Fourier moments, HM—harmonic moments, DF—deep features, and TF—texture features), the name (descriptor column), the feature size (size column), the accuracy obtained with all the tested classifiers (SVM, k-NN, DT, NB, BT columns) and the average accuracy value (AVG column). On the right (AVG cat.), we also report the average accuracy value of each category of features and on the bottom (AVG row), we report the average accuracy per classifier.
CategoryDescriptorSizeClassifiersAVGAVG Cat.
SVMk-NNDTNBBT
CMLM2877.878.277.171.877.576.576.5
CHM2178.077.677.172.977.176.5
CH2M2177.578.077.172.577.176.4
JMZM2177.378.277.170.677.376.176.3
PZM3677.178.477.172.078.276.6
OFMM6677.178.277.172.577.176.4
CHFM6677.180.077.171.377.176.5
PJFM6677.177.877.170.675.775.7
JFM6677.178.078.071.877.876.5
FrJFM6678.277.177.372.577.176.4
BMBFM6677.177.877.174.176.676.576.5
HMRHFM6677.178.478.472.076.676.576.9
EFM12177.178.077.173.976.676.5
PCET12177.378.477.174.878.077.1
PCT6677.178.977.172.777.176.6
PST5577.179.479.672.779.177.6
FrRHFM6677.178.977.172.078.476.7
FrPCET12177.577.177.875.577.177.0
FrPCT6677.179.177.172.577.176.9
FrPST5577.378.279.172.277.176.8
DFAlexNet409677.177.176.877.377.177.176.9
VGG19409678.777.175.777.578.977.6
ResNet50100077.177.375.775.577.176.5
GoogLeNet100078.477.177.172.976.476.4
TFLBP183677.177.377.172.075.575.875.7
HARri5277.177.377.169.777.175.7
- AVG-77.478.077.372.977.2--
Table 6. Results obtained on the COVIDx CT-2A data set with all the analysed features, for which we report the categorisation (category column, CM—Cartesian moments, JM—Jacobian moments, BM—Bessel–Fourier moments, HM—harmonic moments, DF—deep features, and TF—texture features), the name (descriptor column), the feature size (size column), the accuracy obtained with all the tested classifiers (SVM, k-NN, DT, NB, BT columns) and the average accuracy value (AVG column). On the right (AVG cat.), we also report the average accuracy value of each category of features and on the bottom (AVG row), we report the average accuracy per classifier.
Table 6. Results obtained on the COVIDx CT-2A data set with all the analysed features, for which we report the categorisation (category column, CM—Cartesian moments, JM—Jacobian moments, BM—Bessel–Fourier moments, HM—harmonic moments, DF—deep features, and TF—texture features), the name (descriptor column), the feature size (size column), the accuracy obtained with all the tested classifiers (SVM, k-NN, DT, NB, BT columns) and the average accuracy value (AVG column). On the right (AVG cat.), we also report the average accuracy value of each category of features and on the bottom (AVG row), we report the average accuracy per classifier.
CategoryDescriptorSizeClassifiersAVGAVG Cat.
SVMk-NNDTNBBT
CMLM2899.097.294.376.298.893.192.0
CHM2199.096.993.972.098.792.1
CH2M2199.195.391.371.597.590.9
JMZM2198.995.492.272.297.791.391.7
PZM3699.096.793.475.898.492.7
OFMM6698.497.093.073.598.292.0
CHFM6698.397.092.872.598.291.8
PJFM6698.397.192.271.898.391.5
JFM6698.396.992.471.898.191.5
FrJFM6698.297.192.271.898.191.5
BMBFM6698.397.092.373.298.191.891.8
HMRHFM6698.397.192.671.498.191.591.6
EFM12198.397.691.570.398.291.2
PCET12198.097.792.572.198.491.7
PCT6698.497.293.172.298.591.9
PST5598.597.092.271.998.491.6
FrRHFM6698.397.192.671.498.291.5
FrPCET12198.197.692.272.598.391.7
FrPCT6698.597.393.072.398.491.9
FrPST5598.597.092.271.898.391.6
DFAlexNet409697.597.597.497.497.597.596.0
VGG19409699.499.499.499.199.499.3
ResNet50100095.895.895.695.096.595.7
GoogLeNet100092.291.491.590.990.691.3
TFLBP183698.498.095.074.197.992.792.4
HARri5297.596.594.773.997.592.0
- AVG-98.196.893.376.197.9--
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Di Ruberto, C.; Loddo, A.; Putzu, L. On The Potential of Image Moments for Medical Diagnosis. J. Imaging 2023, 9, 70. https://doi.org/10.3390/jimaging9030070

AMA Style

Di Ruberto C, Loddo A, Putzu L. On The Potential of Image Moments for Medical Diagnosis. Journal of Imaging. 2023; 9(3):70. https://doi.org/10.3390/jimaging9030070

Chicago/Turabian Style

Di Ruberto, Cecilia, Andrea Loddo, and Lorenzo Putzu. 2023. "On The Potential of Image Moments for Medical Diagnosis" Journal of Imaging 9, no. 3: 70. https://doi.org/10.3390/jimaging9030070

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop