Next Article in Journal
Construction of Life-Cycle Simulation Framework of Chronic Diseases and Their Comorbidities Based on Population Cohort
Next Article in Special Issue
A Federated Generalized Linear Model for Privacy-Preserving Analysis
Previous Article in Journal
Smart(Sampling)Augment: Optimal and Efficient Data Augmentation for Semantic Segmentation
Previous Article in Special Issue
Using Explainable Machine Learning to Explore the Impact of Synoptic Reporting on Prostate Cancer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Machine Learning Models for Early Stage Detection of Autism Spectrum Disorder

by
Mousumi Bala
1,
Mohammad Hanif Ali
2,
Md. Shahriare Satu
3,
Khondokar Fida Hasan
4 and
Mohammad Ali Moni
5,*
1
Department of Computer Science and Engineering, Eastern University, Ashulia Model Town, Savar 1345, Bangladesh
2
Department of Computer Science and Engineering, Jahangirnagar University, Savar 1342, Bangladesh
3
Department of Management Information Systems, Noakhali Science and Technology University, Sonapur 3814, Bangladesh
4
Centre for Cyber Security Research & Innovation, RMIT University, 124 La Trobe Street, Melbourne, VIC 3000, Australia
5
Artificial Intelligence & Digital Health, School of Health and Rehabilitation Sciences, Faculty of Health and Behavioural Sciences, The University of Queensland, St. Lucia, QLD 4072, Australia
*
Author to whom correspondence should be addressed.
Algorithms 2022, 15(5), 166; https://doi.org/10.3390/a15050166
Submission received: 29 March 2022 / Revised: 3 May 2022 / Accepted: 10 May 2022 / Published: 16 May 2022
(This article belongs to the Special Issue Interpretability, Accountability and Robustness in Machine Learning)

Abstract

:
Autism spectrum disorder (ASD) is a neurodevelopmental disorder that severely impairs an individual’s cognitive, linguistic, object recognition, communication, and social abilities. This situation is not treatable, although early detection of ASD can assist to diagnose and take proper steps for mitigating its effect. Using various artificial intelligence (AI) techniques, ASD can be detected an at earlier stage than with traditional methods. The aim of this study was to propose a machine learning model that investigates ASD data of different age levels and to identify ASD more accurately. In this work, we gathered ASD datasets of toddlers, children, adolescents, and adults and used several feature selection techniques. Then, different classifiers were applied into these datasets, and we assessed their performance with evaluation metrics including predictive accuracy, kappa statistics, the f1-measure, and AUROC. In addition, we analyzed the performance of individual classifiers using a non-parametric statistical significant test. For the toddler, child, adolescent, and adult datasets, we found that Support Vector Machine (SVM) performed better than other classifiers where we gained 97.82% accuracy for the RIPPER-based toddler subset; 99.61% accuracy for the Correlation-based feature selection (CFS) and Boruta CFS intersect (BIC) method-based child subset; 95.87% accuracy for the Boruta-based adolescent subset; and 96.82% accuracy for the CFS-based adult subset. Then, we applied the Shapley Additive Explanations (SHAP) method into different feature subsets, which gained the highest accuracy and ranked their features based on the analysis.

1. Introduction

Autism spectrum disorder (ASD) is a neuro-developmental disorder where it appears in human beings during the first three years [1]. It is basically characterized by several symptoms such as impairments in social interaction, communication, restricted interests, and repetitive behavior [2]. Individuals with ASD face difficulty understanding other’s feelings and thinking. They experience many problems communicating with others. As reported by the World Health Organization (WHO), throughout the world, around 1 in 270 individuals has ASD [3]. Each individual with ASD has unique characteristics, and some have exceptional abilities in visual, academic, and music skills. In this case, the most important steps are required to detect ASD and to ensure proper treatment as early as possible. These steps are helpful to decrease the effects of this disorder and to improve their condition. The symptoms of ASD are identified by different types of observations. However, a significant amount of time and effort are needed where early detection is useful for providing better treatment for ASD patients. Recently, machine learning methods are widely used for analyzing the symptoms of various severe diseases like heart disease, diabetes, and cancer tissues, etc. Therefore, many researchers explored numerous methods [4,5] that are helpful to identify ASD patients and decrease the affects of ASD more precisely.
The objective of this work was to propose a machine learning model that explores ASD of toddlers, children, adolescents, and adults at an early stage as well as to investigate the individual characteristics of them more efficiently. In this model, we generated several feature subsets of toddlers, children, adolescents, and adults using various feature selection techniques. Different classification methods were applied into primary and its feature subsets. Then, their results were compared, and we determined the best classifier for each age group and the feature subset for which the highest results were obtained. Additionally, the performance of these classifiers were investigated using a non-parametric statistical significant test. Then, we interpreted the results of the best feature subsets and selected significant features of ASD using an explainable AI method. The following concise summary of the contribution is given as follows:
  • We proposed an efficient machine learning method that has the potential to identify ASD with high accuracy at an early stage.
  • We concentrated on identifying important feature subsets and explained different features to know how individual features are responsible to generate the best result to diagnose ASD or not.
  • To justify the performance of the classifier, we used a non-parametric statistical method and checked the classifier’s pairwise significance.
  • This method is helpful to identify ASD in a simple and flexible way.
This article is structured as follows: Section 2 contains a literature review on ASD screening approaches. Section 3 represents the working steps of detecting ASD and its characteristics. Section 4 describes the experimental results and an interpretation of these results. Section 5 describes the discussion and conclusion.

2. Literature Review

Many state-of-art works were happened to investigate, classify and explore significant factors of ASD. Thabtah et al. [6,7,8,9] developed a mobile application named ASDTests for data collection related to ASD for toddlers, children, adolescents, and adults. This app was built based on Q-CHAT and AQ-10 tools to predict ASD or not. They collected ASD data using this app and uploaded them into the University of California-Irvine (UCI) Machine Learning (ML) repository. Omar et al. [10] proposed an effective machine learning model where they analyzed AQ-10 and 250 real datasets with Random Forest (RF), Classification and Regression Trees (CART) and Random Forest-Iterative Dichotomiser 3 (ID3). Sharma et al. [11] investigated these datasets by employing CFS-greedy stepwise feature selector and further applied Naïve Bayes (NB), Stochastic Gradient Descent (SGD), K-Nearest Neighbours (KNN), Random Tree (RT), and K-Star (KS) into these datasets. Satu et al. [12] collected some samples of 16–30 years children where several tree based classifiers were used to investigate them and extracted several rules for normal and autism. Erkan et al. [13] analyzed similar datasets by implementing KNN, SVM, and RF where RF showed the best performance to identify ASD. Another study by Thabtah et al. [14] generated several feature subsets of adults and adolescents using Information Gain (IG) and Chi-Squared (CHI) where Logistic Regression (LR) was used to identify ASD from them. Akter et al. [15] gathered toddlers, children, adolescents, and adults datasets and generated some transformed sets. Then, different classifiers were used to analyze them where SVM showed the best performance for the toddler as well as Adaboost provided both children and adult. In addition, Glmboost showed the best outcomes the adolescent dataset. Hossain et al. [16] investigated similar types of datasets and generated subsets using CFS, CHI, IG, One-R and Relief-F methods. Further, they employed LR, Multilayer Perceptron (MLP), Sequential minimal optimization (SMO) into them. Then, SMO showed the best accuracy 91% for child, 99.9% for adolescent, 97.58% for adult datasets. Raj et al. [17] scrutinized these datasets (i.e., excluding toddler) using SVM, LR, NB, and Convolutional Neural Network (CNN) where CNN showed the highest accuracy 98.30% for child, 96.88% for adolescent, and 99.53%, for adult, respectively. Again, Thabtah et al. [18] developed a Rules-based Machine Learning (RML) method for extracting ASD traits where RML gave higher predictive accuracy from other machine learning approaches. Chowdhury et al. [19] provided an association classification technique with seven algorithms where this method showed 97% accuracy to detect ASD. Akter et al. [20] gathered ASD dataset of different age levels and generated some transformed subsets of it. They analyzed them with several classifiers where LR outperformed other classifiers and extracted significant traits. Akter et al. [21] also extracted several autism subtypes using k-means algorithms. Then, they identified different discriminatory factors among them.

3. Methodology

In this work, we used different feature selection methods and generated some feature subsets. Then, some classification algorithms were applied into primary toddler, child, adolescent, and adult datasets and their feature subsets. The performance of different classifiers were investigated to determine which features are more useful to detect ASD from controls. Figure 1 shows the proposed model of ASD detection at early stage.

3.1. Data Preprocessing

Dataset Description: Thabtah et al. [6] from the Nelson Marlborough Institute of Technology developed an autism screening app called ASDTests, which was used for data collection from the target audience (toddlers, children, youths, and adults). It used Q-CHAT-10 and AQ-10 questionnaires (AQ-10 Child, AQ-10 Adolescent, and AQ-10 Adult) to determine ASD risk factors. This app automatically computed its final-score from 0 to 10. It indicates a positive prediction of ASD if the final-score is greater than 6 out of 10. In this work, we used toddler, child, adolescent, and adult ASD datasets (version 2) [22,23,24,25], where the toddler dataset consisted of 18 features; however, the child, adolescent, and adult datasets contained 23 features. These datasets contained the records of 12–36 months (toddlers), 4–11 years (child), 12–16 years (adolescent), and 18 years or greater (adult) age groups. Table 1 shows the details of these datasets, and Table 2 provides the description of individual features of these datasets that were used for analysis.
Data Cleaning: We cleaned these datasets to simplify this model and increased classification accuracy by deleting instances with missing values. After that, we discarded some irrelevant features (i.e., those which are not related to ASD) such as the case, whether they the used app before, the user (who completed the screening), the language, why they had taken the screening, the age description, the screening type, and the score, respectively. The score 7 to 10 was used to classify ASD prediction for the child, adolescent, and adult datasets. A value of 4 or higher was classified as ASD for the toddler dataset. In this work, we selected 16 features for the child, adolescent, and adult datasets and 15 features for the toddler dataset, respectively.

3.2. Implementing Feature Selection Methods

Feature selection is required to identify significant attributes that improve the performance of machine learning models [26]. Many state-of-the-art works were found where various significant features were identified to detect autism more efficiently [27] On the toddler, child, adolescent, and adult datasets, various feature selection techniques such as the Boruta algorithm, Correlation-based Feature Selection with Harmony Search (CFS–Harmony Search), Repeated Incremental Pruning to Produce Error Reduction (RIPPER), and Recursive Feature Elimination (RFE) were used to explore various feature subsets. A brief description about them is given as follows:
  • Boruta algorithm is a wrapper algorithm based on the random forest [28] where it finds the importance of a feature by creating shadow features [29]. It is an extended system where each feature of the given data set is replicated. Then, the values of replicated variables are randomly combined, which are called shadow features. It performs its feature selection process using RF on the extended data set and evaluates the importance of each feature. Additionally, it computes the z-score of real and shadow features. It compares higher z-score values of real features than the maximum z-score value of its shadow features at every iteration. In this process, it constantly eliminates features that are deemed highly unimportant. Finally, the algorithm ends either when all features are approved or rejected or it obtains a particular limit of RF runs [30].
  • Correlation-based Feature Selection (CFS–Harmony Search) is a heuristic function that evaluates the ranks of features based on their correlation [31]. For harmony search, it illustrates the point of intersection with the following parameters: the number of harmonies in memory N, the number of indicator M, the number of possible values of indicator D, the number of optimal indicator i in the harmony memory = E i , and the rate of harmony memory E t . The probability is calculated by the following equation,
    P t ( E ) i = 1 M [ E t E i N + ( 1 E t ) 1 D ]
  • Repeated Incremental Pruning to Produce Error Reduction (RIPPER) is the rule induction algorithm that was introduced by W. Cohen in 1995 [32]. It is generated as a set of if-then-else rules that evolves several iterations of the rule learning algorithm. This model is maintained by three-steps such as grow, prune, and optimize [33] where the evaluation process is denoted as,
    G a i n ( R 0 , R 1 ) = c [ l o g ( T 1 T 1 + F 1 ) l o g ( T 0 T 0 + F 0 ) ]
    where R 0 = the initial rule, R 1 = the rule after adding conjunct, c = the number of true instances covered by R 0 and R 1 , T 0 = the number of true instances covered by R 0 , F 0 = the number of false instances covered by R 0 , T 1 = the number of true instances covered by R 1 , and F 1 = the number of false instances covered by R 1 [34].
  • Recursive Feature Elimination (RFE) is a feature selection technique that discards the least important features recursively. In this process, the initial features are trained where each important feature is acquired through any selected attributes [35]. Then, the least important features are eliminated from the initial feature set. This process is recursively repeated until the desired number of feature subsets are obtained [36]. The steps of RFE are given as follows [37]:
    • Train the classifier.
    • Calculate the score for all features with the ranking.
    • Eliminate the feature with the lowest score.

3.3. Apply Individual Classification Methods

After generating various feature subsets, 30 widely used classifiers were implemented into all of these datasets and their subsets. Some classifiers that produced less than 70% accuracy were eliminated. Thus, the performance of NB, KS, C4.5, CART, SVM, KNN, Bagging (BG), and Random Tree (RT) were considered. Then, we compared their performance and detected the best classifier. In addition, it detects feature subsets for which the classifier shows the highest outcomes. Therefore, a brief description of them are given as follows:
  • Naïve Bayes (NB) is a most constructive probabilistic classifier based on Bayes theorem [38]. It predicts target output more efficiently according to the foundation of the probability of an entity. The formula of Bayes theorem is given as,
    R ( a | b ) = R ( b | a ) R ( a ) R ( b )
    The derivation form for NB is:
    R ( a | b ) = R ( b 1 | a ) R ( b 2 | a ) R ( b n | a ) R ( a )
    Here, R ( a b ) is the subsequent probability of the target class; R ( a ) is the earlier probability of the class; R ( b a ) is the prospect of the predictor specified class; and R ( b ) is the earlier probability of the predictor (see Equation (3)).
  • K-Star (KS) is an instance-based classifier that categorizes samples or instances by differentiating it based on pre-categorized samples [39]. Some similar functions are used to determine the class of test instances. It uses an entropy-based function, which differentiates from other instance-based learners [40].
  • Decision Tree (C4.5) is the extension version of the ID3 algorithm that uses the recursive divide and conquer method to produce the C4.5 decision tree [41]. When unknown data are found, this method predicts a target class by satisfying several conditions. C4.5 uses Information Gain (IG) that calculates the gain ratio by the following equation [42],
    G a i n r a t i o ( A t t r i b u t e ) = G a i n ( A t t r i b u t e ) S p i l t I n f o ( A t t r i b u t e )
  • Classification and Regression Trees (CART) use several classification and regression trees where classification trees are used to accumulate the finite number of unsorted values and determine the prediction errors. Besides, regression trees are employed for grouping sorted or ordered values and determining the prediction error by calculating the root squared difference between target and predicted values [43].
  • K-Nearest Neighbour (KNN) is used for classifying instances based on their nearest neighbors [44]. It generally takes more than one neighbor and determines their distances using the Euclidean method, which is calculated with the following equation [45],
    D = ( m 1 m 2 ) 2 + ( n 1 n 2 ) 2
    where, D is distance between ( m 1 , n 1 ) a n d ( m 2 , n 2 ) points.
  • Support Vector Machine (SVM) generates some vectors to create a decision boundary that separates n-dimensional space into classes. This decision boundary is called a hyperplane. In the general situation, two parallel hyperplanes are generated, which concurrently minimizes the classification error and maximizes the margin of classes. It is called a maximum margin classifier [46].
  • Bagging Classifier (BG) is a parallel ensemble method that generates several random subsets from substitution of the original dataset. Then, we analyzed them using the base classifier and aggregated their predictions by voting [47]. It decreases the variance and correctly predicts the target outcome.
  • Random Tree (RT) is a decision tree where a set of possible trees are randomly generated with K random features. The combination of large sets of random trees is generally produced with accurate predictions more efficiently [48].

3.4. Use Evaluation Metrics for Performance Analysis of Classifiers

The performance measurement is essential for evaluating how well a classification model correctly predicts instances and achieves a desired target [17,49]. The confusion matrix provides a more detailed overview of a predictive model’s performance. It represents which classes are being predicted correctly and incorrectly and shows the measurement of type errors. In the confusion matrix, every instance in a given dataset falls into one of the four categories: True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN) [16]. The above four category data are arranged in a matrix known as the confusion matrix. Table 3 shows the elements of the confusion matrix.
In this work, we used several evaluation metrics including accuracy, kappa statistics, the F1-Score, and AUROC to assess the performance of each classifier. To calculate these metrics, the confusion matrix is required to generate and gather different types of instances from it. These metrics are described briefly as follows:
  • Accuracy: It is a measure of how effective the model is used to predict outcomes [49], in terms of the total number of predictions:
    A c c u r a c y = ( T P + T N ) ( T P + F P + F N + T N )
  • Kappa statistics (Kp): It measures observer agreement for categorical data and expected accuracy and has received considerable attention [50].
    K p = 1 ( 1 p 0 ) ( 1 p e )
  • Precision: It is a measure of true-positive predictions against all retrieved positive instances [51].
    P r e c i s i o n = T P ( T P + F P )
  • Recall: It is a measure of correctly predicted positive observations against all relevant positive classes [51].
    R e c a l l = T P ( T P + F N )
  • F1 score: It is the harmonic average of precision and recall [52]:
    F 1 = 2 ( R e c a l l P r e c i s i o n ) ( R e c a l l + P r e c i s i o n )
  • AUROC: It determines how well true-positive values are isolated from false-positive values [15]:
    T R P = T P ( T P + F N )
    F R P = F P ( F P + T N )

3.5. Determining the Performance of Classifiers Using Statistical Tests

After the classification process, we needed to justify these outcomes using various statistical methods and recheck their performance. In this work, the Wilcoxon Signed-Rank (WSR) method was used to test the statistical significance of the individual classifier. We employed this method into the outcomes of different evaluation metrics in the individual age group. A brief description of the WSR method is given as follows:
  • Wilcoxon Signed-Rank Test is a non-parametric statistical test that is used to compare two independent samples. This method is considered an alternative of the t-test when the population mean is not of interest. The working formula of this method is given as follows:
    W = i = 1 N [ s g n ( x 2 i x 1 i ) . R i ]
    where W denotes test statistics; N indicates the sample size; s g n denotes a sign function; both x 1 i , x 2 i represent the ranked pairs of the two distributions; and R i indicates the rank.

3.6. Interpretation of the Results of Machine Learning Models

Explainable AI is a combination of methods that allows individual users to comprehend the results of machine learning models. In this work, we explored the best performing classifiers which generate the highest results for different feature subsets in the individual age groups. To interpret these outcomes, the SHapley Additive exPlanations (SHAP) method was used to explain which feature vectors are required to generate these predictions. The SHAP method is described in brief as follows:
  • The SHAP method is a game-theoretic process to explain the output of any individual model. This model was developed by Lundberg and Lee [53]. The purpose of this method is to compute the contribution of each feature for an instance’s prediction. It associates optimal credit allocation with a local explanation using Shapley values. The simplest general SHAP values are represented as follows [54]:
    • Select an objective feature function.
    • Calculate the Shapley value for all features.
    • Choose the highest-ranking features.

4. Experimental Results

In the primary stage, this study used the Classification and Regression Training (CARET) package in R for feature selection and classification tasks [55]. However, Boruta, CFS, RIPPER, and RFE methods were implemented to produce numerous feature subsets for different age groups (i.e., toddler, child, adolescent, and adult). Table 4 shows the details of these feature subsets, respectively. Then, different baselines (i.e., primary data) along with their feature subsets were scrutinized by eight classifiers such as NB, BG, CART, KNN, C4.5, KS, SVM, and RT, respectively. In this case, we considered the k-fold cross-validation technique for classification, where the value of k is regarded as 10. After the classification process, a non-parametric statistical WSR test was employed to evaluate the performance of individual classifiers using Knowledge Extraction based on Evolutionary Learning (KEEL) software [56]. Then, SHAP summary plots were generated using the shap package in Python for the best classifier and different subsets. Figure 2 shows the details of illustrations about how to interpret machine learning models and explore importance features for generating significant outputs.

4.1. Generating Several Feature Subsets

Different feature selection methods such as Boruta, CFS, RIPPER, and RFE were applied into four ASD baselines and generated F S b o r , F S c f s , F S r p r , F S r f e , and F S b i c for F S b o r intersecting with F S c f s , respectively [27]. Table 4 shows several feature subsets in different age groups.

4.2. Result Analysis of Accuracy

We applied NB, BG, CART, KNN, C4.5, KS, SVM, and RT into four datasets where the highest 96.67%, 95.48%, 95.48%, and 96.06% accuracies were generated by SVM for toddler, child, adolescent, and adult datasets, respectively. After investigating the performance of these classifiers into the baseline and six feature subsets, we found greater improved accuracy to detect autism. In toddlers, SVM gave the maximum accuracy of 97.82% for F S r p r (see Table 5). Then, we observed that the highest accuracy (99.61%) was calculated by SVM for F S c f s and F S b i c for children (see Table 6). Regarding the case of adolescents, SVM also provided the maximum accuracy of 95.87% for F S b o r (see Table 7). In adults, the highest accuracy (96.82%) was found for F S c f s , which was computed by SVM (see Table 8). Besides, other classifiers such as NB, KS, and KNN provided good accuracy similar to SVM. Figure 3a shows the average accuracy of individual classifiers in each age group. Here, we also observed that SVM represented the best average outcomes among all classifiers.

4.3. Result Analysis of Kappa Statistics

Table 5, Table 6, Table 7 and Table 8 show the outcomes of kappa statistics of NB, BG, CART, KNN, C4.5, KS, SVM, and RT for four ASD groups and their feature subsets. In toddlers, the highest kappa (94.87%) was calculated by SVM for F S r p r . Regarding the case of children, SVM also provided the highest kappa of 99.21% for F S c f s and F S b i c . For adolescents, the maximum result (91.74%) was computed by SVM for F S b o r . In the adult case, the highest kappa (99.59%) was shown by SVM for F S c f s . Like the outcomes of accuracy, NB, KS, and KNN showed the closest kappa to detect autism. Besides, the average kappa statistics of the toddler, child, adolescent, and adult datasets are shown in Figure 3b. In this case, SVM again showed the best average kappa in each age group.

4.4. Result Analysis of F1-Score

The performance of F-measures in each classifier is shown in Table 5, Table 6, Table 7 and Table 8, for ASD baselines and their subsets, respectively. The maximum F1-Score (97.80%) was generated by SVM for F S r p r regarding the toddler dataset. When we observed the findings of the children, SVM also gave the highest f-measure (99.90%) for F S c f s and F S b i c . In adolescents, the best result (95.90%) was generated by SVM for F S b o r . For the adult dataset, SVM again provided the highest outcome (96.90%) for F S c f s . However, NB, KS, and KNN represented good results, which were almost similar to SVM. The average F-measure of the toddler, child, adolescent, and adult datasets of different classifiers are shown in Figure 3c. Further, SVM provided the average highest results in each age group.

4.5. Result Analysis of AUROC

The results of AUROC for the baseline and their feature subsets are shown in Table 5, Table 6, Table 7 and Table 8. In toddlers, NB provided the maximum result of 99.70% for F S b o r . Regarding the children, SVM and KS gave the highest result of 99.60% for F S c f s and F S b i c , respectively. For adolescents, the maximum result of 99.00% was produced by NB for F S b o r . In adults, SVM provided a result of 99.80%, which was the best outcome for F S c f s . AUROC, NB, SVM, KS, and KNN showed good outcomes, whereas the performance of NB was slightly improved over others. The average AUROC of the toddler, child, adolescent, and adult datasets of different classifiers are shown in Figure 3d. In this analysis, NB showed the highest average scores to detect autism.

4.6. Result Analysis of Non-Parametric Statistical Analysis

In this study, the performance of different classifiers were investigated using non-parametric the Wilcoxon Signed-Rank (WSR) test. We observed pairwise WSR values for toddlers (i.e., Table 9, Table 10, Table 11 and Table 12), children (i.e., Table 13, Table 14, Table 15 and Table 16), adolescents (i.e., Table 17, Table 18, Table 19 and Table 20), and adults (i.e., Table 21, Table 22, Table 23 and Table 24). At almost all age levels, SVM is a stable classifier to provide better pairwise statistical significance with other classifiers in order to improve the accuracy, kappa statistics, and f1-measure. Besides, NB sometimes shows better statistical significance for these following metrics. However, this classifier showed better pairwise WSR results with others for AUROC. Some other pairwise outcomes such as C4.5-CART, KNN-CART, KS-BG, KS-CART, KS-KNN, KS-C4.5, KS-RT, CART-RT, KNN-RT, and C4.5-RT showed almost-good outcomes in this test.

4.7. Exploring Significant Feature Sets and Discriminatory Factors of Individual Age Groups

For toddler datasets, SVM represented their maximum result in all metrics except AUROC for F S r p r . When we considered the value of AUROC, NB showed the best result for F S r p r . Then, SVM also showed their highest outcomes for all metrics for F S c f s and F S b i c in child datasets. In addition, KS provided the best AUROC for F S c f s and F S b i c . In adolescent datasets, SVM showed the best accuracy, kappa statistics, and f1-score for F S b o r . However, NB represented the top AUROC for F S c f s . In adult datasets, SVM also produced the highest findings in all metrics for F S c f s . Therefore, it was found that the RIPPER algorithm for toddler, the Boruta algorithm for adolescent, and the CFS method for child and adult datasets were responsible to generate high results in the classification process.
Figure 2 depicts the ranks of SHAP values of toddler’s RIPPER subset, adolescent’s Boruta subset, as well as the subsets of CFS child and adult datasets. The assessment of these values was performed with the best-performing SVM for these individual subsets. For the RIPPER subset of toddlers, the most important discriminatory features were “age group identification,” “Character’s Intention (A7),” “Following relook (A6),” “Pretending Capability (A5),” and “Pointing to the interest (A4)” to detect autism at an early stage. For the CFS subset of children, “Back to the activities (A4),” “Making Friends (A10),” “Finding the character’s intention (A7),” “Understanding someone’s feeling (A9),” and “Noticing sound (A1)” were the most crucial discriminating factors. Besides, “Social activity (A6),” “Indicating toy (A3),” “Pretending Capability (A5),” “Return to work (A4),” and “Developing Relationships (A10)” were the most significant discriminating factors for the Boruta subset of adolescents. In the CFS subset of adults, “Understanding someone’s feelings (A9),” “Pretending Capability (A5),” “Social activity (A6),” “Track activities (A3),” and “Return to work (A4)” are the major discriminating factors to detect autism.

5. Discussion and Conclusions

In this study, we proposed a machine learning model that analyzed ASD datasets of individual age groups (toddlers, children, adolescents, and adults) to detect autism at an early stage. This model used different feature selection methods like Boruta, CFS, RIPPER, and RFE to generate feature subsets. Then, NB, BG, CART, KNN, C4.5, KS, SVM, and RT were employed to classify autism at an early stage. When we evaluated the performance of them, SVM was the most stable classifier to explore the best result for different age groups, respectively. Along with SVM, KS, KNN, and NB showed better results to identify autism, respectively. On the other hand, individual classifiers presented their best performance in F S r p r for toddlers, F S b o r for adolescents, F S c f s for children and adults, and F S b i c for children.
Table 25 shows the comparison of the proposed model with related previous studies. Most of the existing works were worked with version-1 ASD datasets of Thabtah et al. [6]. Besides, they did not properly focus on early detection of autism for toddlers and adolescents. Some works occurred with version-2 adolescent and adult datasets, whereas most of them did not conduct those works with more samples and machine learning approaches. Thabtah et al. [14] used CHI and IG feature ranking methods into primary adolescent and adult datasets to produce the most significant feature subsets. Then, they implemented logistic regression into primary datasets and subsets and observed the highest 99.91% accuracy for the adolescent dataset and a 97.58% accuracy for the adult dataset. In this work, we investigated all version-2 datasets of different age groups of Thabtah et al. [14] where SVM achieved a 97.82% accuracy with F S r p r for the toddler dataset, a 99.61% accuracy with F S c f s and F S b i c for the child dataset, a 95.87% accuracy with F S b o r for the adolescent dataset, and a 96.82% accuracy for the adult dataset for F S c f s . Besides, we applied numerous feature selection and classification methods to justify these baselines and its feature sets more efficiently. In addition, a post hoc statistical significant test and SHAP interpretation method was used to evaluate deeply the outcomes of the proposed models. These types of evaluations did not properly occur in most of the existing works.
In conclusions, the proposed machine learning framework was used to produce more-accurate and efficient results for early detection of ASD at individual age groups. Since diagnosing ASD traits is an expensive and time-consuming procedure, it is often postponed due to the difficulty of detecting autism in children and adolescents. In this process, machine learning models are efficient to detect autism at an early stage very efficiently. However, this model was not trained with various multivariate/dimensional datasets and explored significant attributes. In the future, we will integrate this framework with advanced technologies and develop a more-efficient ASD diagnosis system. This system will be applicable for the preliminary diagnosis of ASD at an early stage with low costs.

Author Contributions

Conceptualization, M.B. and M.H.A.; methodology and software, M.B.; validation and formal analysis, M.H.A., M.S.S. and K.F.H.; resources, M.B. and M.S.S.; data curation, M.B.; writing—original draft preparation, M.B. and M.S.S., writing—review and editing, M.B., M.S.S., K.F.H. and M.A.M.; visualization, M.B.; supervision, M.A.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this paper is available in the references in Section 3.1.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Landa, R.J.; Gross, A.L.; Stuart, E.A.; Faherty, A. Developmental trajectories in children with and without autism spectrum disorders: The first 3 years. Child Dev. 2013, 84, 429–442. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Belmonte, M.K.; Allen, G.; Beckel-Mitchener, A.; Boulanger, L.M.; Carper, R.A.; Webb, S.J. Autism and abnormal development of brain connectivity. J. Neurosci. 2004, 24, 9228–9231. [Google Scholar] [CrossRef] [PubMed]
  3. GBD 2019 Diseases and Injuries Collaborators. GBD 2019 Diseases and Injuries Collaborators Global burden of 369 diseases and injuries in 204 countries and territories, 1990–2019: A systematic analysis for the Global Burden of Disease Study 2019. Lancet 2020, 396, 1204–1222. [Google Scholar] [CrossRef]
  4. Usta, M.B.; Karabekiroglu, K.; Sahin, B.; Aydin, M.; Bozkurt, A.; Karaosman, T.; Aral, A.; Cobanoglu, C.; Kurt, A.D.; Kesim, N.; et al. Use of machine learning methods in prediction of short-term outcome in autism spectrum disorders. Psychiatry Clin. Psychopharmacol. 2019, 29, 320–325. [Google Scholar] [CrossRef] [Green Version]
  5. Hyde, K.K.; Novack, M.N.; LaHaye, N.; Parlett-Pelleriti, C.; Anden, R.; Dixon, D.R.; Linstead, E. Applications of supervised machine learning in autism spectrum disorder research: A review. Rev. J. Autism Dev. Disord. 2019, 6, 128–146. [Google Scholar] [CrossRef] [Green Version]
  6. Thabtah, F.; Kamalov, F.; Rajab, K. A new computational intelligence approach to detect autistic features for autism screening. Int. J. Med. Inform. 2018, 117, 112–124. [Google Scholar] [CrossRef] [PubMed]
  7. Satu, M.S.; Azad, M.S.; Haque, M.F.; Imtiaz, S.K.; Akter, T.; Barua, L.; Rashid, M.; Soron, T.R.; Al Mamun, K.A. Prottoy: A smart phone based mobile application to detect autism of children in Bangladesh. In Proceedings of the 2019 4th International Conference on Electrical Information and Communication Technology (EICT), Khulna, Bangladesh, 17–19 December 2019; pp. 1–6. [Google Scholar]
  8. Akter, T.; Ali, M.H.; Khan, M.I.; Satu, M.S.; Moni, M.A. Machine learning model to predict autism investigating eye-tracking dataset. In Proceedings of the 2021 2nd International Conference on Robotics, Electrical and Signal Processing Techniques (ICREST), Online, 5–7 January 2021; pp. 383–387. [Google Scholar]
  9. Akter, T.; Ali, M.H.; Khan, M.I.; Satu, M.S.; Uddin, M.; Alyami, S.A.; Ali, S.; Azad, A.; Moni, M.A. Improved transfer-learning-based facial recognition framework to detect autistic children at an early stage. Brain Sci. 2021, 11, 734. [Google Scholar] [CrossRef]
  10. Omar, K.S.; Mondal, P.; Khan, N.S.; Rizvi, M.R.K.; Islam, M.N. A machine learning approach to predict autism spectrum disorder. In Proceedings of the 2019 International Conference on Electrical, Computer and Communication Engineering (ECCE), Bazar, Bangladesh, 7–9 February 2019; pp. 1–6. [Google Scholar]
  11. Sharma, M. Improved autistic spectrum disorder estimation using Cfs subset with greedy stepwise feature selection technique. Int. J. Inf. Technol. 2019, 14, 1251–1261. [Google Scholar] [CrossRef]
  12. Satu, M.S.; Sathi, F.F.; Arifen, M.S.; Ali, M.H.; Moni, M.A. Early detection of autism by extracting features: A case study in Bangladesh. In Proceedings of the 2019 International Conference on Robotics, Electrical and Signal Processing Techniques (ICREST), Dhaka, Bangladesh, 10–12 January 2021; pp. 400–405. [Google Scholar]
  13. Erkan, U.; Thanh, D.N. Autism spectrum disorder detection with machine learning methods. Curr. Psychiatry Res. Rev. Former. Curr. Psychiatry Rev. 2019, 15, 297–308. [Google Scholar] [CrossRef]
  14. Thabtah, F.; Abdelhamid, N.; Peebles, D. A machine learning autism classification based on logistic regression analysis. Health Inf. Sci. Syst. 2019, 7, 1–11. [Google Scholar] [CrossRef]
  15. Akter, T.; Satu, M.S.; Khan, M.I.; Ali, M.H.; Uddin, S.; Lio, P.; Quinn, J.M.; Moni, M.A. Machine learning-based models for early stage detection of autism spectrum disorders. IEEE Access 2019, 7, 166509–166527. [Google Scholar] [CrossRef]
  16. Hossain, M.D.; Kabir, M.A.; Anwar, A.; Islam, M.Z. Detecting Autism Spectrum Disorder using Machine Learning. arXiv 2020, arXiv:2009.14499. [Google Scholar]
  17. Raj, S.; Masood, S. Analysis and Detection of Autism Spectrum Disorder Using Machine Learning Techniques. Procedia Comput. Sci. 2020, 167, 994–1004. [Google Scholar] [CrossRef]
  18. Thabtah, F.; Peebles, D. A new machine learning model based on induction of rules for autism detection. Health Inform. J. 2020, 26, 264–286. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Chowdhury, K.; Iraj, M.A. Predicting Autism Spectrum Disorder Using Machine Learning Classifiers. In Proceedings of the 2020 International Conference on Recent Trends on Electronics, Information, Communication & Technology (RTEICT), Bangalore, India, 12–13 November 2020; pp. 324–327. [Google Scholar]
  20. Akter, T.; Khan, M.I.; Ali, M.H.; Satu, M.S.; Uddin, M.J.; Moni, M.A. Improved machine learning based classification model for early autism detection. In Proceedings of the 2021 2nd International Conference on Robotics, Electrical and Signal Processing Techniques (ICREST), Online, 5–7 January 2021; pp. 742–747. [Google Scholar]
  21. Akter, T.; Ali, M.H.; Satu, M.S.; Khan, M.I.; Mahmud, M. Towards Autism Subtype Detection Through Identification of Discriminatory Factors Using Machine Learning. In Proceedings of the International Conference on Brain Informatics, Online, 17–19 September 2021; pp. 401–410. [Google Scholar]
  22. Thabtah, F. Autism Screening Data for Toddlers. 2019. Available online: https://fadifayez.com/ (accessed on 28 March 2022).
  23. Thabtah, F. Autism Screening Data for Child. 2019. Available online: https://fadifayez.com/ (accessed on 28 March 2022).
  24. Thabtah, F. Autism Screening Data for Adolescent. 2019. Available online: https://fadifayez.com/ (accessed on 28 March 2022).
  25. Thabtah, F. Autism Screening Data for Adult. 2019. Available online: https://fadifayez.com/ (accessed on 28 March 2022).
  26. Alzubi, R.; Ramzan, N.; Alzoubi, H. Hybrid feature selection method for autism spectrum disorder SNPs. In Proceedings of the 2017 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB), Piscataway, NJ, USA, 23–25 August 2017; pp. 1–7. [Google Scholar]
  27. Thabtah, F.; Zhang, L.; Abdelhamid, N. NBA game result prediction using feature analysis and machine learning. Ann. Data Sci. 2019, 6, 103–116. [Google Scholar] [CrossRef]
  28. Liaw, A.; Wiener, M. Classification and regression by randomForest. R News 2002, 2, 18–22. [Google Scholar]
  29. Kursa, M.B.; Jankowski, A.; Rudnicki, W.R. Boruta—A system for feature selection. Fundam. Inform. 2010, 101, 271–285. [Google Scholar] [CrossRef]
  30. Kursa, M.B.; Rudnicki, W.R. Feature selection with the Boruta package. J. Stat. Softw. 2010, 36, 1–13. [Google Scholar] [CrossRef] [Green Version]
  31. Howlader, K.C.; Satu, M.; Awal, M.; Islam, M.; Islam, S.M.S.; Quinn, J.M.; Moni, M.A. Machine learning models for classification and identification of significant attributes to detect type 2 diabetes. Health Inf. Sci. Syst. 2022, 10, 2. [Google Scholar] [CrossRef]
  32. Cohen, W.W.; Singer, Y. A simple, fast, and effective rule learner. AAAI/IAAI 1999, 99, 3. [Google Scholar]
  33. Yang, J.; Tiyyagura, A.; Chen, F.; Honavar, V. Feature subset selection for rule induction using RIPPER. In Proceedings of the Genetic and Evolutionary Computation Conference, Berlin, Germany, 15–19 July 1999; Volume 2, p. 1800. [Google Scholar]
  34. Seerat, B.; Qamar, U. Rule induction using enhanced RIPPER algorithm for clinical decision support system. In Proceedings of the 2015 Sixth International Conference on Intelligent Control and Information Processing (ICICIP), Wuhan, China, 26–28 November 2015; pp. 83–91. [Google Scholar]
  35. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  36. Ahammed, K.; Satu, M.S.; Khan, M.I.; Whaiduzzaman, M. Predicting Infectious State of Hepatitis C Virus Affected Patient’s Applying Machine Learning Methods. In Proceedings of the 2020 IEEE Region 10 Symposium (TENSYMP), Online, 5–7 June 2020; pp. 1371–1374. [Google Scholar]
  37. Guyon, I.; Weston, J.; Barnhill, S.; Vapnik, V. Gene selection for cancer classification using support vector machines. Mach. Learn. 2002, 46, 389–422. [Google Scholar] [CrossRef]
  38. John, G.H.; Langley, P. Estimating continuous distributions in Bayesian classifiers. arXiv 2013, arXiv:1302.4964. [Google Scholar]
  39. Witten, I.H.; Frank, E. Data mining: Practical machine learning tools and techniques with Java implementations. Acm Sigmod Rec. 2002, 31, 76–77. [Google Scholar] [CrossRef]
  40. Painuli, S.; Elangovan, M.; Sugumaran, V. Tool condition monitoring using K-star algorithm. Expert Syst. Appl. 2014, 41, 2638–2643. [Google Scholar] [CrossRef]
  41. Alam, F.; Pachauri, S. Detection using weka. Adv. Comput. Sci. Technol. 2017, 10, 1731–1743. [Google Scholar]
  42. Trabelsi, M.; Meddouri, N.; Maddouri, M. A new feature selection method for nominal classifier based on formal concept analysis. Procedia Comput. Sci. 2017, 112, 186–194. [Google Scholar] [CrossRef]
  43. Lee, B.K.; Lessler, J.; Stuart, E.A. Improving propensity score weighting using machine learning. Stat. Med. 2010, 29, 337–346. [Google Scholar] [CrossRef] [Green Version]
  44. Sayad, S. K Nearest Neighbors—Classification. Available online: http://www.saedsayad.com/k_nearest_neighbors.htm (accessed on 28 March 2022).
  45. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction; Springer: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
  46. Keerthi, S.S.; Shevade, S.K.; Bhattacharyya, C.; Murthy, K.R.K. Improvements to Platt’s SMO algorithm for SVM classifier design. Neural Comput. 2001, 13, 637–649. [Google Scholar] [CrossRef]
  47. Lee, S.J.; Xu, Z.; Li, T.; Yang, Y. A novel bagging C4. 5 algorithm based on wrapper feature selection for supporting wise clinical decision making. J. Biomed. Inform. 2018, 78, 144–155. [Google Scholar] [CrossRef]
  48. Ali, J.; Khan, R.; Ahmad, N.; Maqsood, I. Random forests and decision trees. Int. J. Comput. Sci. Issues (IJCSI) 2012, 9, 272. [Google Scholar]
  49. Hasnain, M.; Pasha, M.F.; Ghani, I.; Imran, M.; Alzahrani, M.Y.; Budiarto, R. Evaluating trust prediction and confusion matrix measures for web services ranking. IEEE Access 2020, 8, 90847–90861. [Google Scholar] [CrossRef]
  50. Kalkhan, M.; Reich, R.; Czaplewski, R. Variance estimates and confidence intervals for the Kappa measure of classification accuracy. Can. J. Remote Sens. 1997, 23, 210–216. [Google Scholar] [CrossRef]
  51. Davis, J.; Goadrich, M. The relationship between Precision-Recall and ROC curves. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 233–240. [Google Scholar]
  52. Powers, D.M. Evaluation: From precision, recall and F-measure to ROC, informedness, markedness and correlation. arXiv 2020, arXiv:2010.16061. [Google Scholar]
  53. Lundberg, S.M.; Lee, S.I. A unified approach to interpreting model predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 4768–4777. [Google Scholar]
  54. Fryer, D.; Strümke, I.; Nguyen, H. Shapley values for feature selection: The good, the bad, and the axioms. arXiv 2021, arXiv:2102.10936. [Google Scholar] [CrossRef]
  55. Kuhn, M. The Caret Package. Available online: https://topepo.github.io/caret/ (accessed on 27 March 2019).
  56. Satu, M.; Zoynul Abedin, M.; Khanom, S.; Ouenniche, J.; Shamim Kaiser, M. Application of feature engineering with classification techniques to enhance corporate tax default detection performance. In Proceedings of the International Conference on Trends in Computational and Cognitive Engineering, Online, 21–22 October 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 53–63. [Google Scholar]
  57. McNamara, B.; Lora, C.; Yang, D.; Flores, F.; Daly, P. Machine Learning Classification of Adults with Autism Spectrum Disorder. Available online: http://rstudio-pubs-static.s3.amazonaws.com/383049_1faa93345b324da6a1081506f371a8dd.html (accessed on 29 April 2018).
  58. Bala, M.; Prova, A.A.; Ali, M.H. Prediction of Autism Spectrum Disorder Using Feature Selection and Machine Learning Algorithms. In Proceedings of the International Conference on Computational Intelligence and Emerging Power System, Ajmer, India, 9–10 March 2021; Springer: Berlin/Heidelberg, Germany, 2022; pp. 133–148. [Google Scholar]
Figure 1. The proposed framework for early ASD detection.
Figure 1. The proposed framework for early ASD detection.
Algorithms 15 00166 g001
Figure 2. Analysis of Shapley values for (a) the toddler subset of the RIPPER method, (b) the adolescent subset of the Boruta algorithm, (c) the child subset of the CFS–Harmony Search method, and (d) the adult subset of the CFS method employing the best-performing SVM.
Figure 2. Analysis of Shapley values for (a) the toddler subset of the RIPPER method, (b) the adolescent subset of the Boruta algorithm, (c) the child subset of the CFS–Harmony Search method, and (d) the adult subset of the CFS method employing the best-performing SVM.
Algorithms 15 00166 g002
Figure 3. The average (a) accuracy, (b) kappa statistics, (c) F1, and (d) AUROC of the toddler, child, adolescent, and adult datasets of the classifiers.
Figure 3. The average (a) accuracy, (b) kappa statistics, (c) F1, and (d) AUROC of the toddler, child, adolescent, and adult datasets of the classifiers.
Algorithms 15 00166 g003
Table 1. Data description.
Table 1. Data description.
Dataset NameInstancesAttributesMale/FemaleAge (Years)Average AgeASD/Normal
Toddler Autism dataset105418735/31912–36 (mons)27.86 (mons)735/319
Autism Child Dataset (Version-2)50923363/14611-Apr6.39257/252
Autism Adolescent Dataset (Version-2)24823117/13116-Dec14.04127/121
Autism Adult Dataset (Version-2)111823596/52217–8030.143358/760
Table 2. Features description.
Table 2. Features description.
FeatureTypeDescription
Age (Year)NumberToddlers (months), children, adolescent, and adults (year)
GenderStringMale or female
EthnicityStringList of common ethnicities
Born with jaundiceBooleanWhether the case was born with jaundice
Family member with PDDBooleanWhether any immediate family member has a PDD history
Who is completing the testStringParent, self, caregiver, medical staff, clinician, etc.
Country of residenceStringList of countries
Used the screening app beforeBooleanWhether the user has formerly used the screening app
LanguageStringThe user’s language details
A1: Answer of Q1Binary(0,1)Does your child look at you when you call his/her name? (Toddler)
S/he often notices small sounds when others do not. (Child) S/he
notices patterns in things all the time. (Adolescent) I often notice
small sounds when others do not. (Adult)
A2: Answer of Q2Binary(0,1)How easy is it for you to get eye contact with your child? (Toddler)
S/he usually concentrates more on the whole picture rather than the
small details. (Child, adolescent, adult)
A3: Answer of Q3Binary(0,1)Does your child point to indicate that s/he wants something? (Toddler)
In a social group, s/he can easily keep track of several different people’s
conversations. (Child) In a social group, s/he can easily keep track of
several different people’s conversations. (Adolescent) I find it easy to
do more than one thing at once. (Adult)
A4: Answer of Q4Binary(0,1)Does your child point to share interest with you? (Toddler) S/he finds
it easy to go back and forth between different activities. (Child) If there
is an interruption, s/he can switch back to what s/he was doing very
quickly. (Adolescent) If there is an interruption, I can switch back to
what I was doing very quickly (Adult)
A5: Answer of Q5Binary(0,1)Does your child pretend? (Toddler) S/he does not know how to keep a
conversation going with his/her peers. (Child, Adolescent) I find it easy
to “read between the lines” when someone is talking to me. (Adult)
A6: Answer of Q6Binary(0,1)Does your child follow where you are looking? (Toddler) S/he is good
at social chit-chat (Child, Adolescent). I know how to tell if someone
listening to me is getting bored (Adult)
A7: Answer of Q7Binary(0,1)If you or someone else in the family is visibly upset, does your child
show signs of wanting to comfort them? (Toddler) When s/he is read
a story, s/he finds it difficult to work out the character’s intentions or
feelings (Child). When s/he was younger, s/he used to enjoy playing
games involving pretending with other children (Adolescent). When
I am reading a story I find it difficult to work out the characters’
intentions (Adult).
A8: Answer of Q8Binary(0,1)Would you describe your child’s first words as: (Toddler) When s/he
was in preschool, s/he used to enjoy playing pretending games with
other children (Child). S/he finds it difficult to imagine what it would
be like to be someone else (Adolescent). I like to collect information
about categories of things (Adult).
A9: Answer of Q9Binary(0,1)Does your child use simple gestures? (Toddler). S/he finds it easy to
work out what someone is thinking or feeling just by looking at their
face (Child). S/he finds social situations easy (Adolescent). I find it
easy to work out what someone is thinking or feeling just by looking
at their face (Adult).
A10: Answer of Q10Binary(0,1)Does your child stare at nothing with no apparent purpose? (Toddler).
S/he finds it hard to make new friends (Child, Adolescent). I find it
difficult to work out people’s intentions (Adult).
Screening scoreIntegerIt is based on the scoring algorithm of the screening method used
ClassStringASD or No ASD
Table 3. Confusion matrix for ASD.
Table 3. Confusion matrix for ASD.
TargetPredicted
ASDNo ASD
ASDTPFN
No ASDFPTN
Table 4. The generated feature subsets of individual age groups.
Table 4. The generated feature subsets of individual age groups.
FS MethodSelected Features
Toddler F S b o r BorutaA1, A2, A3, A4, A5, A6, A7, A8, A9, A10, Age, Jaundice, Family_ASD
F S c f s CFSA1, A2, A4, A5, A6, A7, A8, A9, A10, Age, User
F S r p r RIPPERA1, A2, A3, A4, A5, A6, A7, A8, A9, A10, Age, Jaundice
F S r f e RFEA1, A2, A3, A4, A6, A7, A8, A9, A10, Gender, User
F S b i c F S b o r Intersect F S r p r A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, Age
Child F S b o r BorutaA1, A2, A3, A4, A5, A6, A7, A8, A9, A10, Age, Gender, Residence
F S c f s CFSA1, A2, A3, A4, A5, A6, A7, A8, A9, A10
F S r p r RIPPERA1, A2, A3, A4, A5, A6, A7, A8, A9, A10, Jaundice
F S r f e RFEA1, A4, A6, A7, A8, A9, A10, Ethnicity, Jaundice, Family_ASD
F S b i c F S b o r Intersect F S c f s A1, A2, A3, A4, A5, A6, A7, A8, A9, A10
Adolescent F S b o r BorutaA1, A2, A3, A4, A5, A6, A7, A8, A9, A10, Age, Ethnicity, Residence, User
F S c f s CFSA1, A2, A3, A4, A5, A6, A7, A10, Ethnicity
F S r p r RIPPERA1, A2, A3, A4, A5, A6, A7, A8, A10, Age, Gender
F S r f e RFEA1, A2, A3, A4, A5, A6, A8, A10, Age, Gender, Ethnicity, Jaundice, Residence
F S b i c F S b o r Intersect F S c f s A1, A2, A3, A4, A5, A6, A7, A8, A10, Age
Adult F S b o r BorutaA1, A2, A3, A4, A5, A6, A7, A8, A9, A10, Age, Ethnicity, Residence, User
F S c f s CFSA1, A2, A3, A4, A5, A6, A7, A8, A9, A10, Residence
F S r p r RIPPERA1, A2, A3, A4, A5, A6, A7, A8, A9, A10, Age, Residence, User
F S r f e RFEA1, A2, A4, A6, A7, A8, A9, A10, Age, Residence, Used_App_Before
F S b i c F S b o r Intersect F S r p r A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, Age, Residence, User
Table 5. Performance analysis of classifiers for baseline and its generated feature subsets of toddlers.
Table 5. Performance analysis of classifiers for baseline and its generated feature subsets of toddlers.
NBBGCARTKNNC4.5KSSVMRT
Accuracy B a s e l i n e 95.4493.1690.8992.6992.3194.2196.6792.12
F S b o r 95.4594.4092.0394.3092.5995.2597.6292.78
F S c f s 96.3992.5991.1794.9791.7495.1696.6792.97
F S r p r 95.4594.4992.1295.2592.5995.6397.8292.88
F S r f e 95.2593.9292.1294.3092.3195.2595.6393.73
F S b i c 95.1694.2192.1295.9292.7895.9297.6293.45
Kappa Stat. B a s e l i n e 89.5083.8878.3283.2081.8486.6292.2181.46
F S b o r 89.5586.7181.0986.9082.4789.0894.4383.18
F S c f s 91.5982.4179.1588.3880.6088.7892.2083.46
F S r p r 89.8586.9281.3089.0882.4789.9694.8783.44
F S r f e 88.8085.6281.5586.9681.8489.0189.5485.52
F S b i c 88.9186.2881.2790.5782.9290.5994.4384.77
F1-Score B a s e l i n e 95.5093.1090.8092.8092.3094.2096.7092.10
F S b o r 95.5094.4092.0094.4092.6095.3097.6092.80
F S c f s 96.4092.5091.1095.0091.7095.2096.7093.00
F S r p r 95.5094.5092.1095.3092.6095.7097.8092.90
F S r f e 95.2093.9092.1094.4092.3095.3095.6093.80
F S b i c 95.2094.2092.1095.9092.7095.9097.6093.50
AUROC B a s e l i n e 99.5098.0091.7094.2092.5098.8096.0090.70
F S b o r 99.7098.2092.3096.0093.0099.3097.0091.70
F S c f s 99.5097.8091.6096.9092.4099.2095.9091.40
F S r p r 99.6098.2092.3096.6093.0099.4097.209.20
F S r f e 99.1098.2091.3098.5093.2099.2093.7094.20
F S b i c 99.6098.2092.4097.3093.3099.5097.0092.60
Table 6. Performance analysis of classifiers for baseline and its generated feature subsets of children.
Table 6. Performance analysis of classifiers for baseline and its generated feature subsets of children.
NBBGCARTKNNC4.5KSSVMRT
Accuracy B a s e l i n e 93.1280.9483.7188.6189.3987.2295.4880.74
F S b o r 94.6981.9283.6992.9289.7893.1295.8778.58
F S c f s 93.5190.9690.3794.6990.7695.4899.6192.92
F S r p r 93.5190.9689.9892.7389.9893.1298.2390.77
F S r f e 88.9986.0587.2289.1988.9989.1991.3585.65
F S b i c 93.5190.9690.3794.6990.7795.4899.6192.93
Kappa Stat. B a s e l i n e 86.2461.8865.4577.1978.7874.4390.9661.47
F S b o r 89.3863.8567.3985.8579.5786.2391.7457.13
F S c f s 87.0281.9280.7589.3881.5390.9599.2185.85
F S r p r 87.0281.9279.9685.4579.9686.2396.4681.53
F S r f e 77.9972.1074.4678.3877.9978.3782.7171.31
F S b i c 87.0281.9280.7589.3881.5390.9599.2185.85
F1-Score B a s e l i n e 93.0180.9082.7088.6089.0487.2095.5080.70
F S b o r 94.7081.9083.7092.6089.8093.1095.9078.50
F S c f s 93.2090.9090.4094.5090.7095.2099.6092.90
F S r p r 93.5091.0090.0092.7090.0093.1098.2090.80
F S r f e 88.8086.1087.2089.0088.8088.5091.2085.40
F S b i c 93.5091.0090.0494.7090.8095.5099.6092.90
AUROC B a s e l i n e 98.4088.6085.2090.7091.7095.8095.4081.90
F S b o r 99.0089.2085.9094.5093.0098.1095.8079.40
F S c f s 99.2097.8094.0099.0093.6099.6099.6092.90
F S r p r 99.2097.9093.6098.1092.7099.4098.2090.80
F S r f e 96.2095.2091.4094.2092.1096.5091.3087.40
F S b i c 99.2097.8094.0099.0093.8099.6099.6092.90
Table 7. Performance analysis of classifiers for baseline and its generated feature subsets of adolescents.
Table 7. Performance analysis of classifiers for baseline and its generated feature subsets of adolescents.
NBBGCARTKNNC4.5KSSVMRT
Accuracy B a s e l i n e 93.1280.9482.7188.6189.3987.2295.4880.74
F S b o r 94.6981.9283.6992.9289.7893.1295.8778.58
F S c f s 89.9185.0881.0491.9382.6691.5389.5186.69
F S r p r 91.9487.9079.8491.5280.6592.7494.7686.29
F S r f e 87.0975.0078.2288.3081.8588.7089.5178.22
F S b i c 92.7487.9081.0591.9481.0591.1393.9589.11
Kappa Stat. B a s e l i n e 89.2461.8865.4577.1978.7874.4390.9661.47
F S b o r 89.3863.8567.3985.8579.5786.2391.7457.13
F S c f s 79.7870.1462.0883.8465.3183.0079.0273.31
F S r p r 83.8475.7959.6983.0461.3085.4489.5072.55
F S r f e 74.1249.8956.3476.5663.7077.3478.9556.41
F S b i c 85.4475.7962.1183.8562.1082.2087.9078.19
F1-Score B a s e l i n e 93.1080.9082.7088.6089.4087.2095.5080.70
F S b o r 94.7081.9083.7092.9089.8093.1095.9078.50
F S c f s 89.2084.6080.7091.5082.3090.7089.3085.70
F S r p r 91.9087.9079.8091.5080.6092.7094.8086.30
F S r f e 86.1073.5076.7087.6081.5087.7088.4077.50
F S b i c 92.7087.9081.0091.9081.1091.1094.0089.10
AUROC B a s e l i n e 98.4088.6085.2090.7091.7095.8095.4081.90
F S b o r 99.0089.2085.9094.5093.0098.1095.8079.40
F S c f s 93.2096.9082.0095.6081.7097.9089.5087.00
F S r p r 98.6095.7079.6094.6079.8098.3094.7086.30
F S r f e 95.5083.0081.9090.2082.505.2089.3080.10
F S b i c 98.3095.7079.5094.9080.6098.6094.0089.10
Table 8. Performance analysis of classifiers for baseline and its generated feature subsets of adults.
Table 8. Performance analysis of classifiers for baseline and its generated feature subsets of adults.
NBBGCARTKNNC4.5KSSVMRT
Accuracy B a s e l i n e 94.1887.5689.2691.9493.0292.6696.0685.15
F S b o r 94.3887.5789.1793.292.7593.2996.6087.20
F S c f s 95.5287.4789.5396.6993.0297.3199.8287.83
F S r p r 95.3487.5088.9994.9992.6694.4596.4288.90
F S r f e 92.3085.5188.4690.7890.3391.0591.6885.51
F S b i c 95.3487.5088.9994.9992.6694.4596.4288.90
Kappa Stat. B a s e l i n e 86.9371.1675.4581.8683.8183.5290.8966.29
F S b o r 87.3171.2175.3084.5283.2584.8792.1370.77
F S c f s 89.8571.0276.1292.4483.7693.9399.5972.30
F S r p r 89.4671.2174.9088.6083.0587.5291.6774.49
F S r f e 82.4666.1273.2878.7677.5879.5880.2167.11
F S b i c 89.4671.2174.9088.6083.0587.5291.6774.49
F1-Score F S b o r 94.2087.5089.3092.0093.0092.7096.0085.20
F S c f s 96.7090.8092.8097.6094.9098.0099.9091.00
F S r p r 95.4087.5189.0095.0092.6094.5096.4088.90
F S r f e 94.3089.5091.6093.2093.0093.4094.1089.20
F S b i c 95.4087.5189.0095.0092.6094.5096.4088.90
AUROC B a s e l i n e 99.0094.2091.9092.7095.7097.7095.1085.70
F S b o r 99.2094.2192.5093.7095.8098.5095.7086.40
F S c f s 99.5094.4092.7099.1096.1099.7099.8087.00
F S r p r 99.5094.3091.6095.7095.8099.0095.2088.20
F S r f e 97.9092.2090.1091.9092.3097.5088.7085.00
F S b i c 99.5094.3091.6095.7095.8099.0095.2088.20
Table 9. Pairwise WSR test for accuracy of toddlers.
Table 9. Pairwise WSR test for accuracy of toddlers.
NBBGCARTKNNC4.5KSSVMRT
NB 0.0211 **0.0211 **0.0592 *0.0179 **0.26431.00000.0211 **
BG1.0000 0.0211 **1.00000.0211 **1.00001.00000.0592 *
CART1.00001.0000 1.00001.00001.00001.00001.0000
KNN1.00000.14220.0179 ** 0.0211 **1.00001.00000.1787
C4.51.00001.00000.0211 **1.0000 1.00001.00001.0000
KS1.00000.0211 **0.0211 **0.0256 **0.0211 ** 1.00000.1787
SVM0.0211 **0.0211 **0.17870.17870.0211 **0.0211 ** 0.0211 **
RT1.00001.00000.0211 **1.00000.0408 **1.00001.0000
Note: Significant at 5% = **, 10% = *.
Table 10. Pairwise WSR test for kappa statistics of toddlers.
Table 10. Pairwise WSR test for kappa statistics of toddlers.
NBBGCARTKNNC4.5KSSVMRT
NB 0.0211 **0.0211 **0.0592 *0.0211 **0.40171.00000.0211 **
BG1.0000 0.0211 **1.00000.0211 **1.00001.00000.0592 *
CART1.00001.0000 1.00001.00001.00001.00001.0000
KNN1.00000.0592 *0.0211 ** 0.0211 **1.00001.00000.0211 **
C4.51.00001.00000.02111.0000 1.00001.00001.0000
KS1.00000.0211 **0.0211 **0.0211 **0.0211 ** 1.00000.0211 **
SVM0.0211 **0.0211 **0.0211 **0.0211 **0.0211 **0.0211 ** 0.0211 **
RT1.00001.00000.02111.00000.0360 **1.00001.0000
Note: Significant at 5% = **, 10% = *.
Table 11. Pairwise WSR test for F1-measure of toddlers.
Table 11. Pairwise WSR test for F1-measure of toddlers.
NBBGCARTKNNC4.5KSSVMRT
NB 0.0179 **0.0179 **0.0592 *0.0100 ***0.45111.00000.0179 **
BG1.0000 0.0179 **1.00000.0179 **1.00001.00000.5917
CART1.00001.0000 1.00001.00001.00001.00001.0000
KNN1.00000.0591 *0.0211 ** 0.0211 **1.00001.00000.0211 **
C4.51.00001.00000.0179 **1.0000 1.00001.00001.0000
KS1.00000.0211 **0.0211 **0.0256 **0.0211 ** 1.00000.0211 **
SVM0.0211 **0.0211 **0.0211 **0.0211 **0.17870.0211 ** 0.0211 **
RT1.00001.00000.0211 **1.00000.0408 **1.00001.0000
Note: Significant at 1% = ***, 5% = **, 10% = *.
Table 12. Pairwise WSR test for AUROC of toddlers.
Table 12. Pairwise WSR test for AUROC of toddlers.
NBBGCARTKNNC4.5KSSVMRT
NB 0.0149 **0.0179 **0.0211 **0.0211 **0.0592 *0.0211 **0.0211 **
BG1.0000 0.0179 **0.0360 **0.0179 **1.00000.0179 **0.0211 **
CART1.00001.0000 1.00001.00001.00001.00000.3326
KNN1.00001.00000.0211 ** 0.0211 **1.00000.91430.0179 **
C4.51.00001.00000.0179 **1.0000 1.00001.00000.0672 *
KS1.00000.0211 **0.0179 **0.0211 **0.0179 ** 0.0211 **0.0211 **
SVM1.00001.00000.0211 **0.91430.0179 **1.0000 0.0313 **
RT1.00001.00001.00001.00001.00001.00001.0000
Note: Significant at 5% = **, 10% = *.
Table 13. Pairwise WSR test for accuracy of child.
Table 13. Pairwise WSR test for accuracy of child.
NBBGCARTKNNC4.5KSSVMRT
NB 0.0100 ***0.0179 **0.51830.0310 **0.82951.00000.0211 **
BG1.0000 1.00001.00001.00001.00001.00000.6750
CART1.00000.2817 1.00001.00001.00001.00000.5294
KNN1.00000.0179 **0.0179 ** 0.0592 *1.00001.00000.0211 **
C4.51.00000.29450.0310 **1.0000 1.00001.00000.2945
KS1.00000.0179 **0.0179 **0.40240.0592 ** 1.00000.0211 **
SVM0.0149 **0.0179 **0.0179 **0.0179 **0.0179 **0.0179 ** 0.0211 **
RT1.00001.00001.00001.00001.00001.00001.0000
Note: Significant at 1% = ***, 5% = **, 10% = *.
Table 14. Pairwise WSR Test for kappa statistics of children.
Table 14. Pairwise WSR Test for kappa statistics of children.
NBBGCARTKNNC4.5KSSVMRT
NB 0.0100 ***0.0179 **0.51830.0256 **0.82951.00000.0179 **
BG1.0000 1.00001.00001.00001.00001.00000.6668
CART1.00000.2817 1.00001.00001.00001.00000.8295
KNN1.00000.0179 **0.0179 ** 0.0527 *1.00001.00000.0179 **
C4.51.00000.28170.0256 **1.0000 1.00001.00000.2817
KS1.00000.0179 **0.0179 **0.38910.0527 * 1.00000.0179 **
SVM0.0149 **0.0179 **0.0179 **0.0179 **0.0179 **0.0179 ** 0.0179 **
RT1.00001.00001.00001.00001.00001.00001.0000
Note: Significant at 1% = ***, 5% = **, 10% = *.
Table 15. Pairwise WSR test for F1-Measure of children.
Table 15. Pairwise WSR test for F1-Measure of children.
NBBGCARTKNNC4.5KSSVMRT
NB 0.0179 **0.0179 **0.52940.0310 **0.66681.00000.0211 **
BG1.0000 0.0850 *1.00001.00001.00001.00000.6668
CART1.00001.0000 1.00001.00001.00001.00001.0000
KNN1.00000.0211 **0.0211 ** 0.0592 *1.00001.00000.0211 **
C4.51.00000.28170.0211 **1.0000 1.00001.00000.2945
KS1.00000.0211 **0.0179 **0.59040.0935 * 1.00000.0179 **
SVM0.0211 **0.0211 **0.0179 **0.0211 **0.0211 **0.0211 ** 0.0179 **
RT1.00001.00000.13181.00001.00001.00001.0000
Note: Significant at 5% = **, 10% = *.
Table 16. Pairwise WSR Test for AUROC of Child.
Table 16. Pairwise WSR Test for AUROC of Child.
NBBGCARTKNNC4.5KSSVMRT
NB 0.0179 **0.0100 ***0.0179 **0.0211 **0.82950.0850 *0.0211 **
BG1.0000 0.0179 **1.00000.16171.00001.00000.0211 **
CART1.00001.0000 1.00001.00001.00001.00001.0000
KNN1.00000.0527 *0.0179 ** 0.0360 **1.00001.00000.0211 **
C4.51.00001.00000.0211 **1.0000 1.00001.00000.0211 **
KS1.00000.0179 **0.0179 **0.0179 **0.0211 ** 0.0464 **0.0211 **
SVM1.00000.13180.0179 **0.19640.0360 **1.0000 0.0211 **
RT1.00001.00000.0935 *1.00001.00001.00001.0000
Note: Significant at 1% = ***, 5% = **, 10% = *.
Table 17. Pairwise WSR test for accuracy of adolescents.
Table 17. Pairwise WSR test for accuracy of adolescents.
NBBGCARTKNNC4.5KSSVMRT
NB 0.0179 **0.0100 ***0.0179 **0.0211 **0.82950.0850 *0.0211 **
BG1.0000 0.0179 **1.00000.16171.00001.00000.0211 **
CART1.00001.0000 1.00001.00001.00001.00001.0000
KNN1.00000.0527 *0.0179 ** 0.0360 **1.00001.00000.0211 **
C4.51.00001.00000.0211 **1.0000 1.00001.00000.0211 **
KS1.00000.0179 **0.0179 **0.0179 **0.0211 ** 0.0464 *0.0211 **
SVM1.00000.13180.0179 **0.19640.0360 **1.0000 0.0211 **
RT1.00001.00000.0935 *1.00001.00001.00001.0000
Note: Significant at 1% = ***, 5% = **, 10% = *.
Table 18. Pairwise WSR test for kappa statistics of adolescents.
Table 18. Pairwise WSR test for kappa statistics of adolescents.
NBBGCARTKNNC4.5KSSVMRT
NB 0.0211 **0.0211 **0.52940.0211 **0.51831.00000.0211 **
BG1.0000 0.29451.00001.00001.00001.00000.8339
CART1.00001.0000 1.00001.00001.00001.00001.0000
KNN1.00000.0211 **0.0211 ** 0.0360 **0.52941.00000.0211 **
C4.51.00000.40170.0360 **1.0000 1.00001.00000.6750
KS1.00000.0211 **0.0211 **1.00000.0360 ** 1.00000.0211 **
SVM0.03600.0211 **0.0211 **0.09350.0211 **0.0592 * 0.0211 **
RT1.00001.00000.20841.00001.00001.00001.0000
Note: Significant at 5% = **, 10% = *.
Table 19. Pairwise WSR test for F1-Measure of adolescents.
Table 19. Pairwise WSR test for F1-Measure of adolescents.
NBBGCARTKNNC4.5KSSVMRT
NB 0.0211 **0.0211 **0.52940.0211 **0.34881.00000.0211 **
BG1.0000 0.28171.00001.00001.00001.00001.0000
CART1.00001.0000 1.00001.00001.00001.00001.0000
KNN1.00000.0211 **0.0211 ** 0.0360 **0.52941.00000.0211 **
C4.51.00000.29450.0211 **1.0000 1.00001.00000.5294
KS1.00000.0211 **0.0211 **1.00000.0360 ** 1.00000.0211 **
SVM0.0211 **0.0211 **0.0211 **0.0935 *0.0211 **0.0592 * 0.0211 **
RT1.00000.83390.29451.00001.00001.00001.0000
Note: Significant at 5% = **, 10% = *.
Table 20. Pairwise WSR test for AUROC of adolescents.
Table 20. Pairwise WSR test for AUROC of adolescents.
NBBGCARTKNNC4.5KSSVMRT
NB 0.0935 *0.0211 **0.0360 **0.0211 **0.0360 **0.0211 **0.0211 **
BG1.0000 0.0211 **1.00000.0208 **1.00001.00000.0211 **
CART1.00001.0000 1.00001.00001.00001.00001.0000
KNN1.00000.29450.0211 ** 0.0360 **1.00000.82950.0211 **
C4.51.00001.00000.0592 *1.0000 1.00001.00000.6750
KS1.00000.29450.29450.29450.2945 0.29450.2945
SVM1.00000.67500.0211 **1.00000.0211 **1.0000 0.0211 **
RT1.00001.00000.40171.00001.00001.00001.0000
Note: Significant at 5% = **, 10% = *.
Table 21. Pairwise WSR test for accuracy of adults.
Table 21. Pairwise WSR test for accuracy of adults.
NBBGCARTKNNC4.5KSSVMRT
NB 0.0179 **0.0179 **0.0850 *0.0179 **0.28171.00000.0179 **
BG1.0000 1.00001.00001.00001.00001.00001.0000
CART1.00000.0179 ** 1.00001.00001.00001.00000.0179 **
KNN1.00000.0179 **0.0179 ** 0.0767 *1.00001.00000.0179 **
C4.51.00000.0179 **0.0179 **1.0000 1.00001.00000.0179 **
KS1.00000.0179 **0.0179 **0.38910.0313 ** 1.00000.0179 **
SVM0.0313 **0.0179 **0.0179 **0.0179 **0.0179 **0.0179 ** 0.0179 **
RT1.00000.7802 1.00001.00001.00001.0000
Note: Significant at 5% = **, 10% = *.
Table 22. Pairwise WSR test for kappa statistics of adults.
Table 22. Pairwise WSR test for kappa statistics of adults.
NBBGCARTKNNC4.5KSSVMRT
NB 0.0179 **0.0179 **0.0850 *0.0179 **0.28171.00000.0179 **
BG1.0000 1.00001.00001.00001.00001.00001.0000
CART1.00000.0179 ** 1.00001.00001.00001.00000.0179 **
KNN1.00000.0179 **0.0179 ** 0.0767 *1.00001.00000.0179 **
C4.51.00000.0179 **0.0179 **1.0000 1.00001.00000.0179 **
KS1.00000.0179 **0.0179 **0.38910.0313 ** 1.00000.0179 **
SVM0.03130.0179 **0.0179 **0.0179 **0.0179 **0.0179 ** 0.0179 **
RT1.00000.38911.00001.00001.00001.00001.0000
Note: Significant at 5% = **, 10% = *.
Table 23. Pairwise WSR test for F1-Measure of adults.
Table 23. Pairwise WSR test for F1-Measure of adults.
NBBGCARTKNNC4.5KSSVMRT
NB 0.0256 **0.0256 **0.16290.0256 **0.26431.00000.0256 **
BG1.0000 1.00001.00001.00001.00001.00001.0000
CART1.00000.0256 ** 1.00001.00001.00001.00000.0256 **
KNN1.00000.0256 **0.0256 ** 0.09401.00001.00000.0256 **
C4.51.00000.0256 **0.0256 **1.0000 1.00001.00000.0256 **
KS1.00000.0256 **0.0256 **0.78020.0507 * 1.00000.0256 **
SVM0.0507 *0.0256 **0.0256 **0.0256 **0.0256 **0.0126 ** 0.0256 **
RT1.00000.78021.00001.00001.00001.00001.0000
Note: Significant at 5% = **, 10% = *.
Table 24. Pairwise WSR test for AUROC of adults.
Table 24. Pairwise WSR test for AUROC of adults.
NBBGCARTKNNC4.5KSSVMRT
NB 0.0179 **0.0179 **0.0179 **0.0149 **0.0313 **0.0313 **0.0179 **
BG1.0000 0.0179 **1.00001.00001.00001.00000.0179 **
CART1.00001.0000 1.00001.00001.00001.00000.0179 **
KNN1.00000.51830.0179 ** 1.00001.00001.00000.0179 **
C4.51.00000.0100 ***0.0179 **0.2236 1.00000.28170.0179 **
KS1.00000.0149 **0.0100 ***0.0179 **0.0179 ** 0.0313 **0.0179 **
SVM1.00000.19640.0313 **0.66681.00001.0000 0.0179 **
RT1.00001.00001.00001.00001.00001.00001.0000
Note: Significant at 1% = ***, 5% = **.
Table 25. Comparison of proposed model with other previous studies.
Table 25. Comparison of proposed model with other previous studies.
DatasetVersionFeature
Reduction
ReferenceAccuracy
(%)
Kappa.
(%)
F1
(%)
AUROC
(%)
Toddlerv1No[57]----
v1No[58]----
v2No[14]----
v2YesProposed Model97.8294.8797.8099.70
Childv1No[57]----
v1No[58]98.6298.60--
v2No[14]----
v2YesProposed Model99.6199.2199.6099.60
Adolescentv1No[57]----
v1No[58]----
v2No[14]99.91---
v2YesProposed Model95.8791.7495.9099.00
Adultv1No[57]91.74---
v1No[58]99.7399.38--
v2No[14]97.58---
v2YesProposed Model99.8299.5999.9099.80
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Bala, M.; Ali, M.H.; Satu, M.S.; Hasan, K.F.; Moni, M.A. Efficient Machine Learning Models for Early Stage Detection of Autism Spectrum Disorder. Algorithms 2022, 15, 166. https://doi.org/10.3390/a15050166

AMA Style

Bala M, Ali MH, Satu MS, Hasan KF, Moni MA. Efficient Machine Learning Models for Early Stage Detection of Autism Spectrum Disorder. Algorithms. 2022; 15(5):166. https://doi.org/10.3390/a15050166

Chicago/Turabian Style

Bala, Mousumi, Mohammad Hanif Ali, Md. Shahriare Satu, Khondokar Fida Hasan, and Mohammad Ali Moni. 2022. "Efficient Machine Learning Models for Early Stage Detection of Autism Spectrum Disorder" Algorithms 15, no. 5: 166. https://doi.org/10.3390/a15050166

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop