Previous Article in Journal
Comparative Analysis of Real-Time Detection Models for Intelligent Monitoring of Cattle Condition and Behavior
Previous Article in Special Issue
ECG Signal Analysis and Abnormality Detection Application
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ensemble Modeling of Multiple Physical Indicators to Dynamically Phenotype Autism Spectrum Disorder

1
Department of Biomedical Data Science and Department of Pediatrics, Stanford University, Stanford, CA 94305, USA
2
Division of Clinical Informatics & Digital Transformation, Department of Medicine, University of California, San Francisco, CA 94143, USA
*
Authors to whom correspondence should be addressed.
Algorithms 2025, 18(12), 764; https://doi.org/10.3390/a18120764 (registering DOI)
Submission received: 3 October 2025 / Revised: 19 November 2025 / Accepted: 21 November 2025 / Published: 2 December 2025
(This article belongs to the Special Issue Algorithms for Computer Aided Diagnosis: 2nd Edition)

Abstract

Early detection of Autism Spectrum Disorder (ASD), a neurodevelopmental condition characterized by social communication challenges, is essential for timely intervention. Naturalistic home videos collected via mobile applications offer scalable opportunities for digital diagnostics. We leveraged GuessWhat, a mobile game designed to engage parents and children, which has generated over 3000 structured videos from 382 children. From this collection, we curated a final analytic sample of 688 feature-rich videos centered on a single dyad, enabling more consistent modeling. We developed a two-step pipeline: (1) filtering to isolate high-quality videos, and (2) feature engineering to extract interpretable behavioral signals. Unimodal LSTM-based models trained on eye gaze, head position, and facial expression achieved test AUCs of 86% (95% CI: 0.79–0.92), 78% (95% CI: 0.69–0.86), and 67% (95% CI: 0.55–0.78), respectively. Late-stage fusion of unimodal outputs significantly improved predictive performance, yielding a test AUC of 90% (95% CI: 0.84–0.95). Our findings demonstrate the complementary value of distinct behavioral channels and support the feasibility of using mobile-captured videos for detecting clinically relevant signals. While further work is needed to improve generalizability and inclusivity, this study highlights the promise of real-time, scalable autism phenotyping for early interventions.

1. Introduction

Autism Spectrum Disorder (ASD, or autism) is a complex neurodevelopmental condition that affects approximately 1 in 36 children across diverse ethnic, racial, and socioeconomic backgrounds, highlighting its widespread impact and the need for comprehensive strategies to address it [1]. Children with autism face significant challenges in communication, social interactions, repetitive behaviors, and restricted interests, often leading to profound difficulties in everyday functioning and development [2,3]. The impact of autism extends beyond individual health, imposing a substantial economic burden, with an estimated lifetime social cost of approximately USD 3.6 million per affected individual [4].
Early interventions are crucial, as the developmental gap between children with autism and neurotypical (NT) children widens over time [5]. Early diagnosis can lead to improved health outcomes; however, despite the potential for reliable diagnosis as early as 16–24 months, the average age of diagnosis is 4.9 years [6,7]. Diagnostic processes typically involve long waitlists and assessments, resulting in an average delay of two years [8]. Current diagnoses rely on in-person behavioral assessments, which are costly and subjective, lacking definitive medical tests or biomarkers [9]. This subjective process introduces variability and the potential for misdiagnosis influenced by clinician experience, training, and social biases [1,10,11,12,13].
Digital phenotyping, leveraging naturalistic home videos and affordable smartphones, offers a promising approach for faster, more objective diagnosis of autism and other developmental conditions. GuessWhat [14] is a gamified mobile application designed to facilitate the collection of videos enriched with behaviors relevant to autism diagnosis [15] while providing a set of tools that can lead to improved social communication skills for children with autism [16]. Here, we explore the development of multimodal deep learning models that integrate multiple behavioral modalities (eye, face, and head data) to facilitate comprehensive, automated autism diagnosis.

2. Related Works

2.1. Digital Phenotyping of Autism

Computer vision tools have shown promise in identifying multiple phenotypes—such as emotion, eye movement, and posture tracking—in children through video analysis [17,18,19]. However, these tools often lack structured data on children with autism, and predictive models using individual phenotypes have rarely been assessed for fairness or integrated into ensemble models.
Modalities such as eye gaze [20,21,22,23,24,25], head movement [17,26,27], facial expression [24,28,29,30,31], finger movement [32,33,34], and vocalization [35,36] have proven to be valuable for automatically identifying autism. Nonetheless, challenges related to data variability, generalizability across diverse populations, and integration with other diagnostic measures remain.
Building upon these foundational works, GuessWhat provides structured data to address the lack of well-organized datasets for children with autism. The design of GuessWhat was informed by co-design with families of children with autism and validated in prior feasibility and clinical studies to ensure developmental appropriateness, user engagement, and consistent capture of behavioral features across participants [24,35,37,38]. We utilize deep time-series models, such as long short-term memory (LSTM) and gated recurrent units (GRUs), to analyze eye gaze, face, and head features in a multimodal fashion to predict autism. Our filtering and feature engineering pipelines address key challenges such as data variability and generalizability, improving the quality of the extracted features. We also evaluate fairness with respect to age and gender to ensure that our models generalize well across diverse populations. Ultimately, by integrating multiple phenotypes into a single model, we overcome the limitations of single-modality models, thereby improving predictive accuracy and paving the way for a more comprehensive understanding of autism diagnosis.

2.2. Data Fusion for ASD Classification

Recent studies have focused on the potential of data fusion for improving autism classification performance. Perochon et al. [39] recently evaluated a digital autism screening application—Sensetoknow—administered during pediatric well-child visits for children aged 17–36 months. Using XGBoost, they integrated multiple digital phenotypes, achieving an AUROC of 90%, sensitivity of 87.8%, specificity of 80.8%, negative predictive value of 97.8%, and positive predictive value of 40.6%. This app was later validated in another study, where it was remotely administered by caregivers at home using smartphones or tablets, demonstrating similarly high accuracy for autism detection [40]. These findings underscore the potential of combining data sources using tree-based machine learning methods to improve diagnostic outcomes.
Building on the fusion of multiple data types, Han et al. [41] integrated neurophysiological and behavioral data by developing a multimodal diagnostic framework for autism that combines electroencephalogram and eye-tracking data. Their two-step feature learning and fusion model, based on a stacked denoising autoencoder, outperformed unimodal approaches, demonstrating the advantages of combining data types. Wadhera [42] further expanded this approach by incorporating laptop-performance tool data and applying a kernel-based discriminant correlation analysis, which reduced feature redundancy and achieved 92% accuracy in autism classification. These findings highlight the effectiveness of multimodal systems in identifying influential features and clustering individuals into more homogeneous subgroups for the diagnosis of autism.
Our research uses mobile videos captured during at-home gameplay for scalable and dynamic autism data collection [43,44]. By continuously monitoring authentic behavioral data, we aim to facilitate earlier and more accurate diagnoses. We differentiate from these prior works by exploring the utility of multimodal deep learning and feature engineering. By analyzing such video data in real-world settings, we aim to bridge the gap between clinical practice and everyday environments, offering a practical solution for widespread autism screening and diagnosis.

3. Materials and Methods

3.1. Dataset

3.1.1. Data Collection and Type

Data were collected through the mobile game GuessWhat [45], designed to facilitate structured, charades-style interactions between parent and child. As illustrated in Figure 1, the parent holds the phone to their forehead, and the phone’s front camera captures the child acting out prompts (e.g., “happy,” “sad,” “elephant,” “baseball”), with parents opting to share videos after each session, stored securely on HIPAA-compliant servers. A Stanford-approved Video/Photo Release and informed consent were obtained from legal guardians for the use of the images presented in the figures in this manuscript. The study protocol and app were reviewed and approved by the Stanford Institutional Review Board (IRB) in accordance with Stanford, California, United States, and international research guidelines, including the Declaration of Helsinki, prior to the initiation of any study procedures. Informed consent was obtained from parents/legal guardians through the app before study participation. Recruitment, focusing primarily on children with autism, used crowd-sourcing, email, online ads, and the presence of the app store, resulting in a higher percentage of children with ASD than children without ASD (no-ASD). To participate, families were required to be able to read and understand the available translations of the mobile game, have access to a compatible iOS or Android device with internet connectivity, and be an adult parent or guardian of a child aged 1.5 to 17 years old. To safeguard against the potential for self-reporting bias, we required the caregiver to confirm that their child’s autism diagnosis came from a formal medical assessment [46] and asked parents to provide answers to a set of clinical instruments that included the Vineland Adaptive Behavior Scales-Parent Report [47], Social Responsiveness Scale [48], Mobile Autism Risk Assessment [49], Social Communication Questionnaire [50], and Autism Symptom Dimensions Questionnaire [51] to establish a preliminary confirmation of autism diagnosis (or lack thereof); subsequently, each video provided was independently examined in a blinded fashion by a clinician with expertise in autism diagnosis to provide a clinical diagnostic impression. We required the parent-reported diagnosis to match the clinician’s observational diagnosis, to include the video for analysis. This rigorous two-point consensus was used to ensure the ground truth while minimizing the parental burden of participation. To date, GuessWhat has collected over 3000 videos from 382 children, aged 1.5 to 17, with and without autism. We refer to this collection as the GuessWhat Dataset. For conciseness, we report detailed demographic breakdowns by split in Section 3.4.1 Data Splitting and Task Definition. While the consent process allows for participation beginning at 16 months, few videos were collected in the 16–24-month range, and those did not meet the inclusion criteria. Future efforts should aim to capture more data in this early window to support even earlier detection and intervention. That said, given that the average age of autism diagnosis in the U.S. is around 5 years, focusing on children aged 2 to 5 remains clinically relevant and impactful.

3.1.2. Filtering Pipeline

The instability of home videos, often caused by handheld cameras or movement, can create noise and data drift that reduce the models’ performance. Furthermore, some videos may be feature-poor, or the child of interest may be too far away to extract the desired features, etc. Rigorous filtering and feature engineering are necessary to address these limitations and construct a minimally viable training set for our task. Our filtering pipeline can be summarized in three steps (Figure 2).
The criteria for including videos from the overall GuessWhat Dataset in our final filtered dataset were as follows: (1) they must be of high quality, suitable for feature extraction; and (2) each video must focus on the child of interest. To ensure these criteria, we filtered our videos using the Amazon Rekognition Video API for face detection [52], which provides estimates for sharpness, brightness, head pose, facial landmarks, and face size for each video. For the first criterion, we selected videos that guarantee clarity through sharpness and brightness metrics, feature a sufficiently large face for detecting eye-gazing features, have the eyes open for more than 70% of the video’s duration, and maintain the face predominantly facing the camera. To meet the second criterion, we selected videos where the face is proportionally large enough for reliable feature extraction and where the presence of multiple faces is minimal, ensuring that the focus remains on a single child. While the game’s format ensures that the parent playing will not enter the frame during recording sessions, there are instances where the child’s siblings or friends join them onscreen. Thus, we currently adopt a strict approach, discarding videos with a high proportion of frames containing multiple faces, so as to avoid introducing uncertainty about which child the diagnosis refers to. However, a future step will involve incorporating child recognition techniques, allowing us to identify and focus on the child of interest, preventing the loss of potentially useful data. All thresholds are summarized in Appendix A (Table A1). This filtered dataset, referred to as ‘Dataset A’, includes 2123 videos featuring 288 children.

3.1.3. Bias and Imbalances

Our dataset contains a significant imbalance with respect to the ASD and no-ASD classes, with an ASD-to-no-ASD ratio of 17:1 (Figure 2). Furthermore, some users in the ASD class are more represented than others due to more gameplay (Figure 3).
With only 116 videos for no-ASD children, we manually inspected each one to filter out low-quality or non-compliant videos. During this review, 7 videos were excluded due to issues such as the parent playing, the child holding the phone, or siblings appearing sequentially in the same video.
To reduce bias towards a particular child and mitigate the imbalance between ASD and no-ASD, we under-sampled the ASD class by retaining at most two videos per child, prioritizing those with the highest quality (based on sharpness and brightness). Retaining all superusers—who are disproportionately represented in the ASD group—biases the learning process toward a small number of specific individuals, leading to overfitting and near-random generalization performance on unseen children. For no-ASD children (the minority class), we retained all available videos. This balanced dataset, nicknamed ‘Dataset B,’ includes 700 videos (Figure 2). We also tested oversampling no-ASD and adding loss weights to manage class imbalance; however, these strategies did not yield satisfactory model performance, as illustrated in Appendix A (Table A5).

3.2. Feature Extraction

GuessWhat videos have an average frame rate of 28 frames per second (fps) and an average length of 90 seconds (s). We downsampled the Dataset B videos to 10 fps and utilized AWS Rekognition [52] to extract frame-level eye gaze, head, and face features (Figure 4). Feature extraction for a 90 s video sampled at 10 fps costs approximately USD 0.90 per sample. Across the 628 videos processed in this study, this represented a substantial but manageable computational expense. Although this was the most reliable option available to us at the time, newer and more affordable alternatives now offer comparable accuracy and are expected to operate in real time, improving the feasibility of large-scale or mobile deployment. After extraction, we discarded any video with fewer than 15 s of extracted features and obtained a high-quality ‘Dataset C’ comprising 688 videos (Figure 2).

3.3. Data Preprocessing

Mobile data is inherently noisy, requiring robust feature engineering to extract informative sequences for learning. Our feature engineering pipeline, depicted in Figure 5, addresses several challenges.
At the start and end of videos, our features of interest are often missing due to camera stabilization, game initiation, or game completion with the child. To mitigate this, we truncate frames where no face is detected. Within the videos, we identify periods of missing data. Missing eye-gaze data occurs when the child’s eyes are closed, the child is not facing the camera, or the camera angle is misaligned. These instances can be informative, reflecting the child’s interaction challenges or meaningful movements. Conversely, periods without face detection or with poorly centered cameras provide no useful information. To balance these factors, we allow for brief periods of missing face data (up to 2 s) but discard segments with prolonged absence, as they provide little modeling signal. We exclude these uninformative windows and concatenate adjacent informative segments to enhance feature continuity. Our procedures for truncation, window creation, and segment concatenation are detailed in Appendix A (Algorithms A1, A2, and A3, respectively). To reduce the input length, we averaged features every two frames, for an effective frame rate of 5 fps (Figure 5). This downsampling strategy aligns with prior work, which shows that a frame rate of 5 fps is sufficient to preserve key behavioral dynamics in mobile video-based developmental assessments [15,53,54]. We then normalized the features using min–max normalization and replaced missing frames with vectors of −1, allowing us to explicitly encode missingness as part of the temporal data structure.

3.4. Model Training

3.4.1. Data Splitting and Task Definition

The resulting dataset was split at the child level (in opposition to the video level) into training, validation, and test sets to prevent data leakage, especially for children with multiple videos. To ensure fairness and representativeness, we stratified the splits by age group (2–4, 5–8, 9–12) and gender (male, female), as summarized in Table 1. Corresponding video-level statistics are provided in Table 2. We chose this stratified child-level split over k-fold cross-validation because enforcing both child-level independence and balanced demographic distributions across multiple folds is infeasible given the limited number of children in certain age–gender groups. The input of our models consists of a sequence of k frames, where each frame contains a feature vector of size d (d being the dimension of features for a given modality: d { 2 , 7 , 60 } for the eye-gazing, head, and face modalities, respectively). Hence, our input is a multivariate time series denoted as [ X 1 , , X k ] , where X i = [ x 1 i , , x d i ] , and our output is a binary label Y, where Y { 0 , 1 } (0: no-ASD, 1: ASD).

3.4.2. Unimodal Deep Time-Series Models

We trained LSTM and GRU models in PyTorch v1.9.0 (CUDA 11.1) [55] for binary classification, using three input modalities: eye gaze, head pose, and facial expressions. These models accommodate variable-length input sequences, making them well suited to our data. Each model consists of n recurrent layers ( n [ 4 , 8 ] ), each with hidden size h { 16 , 32 , 64 } . These parameters were tuned as part of the hyperparameter search detailed in Appendix A (Table A2). To address class imbalance, we experimented with focal loss [56], but we observed similar performance to that of standard binary cross-entropy. For training stability, we applied early stopping based on validation loss with a patience of 3 epochs and a minimum delta of 0.001.

3.4.3. Fusion and Ensemble Techniques

To enhance predictive performance, we explored multimodal learning through three fusion strategies. In late fusion, we experimented with both (1) averaging the predicted probabilities of unimodal models and (2) concatenating their logit outputs and passing them through a linear layer for final prediction. In intermediate fusion, we (3) concatenated the final hidden states of pretrained unimodal models and fed them into a multi-layer perceptron. All fusion models were trained using binary cross-entropy loss with class weights. We evaluated each fusion strategy across all bimodal combinations and the full trimodal setting. A schematic overview of these approaches is provided in (Appendix A, Figure A3). Finally, for completeness, we also report the performance of a direct early-fusion baseline based on feature concatenation, although this approach is inherently less well suited to our imbalanced feature dimensionalities and the limited size of our dataset.

3.4.4. Hyperparameter Optimization

For hyperparameter optimization, we utilized the Optuna Framework [57], which employs a Bayesian optimization approach using a Tree-structured Parzen Estimator (TPE) to fine-tune our models across various tasks. We selected the macro-averaged (MA) F1 score on the validation set as the optimization objective, because it jointly balances precision and recall while equally weighting all classes, unlike the weighted average, which is dominated by the majority ASD class. The hyperparameter search space is detailed in Appendix A (Table A2). For each model configuration, we ran 40 trials to identify the optimal parameters. The best-performing hyperparameter settings for each model are also reported in Appendix A (Table A4, Table A6 and Table A7).

4. Results

4.1. Effects of Feature Engineering

Table 3 shows the impact of feature engineering on the performance of the unimodal models by comparing their AUC and macro-averaged (MA) F1 scores before and after applying the pipeline described in Figure 5. The eye model shows the most significant improvement. The head model also shows improvement, although its F1 score (MA) decreases slightly, indicating some trade-offs. The face model exhibits minimal improvement, with its AUC slightly increasing and the F1 score (MA) decreasing, showing limited effectiveness of feature engineering.

4.2. Performance Comparison of Our Models

We considered seven distinct input configurations: three unimodal models (eye gaze, head pose, facial expressions), three bimodal combinations, and one trimodal fusion. For each unique input setting and fusion approach, we selected the best model configuration based on the validation macro-averaged F1 score from 40 hyperparameter optimization trials, as described in the Section 3.4.4, Hyperparameter Optimization. The best-performing hyperparameter settings for each model are reported in Appendix A (Table A4). The eye and head models achieved strong predictive power, with test AUCs of 0.86 and 0.78, respectively (Figure 6). The face model had moderate predictive power, with a test AUC of 0.67. Table 4 summarizes the performance metrics across all unimodal models. Confidence intervals were estimated via bootstrapping: we resampled the test set 1000 times and computed the performance metrics on each sample. The 2.5th and 97.5th percentiles define the bounds of the 95% confidence intervals. Confusion matrices for all models are provided in Appendix A (Figure A1).
The eye model outperforms the face and head models, with narrower confidence intervals indicating more consistent performance. Its high F1 score (MA) reflects balanced precision and recall, making it the strongest unimodal model. This is further supported by the confusion matrix (Appendix A, Figure A1), which shows that the eye model achieves high precision for no-ASD (22/24 correctly classified) and recall for ASD (86/112). Integrating all three modalities further improves the predictive power: the late fusion models—particularly the averaging method (AUC: 0.90) and the linear combination method (AUC: 0.84)—achieve strong results (Figure 7). These models also improve no-ASD classification, correctly identifying 23 out of 24 no-ASD instances while maintaining high ASD recall. In contrast, the intermediate fusion and early fusion models perform substantially worse, with a test AUC of 0.55 and 0.50, respectively.
Two-by-two feature combinations also show strong results. For late fusion by averaging, the eye+head and eye+face combinations each achieve an AUC score of 0.87, while the face+head model achieves an AUC score of 0.78 (Figure 8a). With a linear layer, the eye+face combination performs best (AUC: 0.90), followed by eye+head (AUC: 0.82) and face+head (AUC: 0.67) (Figure 8b). The eye+face pairing excels in both methods, leveraging eye gaze and facial features for robust predictions. Table 5 summarizes the best fusion models’ performance, with late fusion models consistently outperforming the intermediate fusion model. The narrower confidence intervals for the late fusion models suggest greater consistency and lower variability.

4.3. Fairness Evaluation

Among the best-performing configurations for each input setting, we selected the three best unimodal models and the best three multimodal models based on validation F1 score (MA) to include in our downstream analyses. We evaluated these models for age and gender sensitivity, excluding geographic locations due to insufficient data. We used demographic parity difference (DPD) and equalized odds difference (EOD), where a lower DPD indicates more evenly distributed positive outcomes across groups, while a lower EOD indicates more evenly distributed error rates. To quantify uncertainty in these fairness metrics, we computed 95% confidence intervals using bootstrapping, as described in Section 4.2. It is essential to note that some demographic subgroups—particularly the 2–4 age group and the female subgroup—are represented by relatively few samples, resulting in wider confidence intervals and less stable fairness estimates.
The eye model demonstrates significant variability across age groups (Table 6), performing moderately for the youngest category (2–4 years), with an AUC of 0.69 and a recall of 0.64, suggesting limited reliability for early-age autism identification. Performance improves considerably for older age categories, particularly 9–12, suggesting better effectiveness for older children. The face model exhibits consistent performance across the younger age groups but demonstrates a notable decline in fairness for the 9–12 age group, with an EOD of 0.63. The head model performs well for the age groups 2–4 and 9–12, with an AUC of approximately 0.80, but moderately for the age group 5–8, with an AUC of 0.71 and significant fairness challenges (EOD: 0.63). The late fusion models across the three modalities combine strengths from each unimodal model, leading to improved overall performance and fairness. The late fusion (average) model shows enhanced accuracy and more equitable outcomes across all age groups (DPD: 0.26, EOD: 0.20), reflecting its superior balance between performance and fairness. While the performance improvements are clear and robust, the fairness metrics for late fusion models still exhibit relatively large bootstrap confidence intervals. This indicates that the observed improvements in fairness may not be consistent, and further investigation is needed to establish statistical significance.
Regarding gender, the eye model performs similarly across gender groups (Table 7), showing balanced fairness metrics (DPD: 0.12, EOD: 0.02). In contrast, the face and head models exhibit more pronounced gender disparities, with lower recall for females and larger fairness differences (DPD: 0.29, EOD: 0.42). This indicates that these models are less effective at identifying autism in females, potentially leading to more missed cases compared to males. The late fusion models show improvements in gender performance and good fairness metrics (DPD: 0.20, EOD: 0.08), with narrower confidence intervals. Given the importance of early diagnosis, the late fusion (average) model is the most balanced and fair option for both age and gender groups. Fairness mitigation techniques applied to the head and face models led to marginal improvements, so we focused on the fairer and more effective late fusion methods.

4.4. Net Benefit Analysis

Net benefit analysis (Figure 9) evaluates the clinical utility of each model across a range of decision thresholds. A decision threshold p t represents the minimum predicted probability at which a clinician would choose to act (e.g., refer a child for further evaluation). Net benefit quantifies how many true positives a model identifies minus the weighted harm of false positives at a given threshold, using the standard decision-curve formula:
Net Benefit = T P n F P n × p t 1 p t
where p t is the threshold probability and n is the total number of observations.
A model provides clinical value at a given threshold if its net benefit exceeds both trivial strategies, “treat all” and “treat none.” The “treat-all” strategy classifies every child as positive, maximizing sensitivity but producing many false positives. The “treat-none” strategy classifies no child as positive, maximizing specificity but missing all true cases. Net benefit evaluates whether a model achieves a more favorable balance between true positives and false positives than these extremes. Models with higher net benefit therefore support a more reliable and practical diagnostic process, guiding early and timely intervention while efficiently allocating resources.
The late fusion (average) model consistently provides higher net benefits across thresholds up to 0.5, while late fusion (eye+face) performs best around a threshold of 0.5, and the eye-only model surpasses others at thresholds above 0.6. The face and head models show lower net benefits. The dataset’s skewed autism prevalence affects the generalizability of these findings. Given the potential consequences of false positives and false negatives, ethical considerations are essential. As this work is upstream of clinical deployment, any future AI use would aim to support—rather than replace—clinical evaluation and would require caution to mitigate risks such as over-reliance on AI and privacy concerns beyond standard PHI.

5. Discussion

This study demonstrates that designing an engineering pipeline and integrating multiple physical indicators into diagnostic models enhances diagnostic precision by generating more informative time-series data and capturing a more comprehensive range of autism-associated signals. This multimodal approach results in models that appear fair, with our observations indicating an equitable and reliable process across diverse demographic cohorts. Further investigation is needed to understand why these models exhibit greater fairness compared to unimodal models. Successfully implementing these predictive models could significantly impact healthcare by optimizing resource allocation, prioritizing children needing further evaluation, and enhancing the efficiency of autism-related services. Additionally, using multimodal time-series data makes the diagnostic process more dynamic, allowing models to adapt to autism’s evolving nature, and thereby supporting effective follow-up care and treatment adjustments. While the present study relied on AWS Rekognition for large-scale feature extraction, we have also implemented on-device face and landmark detection using Google ML Kit [58], which runs in real time on standard smartphones. Given the lightweight size of our behavioral model, these results indicate that end-to-end inference is feasible on mobile devices. The success of these models in autism diagnosis suggests broader potential in other healthcare domains, such as ADHD, highlighting the utility of digital phenotyping and multimodal fusion for improving diagnostics.
Our current models face several limitations. First, camera movement can lead to shifts in the visual field (e.g., the child temporarily leaving the frame). Incorporating device-level accelerometer data and providing real-time user feedback on child distance and identification could mitigate these data drifts, ensuring more reliable predictions. Second, developing an automated preprocessing pipeline is crucial for handling challenges such as detecting when a child is too close or too far from the camera, re-identifying children across frames to handle multiple faces, and maintaining accurate tracking throughout videos. While our filtering and quality-control pipelines ensured high-quality, interpretable inputs, they reduced the dataset to a final sample of 688 videos. This trade-off enhances model robustness but limits statistical power and generalizability, particularly for underrepresented demographic subgroups. A related trade-off exists in our exclusion of “no-face” segments: while prolonged absence of detectable face data often reflects poor video quality and minimal behavioral signal, brief moments when the child turns away or leaves the frame may carry meaningful cues related to social disengagement. Currently, we discard segments with extended face absence, acknowledging the potential loss of useful data.
We emphasize that our results should be interpreted as a proof of concept, and future work will aim to address the aforementioned limitations and expand this study in several directions. Future data collection will explicitly aim for balance and diversity by recruiting participants across multiple community sites, prioritizing underrepresented demographic groups, and incorporating no-ASD children from varied age ranges. The accessibility and scalability of our mobile game provide a unique opportunity to recruit families across diverse socioeconomic and geographic contexts, facilitating inclusive and representative data collection. Additionally, demographic metadata—including age, gender, and skin tone—will be collected to enable fairness auditing and bias correction during model development. In parallel, we plan to integrate the speech modality to capture prosodic and vocal interaction cues that complement facial and head-motion dynamics, enriching the multimodal representation of social behavior. As data collection expands, an age-centric design will enable us to tailor game content and features to different developmental stages, enhancing both engagement and the informativeness of the extracted features. Moreover, prioritizing time-series pretraining will strengthen model robustness. In parallel, developing methods to identify which parts of the sequence contribute most to the model’s decisions will enhance transparency. Together, these steps will make the models more reliable and clinically meaningful. Ultimately, conducting a prospective clinical validation trial will allow for evaluating the reliability and feasibility of the model in real-world diagnostic workflows.

Author Contributions

Conceptualization, M.A.H., A.K., S.S., O.C.M., P.W., and D.P.W.; Data Curation, M.A.H., A.K., S.S., P.A., and M.H.; Formal Analysis, M.A.H., A.K., S.S., P.W., and D.P.W.; Funding Acquisition, K.D. and D.P.W.; Investigation, A.K. and D.P.W.; Methodology, M.A.H., S.S., O.C.M., P.W., and D.P.W.; Project Administration, D.P.W.; Resources, D.P.W.; Supervision, A.K., P.W., and D.P.W.; Validation, A.K. and D.P.W.; Visualization, M.A.H.; Writing—Original Draft, M.A.H. and D.P.W.; Writing—Review and Editing, M.A.H., A.K., S.S., K.D., O.C.M., P.A., M.H., and D.P.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Institutes of Health (R01LM014342, R01LM013364 (Wall), and DP2EB035858 (Washington)) and the Stanford Center for Digital Health. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.

Institutional Review Board Statement

This study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board of Stanford University (protocol code 39562; approved: 21 March 2023).

Informed Consent Statement

Informed consent was obtained from all subjects involved in this study. Written informed consent was obtained from the individuals depicted in the figures in this manuscript.

Data Availability Statement

The datasets analyzed during the current study are not publicly available due to the inclusion of Protected Health Information (PHI). Researchers who wish to gain access may email the corresponding author to request access under a Data Use Agreement with Stanford University.

Acknowledgments

During the preparation of this manuscript, the authors used ChatGPT-4o to assist with text editing, improving clarity, as well as the formatting of figures and tables. All outputs were reviewed and revised by the authors, who take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
ASDAutism Spectrum Disorder
MAMacro-averaged
WAWeighted-averaged

Appendix A

Table A1. Filtering thresholds based on the visual features extracted using AWS for each video. Through manual inspection of a diverse subset of samples, we identified minimum sharpness and brightness levels below which even human reviewers could not reliably extract gaze, head pose, or facial features. These criteria constitute the conditions for a video to be considered.
Table A1. Filtering thresholds based on the visual features extracted using AWS for each video. Through manual inspection of a diverse subset of samples, we identified minimum sharpness and brightness levels below which even human reviewers could not reliably extract gaze, head pose, or facial features. These criteria constitute the conditions for a video to be considered.
Filtering CriteriaVariablesThresholds
QualitySharpness, brightnessSharpness (0–100) > 4;
Brightness (0–100) > 20
Face DetectionNo face proportion,
Multi-face proportion,
Face size
No face detection (0–1) < 0.6;
Multi-face detection (0–1) < 0.3;
Face size (0–100) > 0.01
Eye VisibilityHead pose (pitch, roll, yaw),
Eyes’ confidence
Pitch ( 180 –180) < 45;
Roll ( 180 –180) < 45;
Yaw ( 180 –180) < 45;
Eyes’ confidence (0–100) > 75
Table A2. Hyperparameter search space for unimodal models.
Table A2. Hyperparameter search space for unimodal models.
HyperparameterRange Searched
ModelLSTM, GRU, CNN+LSTM, CNN+GRU
Hidden Size (of LSTM/GRU){16, 32, 64}
Batch Size{32, 48, 64, 100}
Number of Layers[4, 8]
Dropout Probability[0.1, 0.3]
Learning Rate[ 1 × 10 4 , 1 × 10 1 ]
Weight Decay[ 1 × 10 5 , 1 × 10 2 ]
OptimizerAdam
Loss FunctionCross-entropy, focal loss
Table A3. Hyperparameter search space for intermediate fusion models.
Table A3. Hyperparameter search space for intermediate fusion models.
HyperparameterRange Searched
Batch Size{16, 32, 64}
Learning Rate[ 1 × 10 4 , 1 × 10 1 ]
First Hidden Size{128, 192, 256}
Second Hidden Size{32, 64, 128}
Third Hidden Size{32, 64}
OptimizerAdam
Loss FunctionCross-entropy
Table A4. Best hyperparameters for the unimodal models.
Table A4. Best hyperparameters for the unimodal models.
HyperparameterEye GazingHead PoseFacial Landmarks
ModelLSTMLSTMLSTM
Hidden Size643248
Batch Size644848
Number of Layers844
Dropout Probability0.270.190.18
Learning Rate0.030.00030.05
Weight Decay 1 × 10 5 4 × 10 5 4 × 10 5
OptimizerAdamAdamAdam
Loss FunctionCross-EntropyCross-EntropyCross-Entropy
Table A5. Effect of upsampling the no-ASD class in the training set on model performance. Upsampling duplicated up to two additional videos per no-ASD child when available, increasing the number of no-ASD training videos from 65 to 97. Each model is the best of 40 trials, selected by validation macro-averaged F1.
Table A5. Effect of upsampling the no-ASD class in the training set on model performance. Upsampling duplicated up to two additional videos per no-ASD child when available, increasing the number of no-ASD training videos from 65 to 97. Each model is the best of 40 trials, selected by validation macro-averaged F1.
ModelAUC ScoreF1 Score (MA)
Eye (No Upsampling)0.860.73
Eye (Upsampling)0.760.62
Head (No Upsampling)0.780.63
Head (Upsampling)0.600.56
Face (No Upsampling)0.670.63
Face (Upsampling)0.760.61
MA = macro-average. Values in bold denote the best result per metric (before vs. after upsampling).
Table A6. Best hyperparameters for intermediate fusion.
Table A6. Best hyperparameters for intermediate fusion.
HyperparameterValue
Batch Size16
Learning Rate0.07
First Hidden Size256
Second Hidden Size32
Third Hidden Size64
Number of Epochs15
Table A7. Best hyperparameters for late fusion (linear).
Table A7. Best hyperparameters for late fusion (linear).
HyperparameterValue
Batch Size32
Learning Rate 5 × 10 4
Number of Epochs11
Figure A1. Confusion matrices for three unimodal models and three multimodal models: late fusion (average), late fusion (linear), and late fusion (linear) with eye+face inputs. Rows correspond to true labels and columns to predicted labels. A classification threshold of 0.5 was used.
Figure A1. Confusion matrices for three unimodal models and three multimodal models: late fusion (average), late fusion (linear), and late fusion (linear) with eye+face inputs. Rows correspond to true labels and columns to predicted labels. A classification threshold of 0.5 was used.
Algorithms 18 00764 g0a1
Algorithm A1: Truncate Window
Algorithms 18 00764 i001
Algorithm A2: Create Windows
Algorithms 18 00764 i002
Algorithm A3: Concatenate Windows
Algorithms 18 00764 i003
Figure A2. AUC and fairness metrics across age and gender: (a) AUC scores by age category and model. (b) AUC scores by gender category and model. (c) Demographic parity and equalized odds differences by age category. (d) Demographic parity and equalized odds differences by gender category.
Figure A2. AUC and fairness metrics across age and gender: (a) AUC scores by age category and model. (b) AUC scores by gender category and model. (c) Demographic parity and equalized odds differences by age category. (d) Demographic parity and equalized odds differences by gender category.
Algorithms 18 00764 g0a2aAlgorithms 18 00764 g0a2b
Figure A3. Late and intermediate fusion schemes.
Figure A3. Late and intermediate fusion schemes.
Algorithms 18 00764 g0a3

References

  1. Maenner, M.J.; Warren, Z.; Williams, A.R.; Amoakohene, E.; Bakian, A.V.; Bilder, D.A.; Durkin, M.S.; Fitzgerald, R.T.; Furnier, S.M.; Hughes, M.M.; et al. Prevalence and characteristics of autism spectrum disorder among children aged 8 years—Autism and Developmental Disabilities Monitoring Network, 11 sites, United States, 2020. MMWR Surveill. Summ. 2023, 72, 1–14. [Google Scholar] [CrossRef]
  2. Johnson, C.P.; Myers, S.M.; The Council on Children with Disabilities. Identification and evaluation of children with autism spectrum disorders. Pediatrics 2007, 120, 1183–1215. [Google Scholar] [CrossRef]
  3. American Psychiatric Association. DSM-5 Task Force. In Diagnostic and Statistical Manual of Mental Disorders: DSM-5, 5th ed.; American Psychiatric Association: Washington, DC, USA, 2013; Volume 5. [Google Scholar] [CrossRef]
  4. Cakir, J.; Frye, R.E.; Walker, S.J. The lifetime social cost of autism: 1990–2029. Res. Autism Spectr. Disord. 2020, 72, 101502. [Google Scholar] [CrossRef]
  5. Dawson, G.; Zanolli, K. Early intervention and brain plasticity in autism. In Autism: Neural Basis and Treatment Possibilities: Novartis Foundation Symposium 251; Wiley Online Library: Hoboken, NJ, USA, 2003; Volume 251, pp. 266–280. [Google Scholar] [CrossRef]
  6. Lord, C.; Risi, S.; DiLavore, P.S.; Shulman, C.; Thurm, A.; Pickles, A. Autism from 2 to 9 years of age. Arch. Gen. Psychiatry 2006, 63, 694–701. [Google Scholar] [CrossRef] [PubMed]
  7. Shen, L.; Liu, X.; Zhang, H.; Lin, J.; Feng, C.; Iqbal, J. Biomarkers in autism spectrum disorders: Current progress. Clin. Chim. Acta 2020, 502, 41–54. [Google Scholar] [CrossRef] [PubMed]
  8. Gordon-Lipkin, E.; Foster, J.; Peacock, G. Whittling down the wait time: Exploring models to minimize the delay from initial concern to diagnosis and treatment of autism spectrum disorder. Pediatr. Clin. 2016, 63, 851–859. [Google Scholar]
  9. First, M.B. Diagnostic and statistical manual of mental disorders, and clinical utility. J. Nerv. Ment. Dis. 2013, 201, 727–729. [Google Scholar] [CrossRef]
  10. Mazefsky, C.A.; Oswald, D.P. The discriminative ability and diagnostic utility of the ADOS-G, ADI-R, and GARS for children in a clinical setting. Autism 2006, 10, 533–549. [Google Scholar] [CrossRef]
  11. Harrison, A.J.; Long, K.A.; Tommet, D.C.; Jones, R.N. Examining the role of race, ethnicity, and gender on social and behavioral ratings within the Autism Diagnostic Observation Schedule. J. Autism Dev. Disord. 2017, 47, 2770–2782. [Google Scholar] [CrossRef]
  12. Donohue, M.R.; Childs, A.W.; Richards, M.; Robins, D.L. Race influences parent report of concerns about symptoms of autism spectrum disorder. Autism 2019, 23, 100–111. [Google Scholar] [CrossRef]
  13. Kalb, L.G.; Singh, V.; Hong, J.S.; Holingue, C.; Ludwig, N.N.; Pfeiffer, D.; Reetzke, R.; Gross, A.L.; Landa, R. Analysis of race and sex bias in the autism diagnostic observation schedule (ADOS-2). JAMA Netw. Open 2022, 5, e229498. [Google Scholar] [CrossRef]
  14. Kalantarian, H.; Washington, P.; Schwartz, J.; Daniels, J.; Haber, N.; Wall, D. A gamified mobile system for crowdsourcing video for autism research. In Proceedings of the 2018 IEEE International Conference on Healthcare Informatics (ICHI), New York, NY, USA, 4–7 June 2018; IEEE: New York, NY, USA, 2018; pp. 350–352. [Google Scholar]
  15. Kalantarian, H.; Jedoui, K.; Washington, P.; Tariq, Q.; Dunlap, K.; Schwartz, J.; Wall, D.P. Labeling images with facial emotion and the potential for pediatric healthcare. Artif. Intell. Med. 2019, 98, 77–86. [Google Scholar] [CrossRef] [PubMed]
  16. Penev, Y.; Dunlap, K.; Husic, A.; Hou, C.; Washington, P.; Leblanc, E.; Kline, A.; Kent, J.; Ng-Thow-Hing, A.; Liu, B.; et al. A mobile game platform for improving social communication in children with autism: A feasibility study. Appl. Clin. Inform. 2021, 12, 1030–1040. [Google Scholar] [CrossRef] [PubMed]
  17. Zhao, Z.; Zhu, Z.; Zhang, X.; Tang, H.; Xing, J.; Hu, X.; Lu, J.; Qu, X. Identifying autism with head movement features by implementing machine learning algorithms. J. Autism Dev. Disord. 2022, 52, 3038–3049. [Google Scholar] [CrossRef] [PubMed]
  18. Lakkapragada, A.; Kline, A.; Mutlu, O.C.; Paskov, K.; Chrisman, B.; Stockham, N.; Washington, P.; Wall, D.P. The classification of abnormal hand movement to aid in autism detection: Machine learning study. JMIR Biomed. Eng. 2022, 7, e33771. [Google Scholar] [CrossRef]
  19. Sapiro, G.; Hashemi, J.; Dawson, G. Computer vision and behavioral phenotyping: An autism case study. Curr. Opin. Biomed. Eng. 2019, 9, 14–20. [Google Scholar] [CrossRef]
  20. Dawson, G.; Campbell, K.; Hashemi, J.; Lippmann, S.J.; Smith, V.; Carpenter, K.; Egger, H.; Espinosa, S.; Vermeer, S.; Baker, J.; et al. Atypical postural control can be detected via computer vision analysis in toddlers with autism spectrum disorder. Sci. Rep. 2018, 8, 17008. [Google Scholar] [CrossRef]
  21. Chang, Z.; Di Martino, J.M.; Aiello, R.; Baker, J.; Carpenter, K.; Compton, S.; Davis, N.; Eichner, B.; Espinosa, S.; Flowers, J.; et al. Computational methods to measure patterns of gaze in toddlers with autism spectrum disorder. JAMA Pediatr. 2021, 175, 827–836. [Google Scholar] [CrossRef]
  22. Gülü, M.; Büyükbayraktar, F.; Köroğlu, Ç.; Gündoğdu, R.; Akan, A. Machine learning-based prediction of autism spectrum disorder from eye tracking data: A systematic review. Expert Syst. Appl. 2021, 164, 113938. [Google Scholar]
  23. Liaqat, S.; Wu, C.; Duggirala, P.R.; Cheung, S.c.S.; Chuah, C.N.; Ozonoff, S.; Young, G. Predicting ASD diagnosis in children with synthetic and image-based eye gaze data. Signal Process. Image Commun. 2021, 94, 116198. [Google Scholar] [CrossRef]
  24. Varma, M.; Washington, P.; Chrisman, B.; Kline, A.; Leblanc, E.; Paskov, K.; Stockham, N.; Jung, J.Y.; Sun, M.W.; Wall, D.P. Identification of Social Engagement Indicators Associated with Autism Spectrum Disorder Using a Game-Based Mobile App: Comparative Study of Gaze Fixation and Visual Scanning Methods. J. Med Internet Res. 2022, 24, e31830. [Google Scholar] [CrossRef]
  25. Takahashi, Y.; Hashiya, K. Advanced Variations in Eye-Tracking Methodologies for Autism Diagnostics. J. Autism Dev. Disord. 2019, 49, 1234–1250. [Google Scholar]
  26. Martin, K.B.; Hammal, Z.; Ren, G.; Cohn, J.F.; Cassell, J.; Ogihara, M.; Britton, J.C.; Gutierrez, A.; Messinger, D.S. Objective measurement of head movement differences in children with and without autism spectrum disorder. Mol. Autism 2018, 9, 14. [Google Scholar] [CrossRef] [PubMed]
  27. Singh, P.; Pathak, A.; Ganai, U.J.; Bhushan, B.; Subramanian, V.K. Video Based Computational Coding of Movement Anomalies in ASD Children. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Seattle, WA, USA, 17–18 June 2024; pp. 400–409. [Google Scholar]
  28. Krishnappababu, P.R.; Di Martino, M.; Chang, Z.; Perochon, S.P.; Carpenter, K.L.; Compton, S.; Espinosa, S.; Dawson, G.; Sapiro, G. Exploring complexity of facial dynamics in autism spectrum disorder. IEEE Trans. Affect. Comput. 2021, 14, 919–930. [Google Scholar] [CrossRef] [PubMed]
  29. Jiang, M.; Francis, S.M.; Srishyla, D.; Conelea, C.; Zhao, Q.; Jacob, S. Classifying individuals with ASD through facial emotion recognition and eye-tracking. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; IEEE: New York, NY, USA, 2019; pp. 6063–6068. [Google Scholar]
  30. Banerjee, A.; Mutlu, O.C.; Kline, A.; Surabhi, S.; Washington, P.; Wall, D.P. Training and profiling a pediatric facial expression classifier for children on mobile devices: Machine learning study. JMIR Form. Res. 2023, 7, e39917. [Google Scholar] [CrossRef]
  31. Song, C.; Li, J.; Ouyang, G. Early diagnosis of asd based on facial expression recognition and head pose estimation. In Proceedings of the 2022 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Prague, Czech Republic, 9–12 October 2022; IEEE: New York, NY, USA, 2022; pp. 1248–1253. [Google Scholar]
  32. Simeoli, R.; Milano, N.; Rega, A.; Marocco, D. Using technology to identify children with autism through motor abnormalities. Front. Psychol. 2021, 12, 635696. [Google Scholar] [CrossRef]
  33. Marzi, C.; Narzisi, A.; Milone, A.; Masi, G.; Pirrelli, V. Reading Behaviors through Patterns of Finger-Tracking in Italian Children with Autism Spectrum Disorder. Brain Sci. 2022, 12, 1316. [Google Scholar] [CrossRef]
  34. Vabalas, A.; Gowen, E.; Poliakoff, E.; Casson, A.J. Applying machine learning to kinematic and eye movement features of a movement imitation task to predict autism diagnosis. Sci. Rep. 2020, 10, 8346. [Google Scholar] [CrossRef]
  35. Chi, N.A.; Washington, P.; Kline, A.; Husic, A.; Hou, C.; He, C.; Dunlap, K.; Wall, D.P. Classifying autism from crowdsourced semistructured speech recordings: Machine learning model comparison study. JMIR Pediatr. Parent. 2022, 5, e35406. [Google Scholar] [CrossRef]
  36. Plank, I.S.; Koehler, J.C.; Nelson, A.M.; Koutsouleris, N.; Falter-Wagner, C. Automated extraction of speech and turn-taking parameters in autism allows for diagnostic classification using a multivariable prediction model. Front. Psychiatry 2023, 14, 1257569. [Google Scholar] [CrossRef]
  37. Kalantarian, H.; Jedoui, K.; Dunlap, K.; Schwartz, J.; Washington, P.; Husic, A.; Tariq, Q.; Ning, M.; Kline, A.; Wall, D.P.; et al. The performance of emotion classifiers for children with parent-reported autism: Quantitative feasibility study. JMIR Ment. Health 2020, 7, e13174. [Google Scholar] [CrossRef]
  38. Washington, P.; Kalantarian, H.; Kent, J.; Husic, A.; Kline, A.; Leblanc, E.; Hou, C.; Mutlu, O.C.; Dunlap, K.; Penev, Y.; et al. Improved digital therapy for developmental pediatrics using domain-specific artificial intelligence: Machine learning study. JMIR Pediatr. Parent. 2022, 5, e26760. [Google Scholar] [CrossRef]
  39. Perochon, S.; Di Martino, J.M.; Carpenter, K.L.; Compton, S.; Davis, N.; Eichner, B.; Espinosa, S.; Franz, L.; Krishnappa Babu, P.R.; Sapiro, G.; et al. Early detection of autism using digital behavioral phenotyping. Nat. Med. 2023, 29, 2489–2497. [Google Scholar] [CrossRef] [PubMed]
  40. Krishnappa Babu, P.R.; Di Martino, J.M.; Aiello, R.; Eichner, B.; Espinosa, S.; Green, J.; Howard, J.; Perochon, S.; Spanos, M.; Vermeer, S.; et al. Validation of a Mobile App for Remote Autism Screening in Toddlers. NEJM AI 2024, 1, AIcs2400510. [Google Scholar] [CrossRef] [PubMed]
  41. Han, J.; Jiang, G.; Ouyang, G.; Li, X. A multimodal approach for identifying autism spectrum disorders in children. IEEE Trans. Neural Syst. Rehabil. Eng. 2022, 30, 2003–2011. [Google Scholar] [CrossRef] [PubMed]
  42. Wadhera, T. Multimodal Kernel-based discriminant correlation analysis data-fusion approach: An automated autism spectrum disorder diagnostic system. Phys. Eng. Sci. Med. 2024, 47, 361–369. [Google Scholar] [CrossRef]
  43. Washington, P.; Park, N.; Srivastava, P.; Voss, C.; Kline, A.; Varma, M.; Tariq, Q.; Kalantarian, H.; Schwartz, J.; Patnaik, R.; et al. Data-driven diagnostics and the potential of mobile artificial intelligence for digital therapeutic phenotyping in computational psychiatry. Biol. Psychiatry Cogn. Neurosci. Neuroimaging 2020, 5, 759–769. [Google Scholar] [CrossRef]
  44. Washington, P.; Wall, D.P. A review of and roadmap for data science and machine learning for the neuropsychiatric phenotype of autism. Annu. Rev. Biomed. Data Sci. 2023, 6, 211–228. [Google Scholar] [CrossRef]
  45. Kalantarian, H.; Washington, P.; Schwartz, J.; Daniels, J.; Haber, N.; Wall, D.P. Guess What? Towards Understanding Autism from Structured Video Using Facial Affect. J. Healthc. Inform. Res. 2019, 3, 43–66. [Google Scholar] [CrossRef]
  46. Fombonne, E.; Coppola, L.; Mastel, S.; O’Roak, B.J. Validation of autism diagnosis and clinical data in the SPARK cohort. J. Autism Dev. Disord. 2022, 52, 3383–3398. [Google Scholar] [CrossRef]
  47. Sparrow, S.S. Vineland adaptive behavior scales. In Encyclopedia of Clinical Neuropsychology; Springer: Berlin/Heidelberg, Germany, 2011; pp. 2618–2621. [Google Scholar]
  48. Constantino, J.N. Social responsiveness scale. In Encyclopedia of Autism Spectrum Disorders; Springer: Berlin/Heidelberg, Germany, 2013; pp. 2919–2929. [Google Scholar]
  49. Duda, M.; Daniels, J.; Wall, D.P. Clinical evaluation of a novel and mobile autism risk assessment. J. Autism Dev. Disord. 2016, 46, 1953–1961. [Google Scholar] [CrossRef]
  50. Rutter, M.; Bailey, A.; Lord, C. The Social Communication Questionnaire; Western Psychological Services: Los Angeles, CA, USA, 2003. [Google Scholar]
  51. Frazier, T.W.; Dimitropoulos, A.; Abbeduto, L.; Armstrong-Brine, M.; Kralovic, S.; Shih, A.; Hardan, A.Y.; Youngstrom, E.A.; Uljarević, M.; Team, V.B.; et al. Psychometric evaluation of the Autism Symptom Dimensions Questionnaire. Dev. Med. Child Neurol. 2025, 67, 758–769. [Google Scholar] [CrossRef]
  52. Amazon Rekognition Developer Guide. 2023. Available online: https://docs.aws.amazon.com/rekognition/latest/APIReference/Welcome.html (accessed on 4 June 2023).
  53. Hou, C.; Kalantarian, H.; Washington, P.; Dunlap, K.; Wall, D.P. Leveraging video data from a digital smartphone autism therapy to train an emotion detection classifier. medRxiv 2021, 2021-07. [Google Scholar] [CrossRef]
  54. Kalantarian, H.; Jedoui, K.; Washington, P.; Wall, D.P. A mobile game for automatic emotion-labeling of images. IEEE Trans. Games 2018, 12, 213–218. [Google Scholar] [CrossRef]
  55. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. Adv. Neural Inf. Process. Syst. 2019, 32. [Google Scholar]
  56. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  57. Akiba, T.; Sano, S.; Yanase, T.; Ohta, T.; Koyama, M. Optuna: A Next-generation Hyperparameter Optimization Framework. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Anchorage, AK, USA, 4–8 August 2019. [Google Scholar]
  58. Google. ML Kit: Machine Learning for Mobile Developers. Available online: https://developers.google.com/ml-kit (accessed on 11 November 2025).
Figure 1. Illustration of the mobile game GuessWhat. GuessWhat is a charades-style game available for any smartphone device. In a typical game session, the parent holds the smartphone to their forehead and tries to guess the emotion mimicked by the child in response to the prompt shown on the phone’s screen. Each game session typically lasts 90 s.
Figure 1. Illustration of the mobile game GuessWhat. GuessWhat is a charades-style game available for any smartphone device. In a typical game session, the parent holds the smartphone to their forehead and tries to guess the emotion mimicked by the child in response to the prompt shown on the phone’s screen. Each game session typically lasts 90 s.
Algorithms 18 00764 g001
Figure 2. Key filtering steps, leading to 3 subsets containing increasingly filtered portions of the GuessWhat dataset.
Figure 2. Key filtering steps, leading to 3 subsets containing increasingly filtered portions of the GuessWhat dataset.
Algorithms 18 00764 g002
Figure 3. Presence of ‘superusers’. Some children dominate the data, with dozens of videos.
Figure 3. Presence of ‘superusers’. Some children dominate the data, with dozens of videos.
Algorithms 18 00764 g003
Figure 4. Feature extraction scheme.
Figure 4. Feature extraction scheme.
Algorithms 18 00764 g004
Figure 5. Key feature engineering steps.
Figure 5. Key feature engineering steps.
Algorithms 18 00764 g005
Figure 6. ROC curves of the unimodal models.
Figure 6. ROC curves of the unimodal models.
Algorithms 18 00764 g006
Figure 7. ROC curves of the eye and fusion models.
Figure 7. ROC curves of the eye and fusion models.
Algorithms 18 00764 g007
Figure 8. ROC curve results for fusion strategies: (a) average combination of two unimodal models, and (b) linear combination of two unimodal models.
Figure 8. ROC curve results for fusion strategies: (a) average combination of two unimodal models, and (b) linear combination of two unimodal models.
Algorithms 18 00764 g008
Figure 9. Net benefit analysis. Each line on the graph shows a model’s net benefit (y-axis) against threshold values (x-axis). The yellow line (“All”) represents treating all as autism, while the grey line (“None”) represents treating none, serving as baselines.
Figure 9. Net benefit analysis. Each line on the graph shows a model’s net benefit (y-axis) against threshold values (x-axis). The yellow line (“All”) represents treating all as autism, while the grey line (“None”) represents treating none, serving as baselines.
Algorithms 18 00764 g009
Table 1. Number of children by demographic category and diagnostic label (ASD vs. no-ASD) in each split.
Table 1. Number of children by demographic category and diagnostic label (ASD vs. no-ASD) in each split.
DemographicTrainTestVal
ASDNo-ASDASDNo-ASDASDNo-ASD
Age
2–41965381
5–86710175195
9–12617243212
13–17100010
Gender
Male11411355395
Female3412116103
Location
United States652251240
Outside US1600542
Unknown6721215216
Table 2. Number of videos by demographic category and diagnostic label (ASD vs. no-ASD) in each split.
Table 2. Number of videos by demographic category and diagnostic label (ASD vs. no-ASD) in each split.
DemographicTrainTestVal
ASDNo-ASDASDNo-ASDASDNo-ASD
Age
2–43931146141
5–8164244284212
9–12156105610486
13–17400010
Gender
Male2792589138315
Female84402311224
Location
United States1633591550
Outside US3700886
Unknown1636253154213
Table 3. Effects of feature engineering pipeline on model performance. Each model is the best model tuned out of 40 trials, chosen based on validation macro-averaged (MA) F1 score.
Table 3. Effects of feature engineering pipeline on model performance. Each model is the best model tuned out of 40 trials, chosen based on validation macro-averaged (MA) F1 score.
ModelAUC ScoreF1 Score (MA)
Eye (Raw)0.660.66
Eye (After)0.860.73
Head (Raw)0.660.66
Head (After)0.780.63
Face (Raw)0.630.69
Face (After)0.670.63
Values in bold denote the best result per metric for each model (before vs. after feature engineering).
Table 4. Performance metrics for the best unimodal model on each modality. The default classification threshold was 0.5. MA denotes macro-average; WA denotes weighted average.
Table 4. Performance metrics for the best unimodal model on each modality. The default classification threshold was 0.5. MA denotes macro-average; WA denotes weighted average.
MetricEyeFaceHead
AUC Score0.86 [0.79, 0.92]0.67 [0.55, 0.78]0.78 [0.69, 0.86]
Accuracy0.79 [0.72, 0.86]0.75 [0.68, 0.82]0.75 [0.68, 0.82]
Recall (MA)0.84 [0.76, 0.91]0.65 [0.55, 0.76]0.65 [0.55, 0.76]
Recall (WA)0.79 [0.72, 0.86]0.75 [0.68, 0.82]0.75 [0.68, 0.82]
Precision (MA)0.72 [0.65, 0.79]0.62 [0.53, 0.70]0.62 [0.53, 0.70]
Precision (WA)0.89 [0.84, 0.92]0.79 [0.71, 0.86]0.79 [0.71, 0.86]
F1 score (MA)0.73 [0.65, 0.81]0.63 [0.53, 0.71]0.63 [0.53, 0.71]
F1 score (WA)0.82 [0.75, 0.87]0.77 [0.69, 0.83]0.77 [0.69, 0.83]
MA = macro-average; WA = weighted average. Values in bold denote the best score achieved for each metric across models.
Table 5. Performance metrics for the best fusion models. The 95% confidence intervals were obtained by bootstrapping the test set. The default classification threshold was 0.5. MA = macro-average, WA = weighted average.
Table 5. Performance metrics for the best fusion models. The 95% confidence intervals were obtained by bootstrapping the test set. The default classification threshold was 0.5. MA = macro-average, WA = weighted average.
MetricLate Fusion (Eye, Head, Face) AveragingLate Fusion (Eye, Head, Face) LinearLate Fusion (Eye, Face) Linear
AUC Score0.90 [0.84, 0.95]0.84 [0.77, 0.91]0.90 [0.83, 0.95]
Accuracy0.82 [0.75, 0.89]0.84 [0.78, 0.91]0.89 [0.83, 0.93]
Recall (MA)0.88 [0.81, 0.92]0.89 [0.82, 0.94]0.85 [0.76, 0.93]
Recall (WA)0.82 [0.75, 0.89]0.84 [0.78, 0.91]0.89 [0.84, 0.93]
Precision (MA)0.75 [0.67, 0.82]0.76 [0.69, 0.85]0.80 [0.72, 0.89]
Precision (WA)0.90 [0.87, 0.93]0.91 [0.88, 0.94]0.90 [0.85, 0.94]
F1 score (MA)0.77 [0.69, 0.85]0.79 [0.71, 0.88]0.82 [0.74, 0.90]
F1 score (WA)0.84 [0.78, 0.90]0.86 [0.80, 0.92]0.89 [0.83, 0.93]
MA = macro-average; WA = weighted average. Values in bold denote the best score achieved for each metric across models.
Table 6. Summary of model fairness across age groups with confidence intervals. The sample sizes are as follows: 2–4 years (20), 5–8 years (50), and 9–12 years (66).
Table 6. Summary of model fairness across age groups with confidence intervals. The sample sizes are as follows: 2–4 years (20), 5–8 years (50), and 9–12 years (66).
ModelAgeAccuracyRecallPrecisionAUC ScoreF1 ScoreDemographic Parity Diff.Equalized Odds Diff.
Eye Model2–40.70
 [0.50, 0.90]
0.64
 [0.38, 0.87]
0.90
 [0.67, 1.00]
0.69
 [0.44, 0.91]
0.75
 [0.52, 0.92]
0.21 [0.05, 0.48]0.20 [0.10, 0.60]
5–80.74
 [0.62, 0.86]
0.71
 [0.58, 0.85]
0.97
 [0.89, 1.00]
0.81
 [0.69, 0.91]
0.82
 [0.72, 0.91]
9–120.86
 [0.77, 0.94]
0.84
 [0.74, 0.93]
1.00
 [1.00, 1.00]
0.94
 [0.87, 0.99]
0.91
 [0.85, 0.96]
Face Model2–40.80
 [0.60, 0.95]
0.79
 [0.56, 1.00]
0.92
 [0.73, 1.00]
0.79
 [0.45, 1.00]
0.85
 [0.64, 0.97]
0.23 [0.02, 0.32]0.63 [0.27, 1.00]
5–80.74
 [0.62, 0.86]
0.76
 [0.62, 0.88]
0.91
 [0.81, 1.00]
0.75
 [0.52, 0.94]
0.83
 [0.73, 0.91]
9–120.74
 [0.64, 0.83]
0.84
 [0.74, 0.93]
0.85
 [0.75, 0.94]
0.55
 [0.37, 0.72]
0.85
 [0.77, 0.91]
Head Model2–40.80
 [0.60, 0.95]
0.79
 [0.54, 1.00]
0.92
 [0.75, 1.00]
0.82
 [0.61, 0.98]
0.85
 [0.67, 0.97]
0.23 [0.03, 0.34]0.63 [0.27, 1.00]
5–80.74
 [0.60, 0.86]
0.76
 [0.63, 0.88]
0.91
 [0.81, 1.00]
0.71
 [0.49, 0.89]
0.83
 [0.73, 0.92]
9–120.74
 [0.64, 0.83]
0.84
 [0.74, 0.93]
0.85
 [0.75, 0.94]
0.80
 [0.68, 0.90]
0.85
 [0.76, 0.91]
Late Fusion (Average)2–40.75
 [0.55, 0.90]
0.64
 [0.33, 0.87]
1.00
 [1.00, 1.00]
0.93
 [0.77, 1.00]
0.78
 [0.56, 0.93]
0.26 [0.04, 0.50]0.20 [0.07, 0.53]
5–80.80
 [0.68, 0.90]
0.79
 [0.66, 0.90]
0.97
 [0.90, 1.00]
0.84
 [0.69, 0.95]
0.87
 [0.77, 0.94]
9–120.86
 [0.77, 0.94]
0.84
 [0.74, 0.93]
1.00
 [1.00, 1.00]
0.93
 [0.85, 0.98]
0.91
 [0.85, 0.96]
Late Fusion (Linear)2–40.75
 [0.55, 0.90]
0.64
 [0.38, 0.89]
1.00
 [1.00, 1.00]
0.64
 [0.40, 0.88]
0.78
 [0.55, 0.94]
0.27 [0.04, 0.51]0.21 [0.07, 0.53]
5–80.84
 [0.74, 0.94]
0.83
 [0.71, 0.95]
0.97
 [0.90, 1.00]
0.85
 [0.73, 0.94]
0.90
 [0.82, 0.96]
9–120.88
 [0.79, 0.95]
0.86
 [0.76, 0.95]
1.00
 [1.00, 1.00]
0.90
 [0.81, 0.96]
0.92
 [0.86, 0.97]
Late Fusion
(Linear, Eye+Face)
2–40.95
 [0.85, 1.00]
0.93
 [0.77, 1.00]
1.00
 [1.00, 1.00]
0.93
 [0.79, 1.00]
0.96
 [0.87, 1.00]
0.17 [0.02, 0.22]0.38 [0.14, 0.75]
5–80.84
 [0.74, 0.94]
0.88
 [0.78, 0.98]
0.93
 [0.83, 1.00]
0.85
 [0.71, 0.96]
0.90
 [0.83, 0.96]
9–120.91
 [0.83, 0.97]
0.93
 [0.85, 0.98]
0.96
 [0.91, 1.00]
0.93
 [0.85, 0.98]
0.95
 [0.90, 0.98]
Table 7. Summary of model fairness across gender groups with 95% confidence intervals. Sample sizes: female (34), male (102).
Table 7. Summary of model fairness across gender groups with 95% confidence intervals. Sample sizes: female (34), male (102).
ModelGender GroupAccuracyRecallPrecisionAUC ScoreF1 ScoreDemographic Parity Diff.Equalized Odds Diff.
Eye ModelFemale0.82 [0.68, 0.94]0.78 [0.60, 0.95]0.95 [0.83, 1.00]0.81 [0.65, 0.95]0.86 [0.72, 0.96]0.12 [0.00, 0.22]0.02 [0.01, 0.32]
Male0.78 [0.69, 0.86]0.76 [0.66, 0.84]0.99 [0.95, 1.00]0.87 [0.78, 0.93]0.86 [0.80, 0.91]
Face ModelFemale0.68 [0.53, 0.82]0.65 [0.46, 0.84]0.83 [0.63, 1.00]0.67 [0.44, 0.88]0.73 [0.56, 0.86]0.29 [0.02, 0.41]0.42 [0.16, 0.75]
Male0.77 [0.69, 0.85]0.84 [0.76, 0.91]0.89 [0.83, 0.95]0.63 [0.48, 0.78]0.86 [0.81, 0.92]
Head ModelFemale0.68 [0.53, 0.82]0.65 [0.45, 0.85]0.83 [0.65, 1.00]0.70 [0.52, 0.86]0.73 [0.57, 0.86]0.29 [0.02, 0.41]0.42 [0.14, 0.78]
Male0.77 [0.68, 0.85]0.84 [0.76, 0.91]0.89 [0.82, 0.95]0.76 [0.61, 0.88]0.86 [0.80, 0.91]
Late Fusion (Average)Female0.82 [0.68, 0.94]0.74 [0.55, 0.90]1.00 [1.00, 1.00]0.83 [0.65, 0.95]0.85 [0.71, 0.96]0.21 [0.00, 0.27]0.08 [0.01, 0.30]
Male0.82 [0.74, 0.90]0.81 [0.72, 0.89]0.99 [0.96, 1.00]0.91 [0.83, 0.97]0.88 [0.83, 0.93]
Late Fusion (Linear)Female0.85 [0.74, 0.97]0.78 [0.60, 0.93]1.00 [1.00, 1.00]0.79 [0.62, 0.94]0.88 [0.76, 0.97]0.22 [0.01, 0.25]0.08 [0.01, 0.30]
Male0.84 [0.77, 0.91]0.83 [0.74, 0.90]0.99 [0.96, 1.00]0.87 [0.76, 0.91]0.90 [0.85, 0.94]
Late Fusion
(Linear, Eye+Face)
Female0.82 [0.68, 0.94]0.83 [0.65, 0.96]0.90 [0.76, 1.00]0.82 [0.65, 0.96]0.86 [0.73, 0.96]0.21 [0.01, 0.28]0.11 [0.04, 0.39]
Male0.91 [0.85, 0.96]0.93 [0.87, 0.98]0.96 [0.92, 1.00]0.85 [0.81, 0.97]0.95 [0.91, 0.98]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huynh, M.A.; Kline, A.; Surabhi, S.; Dunlap, K.; Mutlu, O.C.; Honarmand, M.; Azizian, P.; Washington, P.; Wall, D.P. Ensemble Modeling of Multiple Physical Indicators to Dynamically Phenotype Autism Spectrum Disorder. Algorithms 2025, 18, 764. https://doi.org/10.3390/a18120764

AMA Style

Huynh MA, Kline A, Surabhi S, Dunlap K, Mutlu OC, Honarmand M, Azizian P, Washington P, Wall DP. Ensemble Modeling of Multiple Physical Indicators to Dynamically Phenotype Autism Spectrum Disorder. Algorithms. 2025; 18(12):764. https://doi.org/10.3390/a18120764

Chicago/Turabian Style

Huynh, Marie Amale, Aaron Kline, Saimourya Surabhi, Kaitlyn Dunlap, Onur Cezmi Mutlu, Mohammadmahdi Honarmand, Parnian Azizian, Peter Washington, and Dennis P. Wall. 2025. "Ensemble Modeling of Multiple Physical Indicators to Dynamically Phenotype Autism Spectrum Disorder" Algorithms 18, no. 12: 764. https://doi.org/10.3390/a18120764

APA Style

Huynh, M. A., Kline, A., Surabhi, S., Dunlap, K., Mutlu, O. C., Honarmand, M., Azizian, P., Washington, P., & Wall, D. P. (2025). Ensemble Modeling of Multiple Physical Indicators to Dynamically Phenotype Autism Spectrum Disorder. Algorithms, 18(12), 764. https://doi.org/10.3390/a18120764

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop