Next Article in Journal
Developing Region-Specific Safety Performance Functions for Intercity Roads in Saudi Arabia
Previous Article in Journal
The Impact of Data Analytics Based on Internet of Things, Edge Computing, and Artificial Intelligence on Energy Efficiency in Smart Environment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of ML Techniques for Assessing Gross Motor Function in Adults Suffering from Cerebral Palsy

by
Jesus Gomez
1,
Jose Luis Jodra
2,*,
Raquel Martinez
3,
Asier Salazar-Ramirez
3,
Maria-Arantzazu Ruescas-Nicolau
4 and
M. Luz Sánchez-Sánchez
4
1
Department of Physiotherapy, Faculty of Physiotherapy, University of Valencia, 46010 Valencia, Spain
2
Department of Computer Architecture and Technology, University of the Basque Country (UPV/EHU), 20018 Donostia, Spain
3
Department of System Engineering and Automation, University of the Basque Country (UPV/EHU), 48013 Bilbao, Spain
4
Physiotherapy in Motion, Multispeciality Research Group (PTinMOTION), Department of Physiotherapy, Faculty of Physiotherapy, University of Valencia, 46010 Valencia, Spain
*
Author to whom correspondence should be addressed.
Appl. Sci. 2026, 16(1), 226; https://doi.org/10.3390/app16010226 (registering DOI)
Submission received: 15 November 2025 / Revised: 12 December 2025 / Accepted: 23 December 2025 / Published: 25 December 2025
(This article belongs to the Section Biomedical Engineering)

Abstract

Cerebral palsy (CP) is a neuromotor disorder that affects movement and posture, causing functional limitations and musculoskeletal deformities that persist into adulthood. Variability in motor expression makes identifying the functional level a clinical challenge; achieving greater accuracy in this assessment enables detection of risk factors for postural deterioration. This study analysed a sample of 56 adults with CP to evaluate the ability of different machine learning (ML) models to discriminate between levels IV and V of the Gross Motor Function Classification System (GMFCS), using a total of 78 clinical variables such as spasticity, range of motion, deformities, and postural asymmetries. Different supervised learning models were compared, and a relabelling procedure was applied to improve classification consistency. The results indicate that there is valuable information in the collected standardised variables for classifying levels IV and V. The best performances were subsequently obtained by the neural network and the linear logistic regression, achieving the latter, which has explanatory properties interesting for clinicians, accuracy, F1-score and AUC metrics of 92.83%, 93.44% and 99.35%, respectively. These findings suggest that ML could emerge as a useful and explanatory tool for functional assessment of CP in adults and for the design of personalised rehabilitation strategies.

1. Introduction

Cerebral palsy (CP) is a heterogeneous group of developmental disorders that affect movement, muscle coordination and postural control, resulting in varying degrees of functional impairment [1,2]. It is a permanent condition caused by lesions or abnormalities in the developing brain, which alter the maturation of motor circuits and lead to lifelong limitations in activity and social participation [3].
CP is linked to a variety of factors that can occur during the prenatal, perinatal and postnatal stages. These include congenital malformations, intrauterine growth restriction, multiple pregnancies, infections during the foetal or neonatal period, asphyxia at birth, premature birth, untreated maternal hypothyroidism, perinatal strokes, and thrombophilias. Among all these factors, premature birth is considered to be the most common and has the greatest impact on the risk of developing this condition [4,5].
The symptomatology of CP is very diverse and includes motor, sensory, cognitive and speech impairments, epilepsy, and musculoskeletal problems [3,6]. From a motor perspective, the spastic syndromes are the most common (frequently unilateral or bilateral) and can affect postural control and body alignment [7]. Over time, these impairments often lead to postural asymmetries, musculoskeletal contractures and deformities that negatively impact mobility, comfort, independence and quality of life in adulthood [8,9,10,11].
The functional assessment of people with CP is usually performed according to the Gross Motor Function Classification System (GMFCS) [12]. This system consists of a widely accepted scale that classifies gross motor function into five levels: the higher the level, the greater the impact on the person’s mobility. The first level, level I, includes those people who can walk without limitations. Then, level II refers to those who walk with speed or balance restrictions. Level III includes those who require technical aids or a wheelchair to traverse long distances. The next level, IV, includes people with very limited mobility who need assistance or devices to stand or move. Finally, level V corresponds to those who are completely dependent on help for antigravity posture and mobility.
Although the literature about CP in children is extensive, research focusing on adults with CP remains limited. As life expectancy improves, adults with CP face new clinical challenges: progressive loss of motor function, increasing skeletal deformities, chronic pain and reduced social participation [13]. Nevertheless, studies exploring the determinants of posture and body alignment in adulthood are still scarce. Research such as that by Rodby-Bousquet et al. [14,15] described the prevalence of postural asymmetries in adults with CP and their strong association with elevated GMFCS levels. They observed that as GMFCS levels increased, so did the likelihood of bodily asymmetries and the inability to maintain or change posture [15].
In adults with CP, postural control represents a multifactorial and dynamic phenomenon influenced by neuromuscular variables (such as spasticity and range of motion), biomechanical variables, postural variables, the use of assistive devices, and the person’s sitting, lying down or standing positioning habits. Traditionally, these relationships have been analysed using conventional statistical methods, such as correlations or logistic regressions [15,16].
In this context, machine learning (ML) techniques are emerging as promising tools in rehabilitation and clinical research. These technologies permit analysing clinical and functional data, modelling multidimensional relationships and identifying patterns in data. Although there have been significant advances in different applications within the field of CP [17,18,19], its implementation in the adult population remains limited. However, the results obtained by studies focusing on this population suggest that ML techniques have promising potential for improving functional assessment, customising therapeutic intervention and optimising clinical decision-making processes.
More precisely, for performing the functional assessment of adults with CP, it is necessary to take into account data belonging to different natures, such as age and sex, GMFCS level, degree of spasticity, range of motion, the use of assistive devices, and the time that the patient could spend in certain postures, among others. In this sense, ML algorithms can help integrate multiple sources of information for predicting complex clinical outcomes, such as postural quality or the presence of musculoskeletal deformities. Accordingly, these models can emerge as clinical support for assessing postural risk and, in this way, help to promote the personalisation of interventions in adults with CP.
This work provides a novel contribution to the field of CP through the application of ML to standardised clinical variables to discriminate between GMFCS levels IV and V. The study analyses the feasibility of addressing this classification problem using variables from different clinical domains related to postural control: spasticity, range of motion, musculoskeletal deformities, and postural asymmetries. Moreover, the study focuses on the understudied population of adults suffering from CP and reinforces the literature on this topic. In addition, the paper presents a relabelling procedure that facilitates the identification of borderline cases that could have been mislabelled due to subjective observation and improves the consistency of functional classification. Finally, the work analyses the explainability of the trained ML models and highlights the most influential variables in the classifier’s decisions, providing useful insights for clinical decision-making and the personalisation of rehabilitation interventions.
The rest of the article is organised as follows: Section 2 presents the dataset used for the experiment, providing a detailed explanation of all variables, as well as the filtering and preprocessing applied to obtain the dataset version used in the article. Next, Section 3 describes the feature selection process and the experimentation to determine the GMFCS level of the patients in the dataset. In addition, it also explains the relabelling process proposed for some doubtful cases and presents the classification results obtained from the relabelled dataset. Later, the section also analyses the explainability of the best explainable classifier’s decisions, seeking to give professionals a reason on which they can base their assessments. Finally, the results of the previous section and the conclusions drawn from the article will be discussed in Section 4.

2. Materials and Methods

This section presents the dataset used for this work’s experimentation. First, Section 2.1 will give a detailed explanation of the variables in the dataset, describing what they represent, how they were scored, and the system used for naming them. Then, Section 2.2 will briefly describe the filtering applied to the dataset to remove some variables, as well as the preprocessing to prepare it for the ML algorithms.

2.1. Description of the Dataset

As mentioned in the introduction, this study focuses on the functional assessment of adults suffering from CP. To do so, the study uses data collected by four physiotherapists of the Valencian Association of Cerebral Palsy (AVAPACE) centres in Valencia (Spain), with between 5 and 20 years of experience in the field of treating patients suffering from CP.
Concerning the sampled population, originally the data were collected from 58 adult users during functional and positional assessment tests. Nevertheless, it was impossible for the team to fully collect two of the registers. Thus, the data of those two users were discarded, and the dataset ended up having records of 56 users, who were aged between 19 and 74 (mean: 39.04 years; standard deviation: 14.11 years), 24 of them being female and the other 32 male. Before taking any test, all participants received an explanation of the research and gave their informed consent for using their data in this study, as approved by the Ethics Committee of Research in Humans of the Ethics Commission in Experimental Research of the University of Valencia (protocol code FIS-2837811).
To ensure consistent inter-rater assessments of the participants, before beginning with the study and aiming to reduce possible inter-rater biases, the researchers shared a manual with the physiotherapists in which the assessment protocol was explained. Moreover, physiotherapists were also given a seminar about how to apply the protocol. Finally, to further ensure inter-rater consistency, the assessments of the first data collection day were performed along with the four physiotherapists together, so that a consensus was reached for the evaluation criteria. In addition, concerning the consistency and coherence in the assessment of the GMFCS level, the scale itself has been reported to have a high degree of inter-rater agreement in the literature ([20,21]).
The dataset collected by the physiotherapists originally consisted of 94 independent variables from six domains: clinical record, postural quality, postural ability, the range of passive movement, the presence of certain bodily asymmetries and the muscle tone or degree of spasticity. Concerning the completeness of the dataset, no imputation method had to be applied, as it had been possible to fully collect the 94 variables for all the 56 participants. Apart from those independent variables, the dataset also contains each participant’s level in the GMFCS according to the empirical observations made by the physiotherapists, which is independent of all other variables in the dataset that do not directly measure gross motor function. This last variable corresponds to the class variable to identify with the ML analysis proposed in this work. The detailed explanation of the variables (both independent and class variables) will be given in the following paragraphs.
To begin with, the dataset’s first three variables correspond to the subject’s clinical record: the sex, the primary medical diagnosis (named DX_P in the dataset) and the subtype of CP (named CP_SUB in the dataset). The sex variable was considered binary, using 0 for female and 1 for male subjects. Then variable DX_P was ranged from 0 to 5, with the following meanings: 0 stands for cerebral palsy, 1 for infant CP, 2 for diparesis due to congenital encephalopathy, 3 for perinatal CP, 4 for intrauterine encephalopathy, and 5 for encephalopathy. The third and last variable, CP_SUB, included the subtype of CP that was diagnosed by the doctors (or, in the case of not having a subtype, it was assessed by the physiotherapists). This variable ranged from 0 to 6, values which, in increasing order, stand for spastic hemiplegia, spastic diplegia, spastic tetraplegia, dyskinetic or athetoid CP, ataxic CP, hypotonic athetoid CP and mixed CP.
The dataset’s second set of variables corresponds to postural quality, which was assessed using the Posture and Postural Ability Scale (PPAS) [22]. This scale is a tool to measure both body alignment and the ability to maintain a stable position. In this study, the scale was applied to four functional positions (standing, sitting, supine and prone) assessed in both the frontal and sagittal planes. Depending on the functional position analysed, the postural assessment considered different bodily items: head, trunk, pelvis, hips, knees, legs, feet and arms. Another item assessed whether the person’s weight was evenly distributed for each functional position. For example, in the sitting position, this item would check whether both glutei and legs were evenly resting on the chair, supporting the weight equally on both sides. Finally, the last item assessed the overall quality of the posture for a position, understood as the sum of all the bodily items considered for that functional position (including the item for the weight distribution).
Nevertheless, not all items were evaluated in each position; the specific selection of items varied depending on the position adopted and the plane of observation. A total selection of six body and weight items, plus a seventh corresponding to the overall score for the selected position, was always maintained. Each item scored either 0 or 1, where 0 indicates the presence of asymmetry or deviation from the midline of the body, and 1 reflects symmetrical and adequate alignment. Concerning each position’s overall score, they could vary between 0 (complete asymmetry) and 6 (optimal postural alignment).
Thus, according to the previous explanation, the dataset contained a total of 56 postural quality-related PPAS features (named PPASPQ_#, with the # being an identifier numeral). The following table (Table 1) gives a summarised explanation of the bodily items and functional positions represented by each of the PPASPQ_# features.
In addition to the features associated with postural quality, the PPAS also incorporates the “Postural Ability Level” (PAL) component. This second component constitutes the third block of characteristics in the dataset and assesses the individual’s ability to maintain or modify their posture. This component is scored on a scale from 1 to 7, where 1 represents a total inability and 7 indicates complete independence. To be consistent with the naming methodology used for the PPASPQ features, the features related to the PAL have been named PPASPAL_position (see Table 2).
The fourth group of variables identifies contractures and articulation limitations associated with postural asymmetries. To this end, the physiotherapists measured the subjects’ passive range of motion (ROM) using a universal goniometer according to the standards of the American Academy of Orthopaedic Surgeons [23,24]. The shoulder, elbow, hip, knee and ankle joints’ ROMs were evaluated bilaterally, recording the maximum extension achieved without causing pain. The ROM was considered limited when the hip’s, knee’s, or elbow’s extension was less than 0°, or when ankle dorsiflexion did not reach the neutral position. Thus, these variables were scored either 1 or 0 for the cases with or without any joint limitation, respectively.
In addition to those variables, and also linked to the ROM, the physiotherapists assessed pelvic tilt. This variable analyses the alignment of the anterior superior iliac spines in the frontal plane. It was considered that there was a pelvic asymmetry when one anterior superior iliac spine was visibly higher or lower than the contralateral one, which indicates a deviation from the midline and an alteration in pelvic position. This variable was included in the dataset under the name INPED and coded binary (1 = pelvic tilt, 0 = aligned pelvis).
In total, the dataset contained 11 features linked to ROM measurements. 10 of them correspond to joint limitations and were named according to the LIMIT_xy system, where x referred to the joint selected for the measurement and y referred to the body side. For example, following the guide of Table 3, the variable measuring the ROM limitation of the left elbow would be named as LIMIT_EL. The remaining variable, INPED, has a postural character and does not follow the aforementioned naming system because it assesses pelvic symmetry in the frontal plane rather than a specific joint limitation.
Following with the next set of variables, it relates to postural asymmetries of the different bodily segments (see Table 4). These asymmetries were determined through direct clinical observation of the alignment of those segments in the three positions. Specifically, the physiotherapists analysed the presence of dorsal hyperkyphosis or lumbar hyperlordosis and the relative position of the pelvis, lower limbs, and upper limbs with respect to the body’s midline. The presence or absence of asymmetry was recorded as 1 or 0 for each segment, complementing the quantitative information obtained with the PPAS. In addition, scoliosis was recorded based on medical history, taking into account the Cobb angle [25] and, when necessary, radiographic review of the medical history. Scoliosis was considered present (value 1) when lateral curvature of the spine with visible vertebral rotation was observed or when the patient had undergone spinal arthrodesis for this reason.
The last and sixth group of variables corresponds to muscle tone, which was assessed using the Modified Ashworth Scale (MAS) [26]. The MAS is a tool that quantifies the increase in resistance to passive movement as an indicator of the degree of spasticity. Measurements were taken in a relaxed position and at a constant speed, covering different muscle groups: psoas, glutei, adductors, quadriceps, hamstrings, tibialis and calves. The scale assigns scores from 0 to 4, with an intermediate category (1+), where 1 represents a slight increase with minimal resistance at the end of the range of motion, 1+ indicates slight resistance at the beginning and during less than half of the range, 2 reflects a more marked increase in most of the range but with ease of movement, 3 corresponds to a considerable increase that hinders passive movement, and 4 indicates complete stiffness in flexion or extension. A score of 0 is interpreted as no spastic response. However, as the variable was to be used with ML, its values had to be discrete numbers. As 1+ is not a number, value 2 was used to represent the scale’s 1+ level and the scale scores from then on were represented with one more point with respect to their original scale value (i.e., value 2 in the scale is represented with 3 in the variable). Similarly to the ROM limit features, these features have been named according to the MAS_xy system (see Table 5), where x refers to the muscle group and y to the body side.
Finally, the class variable (named as GMFCS_LEV) corresponds to the level of gross motor function, which was classified using the expanded and revised version of the GMFCS, the GMFCS-E&R [27]. Although this scale was originally designed for children up to 18 years of age, this study applied it to an adult population, exclusively focusing on users classified as levels IV and V. Of these, 27 were at level IV and 29 at level V. On the one hand, level IV was assessed when the person had the functional ability to stand actively, requiring little assistance. On the other hand, for level V, the standing posture is achieved passively with the help of two people who provide physical support, with minimal or residual active participation by the user. Although there are five scale levels, the authors decided to focus on these two classes because of their physical similarities and because they are the most limiting and because they have the greatest impact on the patients themselves, their relatives, the healthcare system and society in general.

2.2. Preprocessing the Dataset

After identifying all variables, the next step was to preprocess the dataset and prepare it for the classification stage, in which the GMFCS level of the users would be determined using ML techniques.
As explained, the original dataset contained 94 variables. Nevertheless, not all variables could be used as features with the ML algorithms. This work focuses on the problem of discriminating between GMFCS levels IV and V, which, respectively, represent users able or not to stand on their own or with very little help. This way, it would not make sense to consider the PPASPQ variables collected in the standing position, as they implicitly carry information about the user’s class. Therefore, the first 14 PPASPQ variables were filtered out from the database. Also, the PPASPAL_STAND variable was discarded for the same reason.
Then, the researchers standardised the remaining independent variables using z-score normalisation. With this, variables with bigger ranges are prevented from dominating the decisions of the ML, as all are transformed to have 0 and 1 mean and standard deviation values, respectively. The formula for applying the standardisation is presented in Equation (1), where Z i is the standardised value for the variable sample, X i is the variable sample, μ is the mean value of the variable and σ is the standard deviation of the variable.
Z i = X i μ σ
In summary, after the aforementioned steps, the dataset consisted of 78 independent variables plus the class variable. This way, the dataset was ready for addressing the proposed classification problem using ML algorithms (this version of the dataset is available as Supplementary Material attached to this article).

3. Results

Having so far presented both the dataset and the proposed classification problem, the work in this paper went through five different stages. First, the researchers carried out a preliminary analysis to see the discriminant capacity of the dataset’s variables for classifying the dependent variable and ranked them with the intention of having a first impression on whether the collected variables could be used for discriminating between GMFCS levels IV and V. Then, in a second stage, the researchers swept different subspaces to build classification systems for an initial solution to the problem proposed in this paper.
The third stage consisted in analysing the possible borderline cases to detect whether there was any possible mislabelled or ambiguous case. In order to obtain a consistent and trustworthy result, this was achieved using both statistical methods (principal component analysis, PCA) and unsupervised learning methods.
In the fourth stage, the dataset was automatically relabelled according to the previously mentioned analysis, and, after taking the relabelled cases to the experts for approval, the researchers generated new classification models that obtained better results compared to those achieved with the dataset without being relabelled.
Finally, in clinical fields, knowing why a decision has been taken is almost as important as taking the correct decision. That is why the fifth stage consisted in analysing the explainability of the decisions taken in the previous stage by calculating the Shapley values and the coefficients of the best explainable algorithm used in this work: the linear logistic regression.
The following Section 3.1, Section 3.2, Section 3.3, Section 3.4 and Section 3.5 will, respectively, explain in more detail all five stages, presenting the results obtained in each one.

3.1. Preliminary Analysis of the Relevance of the Features

First, the researchers decided to analyse the convenience of reducing the original dataset to a smaller subset of features. To do so, three different methods were used to measure the relation between the features and the GMFCS_LEV class variable. Whereas two of these methods correspond to automatic feature selection techniques (chi-square and minimum redundancy maximum relevance, Mrmr, both in their respective fscchi and fscmrmr Matlab implementations), the third belongs to Kendall’s univariate correlation analysis.
The values obtained from the analysis are shown in Table 6, which lists the 20 features that were either selected or more strongly correlated with the class variable depending on the method used. The table indicates that the variables more strongly linked to the class variable were gathered in a rather stable subset of features for the three methods. In this sense, using the features of the chi-square column as the reference, the cells highlighted in grey indicate that the feature has also been marked among the most relevant ones by either of the other (or both) two methods.
Seeing the consistency of the results of this initial feature selection analysis, the researchers decided to carry out a feature sweep to find out how many variables should be considered for the classification.

3.2. Classification

As mentioned before, the next step of the work consisted in sweeping the dataset to see how classifiers performed in different dimensional feature subspaces. More precisely, the preprocessed version of the dataset consisted of 78 features for a total of 56 instances (24 female and 32 male), 27 of them belonging to GMFCS level IV and the remaining 29 to level V.
For the experimentation presented in this section, the researchers opted for training nine classifiers belonging to different paradigms of machine learning [28,29]: discriminant analysis (Discr), decision trees (Tree), k-nearest neighbours (KNN), boosting (Boost), bagging (Bag), feed-forward fully connected neural network (Net), support vector machines (Svm), linear logistic regression (Lin) and Gaussian kernel classifiers using vector machines (Kern). All algorithms were implemented with Matlab’s default settings using a 10-run 3-fold cross-validation methodology [30] and the chi-square feature selection method to select the most relevant features in each fold.
The following Table 7, Table 8 and Table 9 respectively present the average accuracy, F1 and AUC algorithm performance metrics [31] obtained after 10 runs for the subspace sweeping process. It must be noted that although the experimentation was conducted using all possible dimension values, for the sake of clarity, Table 7, Table 8 and Table 9 only present the values between 5 and 78 dimensions, with a stride of five variables. Concerning the meaningfulness of the values in these tables, they have been highlighted differently depending on the meaning: values in bold refer to the best performance obtained by each classifier, the light grey cell corresponds to the best performance obtained by the best explainable algorithm (Lin, in this case), and the dark grey cell corresponds to the best performance in the whole table. Finally, seeking to reduce the number of features needed to obtain a good classification result, underlined values represent the best performance that can be obtained with fewer features if the best performance of each classifier is reduced by 2%, which would approximately correspond to missing one instance out of the 56.
To begin with the analysis of the results of Table 7, Table 8 and Table 9, a quick overview shows that the overall performances score around 70% and 75%, with certain performance peaks and troughs close to 80% and 60%, respectively. These results indicate that the features of the dataset contain useful information for approaching the GMFCS level IV and V classification problem.
Then, a deeper analysis of the tables shows that whereas the KNN was the best-performing algorithm for accuracy and F1 metrics (78.72% in Table 7 and 79.52% in Table 8, respectively), the bagging obtained the best AUC (80.83%, Table 9) in the scenario using the original dataset. Also, it can be noted that, even if it needed all the features to obtain its best performance, it would be possible to obtain a reduction to 35 or 50 features (depending on whether a loss of 2% is assumed for the accuracy and AUC or the F1, respectively) and still achieve over 77% for the three performance metrics. Looking at the tables from the perspective of receiving an explanation about the decisions made by the algorithms, both the Tree and Lin obtained their best performances around the five-feature subspace, where Lin was the best among them.
Finally, should the reader desire to consult other performance metrics, Table A1, Table A2 and Table A3 presenting the precision, recall and specificity scores for this scenario are available in Appendix A.

3.3. Relabelling

After obtaining promising results from the initial approach, and considering that the physiotherapists label the patients based on observations with a certain degree of subjectivity for the borderline cases, the team decided to search for cases located in an ambiguous region between classes IV and V. In this sense, if such cases existed, detecting and correcting them would positively impact the classifiers’ results.
Accordingly, the team decided to do a principal component analysis (PCA) of all instances to obtain a graphical representation of their location in a scatter plot using the first two principal components of the classification space as the chart axes. This representation would give the team a different view of how the subjects are distributed from the perspective of the classification space, which relies on the variables listed in Section 2 and differs from the observations made by the physiotherapists for assessing the GMFCS level. The result of this analysis is depicted in Figure 1, where the blue and red dots correspond to the instances belonging to classes IV and V, respectively. In addition, to make the interpretation of the graph easier, the centroids of both classes have been marked with a rhombus of the same colour.
After examining Figure 1, the team realised that there were 9 subjects located closer to the centroid of the opposite class. In this sense, from the perspective of the PCA, it would make more sense if instances 27, 28, 30, 31 and 56 belonged to class V instead of IV. Likewise, instances 24, 35, 39 and 41 would better fit into class IV. Thus, the team thought that these nine cases could be good candidates for being taken back to the experts so that they could further analyse the given label and, if they considered them borderline cases, whether those labels could belong to the other class.
However, before asking the experts to reevaluate those cases, the team proceeded with a deeper analysis of the situation to ensure the stability of their relabelling proposal. In this sense, two approaches were followed to determine whether it was worth asking the experts. On the one hand, the first approach consisted in grouping all the features according to their nature and then calculating and comparing for each feature group the centroids of the borderline cases and the whole set of each class. On the other hand, the team used unsupervised learning algorithms to validate the relabelling proposal. More precisely, the team used two algorithms coming from two different paradigms: K-Means and Hierarchical clustering. In this case, the parameter corresponding to the number of groups (K) was set to 2 for both algorithms, i.e., the number of categories handled by the class variable GMFCS_LEV.
The following table (Table 10) relates to the first approach and presents the aggregated centroids for each feature group (MAS, ESCOL…, LIMIT, PPASPAL and PPASPQ) along with the labels. Looking at the table’s rows, those identified as TOT correspond to all the instances of each class. Then, the other two rows correspond to the patterns subject to a possible class change (from class IV to V and vice versa). Looking at the table, it becomes clear that the centroids of the second row’s five cases, originally labelled as class IV, are much closer to the third row’s centroids (class V). In the same manner, the centroids of the last row’s four cases (class V) are much closer to the ones of the first row (class IV). This phenomenon takes place for all feature groups, and therefore, the idea that it might be worth reconsidering the class of those nine borderline cases is reinforced.
Regarding the clustering approach, the columns of Table 11 present the clusters generated by both the Hierarchical clustering and K-Means algorithms. In particular, the table shows the cases of the minority class in each cluster, which can be understood as the cases that each clustering algorithm would suggest labelling as belonging to their opposite class. In addition, the cells highlighted in bold correspond to cases that both algorithms considered more similar to the other class they had previously been labelled as (the intersection cases). For instance, whilst subject 19 was only included in Cluster 1 by the hierarchical algorithm, the two algorithms decided that subject 24 would better fit in Cluster 1 (whose majority class is class IV).
Bearing in mind the results of Table 10 and Table 11, the authors saw that both approaches were concordantly suggesting that nine cases could better fit in the opposite class they had been labelled into. Moreover, those nine cases are consistent with the instances proposed by the PCA scatter plot of Figure 1.
Accordingly, after seeing what the analysis based on automated techniques suggested, the team queried the experts whether it would make sense to swap the class of subjects 27, 28, 30, 31 and 56 from IV to V according to the criteria they used for labelling these subjects. Similarly, they proposed changing subjects 24, 35, 39 and 41 from class V to IV. In both cases, the experts answered that those subjects fell within the limit of belonging to one class or the other and that reconsidering the class of those subjects would be pertinent.

3.4. Classification After the Automatic Relabelling

Having received a positive answer to their proposal, the researchers proceeded to relabel the aforementioned nine cases and repeated the algorithm training experimentation with the modified dataset. The accuracy, F1 and AUC performances achieved by the same nine algorithms and using the same methodology as before are respectively shown in the following tables (Table 12, Table 13 and Table 14). Again, the values of the tables have been highlighted as mentioned in Section 3.2.
Starting with a general analysis of Table 12, Table 13 and Table 14, it is possible to conclude that using the relabelled dataset had an overall positive impact on the classifiers’ performance for all three performance metrics. This phenomenon might have its roots in the noise that the mislabelled instances were introducing into the dataset. This way, the overall performance of the classifiers increased around 10% to values surrounding 80% and 85%, with peak performances over 90%. Thus, these good results reinforce the hypothesis that it is possible to discriminate between GMFCS levels IV and V using the standardised clinical variables collected in this dataset.
If the tables are analysed in more detail, the reader will realise that when using the relabelled dataset, it was the Net which obtained the best performances for the three metrics: 93.68% for accuracy, 94.14% for F1 and 99.39% for AUC. These optimal results were obtained using the entire dataset or a subspace of 75 features.
However, bearing in mind that clinicians often desire a reason for taking their decisions and that the Net is not an explainable classifier, it is interesting to take a look at the good performances achieved by the Lin classifier (with explanatory properties). The performances scored by this classifier are less than 1% away from those obtained by the Net (92.83%, 93.44% and 99.35% for accuracy, F1 and AUC values, respectively). Furthermore, taking a loss of 2% in Lin’s AUC as reference (97.58%), it is possible to achieve over 89% scores for the three metrics using a much smaller subspace of 30 features. Similarly, being more conservative with overall loss of performance but still trying to compensate for the possible curse of dimensionality, the Lin classifier would achieve scores over 91% for the three metrics when using a reduced subspace of 50 features.
To conclude with this subsection, the authors want to note that, as they did with the results of Section 3.2, it is possible to find the precision, recall and specificity scores for this second scenario in Table A1, Table A2 and Table A3 included in Appendix A.

3.5. Explainability Analysis

It is clear that the results obtained using the relabelled dataset are better than those obtained with the original. Nevertheless, this improvement comes at the cost of expanding the problem, needing to use more variables to obtain the best performances, just as mentioned in the previous Section 3.4. For this reason, and with the intention of supporting the physiotherapists in their work, the authors chose to extract explanatory information about which variables are the most important, from a discriminative point of view, in this new scenario after the relabelling.
With this in mind, the analysis presented in this section will focus on the Lin classifier, which obtained the second-best performance with the relabelled dataset and which, in contrast to the Net, has explanatory properties. To do such an analysis of the explainability of the classification, the researchers used two different sources of information to verify the stability of the most significant variables. In order to make maximum use of the information in the dataset, all subjects were used for these analyses, since the aim was not to estimate the performance of the system in operation, as in the previous tables.
On the one hand, given that the Lin classifier induces a coefficient per variable to perform the classification, the team analysed the features with the highest absolute coefficient values, as they have the greatest influence on the classification result. Depending on whether the feature is directly or inversely proportional to an increase in the pathology, the sign of the coefficient will be positive or negative. Table 15 shows the 10 variables with the highest coefficients in the model.
As seen in Table 15, the variables in the PPASPQ domain have negative coefficients. This makes sense, as if subjects are able to maintain these postures (value 1 means that they can), they are classified as class IV (less severe pathology). Inversely, the contribution of variables involving spasticity (MAS) and range of movement limitations (ROM) points to class V with positive values. As shown in the table, the three sources of information provide discriminating power to the system, so they can be taken into account by specialists when assessing gross motor function.
On the other hand, as a second means of analysing the explanation, the team opted to calculate the Shapley values ([32], SHapley Additive exPlanations) for this problem. Specifically, we applied the Matlab implementation [33] to extract the Shapley values. The values for the 10 features that contribute most to the classification for the Lin classifier are presented in Figure 2. The horizontal axis of the figure represents the contribution of each feature to a classification towards class V if it is positive (right side) or to class IV if it is negative (left side). The colour of the dots reflects whether the contribution is with high feature values (yellow) or with low feature values (blue). Whilst some variables are binary (0/1, which is why they appear only in yellow and blue), others have several categories (hence the scale of values).
As can be seen in Figure 2, the 10 most relevant features belong to three different natures: postural ability and quality (PPASPQ and PPASPAL), ROM range of motion (LIMIT_xy), and spasticity (Modified Ashworth Scale, MAS). Features PPASPQ_33, PPASPAL_PRON, PPASPQ_52, PPASPAL_SUP, PPASPQ_19, PPASPQ_47, and PPASPQ_39 (see Table 1 and Table 2) contribute in such a way that lower values (i.e., they cannot hold certain postures in a stable manner) bring them closer to class V (positive horizontal axis). Likewise, their higher values (in yellow) point to class IV (negative horizontal axis), that is, their greater functional autonomy allows them to hold these postures in a stable manner. On the contrary, variables related to ROM and spasticity have a symmetrical behaviour. Specifically, LIMIT_SR and LIMIT_ER point towards class V with high values (yellow), as they imply a greater level of limitation. The same tendency applies to variable MAS_PL, which behaves in the same way.
Having performed both types of analysis and comparing the coefficients of the Lin classifier at its best performance with the most discriminating according to the Shapley values, it can be seen that the results are consistent with an 80% intersection (variables that coincide in both analyses are marked in the grey table (Table 15)). Likewise, it could be concluded that relabelling the nine subjects made it possible to avoid inconsistencies in the values of certain features that were interfering with the classification. Thanks to this, the performance of the Lin classifier improved several points, and also allowed other types of features to emerge as discriminants for solving the problem (coming from ROM and MAS domains).

4. Discussion

The application of ML in CP has traditionally been focused on the paediatric population and the use of instrumental or self-reported data. On the one hand, studies such as that by Ahmadi et al. [19] used supervised models, including random forest and support vector machines, to recognise patterns of physical activity in children based on accelerometer signals. Additionally, Schwartz and Munger [34] used random forest to predict GMFCS levels based on functional questionnaires. Although both studies showed artificial intelligence’s potential for automating functional assessment, their models were limited to young subjects and relied on indirect measures. On the other hand, recent reviews by Nahar et al. [18] and Balgude et al. [17] have highlighted the scarcity of studies made with adults and the lack of integration of objective clinical variables.
From the perspective of physiotherapy, even if there are studies that link asymmetries and range of motion limitations to the GMFCS level [14], the present study has added to the analysis variables that quantify spasticity alongside variables that determine asymmetries. In this line, the explanability analysis of the present paper has pointed out the discriminating power of three variables of those natures (LIMIT_SR, LIMIT_ER, MAS_PL) and thus reinforces what was proposed in the aforementioned research.
Concerning deformities such as wind-swept hips and scoliosis, other studies observed how these appear at high GMFCS levels [35]. In addition to those mentioned above, the study presented in this paper also included other variables that cause misalignment due to spinal deformities: dorsal hyperkyphosis, lumbar hyperlordosis, pelvic tilt, and anteversion/retroversion. However, during the present work, none of these variables was found among the subgroup that most influenced the classification of the GMFCS level.
In terms of the ability to predict variables that can determine the GMFCS level, the authors have not found any research seeking this association using automated systems. In this work, the team found a subset of variables linked to bodily asymmetries to be highly correlated and stable in determining the GMFCS class. This idea suggests that characteristics referring to postural asymmetries are very decisive in determining the GMFCS level, more than variables linked to muscle tone or deformities. In this sense, previous studies already linked asymmetries to high GMFCS levels [14], which reinforces the results presented in this paper.
Taking an overall look at the results presented in this paper, it can be concluded that there is a relationship between the functional assessment of class IV and V patients and the information that can be gathered from the three areas mentioned before (postural quality and ability, ROM and MAS). In addition, some of the variables appear to be more discriminating than others and may provide experts with information to perform a more accurate assessment. In addition, these conclusions have made the authors think that the use of ML algorithms could emerge as a useful tool to make an early prognosis of the patient’s possible evolution. Nevertheless, this study used cross-sectional data from subjects at a single time point. This is a limitation of the work that prevents the authors from analysing the prognosis capability of the solution. However, this limitation opens the line that proposes collecting data from the same subjects in the future to analyse the predictive capability of the approach presented in this article.
With regard to the relabelling process proposed in the paper, it suggests the possibility of evaluating possible errors in the GMFCS assessment. As this scale seeks to classify individuals based on their gross motor function, some subjects might fall on the borderline between one score and another due to aspects of functionality that a broad scale does not capture so finely. Although previous studies have shown that the GMFCS has good concordance between measurements [36], the relabelling procedure proposes a new way to better classify those subjects who fall within a range close to two possible scores. Anyway, due to the limited size of the studied sample (a problem that happens more often than desired when working with clinical data), this study was forced to test the efficiency of the relabelling with the whole relabelled dataset and not with an independent partition. This is a weakness that is going to be addressed by the team in the future by collecting a bigger sample that will permit testing the relabelling approach with an independent data partition. Furthermore, having more data will also open the line of optimising the classifiers’ hyperparameters, an approach that had a limited scope due to the scarcity of data at the current state of the research.
In this line, the improved results obtained after the relabelling process support the idea that subjects evaluated in borderline areas could be assigned to different scores from the one assigned in a first evaluation. This is particularly relevant for improving the good levels of concordance presented by the scale and allowing for clearer assessment that results in better experimental practice. It is important to bear in mind that health professionals often use the GMFCS score to characterise the functional capabilities of a patient, and, based on that score, they plan different rehabilitation plans, propose realistic and feasible objectives, and assign specific resources to that patient. A clear example of this can be found in wheelchairs, where those designed for GMFCS classes IV and V are completely different in terms of offered functionalities and costs.
In summary, this study expands the application of ML techniques to real standardised clinical and functional data (spasticity, range of motion, deformities, and postural asymmetries) obtained from adults with CP, using a relabelling procedure that improves the accuracy and consistency of GMFCS classification. This approach introduces a novel and clinically relevant perspective by transferring artificial intelligence from the paediatric experimental setting to functional and postural assessment in adulthood, providing a potentially useful tool for the customisation of therapeutic interventions.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/app16010226/s1. The preprocessed dataset is attached to this article as a supplementary file and can be downloaded from the journal’s website.

Author Contributions

Conceptualization, J.G. and R.M.; methodology, J.L.J. and M.-A.R.-N.; validation, M.-A.R.-N. and M.L.S.-S.; formal analysis, A.S.-R.; investigation, J.G. and J.L.J.; resources, J.G., M.-A.R.-N. and M.L.S.-S.; data curation, R.M., J.L.J. and A.S.-R.; writing—original draft preparation, A.S.-R.; writing—review and editing, A.S.-R.; visualization, J.G. and R.M.; supervision, J.L.J., A.S.-R. and M.L.S.-S.; project administration, M.-A.R.-N. and M.L.S.-S.; funding acquisition, R.M., J.L.J. and A.S.-R. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially funded by grant PID2021-123087OB-I00 funded by MICIU/AEI /10.13039/501100011033 and ERDF, a way of making Europe; and by the Department of Education, Universities and Research of the Basque Government (ADIAN, IT-1437-22).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Ethics Committee of Research in Humans of the Ethics Commission in Experimental Research of the University of Valencia (protocol code FIS-2837811, approved on the 14th of September of 2023).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The original contributions presented in this study are included in the article/Supplementary Material. Further inquiries can be directed to the corresponding author.

Acknowledgments

The authors would like to thank AVAPACE for their help with the data collection, as well as the participants in the experiment and their relatives for their accessibility and patience.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
CPCerebral Palsy
GMFCSGross Motor Function Classification System
MLMachine Learning
PPASPosture and Postural Ability Scale
PALPostural Ability Level
ROMRange Of Motion
MASModified Asworth Scale
MRMRMinimum Redundancy Maximum Relevance
PCAPrincipal Component Analysis
AUCArea Under the Curve

Appendix A

This appendix respectively presents the performance precision, recall and specificity metrics achieved by the classifiers using both the original dataset (Table A1, Table A2 and Table A3) and the relabelled dataset (Table A4, Table A5 and Table A6).
Table A1. Precision metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the original dataset.
Table A1. Precision metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the original dataset.
5101520253035404550556065707578
Discr74.3870.4267.3469.0266.8664.6965.0564.4063.2561.6660.6562.1663.0560.6464.0963.30
Tree74.1270.0070.0367.9668.0965.8567.8667.8568.0968.0467.8167.7968.6267.5767.6170.41
KNN72.2570.5877.2775.5175.6976.9879.2179.1079.8682.4781.0578.6877.6175.0372.6378.67
Boost70.5371.0568.5968.1272.5170.3971.6869.4070.1472.0370.5269.1970.0770.6768.7071.44
Bagg72.4972.6473.5771.6674.0671.6174.8475.3976.4577.6677.0373.4873.1972.1772.8975.14
Net70.4972.5369.7371.9472.3772.9572.7972.1673.4371.6969.9873.1071.9871.4670.7972.99
SVM74.1071.2468.0469.8965.0165.4167.2864.9667.6868.4868.2167.9866.9566.5666.1665.76
Lin75.2671.7470.4369.4569.1668.8669.3567.2070.3071.0968.9269.5270.2569.3469.2669.43
Kern73.3675.2573.5365.9063.3556.1054.8457.0752.7755.5251.0751.8853.1753.2551.1451.25
Table A2. Recall metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the original dataset.
Table A2. Recall metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the original dataset.
5101520253035404550556065707578
Discr75.8271.3371.0068.0465.9367.4464.6363.1965.5262.2264.1963.3765.4463.6366.3365.00
Tree72.1168.1967.2667.5267.8170.9669.5968.8569.8270.4471.1170.4169.0767.6767.3368.07
KNN67.5668.5272.4172.0076.1976.2276.4874.4173.1174.8175.1578.0476.5977.9676.2281.59
Boost72.6768.0469.6368.5271.6768.8569.6766.8569.7072.4469.6771.3069.0470.7871.1169.00
Bagg73.6770.8972.0070.4871.5269.1569.1568.1169.5669.5668.8571.1969.5270.5271.5270.56
Net75.2268.7870.0471.5271.8568.5671.3369.2671.3070.7069.2268.1973.4171.7068.4466.11
SVM75.1970.5970.5667.9369.8568.8568.5666.9369.0069.5267.8567.1167.8965.0466.5666.56
Lin76.4472.6370.2671.3371.5668.8173.4170.5673.7870.6369.9369.8970.7470.3070.9671.00
Kern70.3762.6757.6755.6757.1154.4462.2668.4465.2271.8567.3765.8170.1576.3772.4171.37
Table A3. Specificity metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the original dataset.
Table A3. Specificity metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the original dataset.
5101520253035404550556065707578
Discr70.7466.6762.2265.9363.3359.2661.8561.4857.7858.1555.1957.7857.7857.4159.6359.26
Tree71.1167.4166.6765.1964.4458.8960.7461.4861.1160.7460.3760.7462.2261.4861.8565.19
KNN70.0067.7875.9374.0772.9674.8177.7878.1578.8980.7478.5275.9374.8171.1168.1575.56
Boost65.9368.5264.8165.1969.2667.7869.2667.4167.7868.5265.9363.7065.5666.6764.4468.89
Bagg67.7870.7470.3769.2671.8570.3773.3374.8175.9377.4176.6771.8571.8570.0070.7474.44
Net64.0770.7466.6768.8969.2670.0069.6370.3771.4869.2667.7871.1166.6768.8968.5273.33
SVM71.1168.8964.0767.0458.8960.0062.5960.3763.7065.5665.5665.1962.5964.4462.9662.96
Lin71.8568.8967.0465.9364.4464.8164.4462.9665.9368.1565.9365.9366.3065.5665.5665.56
Kern70.7475.5675.9369.2661.1152.5944.8142.9636.6737.7829.6334.4433.7027.7826.3025.93
Table A4. Precision metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the relabelled dataset.
Table A4. Precision metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the relabelled dataset.
5101520253035404550556065707578
Discr82.2081.3081.3576.8674.3475.1970.1870.0170.3375.8779.2981.3485.6988.0286.9086.47
Tree85.1786.7490.0290.5989.6389.5989.5689.0489.0189.6489.1187.8087.5187.4786.9384.71
KNN83.5387.2983.5286.4586.7887.6886.6987.6286.5486.1385.0287.3186.2484.7386.2586.42
Boost83.4187.9489.1493.6691.3289.9991.3789.2588.6789.6989.3686.6986.8287.2586.6783.76
Bagg85.8687.1090.4893.3393.6593.1393.5792.4091.6691.0090.6590.2189.9089.9090.2689.58
Net84.3785.7084.9888.4288.2191.6494.5293.2193.0194.0292.1594.1992.8993.5892.5294.58
SVM83.0383.9883.1585.2287.6190.2590.9789.6890.5090.3991.0490.8089.8889.4788.6991.47
Lin82.9585.9986.0386.8488.7291.5190.6191.3590.9792.8191.3391.6991.8991.3892.0193.38
Kern86.9688.2381.2979.1569.3163.6961.6251.8352.2054.1354.7054.0352.7453.5554.8452.80
Table A5. Recall metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the relabelled dataset.
Table A5. Recall metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the relabelled dataset.
5101520253035404550556065707578
Discr87.3386.0084.0080.0078.3373.6773.6771.0076.6779.0082.0087.6788.0090.0089.3388.33
Tree83.3381.3383.6785.6785.6786.0086.6784.6783.3383.6784.6785.0084.6784.3384.6786.00
KNN84.3387.6788.0089.3392.3391.0091.6792.6791.6790.6791.6790.6791.3392.0094.0093.00
Boost84.6782.6786.6786.0087.6788.0090.3389.0087.6787.3387.6788.3388.6788.3388.0088.33
Bagg87.6786.6788.6790.3389.6791.0089.6789.6789.6791.0090.3390.6790.0090.6791.0091.33
Net85.0085.0085.3386.3388.0088.0089.0091.0091.6790.6791.6793.0094.0092.6795.3394.67
SVM87.3384.6785.0087.3390.0090.6789.0090.3391.3392.3392.3392.3393.3392.6793.6793.67
Lin86.6788.0089.0089.0090.0089.6790.0091.3392.3392.0092.6793.3393.3393.3393.6794.33
Kern82.0069.3361.6759.6761.3366.3371.0075.0076.6782.6778.6779.3383.6787.3390.0083.00
Table A6. Specificity metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the relabelled dataset.
Table A6. Specificity metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the relabelled dataset.
5101520253035404550556065707578
Discr76.0274.4475.5170.8367.3169.3162.6963.8461.6269.3173.6174.5880.8884.4482.8782.45
Tree81.0683.2987.9288.7086.8586.8586.8187.0887.0887.5086.6784.0783.6683.6682.5579.72
KNN78.7083.7578.3882.1882.4183.6682.1883.6182.0882.0880.4683.9482.0479.8181.3481.76
Boost78.0185.9785.6092.7389.3187.8289.3586.6785.8386.7686.4481.8181.7682.5981.4878.29
Bagg81.4483.3387.1391.7692.1391.3992.0890.5689.3588.6188.1987.4587.0487.0487.4186.25
Net79.4482.8781.0285.4684.7789.4493.1591.6790.7492.3190.0992.4590.7491.5390.0992.59
SVM77.5579.1778.3380.6583.6687.1388.1986.7187.4187.0888.1587.7386.1685.6984.6888.15
Lin77.5581.4881.4883.1985.1988.7087.5988.7087.8790.5688.6188.9488.9488.4789.2191.16
Kern83.6186.4478.1574.4963.3851.1642.1320.3218.6619.3524.1221.6713.1012.0813.9814.35

References

  1. Rosenbaum, P.; Paneth, N.; Leviton, A.; Goldstein, M.; Bax, M.; Damiano, D.; Dan, B.; Jacobsson, B. A report: The definition and classification of cerebral palsy April 2006. Dev. Med. Child Neurol. Suppl. 2007, 49, 480. [Google Scholar]
  2. Oskoui, M.; Coutinho, F.; Dykeman, J.; Jetté, N.; Pringsheim, T. An update on the prevalence of cerebral palsy: A systematic review and meta-analysis. Dev. Med. Child Neurol. 2013, 55, 509–519. [Google Scholar] [CrossRef]
  3. Sadowska, M.; Sarecka-Hujar, B.; Kopyta, I. Cerebral Palsy: Current Opinions on Definition, Epidemiology, Risk Factors, Classification and Treatment Options. Neuropsychiatr. Dis. Treat. 2020, 16, 1505–1518. [Google Scholar] [CrossRef]
  4. Stavsky, M.; Mor, O.; Mastrolia, S.A.; Greenbaum, S.; Than, N.G.; Erez, O. Cerebral Palsy—Trends in Epidemiology and Recent Development in Prenatal Mechanisms of Disease, Treatment, and Prevention. Front. Pediatr. 2017, 5, 21. [Google Scholar] [CrossRef]
  5. Hankins, G. Defining the pathogenesis and pathophysiology of neonatal encephalopathy and cerebral palsy. Obstet. Gynecol. 2003, 102, 628–636. [Google Scholar] [CrossRef]
  6. Mathewson, M.A.; Lieber, R.L. Pathophysiology of Muscle Contractures in Cerebral Palsy. Phys. Med. Rehabil. Clin. N. Am. 2015, 26, 57–67. [Google Scholar] [CrossRef]
  7. Albright, A.L. Spasticity and movement disorders in cerebral palsy. Child’s Nerv. Syst. 2023, 39, 2877–2886. [Google Scholar] [CrossRef]
  8. Holmes, C.; Brock, K.; Morgan, P. Postural asymmetry in non-ambulant adults with cerebral palsy: A scoping review. Disabil. Rehabil. 2018, 41, 1079–1088. [Google Scholar] [CrossRef]
  9. Fulford, G.E.; Brown, J.K. Position as a Cause of Deformity in Children with Cerebral Palsy. Dev. Med. Child Neurol. 1976, 18, 305–314. [Google Scholar] [CrossRef] [PubMed]
  10. Ágústsson, A.; Sveinsson, T.; Pope, P.; Rodby-Bousquet, E. Preferred posture in lying and its association with scoliosis and windswept hips in adults with cerebral palsy. Disabil. Rehabil. 2018, 41, 3198–3202. [Google Scholar] [CrossRef] [PubMed]
  11. Morrell, D.S.; Pearson, J.M.; Sauser, D.D. Progressive Bone and Joint Abnormalities of the Spine and Lower Extremities in Cerebral Palsy. RadioGraphics 2002, 22, 257–268. [Google Scholar] [CrossRef]
  12. Palisano, R.; Rosenbaum, P.; Walter, S.; Rossell, D.; Wood, E.; Galuppi, B. Development and reliability of a system to classify gross motor function in children with cerebral palsy. Dev. Med. Child Neurol. 1997, 39(4), 214–223. [Google Scholar] [CrossRef] [PubMed]
  13. Haak, P.; Lenski, M.; Hidecker, M.; Li, M.; Paneth, N. Cerebral palsy and aging. Dev. Med. Child Neurol. 2009, 51, 16–23. [Google Scholar] [CrossRef]
  14. Rodby-Bousquet, E.; Czuba, T.; Hägglund, G.; Westbom, L. Postural asymmetries in young adults with cerebral palsy. Dev. Med. Child Neurol. 2013, 55, 1009–1015. [Google Scholar] [CrossRef] [PubMed]
  15. Rodby-Bousquet, E.; Agustsson, A. Postural Asymmetries and Assistive Devices Used by Adults With Cerebral Palsy in Lying, Sitting, and Standing. Front. Neurol. 2021, 12, 758706. [Google Scholar] [CrossRef]
  16. Cankurtaran, D.; Abidin, N.; Unlu Akyuz, E.; Tezel, N. An investigation into the factors which affect the sitting balance of non-ambulatory children with cerebral palsy. Neurol. Asia 2021, 26, 575–583. [Google Scholar] [CrossRef]
  17. Balgude, S.D.; Gite, S.; Pradhan, B.; Lee, C.W. Artificial intelligence and machine learning approaches in cerebral palsy diagnosis, prognosis, and management: A comprehensive review. PeerJ Comput. Sci. 2024, 10, e2505. [Google Scholar] [CrossRef] [PubMed]
  18. Nahar, A.; Paul, S.; Saikia, M. A systematic review on machine learning approaches in cerebral palsy research. PeerJ 2024, 12, e18270. [Google Scholar] [CrossRef]
  19. Ahmadi, M.; O’Neil, M.; Fragala-Pinkham, M.; Lennon, N.; Trost, S. Machine learning algorithms for activity recognition in ambulant children and adolescents with cerebral palsy. J. Neuroeng. Rehabil. 2018, 15, 105. [Google Scholar] [CrossRef]
  20. McDowell, B.C.; Kerr, C.; Parkes, J. Interobserver agreement of the Gross Motor Function Classification System in an ambulant population of children with cerebral palsy. Dev. Med. Child Neurol. 2007, 49, 528–533. [Google Scholar] [CrossRef]
  21. Benedict, R.E.; Patz, J.; Maenner, M.J.; Arneson, C.L.; Yeargin-Allsopp, M.; Doernberg, N.S.; Van Naarden Braun, K.; Kirby, R.S.; Durkin, M.S. Feasibility and reliability of classifying gross motor function among children with cerebral palsy using population-based record surveillance. Paediatr. Perinat. Epidemiol. 2010, 25, 88–96. [Google Scholar] [CrossRef] [PubMed]
  22. Rodby-Bousquet, E.; Ágústsson, A.; Jónsdóttir, G.; Czuba, T.; Johansson, A.; Hägglund, G. Interrater reliability and construct validity of the Posture and Postural Ability Scale in adults with cerebral palsy in supine, prone, sitting and standing positions. Clin. Rehabil. 2012, 28, 82–90. [Google Scholar] [CrossRef]
  23. Moromizato, K.; Kimura, R.; Fukase, H.; Yamaguchi, K.; Ishida, H. Whole-body patterns of the range of joint motion in young adults: Masculine type and feminine type. J. Physiol. Anthropol. 2016, 35, 23. [Google Scholar] [CrossRef] [PubMed]
  24. American Academy of Orthopaedic Surgeons; British Orthopaedic Association. Joint Motion: Method of Measuring and Recording; Edinburgh British Orthopaedic Association: Edinburgh, UK, 1965. [Google Scholar]
  25. Langensiepen, S.; Semler, O.; Sobottke, R.; Fricke, O.; Franklin, J.; Schönau, E.; Eysel, P. Measuring procedures to determine the Cobb angle in idiopathic scoliosis: A systematic review. Eur. Spine J. 2013, 22, 2360–2371. [Google Scholar] [CrossRef]
  26. Harb, A.; Margetis, K.; Kishner, S. Modified Ashworth Scale. Internet, 2025. Available online: https://www.ncbi.nlm.nih.gov/books/NBK554572/ (accessed on 6 November 2025).
  27. Palisano, R.; Rosenbaum, P.; Bartlett, D.; Livingston, M. Content validity of the expanded and revised Gross Motor Function Classification System. Dev. Med. Child Neurol. 2008, 50, 744–750. [Google Scholar] [CrossRef]
  28. Mitchell, T.M. Machine Learning, 1st ed.; McGraw-Hill, Inc.: New York, NY, USA, 1997. [Google Scholar]
  29. Shalev-Shwartz, S.; Ben-David, S. Understanding Machine Learning: From Theory to Algorithms; Cambridge University Press: Cambridge, MA, USA, 2014. [Google Scholar] [CrossRef]
  30. Purushotham, S.; Tripathy, B.K. Evaluation of Classifier Models Using Stratified Tenfold Cross Validation Techniques. In Global Trends in Information Systems and Software Applications; Springer: Berlin/Heidelberg, Germany, 2012; pp. 680–690. [Google Scholar] [CrossRef]
  31. Sammut, C.; Webb, G. (Eds.) Encyclopedia of Machine Learning; Springer: New York, NY, USA, 2010. [Google Scholar] [CrossRef]
  32. Lundberg, S.; Lee, S.I. A Unified Approach to Interpreting Model Predictions. arXiv 2017, arXiv:1705.07874. [Google Scholar] [CrossRef]
  33. Matlab. Shapley—Shapley Values. Internet, 2025. Available online: https://es.mathworks.com/help/stats/shapley.html (accessed on 14 November 2025).
  34. Schwartz, M.H.; Munger, M.E. Using machine learning to overcome challenges in GMFCS level assignment. Gait Posture 2018, 63, 290–295. [Google Scholar] [CrossRef]
  35. Porter, D.; Michael, S.; Kirkwood, C. Patterns of postural deformity in non-ambulant people with cerebral palsy: What is the relationship between the direction of scoliosis, direction of pelvic obliquity, direction of windswept hip deformity and side of hip dislocation? Clin. Rehabil. 2007, 21, 1087–1096. [Google Scholar] [CrossRef] [PubMed]
  36. Jahnsen, R.; Aamodt, G.; Rosenbaum, P. Gross Motor Function Classification System used in adults with cerebral palsy: Agreement of self-reported versus professional rating. Dev. Med. Child Neurol. 2006, 48, 734. [Google Scholar] [CrossRef]
Figure 1. Scatter plot of all the instances according to the first two principal components. Blue and red colours correspond to classes IV and V respectively. The arrows point from the instances to the centroid which is closer to them for those cases that fall closer to the opposite class.
Figure 1. Scatter plot of all the instances according to the first two principal components. Blue and red colours correspond to classes IV and V respectively. The arrows point from the instances to the centroid which is closer to them for those cases that fall closer to the opposite class.
Applsci 16 00226 g001
Figure 2. Shapley plot for the 10 features with the highest contribution to the classification according to the chi-square method for the logistic linear regression (Lin) classifier.
Figure 2. Shapley plot for the 10 features with the highest contribution to the classification according to the chi-square method for the logistic linear regression (Lin) classifier.
Applsci 16 00226 g002
Table 1. Summary of the features included in the dataset related to the Positional Quality of the Posture and Postural Ability Scale (PPASPQ).
Table 1. Summary of the features included in the dataset related to the Positional Quality of the Posture and Postural Ability Scale (PPASPQ).
FeatureFunctional PositionPlane and Items
PPASPQ_1-7standingfrontal 1
PPASPQ_8-14standingsagittal 2
PPASPQ_15-21sittingfrontal 1
PPASPQ_22-28sittingsagittal 2
PPASPQ_29-35supinefrontal 1
PPASPQ_36-42supinesagittal 3
PPASPQ_43-49pronefrontal 1
PPASPQ_50-56pronesagittal 4
1 Items: head, trunk, pelvis, legs, arms, weight and overall position score (sum). 2 Items: head, trunk, pelvis, hips, knees, feet and overall position score (sum). 3 Items: head, trunk, pelvis, legs, feet, weight and overall position score (sum). 4 Items: head, pelvis, hips, knees, arms, weight and overall position score (sum).
Table 2. Summary of the Postural Ability Level PPAS (PPASPAL) features included in the dataset.
Table 2. Summary of the Postural Ability Level PPAS (PPASPAL) features included in the dataset.
FeaturePostural Ability
PPASPAL_STANDstanding
PPASPAL_SITsitting
PPASPAL_SUPsupine
PPASPAL_PRONprone
Table 3. Summary of the features related to the passive range of motion (ROM) limitations included in the dataset.
Table 3. Summary of the features related to the passive range of motion (ROM) limitations included in the dataset.
FeatureDefinition
LIMIT_xyx = Shoulder, Elbow, Hip, Knee, Ankle
y = Left, Right
INPEDPelvic tilt
Table 4. Summary of the features related to bodily asymmetries included in the dataset.
Table 4. Summary of the features related to bodily asymmetries included in the dataset.
FeatureDefinition
ESCOLScoliosis
CIFDorsal hyperkyphosis
LORDLumbar hyperlordosis
ROT_PPelvis rotation
ANT_RETAnteversion/retroversion alterations
Table 5. Summary of the Modified Ashworth Scale (MAS) features included in the dataset.
Table 5. Summary of the Modified Ashworth Scale (MAS) features included in the dataset.
FeatureDefinition
MAS_xyx = Psoas, Gluteus, Adductor, Quadriceps, Hamstring, Tibialis, Calf
y = Left, Right
Table 6. Ranked results of the most discriminant variables (feature selection methods and correlation) that point out the convenience of applying feature selection.
Table 6. Ranked results of the most discriminant variables (feature selection methods and correlation) that point out the convenience of applying feature selection.
Chi-SquareMrmrKendall
PPASPQ_53PPASPQ_35PPASPAL_SIT
PPASPQ_47PPASPQ_19PPASPQ_53
PPASPQ_52LIMIT_ELPPASPQ_47
PPASPAL_SITLIMIT_ALPPASPQ_52
PPASPQ_15PPASPAL_SITPPASPQ_56
PPASPQ_39ESCOLPPASPQ_15
PPASPQ_30PPASPQ_31PPASPQ_35
PPASPQ_43PPASPQ_47PPASPQ_39
PPASPQ_29PPASPQ_16PPASPQ_49
PPASPQ_54PPASPQ_53PPASPQ_42
PPASPQ_34PPASPQ_30PPASPQ_29
PPASPQ_38PPASPQ_43PPASPQ_52
PPASPQ_46PPASPQ_15PPASPAL_PRON
PPASPQ_56PPASPQ_36PPASPQ_29
PPASPQ_21MAS_GRPPASPQ_54
PPASPAL_PRONPPASPQ_23PPASPQ_34
PPASPQ_32PPASPQ_39PPASPQ_38
PPASPQ_50PPASPQ_30PPASPQ_46
PPASPQ_31PPASPQ_46PPASPQ_28
PPASPQ_35LORDPPASPQ_21
Table 7. Accuracies for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the original dataset.
Table 7. Accuracies for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the original dataset.
5101520253035404550556065707578
Discr73.3769.0966.7966.9964.6963.5063.2962.3961.7260.3159.8260.6561.7360.6563.1762.24
Tree71.5767.8766.9766.3966.2365.1965.4165.4065.7365.8966.0865.9165.9164.8064.8066.78
KNN68.7568.2374.1073.0374.6875.5977.1376.2675.9277.6776.7776.9875.7074.6472.3178.72
Boost69.4468.3467.3066.9670.5168.3369.4867.1468.7370.5367.8967.7067.3968.7967.9068.97
Bagg70.8770.8871.2369.9371.7169.7671.2171.3972.6873.3772.6471.5670.6870.3371.1972.47
Net69.8469.8268.4470.2970.7069.3070.5169.8271.4069.9968.5569.6470.2070.3568.5369.58
SVM73.2069.8167.4767.5164.6264.6265.7063.7566.4167.6266.7566.2065.3864.8164.8264.80
Lin74.2670.8968.7668.7468.1866.9769.1166.9370.0069.4368.0067.9968.6068.0468.3868.35
Kern70.5468.9166.4262.2559.0853.6253.9556.0451.4155.3449.0550.5352.3752.8150.0349.15
Values in bold refer to the best performance obtained by each classifier, the light grey cell corresponds to the best performance obtained by the best explainable algorithm, the dark grey cell corresponds to the best performance in the whole table and underlined values represent the best performance that can be obtained with fewer features if the best performance of each classifier is reduced by 2%.
Table 8. F1 metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the original dataset.
Table 8. F1 metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the original dataset.
5101520253035404550556065707578
Discr74.2469.8068.4167.6765.8665.4364.1062.5463.6861.3161.4461.8363.0661.4064.2463.08
Tree71.7867.7766.9666.4566.5867.1866.3166.0566.8166.8767.1666.8066.4365.2265.0966.74
KNN68.1868.4374.1172.8875.4176.0177.4576.1875.7177.5777.1477.7576.4375.9173.8379.52
Boost70.5468.2968.3667.6871.0568.6869.8166.7069.0171.3568.7969.0868.2869.5668.8068.23
Bagg71.9770.6671.5470.1771.7269.4470.1870.0271.3471.9671.3071.0670.0169.9770.9471.44
Net71.6569.5269.0170.8871.4469.4071.1669.8171.3970.4168.5469.2671.1770.9868.8168.45
SVM73.6670.0368.5167.7166.7966.1966.8664.9467.5668.1867.2266.8166.5864.9265.6665.45
Lin75.0971.5169.2269.6669.3567.7370.7268.0971.2270.1068.7268.8069.5568.9169.3069.50
Kern70.4266.0962.5559.7957.0353.7255.8261.0157.7162.0057.4657.2559.9462.1959.5858.57
Values in bold refer to the best performance obtained by each classifier, the light grey cell corresponds to the best performance obtained by the best explainable algorithm, the dark grey cell corresponds to the best performance in the whole table and underlined values represent the best performance that can be obtained with fewer features if the best performance of each classifier is reduced by 2%.
Table 9. AUC metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the original dataset.
Table 9. AUC metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the original dataset.
5101520253035404550556065707578
Discr79.0571.5068.3368.0664.4664.0663.3563.6762.7660.6461.3064.4764.3765.9866.5367.33
Tree73.5771.4769.8769.0767.5169.3369.1768.6469.1768.4569.0968.8268.6967.7467.8369.38
KNN68.7868.1574.1773.0474.5775.5277.1376.2876.0077.7876.8376.9875.7074.5472.1978.57
Boost71.5075.3275.3574.9377.3578.0277.2977.4078.0079.0378.4878.3977.9778.3578.7277.70
Bagg77.3777.8479.4278.3878.9079.7379.7880.1880.8380.8180.2479.9480.0979.9879.8280.54
Net68.3871.2872.6174.2577.0277.1977.7379.2678.9477.1575.9175.6674.1974.4773.8277.69
SVM77.0174.5871.7569.4269.8471.1171.0170.8173.4573.0772.7372.2171.4271.2471.0772.52
Lin80.7477.7976.9475.1676.5276.4776.2576.2678.6177.6476.0175.3874.1373.6874.2276.89
Kern77.2679.5876.3970.0765.1561.1861.5659.1852.1956.2947.9647.7252.3552.7351.3350.82
Values in bold refer to the best performance obtained by each classifier, the light grey cell corresponds to the best performance obtained by the best explainable algorithm, the dark grey cell corresponds to the best performance in the whole table and underlined values represent the best performance that can be obtained with fewer features if the best performance of each classifier is reduced by 2%.
Table 10. Aggregated centroids and class for each feature group of different nature (MAS_xy, asymmetries, LIMIT_xy, PPASPAL_position and PPASPQ_#).
Table 10. Aggregated centroids and class for each feature group of different nature (MAS_xy, asymmetries, LIMIT_xy, PPASPAL_position and PPASPQ_#).
SubjectMAS_xyAsymmetries (ESCOL, CIF, ...)LIMIT_xyPPASPAL_positionPPASPQ_# 2GMFCS_LEV
TOT (IV)171.6761527IV
5 cases 1392.208710IV
TOT (V)232.66778V
4 cases 1132.0051421V
1 Cases subject to a possible change in class. 2 Only overall position scores were considered, with # numerals: 21, 28, 35, 42, 49 and 56.
Table 11. Subjects proposed for a class change by the unsupervised learning algorithms. The cases marked in bold correspond to the intersection cases, i.e., the cases that both algorithms agreed that would better fit into the other class.
Table 11. Subjects proposed for a class change by the unsupervised learning algorithms. The cases marked in bold correspond to the intersection cases, i.e., the cases that both algorithms agreed that would better fit into the other class.
Hierarchical ClusteringK-Means
(10 Class Change Proposals)(13 Class Change Proposals)
Cluster 1Cluster 1
(IV:22—V:5)(IV:18—V:4)
19-
2424
3535
3939
4141
Cluster 2Cluster 2
(IV:5—V:24)(IV:9—V:25)
-1
-9
-18
2727
2828
3030
3131
-47
5656
Table 12. Accuracies for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the relabelled dataset.
Table 12. Accuracies for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the relabelled dataset.
5101520253035404550556065707578
Discr82.0180.6080.0175.6873.2271.6468.5667.6869.6174.4778.0481.5884.6387.3286.2685.51
Tree82.1782.1385.5486.9986.1186.2986.6385.7184.9985.3785.5784.5084.1283.9483.6083.03
KNN81.6485.7883.4485.9187.7087.5187.2088.4387.1786.6386.4487.5186.9886.2788.0687.70
Boost81.4684.1286.0689.0888.4087.8889.8387.8986.7787.0087.0285.2385.3985.5884.8783.58
Bagg84.6985.0687.8590.9190.7191.0590.6989.9989.4589.8089.2689.0788.5588.9089.2588.90
Net82.3583.9583.2585.9086.4488.6390.9291.2891.2391.4290.9192.6992.4792.1192.8893.68
SVM82.7382.0481.8184.1486.9588.9588.5788.5989.4789.8490.3690.1889.9989.4289.4691.06
Lin82.3684.9085.4286.2787.7089.1388.8090.0490.2091.2790.7291.2491.2491.0391.5792.83
Kern82.7077.1369.0766.2561.9259.0657.4549.6049.6753.2053.2552.5350.7852.3754.6751.07
Values in bold refer to the best performance obtained by each classifier, the light grey cell corresponds to the best performance obtained by the best explainable algorithm, the dark grey cell corresponds to the best performance in the whole table and underlined values represent the best performance that can be obtained with fewer features if the best performance of each classifier is reduced by 2%.
Table 13. F1 metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the relabelled dataset.
Table 13. F1 metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the relabelled dataset.
5101520253035404550556065707578
Discr84.0482.8281.8277.6375.5273.4071.0169.7572.7976.4479.9283.9186.0988.2587.4286.67
Tree83.1282.5585.8987.3986.5986.7387.0585.9085.1885.4785.7785.0884.7384.5384.3984.24
KNN83.0386.7985.1787.1488.9988.6788.4789.6288.5187.8787.8188.4688.2287.7689.4889.11
Boost83.0584.6786.9889.1388.7488.3690.2088.4087.2887.4887.5086.2886.5186.6186.0784.98
Bagg86.0886.2788.7091.3791.0291.3590.9390.3589.9590.4789.9089.8689.2689.6590.0389.82
Net84.0184.8384.3486.5587.3689.0891.2191.7391.8691.9291.5593.1693.0992.5593.4594.14
SVM84.4983.5483.2285.3688.0189.7489.2189.4090.2690.8091.2291.0190.9990.4390.5591.96
Lin84.1386.2386.7087.4088.7589.8389.5490.7891.0791.9191.5291.9892.0391.8292.2193.44
Kern83.4875.4666.5563.5761.3862.1963.0961.0261.7665.0963.9563.7464.4166.2267.8564.28
Values in bold refer to the best performance obtained by each classifier, the light grey cell corresponds to the best performance obtained by the best explainable algorithm, the dark grey cell corresponds to the best performance in the whole table and underlined values represent the best performance that can be obtained with fewer features if the best performance of each classifier is reduced by 2%.
Table 14. AUC metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the relabelled dataset.
Table 14. AUC metrics for dimensions between 5 and 78, with a stride of 5 dimensions using the chi-square method for ordering the most relevant features from the relabelled dataset.
5101520253035404550556065707578
Discr89.8887.4384.1379.5578.0174.6869.4368.0372.4879.6483.6189.9692.2294.7894.2194.19
Tree82.7983.3886.5087.6087.0287.1487.3486.2285.6786.0886.6085.4785.0084.8384.4484.25
KNN81.5285.7183.1985.7587.3787.3386.9288.1486.8886.3886.0687.3086.6985.9187.6787.38
Boost87.9592.0894.1796.6295.3495.8196.6995.3495.2594.9194.3393.4193.4793.1992.8591.15
Bagg90.5994.2695.9997.2697.2297.3997.5297.3497.0497.2997.3597.3597.2097.0796.9797.05
Net84.5090.8393.6395.6496.7497.4597.6997.9298.1998.4498.0698.8499.1298.9799.3999.09
SVM87.1289.8591.8393.8695.5797.2097.1497.5597.7098.1498.1098.1898.4898.5598.1398.49
Lin91.7493.2395.0995.7796.7497.5897.7898.0898.3698.5998.6398.8999.0899.1998.8599.35
Kern88.5286.0781.3277.7873.4866.8663.6249.1846.0047.6951.3746.1447.2951.6851.9851.28
Values in bold refer to the best performance obtained by each classifier, the light grey cell corresponds to the best performance obtained by the best explainable algorithm, the dark grey cell corresponds to the best performance in the whole table and underlined values represent the best performance that can be obtained with fewer features if the best performance of each classifier is reduced by 2%.
Table 15. The 10 most significative coefficients of the linear logistic regression (Lin) classifier.
Table 15. The 10 most significative coefficients of the linear logistic regression (Lin) classifier.
FeatureCoefficient Value
PPASPQ_33−0.538
PPASPAL_PRON−0.368
PPASPAL_SUP−0.352
LIMIT_SR0.324
PPASPQ_52−0.320
MAS_PL0.296
PPASPAL_SIT−0.292
LIMIT_ER0.292
MAS_PR0.291
PPASPQ_19−0.286
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gomez, J.; Jodra, J.L.; Martinez, R.; Salazar-Ramirez, A.; Ruescas-Nicolau, M.-A.; Sánchez-Sánchez, M.L. Application of ML Techniques for Assessing Gross Motor Function in Adults Suffering from Cerebral Palsy. Appl. Sci. 2026, 16, 226. https://doi.org/10.3390/app16010226

AMA Style

Gomez J, Jodra JL, Martinez R, Salazar-Ramirez A, Ruescas-Nicolau M-A, Sánchez-Sánchez ML. Application of ML Techniques for Assessing Gross Motor Function in Adults Suffering from Cerebral Palsy. Applied Sciences. 2026; 16(1):226. https://doi.org/10.3390/app16010226

Chicago/Turabian Style

Gomez, Jesus, Jose Luis Jodra, Raquel Martinez, Asier Salazar-Ramirez, Maria-Arantzazu Ruescas-Nicolau, and M. Luz Sánchez-Sánchez. 2026. "Application of ML Techniques for Assessing Gross Motor Function in Adults Suffering from Cerebral Palsy" Applied Sciences 16, no. 1: 226. https://doi.org/10.3390/app16010226

APA Style

Gomez, J., Jodra, J. L., Martinez, R., Salazar-Ramirez, A., Ruescas-Nicolau, M.-A., & Sánchez-Sánchez, M. L. (2026). Application of ML Techniques for Assessing Gross Motor Function in Adults Suffering from Cerebral Palsy. Applied Sciences, 16(1), 226. https://doi.org/10.3390/app16010226

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop