Next Article in Journal
Framework for Planning and Evaluation of Nature-Based Solutions for Water in Peri-Urban Areas
Next Article in Special Issue
Developing Capacities to Lead Change for Sustainability: A Quasi-Experimental Study of Simulation-Based Learning
Previous Article in Journal
Assessing Resource Efficiency of City Neighbourhoods: A Methodological Framework for Structuring and Practical Application of Indicators in Urban Planning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ILS Validity Analysis for Secondary Grade through Factor Analysis and Internal Consistency Reliability

1
School of Electrical Engineering and Computer Science, National University of Sciences and Technology (NUST), Islamabad 44000, Pakistan
2
Department of Information Technology, Sydney International School of Technology and Commerce, Sydney, NSW 2000, Australia
3
College of Computer Engineering and Sciences, Prince Sattam Bin Abdulaziz University, Al-Kharj 11942, Saudi Arabia
*
Authors to whom correspondence should be addressed.
Sustainability 2022, 14(13), 7950; https://doi.org/10.3390/su14137950
Submission received: 13 May 2022 / Revised: 10 June 2022 / Accepted: 13 June 2022 / Published: 29 June 2022
(This article belongs to the Special Issue Innovative Teaching and Learning in Education for Sustainability)

Abstract

:
In differentiated learning, the teacher needs to be aware of the learning styles of students in the classroom to accommodate specific learning preferences, e.g., the Felder–Silverman learning style model. The corresponding instrument, i.e., the Felder–Silverman Index of Learning Style (ILS), was designed to assess learning styles, specifically for engineering students. The ILS has been tested at the middle school level to identify the learning styles; however, validity/reliability had not been established in earlier studies with large samples. The focus of this study was to identify the validity and reliability of an ILS instrument for middle school students ( N = 450 ) by investigating the factor structure through factor analysis. This includes internal consistency reliability and constructing validity report of the ILS. An exploratory and confirmatory factor analysis was undertaken to investigate the factor structure to establish validity. As a result of the study, the reliability of the instrument was established. Five-factors emerged through exploratory factor analysis (EFA), which were subjected to confirmatory factor analysis (CFA). The outcome provided five-factors (i.e., Comparative Fit Index (CFI), Tucker–Lewis Index (TLI), Root Mean Square Error of Approximation (RMSEA), Standardized Root Mean Residual (SRMR), and Goodness of Fit (GFI)), out of which four factors were related to the four dimensions of the Felder–Silverman model, and the fifth factor was related to the association of sensing/intuitive and sequential/global dimensions of the model, which is in agreement with the theoretical construct of ILS. As a result of CFA, ILS entailing 24 items indicates a good fit with five-factor structure. C F I = 0.922 ; T L I = 0.927 ; R M S E A = 0.026 ; S R M R = 0.585 ; G F I = 0.911 ; X 2 = 277 ; d f = 42 ; p = 0.60 . This study suggests that the ILS for the secondary-grade students needs to be revised with fewer items to improve the reliability, as supported by empirical evidence through the EFA and CFA.

1. Introduction

Differentiated instruction/learning is a process where the instructor adapts several strategies according to the learning needs and preferences of individual students, in a heterogeneous classroom. Hence, with such strategies, the instructors/learners strive to achieve various characteristics with the aim of developing social, moral, and intellectual abilities that allow students to be more integrated in modern communities [1,2]. The student’s individual preferences have been interpreted in several ways, e.g., adapting instruction and monitoring each student’s progress [3] and student’s ability preferences [4], in which case the teaching method is varied while monitoring their progress and making decisions based on historical data. Similarly, there is a reactive response differentiation approach where individual students are recognized based on their background knowledge, language, and interests and are responded to accordingly [5]. In addition, the authors of [6] created a five-dimensional differentiation framework, i.e., teaching arrangements, learning environment, teaching methods, support materials, and assessment. Students’ other capabilities include their verbal abilities, family environment, social skills, and extra co-curricular activities [7].
The Felder–Silverman Learning Style (FSLSM) model is centered around the students’ preferences, i.e., presenting students with their distinctive learning styles. On the other hand, the teachers must deliver materials that address the various learning preferences of the students. There are various dimensions of learning preferences: An active learner can learn best by actively working and trying/applying the material, an intuitive learner can learn via abstract learning material, and some learners can best learn with visual effects, e.g., diagrammatic illustrations. Similarly, there are categories of learners that sequentially learn with linear learning progress; in contrast, other learners can consume learning material in abundance and use a holistic process. According to [8], personalization and differentiation should be used to meet the preferences and goals of individual students while at the same time authorizing them to actively participate in the design of their learning. A comprehensive article showed how personalized learning [9] and differentiated learning [10] could be applied in the classroom. It is evident that students with a strongly specific learning style may face learning complications if the teaching style is incompatible with their preferred learning style [11,12]. The authors of [13] carried out a study with students who attended an online course. Those with adaptive learning styles achieved significantly better results compared to those students whose preferred learning style was not catered for.
The Felder–Silverman Index of Learning Style (ILS) is a 44-item questionnaire [12] developed by Solomon and Felder for identifying learner’s learning styles. It is a renowned and mostly used among researchers and educationists. There are static and automatic approaches to analyzing the resultant questionnaire responses to detect various learning styles [14]. A static approach for learning style detection is carried out by filling out the questionnaires [15,16,17]. It is possible that some of the questionnaires will not be filled in properly, which may lead to inaccurately detecting the learning styles [18,19,20]. The other approaches are based on data-driven methods to build classification models that use the sample data and ILS. Henceforth, various classification techniques have widely been used to detect learning styles, e.g., using neural networks [21,22,23], decision trees [24,25,26], and Bayesian networks [27,28,29,30]. The authors of [31] provided a detailed overview of how to analyze the response data of the ILS questionnaire for identifying a group of learners’ preferences for various dimensions, in addition to verifying the validity and reliability of the used instrument, while these studies support the reliability and validity of ILS, there still exist some open issues related to dependencies between learning styles and a learner’s undeveloped dimension(s) that further needs investigations [32].
We carried out a comprehensive study on the ILS’s validity and reliability using the data collected from three different secondary schools who voluntarily participated. The students were 11–16 years old, from grades 6 to 10 inclusive. This study was carried out with 450 students (among them, 29 did not complete the questionnaire; hence, only 421 surveys were considered in our study). We obtained signed consent from their parents and verbal consent form the participants. We first randomly split the data into two halves and then applied the EFA (on 201 data samples) and CFA (on 200 data samples) to two samples. Note that the authors of [33] suggest that arguably, up to 150 samples are considered acceptable; hence, our sample size was adequate. Next, we examined the suitability of sampled data for factor analysis through the Kaiser–Meyer–Olkin test and the Bartlett test of sphericity [34]; however, we used small eigenvalues for this method (e.g., <1.0). Henceforth, we explored a five-factor solution on the basis of Catell’s scree plot [34] for a range of eigenvalues. In addition, we note that using this method, the cumulative variance may decrease with a range of eigenvalues; hence, we used oblique rotation instead of Varimax rotation [35] and compared their results for reliability. Furthermore, to best fit our model, we used chi-square (X2), Root Mean Square Error of Approximation (RMSEA), Goodness of Fit (GFI) and Standardized Root Mean Residual (SRMR) for absolute fit indices; and Comparative Fit Index (CFI) and Tucker–Lewis Index (TLI) for incremental fit indices.
Our exploratory analysis shows an increased number of factors when we used the Kaiser’s criterion method. Hence, we used the Catell’s scree test that extracted the five-component solution with a total of 31.55% variance contributing to the five factors, respectively, of 9.128%, 7.841%, 6.377%, 4.773% and 3.795% variance. During our analysis, we examined that the five factors analysis presented better results compared to 16 factors solutions where all the four dimensions were successfully loaded, i.e., Active/Reflective (AR), Sequential/Global (SG), Visual/Verbal (VV), and Sensing/Intuitive (SI) (detailed discussions over these Felder–Silverman Learning Style Model dimensions can be found in Section 2.1). Furthermore, for absolute fit indices, we assess the X2, RMSEA, GFI, and SRMR models, whereases for incremental fit indices, we tested the CFI and TLI models. For CFA, our analysis resulted in good fit indices of C F I = 0.759 ; T L I = 0.745 ; R M S E A = 0.015 ; S P R M R = 0.067 ; G F I = 0.815 ; p ( 0.05 ) = 0.000 ; X 2 / d f = 2.185 for diverse settings of various fitting models. Our investigations suggest that, if ILS is used for identification of learning styles of secondary-grade students, it needs to be revised with reduced items as supported by empirical evidence provided by EFA, where several items have cross loadings, do not load well (i.e., > 0.3 ) and through CFA in the present study. Moreover, construct validity through factorial validity also needs to be verified using CFA along with EFA.
Our main contributions in this paper are:
  • To carry out a comprehensive study on the ILS validity and reliability using the data collected from different secondary schools.
  • A comprehensive evaluation of psychometric analysis of ILS to investigate the learning styles of secondary-grade students.
  • To explore the internal consistency reliability and construct validity using exploratory and confirmatory factor analysis of the Felder–Silverman’s ILS for the secondary grade students.
  • The use of various fitting models to designate the best fit for the collected data.
  • To inspect the average item correlation and exploratory and confirmatory factor analysis for the statistical evidence of validity and reliability of ILS for secondary-grade students.
We organize the paper in the following sections. Section 2 presents the background on Felder–Silverman Learning Style Model and ILS. Section 3 presents literature work. We present our methodology in Section 4. Results are dicussed in Section 5 and finally we conclude in Section 6.

2. Background

For many years, it has been claimed that understanding of learning styles has an effective role in the learning practices of the educators and students. Learning styles advocate Differentiated Instruction (DI) where each student to process information in the way they process information. Several learning style models have been introduced, among which, model proposed by Felder and Silverman is most commonly used by teachers, students and researchers.

2.1. Felder–Silverman Learning Style Model

The Felder–Silverman experiential, phenomenological and sensory model was initially presented in 1988, designed to explore the learning preferences of higher education students, specifically the engineering major. According to this model, each student has specific learning styles over multiple bipolar axes. Educators need to understand these specific learning preferences of the students to accommodate their learning needs [36]. The model categorizes learner’s preferences in bipolar dimensions, which are Active (learning by doing and working with other) vs. Reflective (learning by thinking or reflecting and working alone), Sensing (practical, concerned with procedure and facts) vs. Intuitive (creative, concerned with theory and concepts), sequential (prefer stepwise information) vs. Global (learn holistically, understand the whole theme), Visual (prefer visual representation of information) vs. Verbal (prefer spoken or written information), and Deductive (general to specific presentation) vs. Inductive (specific to general presentation).
The designed model is not entirely original, but derived from other experiential, phenomenological and sensory models. For example, Active/Reflective (AR) dimension was derived from Kolb’s experiential model [37], Sensing/Intuitive (SI) was derived from Jung’s theory [38], and Visual/Verbal (VV) dimension was taken form Dunn and Dunn model [11]. The Sequential/Global (SG) dimension proposed by Felder–Silverman model was original. The psychometric assessment instrument associated with the model is called index of learning style (ILS).

2.2. Index of Learning Styles (ILS)

Index of learning style is the diagnostic tool, comprising 44 items, to explore the Felder–Silverman model along four bipolar dimensions, i.e., AR, VV, SI and SG. It was designed by Richer. M. Felder and Barbara A. Solomon in 1991 [39]. The bipolar dimensions AR, VV, SI and SG are related to the process, input, perception and understanding of information, respectively. Deductive/Inductive (DIn) dimension was not included in the assessment tool because it was claimed that the educators need to design instruction in for both DIn for DI (i.e., Differentiated Instruction). Each characteristics of the ILS is allocated in opposing learning preference.
The ease of access due to free availability of the online version of ILS is one of the main reasons that it is widely utilized to assess leaning styles of students for various disciplines in higher education. Although, the model and the instrument was designed to identify the learning preferences of engineering students, however, it is also used for secondary-grade students [40,41]. The scoring of the instrument is automatic for online version, even the paper pencil test scoring is also very easy and takes less than two minutes. Each of the four bipolar dimensions has 11 questions, thus a total of 44 questions. The answer to a question of one of the two poles of applicable bipolar modality is according to the set pattern. This means that questions are designed in a way that students have to choose one option out of two (either ‘a’ or ‘b’). The pattern for scoring is that if answer is ‘a’, then it will be scored for Pole A and if the answer is ‘b’, then the score will be for Pole B. For example, if bipolar dimension is Active Reflective (AR) and response ‘a’, the score will be for active and if response is ‘b’, the score will be for Reflective. The scores are complementary so they sum of scores is 11. The difference in the cumulative scores of ‘a’ and ‘b’ aids to distinguish the preference in the given modality. Each pole scores can range from 0 to 11. The students can be categorized in neutral, moderate or strong preferences of the dimensions based on the scoring. 1 and 3 reflects a neutral along the bipolar continuum, 5 and 7 reflects a moderate inclination whereas 9 and 11 show a strong inclination.
Rahadian and Budiningsih [40] recommended the Felder–Silverman [36] model to recognize the preferences of secondary-grade students, so that the teaching and medium of instruction can be designed for personalized leaning (2018). However, the instrument was designed to identify learning preferences of students of higher education, specifically engineering students [36]. To identify learning style of middle school students, ILS tool need to be tested for it’s reliably and validity for middle school students. The measurement of validity and reliability is essential before the use of instrument for research study. These measurements are required beforehand for the interpretation of study findings [42]. Although there are several studies, which report the validity and reliability of ILS for higher education [17,31,35,43,44,45,46,47], however, we could not find reliability and validity report on middle school students. Thus, the present study has been conducted purposely to explore the most essential prospective of the instrument for secondary-grade students.

2.3. Assessment Tools

Internal Consistency Reliability: Internal consistency is an essential assessment tool to examine the uniformity of construct, whereas, reliability information illustrates the correlation among the items of the tool and is reported using the Cronbach’s alpha value [48]. It also measures the stability, consistency, and accuracy of the instrument [49]. If the scale has a high correlation between the items, the alpha ( α ) value increases accordingly. Reliability is a necessary component for the instrument to be standardized, but it is not sufficient for the validity of that instrument.
Construct Validity: Validity measures the degree to which the instrument provides consistent results and the amount to which the test predicts other observed variables of interest [50]. According to [50] “construct validity is involved whenever a test is to be interpreted as a measure of some attribute or quality which is not operationally defined” and factor analysis is used to assess whether the test measures hypothesized learning ability. The factor analysis is the set of statistical approaches used to assess the relationships between a group of observed variables that are assessed through a number of items or questions.
Exploratory Factor Analysis: Exploratory Factor Analysis (EFA) is used to recognize the association between the variables to comprehend the underlying constructs [51]. The steps include examining the measure of the association and suggestion of factorability of the variables [52], extracting a set of factors from the correlation matrix, determining the number of factors using the Kaiser Criterion method and the scree plot method, rotation of factors to increase interpretability, and lastly, comprehension of results and identification of nature of factors [53].
Confirmatory Factor Analysis: The EFA specifies the essential quantity of factors, then the confirmatory factor analysis (CFA) confirms how well the analyzed variables characterize a smaller number of construct [49]. To cross-validate the factor structure, CFA is used to analyze the adequacy of the model. The adequacy of the model is calculated initially with Chi-square test statistics and then fit indices [53].

3. Literature Review

According to Felder, students have different characteristics and strengths: some are more inclined reflection and work alone, and others like to work in a group and prefer discussion and brainstorming. Some learners wish to learn through facts; others are more comfortable with the theoretical explanations. Few in the classroom prefer to learn through visual representation of information and others learn tends to learn through verbal or written explanation [36]. M. Felder and Barbara A. Solomon develop psychometric assessment tools called Index of Learning style in 1991 to assess the learning preferences of engineering students based on Felder–Silverman’s model. The model explores the learner styles of the 44-item survey. Index of the learning style is a concise and comprehensive tool, freely available online to assess learning style. Thus, it is widely used to explore learning styles of engineering students [17,35,43,47] and higher education students from other fields of studies [17,43,45,47,54,55]. Every year several educators and students use free of cost web-based version of ILS [43].
When the psychometric tool is used for a research study with a new population, the assessment of validity is fundamental requirement [42]. Felder and Spurlin [31] states that several studies used ILS for assessing psychological, affective, and cognitive behavior in terms of their learning preferences. The authors in [35] examined the relationship between learning styles and performance in hypermedia-assisted learning environment and provided the statistical evidence of ILS reliability through test–retest reliability, internal consistency reliability, and validity through exploratory factor analysis for engineering students. The study recommended that ILS identify separate qualities as predicted theoretically and established the reliability through alpha coefficient, however, the resultant five factors of EFA had cross-loadings among SI and SG domain [35]. The association among the dimensions of Felder–Silverman model is claimed to be expected as it is suggested by Felder and Spurlin [31]. The relationship between SI and SG is consistent with basic theory of ILS [31].
Relatively, ref. [43], identified reliability and validity of ILS and assessed the learning styles regarding the field of studies (i.e., engineering, liberal arts, and education students) and gender. The collective samples from all three fields of studies were subjected to analyze internal consistency reliability and factorial analysis through EFA, ref. [43]. The alpha coefficient provided the evidence of the reliability of ILS and eight-factor solution estimated through EFA combined with reliability assessment was said claimed as evidence of construct validity [43]. However, a few items did not load well and items deletions were suggested to improve reliability with the concern of issues in ILS-scoring [43]. Then, ILS was modified by [44] in 2007, using five option response scale, assessed for reliability and validity. The construct validity was analyzed through direct student feedback about their learning preferences [44]. It was established that alteration in scale does not transform the factor structure but has a positive impact on scale reliability [44] as recommended in [43]. ILS was also used to identify learning the styles of medical students [42]. When they tested the temporal stability, factorial structure, and reliability, it was established that ILS is an appropriate tool to examine the learning styles of medical students. Though the reliability was found moderate to high, the construct validity was established through EFA and few items (i.e., 17, 32, and 40) did not load well on any factor [42]. The translated version of ILS has also been used to assess learner preferences of students [17,45].
Wang and Medori [17] contributed to validating the Mandarin translation of ILS] by presenting the evidence of internal consistency reliability, inter-scale correlation. They used the Felder–Silverman’s model to recognize the learning styles of language students ( N = 138 ) and ( N = 60 ) engineering students. The evidence of internal consistency for the mandarin version of ILS satisfies the acceptability criteria; however, to determine to construct validity, learning style trends were associated with gender and field of study [17]. This does not provide the evidence of construct validity using the scientific method suggest by [48]. The authors in [48] state that construct validation is a scientific method and can be carried out by analyzing the reliability and factorial structure. Kaliská [45] provided the evidence of test–retest stability, through Pearson’s correlation coefficient for ILS translated version in Slovakia but did not provide the evidence of construct validity as recommended by [48]. The non-significant correlation was reported for two (AR, SG) out of four dimensions. These findings may be caused because of insignificant sample size (37), poor translation of ILS, and weak motivation of participants [45]. Linear discriminant analyses were used to assess the representative features of identified learning styles through ILS [47]. The authors in [47] state that the identified description of learning style dimensions was found more significant for incorporation in technology-based education. They used a hybrid approach to analyze the dimensions of the model for technological education and online environment in 2007.
Rinehart, Sharkey, and Kahl [54] examined the association of learning styles and responsibilities and professional positions of librarians. Their result showed the statistical significance in terms of the relationship between the learning styles of librarians and their responsibilities [54], but the instrument was not developed to assess the learning styles of librarians but explicitly for engineers [36] and it is vital to assess the validity of tool before interpreting the findings of the study when the study is contextually sensitive [48]. Similarly, another study carried out on higher education students of mathematics, interpret their findings without reporting the validity of the tool for the said sample [55]. They assessed learning styles using ILS, to examine the association of attitude and learning style concerning mathematics. When a tool is used in a new context or population its reliability and validity report is required. The authors in [54] mentioned that respondents (librarians) do not find the ILS appropriate in their context as it has various items designed for understanding the learning preferences of students in a classroom. In these studies, ref. [40,55], ILS has been used to assess the learning preferences of higher education students, but [41] used ILS to estimate difference in learning style of monolingual and bilingual secondary-grade students after providing the reliability evidence through test retest reliability. Although [41] found strong test-rest reliability as shown in the Table 1, but [48] review of the use of psychometric measures revealed that construct validity needs to be a fundamental prerequisite for interpreting the result of study for a new population or context and solely reporting alpha coefficient is not sufficient.
Comparably, ref. [40] suggested that ILS have helped when applying instructions with various adaptations to assess the finest strategy by referring to the studies carried out on tertiary level [58,59]. The authors in [40,41] used ILS to examine the relationship of learning styles with instructional strategy and media on secondary-grade students without providing any evidence of reliability alpha coefficient or construct validity. This practiced is frowned upon. According to [48,49], provision of evidence of structural validity of the instrument is prerequisite, for replication study or if the study is carried out on new population. ILS was used to determine learning styles of secondary-grade students in a recent study by [40,41] without providing empirical evidence of reliability and validity on said population. The present study, therefore, inspected the average item correlation and exploratory and confirmatory factor analysis to provide the statistical evidence of reliability and validity of ILS for secondary-grade students.

4. Methodology

The index of learning styles (ILS) is widely used to identify the learning preferences of the students. Although the instrument is designed primarily for the engineering students, in a few recent studies, it has also been used to determine the learning styles of the secondary-grade students [40,41]. The evidence of the validity of existing tools, specifically, construct validity, needs to be examined when existing tool is used in a new context [48]. Since ILS has not been validated to investigate the learning styles of the secondary-grade students, this study examines the psychometric analysis of the ILS as suggested by [48]. This study explores the internal consistency reliability and the construct validity using exploratory and confirmatory factor analysis of the Felder–Silverman’s ILS for the secondary-grade students.

4.1. Participants

The study for the validity and reliability of ILS was carried out on the data collected from three secondary schools ( N = 450 ). The age of the students was between 11 and 16 years, and the grades of the students were between 6 and 10. Signed consent forms were obtained from the parents before the study, and verbal consent was taken from the students. Thus, participation from all the students in this study was voluntary. It was made clear to the students that if they did not wish to participate, they would not be exposed to any academic deficiency. Some of the students ( N = 29 ) did not complete the survey; therefore, all such data were excluded from the study. Only the valid dataset ( N = 421 ) was used for the analysis.

4.2. Research Design

To validate the index of learning style (ILS) instrument for the secondary-grade students, internal consistency reliability and construct validity were examined through a step-wise approach. Average items correlation was identified by Cronbach’s alpha coefficient to find the internal consistency reliability of an instrument. The construct validity of the tool was identified using Structural or factorial validity (exploratory and confirmatory factor analysis). The researcher should provide the evidence of construct validity; even the reliability of the instrument has been established [48]. ILS has been only validated by various researchers using EFA [17,31,35,43,44,45]; however, researchers have argued that factorial validity also needs to be verified using CFA [49]. Exploratory Factor Analysis (EFA) provides the necessary amount of factors to represent the data. It is a measure to examine the dimension of a group of items [49]. Confirmatory factor analysis confirms how well the estimated variables represent the small number of constructs and approves the structural model of the tool [49].
The instrument was administered in the mid of the session of the academic year 2018. The paper–pencil form of instruments was used for the study. This made it easy for the students and teachers (who helped in conducting the survey) as they were able to obtain the data in their classes and did not require going to the computer lab. Initially, the scale scoring was carried out manually on the form, and later the data was digitized to be used in Statistical Package for Social Sciences (SPSS) version 20. For construct validity and internal consistency reliability, 421 valid ILS surveys were utilized. The sample was split randomly in half to analyze factorial structure to minimize the chance of overanalyzing the data. The EFA and CFA were carried out on separated sub-samples. When sample was randomly split, it reduces the sample size for EFA ( N = 201 ) and CFA ( N = 220 ), but ref. [60] suggests small sample size for factor analysis. According to [33], the sample size up to 150 is considered acceptable; hence the sample size was adequate for the present study.
The suitability of data was first, examined for factor analysis through Kaiser–Meyer–Olkin and Bartlett’s test of Sphericity [34]. The data analysis for EFA was conducted on SPSS 25.0 software. Kaiser’s criterion method with Varimax rotational was used first to retain a number of factors. Kaiser’s criterion method with Eigen values less than 1.0 sometimes extracts too many factors [34]. Thus, five factors solution was investigated based on Catell’s scree plot test, which often extracts very few factors. The scree plot test implicates plotting each of the eigenvalues of the factor and examining the plot to look for a point where the shape of the curve alters the direction and becomes horizontal [34]. Both Kaiser’s criterion method and scree plot were used to extract factors and the results were assessed for consistency. When factors are extracted using scree plot tests, the cumulative variance may decrease. Then Oblique rotation instead of Varimax rotation was used since different factors were expected to correlate as suggested by [35] and oblique rotation helps provide a clear pattern than Varimax rotation. The solution attained through both methods was compared for consistency.
CFA was conducted on the second sub-sample ( N = 220 ) to cross-validate the factor structure identified by EFA using AMOS version 26.0. This technique helps to confirm the structural model of the tool. In CFA, a model that consists of a number of items and the number of factors as retained from EFA was assessed. KMO and Bartlett’s Test of Sphericity was examined to test the suitability of data and then the goodness-of-fit-indices were examined.
For absolute fit indices that establish how well a priori model fits sample data initially, a model chi-square statistics investigation is required to calculate that the model fits exactly in the population (i.e., p > 0.05 ). However, Chi-Square statistic is sensitive to the sample size (i.e., it rejects the large sample size and lacks power if a small sample size is used) [61] and assessing CFA for the instrument with several items (i.e., N = 44 ), Chi-square statistics must not be seen as standard and only goodness of fit index [53,62]. For, absolute fit indices, model chi-square (X2), RMSEA Root Mean Square Error of Approximation), GFI (Goodness of Fit), and SRMR (Standardized Root Mean Residual) were assessed, to determine the indication of the amount at which the proposed theory fits the data [61]. For incremental fit indices, CFI (Comparative Fit Index) and TLI (Tucker–Lewis Index) were examined. The study strives for a value of CFI (Comparative Fit Index) GFI (Goodness of Fit) and TLI (Tucker–Lewis Index) around 0.90 and RMSEA (Root Mean Square Error of Approximation) less than 0.05, and SRMR (Standardized Root Mean Residual) value less than 0.08 [33,61]. RMSEA determines how well the proposed model works; it tests whether unfamiliar but optimally selected factor estimates would fit the population covariance matrix. SRMR provides the square root of the difference between residuals of the sample covariance matrix and hypothesized covariance model [61]. CFI compares the dataset covariance matrix with the null model and it is not sensitive to sample size [62]. GFI estimates the proportion of variance that is accounted for by calculated population covariance [61].

5. Findings

Internal Consistency Reliability

The study analyzed the internal consistency reliability of ILS on secondary-grade students with 421 valid instruments. The analysis was not conducted on the missing data hence, cases ( N = 29 ) with missing responses of participants were eliminated from the data analysis. Table 2 provides the Cronbrach value of each bipolar dimension of Felder’s learning style model for secondary-grade students. It was found that Cronbach’s alpha value was between 0.518 and 0.618. Table 1 provides the comparison of Cronbach’s alpha value of the present study with other studies, which shows that the reliability report is quite similar. However, the rest of the studies were carried out on higher education students and the present study finds the reliability of ILS for secondary-grade students.
The internal consistency reliability of the index of learning style was identified through Cronbach’s alpha then construct validity must be explicitly reported as suggested by [48]. To establish the construct validity, Structural or factorial validity (exploratory and confirmatory factor analysis) was carried out. The dataset was randomly split to minimize the chance of the over-analyzing of the data and EFA was analyzed on the first subset of data. The exploratory factor analysis was carried out to extract the factors using principal component analysis (PCA). ILS is a psychometric test developed to identify the learning preferences of higher education students specifically engineering students. Its construct validity has been established for only higher-grade students by exploratory factor analysis (EFA) in various studies. The present study analyzed the EFA of ILS for secondary-grade students for the first time because the tool has been used in secondary-grade by [40,41]. The data analysis for EFA, Kaiser–Meyer–Olkin (KMO), and Bartlett’s Test of sphericity, was performed on SPSS 25.0.
The splitting of data reduced the sample size for EFA that is ( N = 201 ). According to Pallant, a sample size of up to N = 150 cases can be sufficient if the solution has high loading markers variable [34], however, Kyriazos [60] recommends a small sample size for factor analysis. Kaiser–Meyer–Olkin and Bartlett’s Test of Sphericity was performed on the first half of the random split dataset to assess the suitability of data for EFA [34]. The KMO measures the sample adequacy, its index ranges from 0 to 1 with 0.6 suggested as good for factor analysis [63]. The estimated Kaiser–Meyer–Olkin (KMO) value was 0.630 and Bartlett’s Test of Sphericity was significant ( p = 0.000 ), where p < 0.05 is considered significant [53]. Hence, the given dataset was suitable for EFA.
Two techniques (i.e., Kaiser’s criterion and Catell’s scree test) were used to retain the number of factors and then were examined for consistency. First, Kaiser’s criterion method was used to extract factors, that help in deciding the number of factors that can be retained using the eigenvalue of a factor [34]. The factors were extracted by identifying items that loaded greater than 0.3 as suggested by Tabachnick and Fidell [53], that in order to inspect the factorability, a correlation coefficient over 0.3 should be considered. Loading less than ± 0.3 is considered minimal value as a rule of thumb [34]. This sixteen-factor solution was estimated with a total variance of 62.515% and the distribution of loading above 0.3 is shown in Table 3. However, in a study carried out by [35], a total variance was 54.1% when 14 factors were extracted using the Kaiser criterion method [63]. In the present study, a 16-factor solution, ten items—1, 5, 9, 13, 21, 25, 29, 33, 37 and 41—have high loading, i.e., greater than 0.4 in factor 1 and; one item, 17, does not load well (i.e., less than 0.3) on any factor. However, cross-loadings were observed on factors 6, 9, and 10 for items 5, 25, and 29, respectively. All of these items were related to the active-reflective (AR) dimension. Ten Items (i.e., 8, 12, 16, 20, 24, 28, 32, 36, 40 and 44) related to sequential-global (SG) dimension loaded on factor 2 but item 32 also cross-loaded on factors 14 and 15. One item, 4, loaded on factor 5 with cross-loading on factors 9 and 13. More cross-loading was observed in SG on factors 7, 10, and 16 for items 12, 20, 8, respectively.
For the visual-verbal (VV) dimension, ten items (i.e., 3, 8, 11, 15, 19, 23, 27, 31, 35, 39 and 43) loaded on factor 3. One item, 43, loaded on factor 7. Several cross-loadings were observed (i.e., item 3 also loaded on factors 4 and 11, item 8 on factors 6, item 11 on factors 5, item 15 on factors 14, Item 19 on factors 4 and 14, item 23 on factors 4, 8, 10, item 27 on factors 4, item 35 on factors 6 and 13, item 39 on factor 5, item 43 on factor 7. Five items (i.e., 2, 18, 30, 34, and 42) in the sensing-intuitive (SI) dimension loaded on factor 4, though cross-loadings for item 2 are on factors 7 and item 42 on factor 5 and 9. Items 14, 22, and 38 loaded on factor 3 with cross loadings of item 14 on factor 8, 22 on factor 12, item 38 on factors 6, 9, and 15. Item 6 loaded on 6, item 10 loaded on factors 5 and 6, and item 26 loaded on factors 5 and 7. The resulting table implicates that three of the four scales (i.e., active-reflective, sequential-global and visual-verbal) were relatively orthogonal with AR largely loading on factor 1, SG on factor 2, and VV on factor 3. However, SI loaded on several factors. However, there were a lot of cross-loadings observed in the 16-factor solution, as shown in Table 3.
Additionally, the Kaiser’s criterion method sometimes extracts too many factors; thus, it was vital to look at the scree plot to retain an appropriate number of factors as suggested in [34,35]. Therefore, the 5 factors were extracted using Catell’s Scree test. All factors above the elbow were retained because these factors deliver the most to the validation of variance in the data set. Then, five factors were examined, and the associated Scree plot is shown in Figure 1. These five-component solution explained a total of 31.55% variance with components 1, 2, 3, 4 and 5 contributing 9.128%, 7.841%, 6.377%, 4.773% and 3.795%, respectively. However, in [35], the total variance was 28.9% when five factors were reduced form a 14-factor solution. Next, the factors were rotated to have a pattern of loading with better and easier interpretation. Oblique rotation was used instead of Varimax rotation to avoid the overlap as suggested by [35]. Moreover, Oblimin rotation provides information about the degree of correlation between the factors [34]. The five-factor solution is provided in Table 4.
The solution greater than 0.4 is considered as high loading, because the correlation coefficient ± 0.40 is considered important and ± 0.50 is considered significant [34].The clear pattern can be observed with high loadings ( > ± 0.40 ) through Table 4 for each dimension. The loading ( > 0.3 ) of five-factor solution of ILS for Secondary grade students is identified in ‘bold’ in Table 4 and the distribution of loading ( > 0.3 ) is also provided in Table 5 to have easier comparison with 16 factor solution provided in Table 3. Item 17 does not load well on any factor and highlighted by underlining in Table 4 and Table 6.
In Table 5, well-defined pattern can be seen in 5 factor solution as compared to 16 factor solution. All four dimensions load predominantly (i.e., AR on factor 1, SG on factor 2 VV on factor 3, and SI on factor 4). However, three items of SI and one item of SG loaded on factor 5 and VV (i.e., two items) somewhat overlap with SI. Subsequently, items were reviewed concerning factor loadings to determine the nature of factors, and are summarized in Table 6.
The item: 17 did not load well is given below:
When I start a homework problem, I am more likely to: (a) Start working on solution immediately; (b) Try to fully understand the problem first.
Item 17 of the AR scale requires students to choose one of the two options in the given context. The students were asked to generalize their approach to doing their homework however the approach to doing homework is subjective with respect to discipline. This item may need revision if ILS needs to be administered to identify the leaning preferences of secondary-grade students. A slight revision of the item with minor word changes would compensate for the weakness, which is also supported by [35]. For construct validity, factor analysis using EFA joined with the evaluation of reliability is provided, and according to [35], this provides the evidence of construct validity. The finding shows strong evidence for AG and SG with ten items loaded on a single factor in both 16-factor solution and 5-factor solution with appropriate Cronbach’s alpha value. Evidence of construct value is also good for VV and SI as most of the items load on single factor. However, few items of VV were shared with SI and one item of SG is shared with SI, which implies that dimensions are not orthogonal and these dimensions have somewhat association with each other and these findings are similar to [31,35,43,44,56,57]. However, they provided evidence of construct validity for tertiary grade students. The association between these scales is consistent with the underlying theory of ILS and it does not impact the validity of tool [31].
Next, The five factors and corresponding 44 items were subjected to CFA. The second part of the randomly split sample ( N = 220 ) was analyzed through CFA to cross-validate the factorial validity identified by EFA. The suitability of the data was checked before analysis through Kaiser–Mayer–Oklin (KMO) (i.e., 0.628) and significance was assessed through Bartlett’s test of sphericity. The KMO was 0.628 and Bartlett’s Test of sphericity reached significance ( X 2 = 0.628 , d f = 946 , p = 0.000 ). This indicates that existing commonalities are appropriate in the manifest variables in order to conduct factor analysis [34].
The path diagram (Figure 2) displays the standardized regression weight for the common factors and each of the indicators. For absolute fit indices that establish how well a priori model fits sample data initially, a model chi-square statistics examination is required to assess the model fit exactly in the population (i.e., p > 0.05 ). However, Chi-Square statistic is sensitive to the sample size (i.e., it rejects the large sample size and lacks power if a small sample size is used) [61] and with the complex model as in the case of ILS Model 1 (Figure 2) and assessing CFA for the instrument with several items (i.e., N = 44 ), Chi-square statistics must not be seen as valid and only goodness-of-fit index [53]. Hence, for, absolute fit indices, model chi-square (X2), RMSEA, GFI, and SRMR were assessed, to determine the indication of the amount at which the proposed theory fits the data. For well-fitting model, RMSEA value less than 0.05, GFI greater than 0.90, and SRMR below 0.08 are needed. For incremental fit indices, CFI and TLI were examined and strived for values less than 0.90 [61].
The result of CFA points to moderate but still insufficiently high fit indices ( C F I = 0.759 ; T L I = 0.745 ; R M S E A = 0.015 ; S P R M R = 0.067 ; G F I = 0.815 ; p ( 0.05 ) = 0.000 ; X 2 / d f = 2.185 ). The fit indices are contradicting each other. Although the RMSEA, SPRMR, and X2/df provide a good fit for model 1 but, CFI, TLI, and X2 significantly suggest that model- 1 does not fit well. In order to achieve a good fit, changes were made to model 1 and re-tested. Items (i.e., SG44, AR25, SG28, SG8, AR37, VV31, AR9, SI2, VV23, VV43, SI42) were removed as the modification indices indicated that the error covariance must be included among these test scores and test scores of the other scale [33]. Standardized residual covariance matrix was also examined for re-specification of the model. The items (i.e., AR17, SI26, VV35, VV39, SG16, AR33, SI28) with a value above one was deleted. The nested model is provided in Figure 3.
This leads to new five factor structure with 28 items and with well model fit ( C F I = 0.922 ; T L I = 0.927 ; R M S E A = 0.026 ; S R M R = 0.585 ; GFI= 0.911 ; X 2 = 277 ; d f = 242 ; p = 0.60 ; X 2 / d f = 2.62 ). The several item reduced impact the scoring of ILS but it does not impact the underlying theory of Felder’s Learning style model.

6. Conclusions

This study examined the internal consistency reliability and construct validity of the ILS for the secondary-school students. The internal consistency of the ILS was found appropriate since the Cronbach’s alpha value of all the items was greater than or equal to 0.5, which is the acceptable threshold to ensure the reliability of the instrument; this aligned with the previous studies for assessing learning styles for higher education students. A 16-factor solution was achieved through the Varimax rotational method but then Scree test was used to extract five factors using the oblique rotational method, except for item (AR17) which loaded poorly ( < 0.03 ). The five-factor model and corresponding 44 items achieved as a result of EFA were analyzed through CFA with randomly split data. The CFA of ILS has not been identified to validate the instrument by any of the research studies. We found that ILS model-1and nested ILS model-2 (Figure 3) also has similar constructs. Since the ILS model was not a sufficient fit ( C F I = 0.759 ; T L I = 0.74 5; R M S E A = 0.015 ; S P R M R = 0.067 ; G F I = 0.815 ; p ( 0.05 ) = 0.000 ; X 2 / d f = 2.185 ), the re-specification of model was carried out by removal of items to improve the fit. By various combination various items, a good-fit model was achieved with indices ( C F I = 0.922 ; T L I = 0.927 ; R M S E A = 0.026 ; S R M R = 0.585 ; G F I = 0.911 ; X 2 = 277 ; d f = 242 ; p = 0.60 ; X 2 / d f = 2.62 ). The present study suggests that, if ILS is further used for the identification of learning styles of secondary-grade students, it needs to be revised with reduced items as supported by empirical evidence provided by EFA, where several items have cross loadings, do not load well (i.e., > 0.3 ) and through CFA. We are planning to further extend this study over a larger sample size; in addition, the revised instrument with a reduced number of items needs to be tested for reliability and construct validity CFA along with EFA. Furthermore, future studies on external validity and the test–retest reliability of ILS for secondary grade are desirable, since the validity and reliability assessed in this study were limited to construct validity and internal consistency reliability.

Author Contributions

Conceptualization: M.A.M. & K.K.; Methodology: M.A.M., K.K. & Z.S.; Investigation: M.A.M., K.K., I.U. & Z.S.; Resources: M.A.M. & K.K.; Writing—original draft preparation: M.A.M., K.K. & I.U.; Writing—review and editing: M.A.M., K.K., Z.S., A.B. & M.M.; Supervision: K.K., Z.S., A.B. & M.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

This submission does not include human or animal research.

Data Availability Statement

There is no data associated with this article.

Conflicts of Interest

The authors do not declare any conflict of interests.

References

  1. Buble, M.; Zelic, M. The Modern Model of Education as a Driver of the Development of Entrepreneurial Thinking and Learning. Econ. Soc. Dev. 2020, 70–76. [Google Scholar]
  2. Stropnik Kunič, N. Individualization and differentation as a model of new communication in the learning process. Informatologia 2012, 45, 44–52. [Google Scholar]
  3. Roy, A.; Guay, F.; Valois, P. The big-fish–little-pond effect on academic self-concept: The moderating role of differentiated instruction and individual achievement. Learn. Individ. Differ. 2015, 42, 110–116. [Google Scholar] [CrossRef]
  4. Roy, A.; Guay, F.; Valois, P. Teaching to address diverse learning needs: Development and validation of a differentiated instruction scale. Int. J. Incl. Educ. 2013, 17, 1186–1204. [Google Scholar] [CrossRef]
  5. Hall, T. Differentiated Instruction. 2002. Available online: http://www.dr-hatfield.com/educ342/Differentiated_Instruction.pdf (accessed on 12 May 2022).
  6. Roiha, A.; Polso, J. The 5-dimensional model: A tangible framework for differentiation. Pract. Assess. Res. Eval. 2021, 26, 20. [Google Scholar]
  7. Marjanovič Umek, L. Diferenciacija v šoli: Enako ali različno za različne otroke. Sodob. Pedagog. 2007, 58, 108–127. [Google Scholar]
  8. Available online: https://edulastic.com/blog/personalized-learning-differentiated-instruction/ (accessed on 12 May 2022).
  9. Available online: https://edulastic.com/blog/personalized-learning/ (accessed on 12 May 2022).
  10. Available online: https://www.prodigygame.com/main-en/blog/differentiated-instruction-strategies-examples-download/ (accessed on 12 May 2022).
  11. Felder, R.M.; Silverman, L.K. Learning and teaching styles in engineering education. Eng. Educ. 1988, 78, 674–681. [Google Scholar]
  12. Soloman, B.A.; Felder, R.M. Index of Learning Styles Questionnaire; NC State University: Raleigh, NC, USA, 2005; Volume 70, Available online: https://www.webtools.ncsu.edu/learningstyles/ (accessed on 12 May 2022).
  13. Bajraktarevic, N.; Hall, W.; Fullick, P. Incorporating learning styles in hypermedia environment: Empirical evaluation. In Proceedings of the Workshop on Adaptive Hypermedia and Adaptive Web-Based Systems, Phoenix, AZ, USA, 26 August 2003; pp. 41–52. [Google Scholar]
  14. Feldman, J.; Monteserin, A.; Amandi, A. Automatic detection of learning styles: State of the art. Artif. Intell. Rev. 2015, 44, 157–186. [Google Scholar] [CrossRef]
  15. Ovariyanti, A.S.; Santoso, H.B. An adaptation of the Index of Learning Style (ILS) to Indonesian version: A contribution to Index of Learning Style (ILS), validity and reliability score. In Proceedings of the 2016 International Conference on Advanced Computer Science and Information Systems (ICACSIS), Malang, Indonesia, 15–16 October 2016; pp. 129–134. [Google Scholar]
  16. Švarcová, E.; Jelínková, K. Detection of learning styles in the focus group. Procedia-Soc. Behav. Sci. 2016, 217, 177–182. [Google Scholar] [CrossRef] [Green Version]
  17. Wang, J.; Mendori, T. The reliability and validity of felder-silverman index of learning styles in mandarin version. Inf. Eng. Express 2015, 1, 1–8. [Google Scholar] [CrossRef]
  18. Raleiras, M.; Nabizadeh, A.H.; Costa, F.A. Automatic learning styles prediction: A survey of the State-of-the-Art (2006–2021). J. Comput. Educ. 2022, 9, 1–93. [Google Scholar] [CrossRef]
  19. Ngatirin, N.R.; Zainol, Z. Automatic Detection of Learning Styles: A Decade Review on Data-driven Approaches. J. Phys. 2021, 1997, 012001. [Google Scholar] [CrossRef]
  20. Ndognkon Manga, M.; Fouda Ndjodo, M. An Approach for Non-deterministic and Automatic Detection of Learning Styles with Deep Belief Net. In Intelligent Computing; Springer: Berlin/Heidelberg, Germany, 2021; pp. 427–452. [Google Scholar]
  21. Amir, E.S.; Sumadyo, M.; Sensuse, D.I.; Sucahyo, Y.G.; Santoso, H.B. Automatic detection of learning styles in learning management system by using literature-based method and support vector machine. In Proceedings of the 2016 International Conference on Advanced Computer Science and Information Systems (ICACSIS), Malang, Indonesia, 15–16 October 2016; pp. 141–144. [Google Scholar]
  22. Bernard, J.; Chang, T.W.; Popescu, E.; Graf, S. Learning style Identifier: Improving the precision of learning style identification through computational intelligence algorithms. Expert Syst. Appl. 2017, 75, 94–108. [Google Scholar]
  23. El Aissaoui, O.; El Alami, Y.E.M.; Oughdir, L.; El Allioui, Y. Integrating web usage mining for an automatic learner profile detection: A learning styles-based approach. In Proceedings of the 2018 International Conference on Intelligent Systems and Computer Vision (ISCV), Fez, Morocco, 2–4 April 2018; pp. 1–6. [Google Scholar]
  24. Sheeba, T.; Krishnan, R. Prediction of student learning style using modified decision tree algorithm in e-learning system. In Proceedings of the 2018 International Conference on Data Science and Information Technology, Singapore, 20–22 July 2018; pp. 85–90. [Google Scholar]
  25. Karagiannis, I.; Satratzemi, M. An adaptive mechanism for Moodle based on automatic detection of learning styles. Educ. Inf. Technol. 2018, 23, 1331–1357. [Google Scholar] [CrossRef]
  26. Maaliw, R.R., III; Ballera, M.A.; Ambat, S.C.; Dumlao, M.F. Comparative analysis of data mining techniques for classification of student’s learning styles. In Proceedings of the International Conference on Advances in Science, Engineering and Technology (ICASET-17), Qingdao, China, 20–22 May 2017; pp. 18–19. [Google Scholar]
  27. Xing, W.; Li, C.; Chen, G.; Huang, X.; Chao, J.; Massicotte, J.; Xie, C. Automatic assessment of students’ engineering design performance using a Bayesian network model. J. Educ. Comput. Res. 2021, 59, 230–256. [Google Scholar] [CrossRef]
  28. El Aissaoui, O.; El Madani, Y.E.A.; Oughdir, L.; El Allioui, Y. Combining supervised and unsupervised machine learning algorithms to predict the learners’ learning styles. Procedia Comput. Sci. 2019, 148, 87–96. [Google Scholar] [CrossRef]
  29. El Mezouary, A.; Hmedna, B.; Omar, B. An evaluation of learner clustering based on learning styles in mooc course. In Proceedings of the 2019 International Conference of Computer Science and Renewable Energies (ICCSRE), Agadir, Morocco, 22–24 July 2019; pp. 1–5. [Google Scholar]
  30. Kika, A.; Leka, L.; Maxhelaku, S.; Ktona, A. Using data mining techniques on Moodle data for classification of student? S learning styles. In Proceedings of the International Academic Conferences, Rome, Italy, 19–21 November 2019; p. 9211567. [Google Scholar]
  31. Felder, R.M.; Spurlin, J. Applications, reliability and validity of the index of learning styles. Int. J. Eng. Educ. 2005, 21, 103–112. [Google Scholar]
  32. Viola, S.R.; Graf, S.; Leo, T. Investigating relationships within the index of learning styles: A data driven approach. Interact. Technol. Smart Educ. 2007, 4, 7–18. [Google Scholar] [CrossRef] [Green Version]
  33. Khine, M.S. Application of Structural Equation Modeling in Educational Research and Practice; Springer: Berlin/Heidelberg, Germany, 2013; Volume 7. [Google Scholar]
  34. Pallant, J. SPSS Survival Manual; McGraw-Hill Education: London, UK, 2013. [Google Scholar]
  35. Zywno, M. A contribution to validation of score meaning for Felder Soloman’s Index of Learning Styles. In Proceedings of the 2003 Annual Conference, Boston, MA, USA, 24–28 July 2003; pp. 8–31. [Google Scholar]
  36. Felder, R.M. Matters of style. ASEE Prism 1996, 6, 18–23. [Google Scholar]
  37. Kolb, D.A. Experiential Learning: Experience as the Source of Learning and Development; FT Press: Upper Saddle River, NJ, USA, 2014. [Google Scholar]
  38. Jung, C.G. Collected Works of CG Jung, Volume 8: Structure & Dynamics of the Psyche; Princeton University Press: Princeton, NJ, USA, 2014; Volume 47. [Google Scholar]
  39. Solomon, B.; Felder, R. Index of Learning Styles. 1999. Available online: https://www.researchgate.net/publication/228403640_Index_of_Learning_Styles_Questionnaire (accessed on 12 May 2022).
  40. Rahadian, R.B.; Budiningsih, C.A. What are the suitable instructional strategy and media for student learning styles in middle schools? arXiv 2018, arXiv:1801.05024. [Google Scholar] [CrossRef]
  41. Esfandabad, H.S. A comparative study of learning styles among monolingual (Persian) and bilingual (Turkish-Persian) secondary school students. Procedia-Soc. Behav. Sci. 2010, 5, 2419–2422. [Google Scholar]
  42. Hosford, C.C.; Siders, W.A. Felder-Soloman’s index of learning styles: Internal consistency, temporal stability, and factor structure. Teach. Learn. Med. 2010, 22, 298–303. [Google Scholar] [CrossRef]
  43. Felder, R.; Litzinger, T.; Lee, S.H.; Wise, J. A study of the reliability and validity of the felder soloman index of learning styles. In Proceedings of the 2005 Annual Conference, Chania, Greece, 24–26 September 2005; pp. 10–95. [Google Scholar]
  44. Litzinger, T.A.; Lee, S.H.; Wise, J.C.; Felder, R.M. A psychometric study of the index of learning styles©. J. Eng. Educ. 2007, 96, 309–319. [Google Scholar] [CrossRef]
  45. Kaliská, L. Felder’s learning style concept and its index of learning style questionnaire in the Slovak conditions. GRANT J. 2012, 1, 52–56. [Google Scholar]
  46. Graf, S.; Viola, S.R.; Kinshuk, T.L. Representative characteristics of felder-silverman learning styles: An empirical model. In Proceedings of the IADIS International Conference on Cognition and Exploratory Learning in Digital Age (CELDA 2006), Barcelona, Spain, 8–10 December 2006; pp. 235–242. [Google Scholar]
  47. Graf, S.; Viola, S.R.; Leo, T.; Kinshuk. In-depth analysis of the Felder-Silverman learning style dimensions. J. Res. Technol. Educ. 2007, 40, 79–93. [Google Scholar] [CrossRef] [Green Version]
  48. Flake, J.K.; Pek, J.; Hehman, E. Construct validation in social and personality research: Current practice and recommendations. Soc. Psychol. Personal. Sci. 2017, 8, 370–378. [Google Scholar] [CrossRef]
  49. De Souza, A.C.; Alexandre, N.M.C.; de Guirardello, E.B. Psychometric properties in instruments evaluation of reliability and validity. Epidemiol. E Serv. Saude 2017, 26, 649–659. [Google Scholar]
  50. Cronbach, L.J.; Meehl, P.E. Construct validity in psychological tests. Psychol. Bull. 1955, 52, 281. [Google Scholar] [CrossRef] [Green Version]
  51. Leech, N.; Barrett, K.; Morgan, G.A. SPSS for Intermediate Statistics: Use and Interpretation; Routledge: London, UK, 2013. [Google Scholar]
  52. Beavers, A.S.; Lounsbury, J.W.; Richards, J.K.; Huck, S.W.; Skolits, G.J.; Esquivel, S.L. Practical considerations for using exploratory factor analysis in educational research. Pract. Assessment, Res. Eval. 2013, 18, 6. [Google Scholar]
  53. Tabachnick, B.G.; Fidell, L.S.; Ullman, J.B. Using Multivariate Statistics; Pearson: Boston, MA, USA, 2007; Volume 5. [Google Scholar]
  54. Rinehart, A.; Sharkey, J.; Kahl, C. Learning style dimensions and professional characteristics of academic librarians. Coll. Res. Libr. 2015, 76, 450–468. [Google Scholar] [CrossRef] [Green Version]
  55. Middleton, K.; Ricks, E.; Wright, P.; Grant, S.; Middleton, K.; Ricks, E.; Wright, P.; Grant, S. Examining the relationship between learning style preferences and attitudes toward Mathematics among students in higher education. Inst. Learn. Styles J. 2013, 1, 1–15. [Google Scholar]
  56. Van Zwanenberg, N.; Wilkinson, L.; Anderson, A. Felder and Silverman’s Index of Learning Styles and Honey and Mumford’s Learning Styles Questionnaire: How do they compare and do they predict academic performance? Educ. Psychol. 2000, 20, 365–380. [Google Scholar] [CrossRef]
  57. Livesay, G.; Dee, K.; Nauman, E.; Hites, L., Jr. Engineering student learning styles: A statistical analysis using Felder’s Index of Learning Styles. In Proceedings of the Annual Conference of the American Society for Engineering Education, Montreal, QC, Canada, 5–8 June 2002. [Google Scholar]
  58. Franzoni, A.L.; Assar, S.; Defude, B.; Rojas, J. Student learning styles adaptation method based on teaching strategies and electronic media. In Proceedings of the 2008 Eighth IEEE International Conference on Advanced Learning Technologies, Tartu, Estonia, 1–5 July 2008; pp. 778–782. [Google Scholar]
  59. Psycharis, S.; Botsari, E.; Chatzarakis, G. Examining the effects of learning styles, epistemic beliefs and the computational experiment methodology on learners’ performance using the easy java simulator tool in stem disciplines. J. Educ. Comput. Res. 2014, 51, 91–118. [Google Scholar] [CrossRef]
  60. Kyriazos, T.A. Applied psychometrics: Sample size and sample power considerations in factor analysis (EFA, CFA) and SEM in general. Psychology 2018, 9, 2207. [Google Scholar] [CrossRef] [Green Version]
  61. Hooper, D.C.J.; Mullen, M. Structural equation modelling: Guidelines for determining model fit. Electron. J. Bus. Res. Methods 2008, 6, 53–60. [Google Scholar]
  62. Hatcher, L.; O’Rourke, N. A Step-by-Step Approach to USING SAS for Factor Analysis and Structural Equation Modeling; Sas Institute: Cary, NC, USA, 2013. [Google Scholar]
  63. Kaiser, H.F. A second generation little jiffy. Psychometrika 1970, 35, 401–415. [Google Scholar] [CrossRef]
Figure 1. The Scree plot for the exploratory factor analysis of ILS items.
Figure 1. The Scree plot for the exploratory factor analysis of ILS items.
Sustainability 14 07950 g001
Figure 2. ILS Model.
Figure 2. ILS Model.
Sustainability 14 07950 g002
Figure 3. ILS nested Model 2.
Figure 3. ILS nested Model 2.
Sustainability 14 07950 g003
Table 1. Comparison of the Cronbach’s alpha value for the present study and the prior work for 11 items.
Table 1. Comparison of the Cronbach’s alpha value for the present study and the prior work for 11 items.
StudyNARSIVVSG
[35]560.590.700.630.53
[17]1980.510.630.640.51
[56]5720.510.650.560.41
[31]5840.700.760.690.55
[57]2420.560.720.620.54
[21]5000.870.770.770.61
[42]3580.630.760.640.62
Present4210.610.520.550.62
Table 2. Cronbach’s alpha values for index of learning style of the secondary-grade students for internal consistency reliability.
Table 2. Cronbach’s alpha values for index of learning style of the secondary-grade students for internal consistency reliability.
Learning StyleValid CasesItemsMeanVarianceCronbach Alpha ( α )
AR421117.346.450.61
SI421115.504.920.52
VV421116.954.840.55
SG421116.905.700.62
Table 3. The amount of factor loading (>0.3) in rotated component matrix of the 16-factor solution. The rotation method used is oblivion with Kaiser Normalization.
Table 3. The amount of factor loading (>0.3) in rotated component matrix of the 16-factor solution. The rotation method used is oblivion with Kaiser Normalization.
Learning StylesFactors
12345678910111213141516
AR10 1 11
SG 10 11 1 11 1111
VV 1042211 11 12
SI 3533212 1 1
Table 4. The rotated component matrix of the 5-factor solution. The Rotation method used is Oblimin with Kaiser Normalization, Rotation converged in 11 iterations.
Table 4. The rotated component matrix of the 5-factor solution. The Rotation method used is Oblimin with Kaiser Normalization, Rotation converged in 11 iterations.
 ComponentsComponents
 12345 12345
AR10.6360.137−0.0090.0760.054SG80.1170.5030.109−0.1190.129
AR50.595−0.0190.115−0.0160.163SG12−0.0140.407−0.1670.0270.186
AR90.6840.104−0.117−0.127−0.079SG160.1010.5600.064−0.018−0.124
AR130.6330.1110.096−0.0260.072SG200.1550.562−0.0460.0490.006
AR210.479−0.066−0.0610.061−0.298SG240.0220.7260.0390.0290.007
AR250.5510.151−0.084−0.0690.109SG28−0.0840.512−0.1930.2230.115
AR290.534−0.004−0.068−0.076−0.099SG32−0.0800.557−0.0010.029−0.112
AR330.610−0.080−0.022−0.0570.069SG360.1000.600−0.0630.182−0.170
AR370.666−0.0030.169−0.038−0.195SG40−0.1120.5130.024−0.018−0.155
AR410.474−0.0040.1420.039−0.108SG440.1480.411−0.0360.128−0.279
AR170.090.218−0.0790.1970.155SG40.104−0.1290.1700.0520.357
VV3−0.0550.0610.407−0.0570.274SI2−0.151−0.043−0.0020.4420.094
VV70.069−0.1100.4800.2400.077SI60.0390.1790.1570.310−0.213
VV11−0.142−0.1330.5600.058−0.200SI14−0.1500.0540.2180.4500.178
VV15−0.177−0.0470.4770.1420.040SI180.145−0.0300.0250.480−0.173
VV190.0850.1200.483−0.0140.084SI22−0.0710.2080.1710.3370.003
VV270.0520.0170.600−0.0880.130SI30−0.1800.046−0.0240.4620.269
VV310.082−0.1300.6290.079−0.157SI34−0.0700.098−0.1440.550−0.084
VV350.0530.0700.3190.1250.126SI38−0.064−0.0380.1410.3640.120
VV390.240−0.0230.3210.262−0.257SI100.054−0.1150.1450.1410.559
VV230.1130.0620.1260.5040.128SI26−0.122−0.0310.0210.2970.425
VV430.1530.0730.0440.367−0.248SI42−0.1120.072−0.2990.245−0.345
Table 5. The distribution of loading ( > 0.3 ) of the 5-factor solution. The rotation method used is Oblimin with Catell’s scree test. Rotation converged in 11 iterations.
Table 5. The distribution of loading ( > 0.3 ) of the 5-factor solution. The rotation method used is Oblimin with Catell’s scree test. Rotation converged in 11 iterations.
Learning StylesFactors
12345
AR10
SG 10 1
VV 92
SI 83
Table 6. Factors in the five factor solution.
Table 6. Factors in the five factor solution.
FactorsItemsLabelFactors Explained
11, 5, 9, 13, 21, 25, 29, 33, 37, 41, 17ARAction first or reflection first
217, 8, 12, 16, 20, 24, 28, 32, 36, 40, 44SGLinear VS sequential or random VS holistic thinking
33, 8, 11, 15, 19, 27, 31, 35, 39VVInformation format preferred as input or memory
423, 43, 2, 6, 14, 18, 22, 30, 34, 37SIInformation format preferred and preference of concrete or abstract information
54, 10, 26, 42SI-SGConceptual VS factual and detail VS theme
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mirza, M.A.; Khurshid, K.; Shah, Z.; Ullah, I.; Binbusayyis, A.; Mahdavi, M. ILS Validity Analysis for Secondary Grade through Factor Analysis and Internal Consistency Reliability. Sustainability 2022, 14, 7950. https://doi.org/10.3390/su14137950

AMA Style

Mirza MA, Khurshid K, Shah Z, Ullah I, Binbusayyis A, Mahdavi M. ILS Validity Analysis for Secondary Grade through Factor Analysis and Internal Consistency Reliability. Sustainability. 2022; 14(13):7950. https://doi.org/10.3390/su14137950

Chicago/Turabian Style

Mirza, Munazza A., Khawar Khurshid, Zawar Shah, Imdad Ullah, Adel Binbusayyis, and Mehregan Mahdavi. 2022. "ILS Validity Analysis for Secondary Grade through Factor Analysis and Internal Consistency Reliability" Sustainability 14, no. 13: 7950. https://doi.org/10.3390/su14137950

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop