Next Article in Journal
The Creativity Diamond—A Framework to Aid Creativity
Previous Article in Journal
Influence of Teaching Styles on the Learning Academic Confidence of Teachers in Training
Previous Article in Special Issue
The Behavioral, Emotional, and Social Skills Inventory (BESSI): Psychometric Properties of a German-Language Adaptation, Temporal Stabilities of the Skills, and Associations with Personality and Intelligence
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Big Five-Based Multimethod Social and Emotional Skills Assessment: The Mosaic™ by ACT® Social Emotional Learning Assessment

1
ACT, Inc., Iowa City, IA 52243, USA
2
Research and Assessment Design (RAD) Science Solution, Philadelphia, PA 19146, USA
*
Author to whom correspondence should be addressed.
J. Intell. 2022, 10(4), 72; https://doi.org/10.3390/jintelligence10040072
Submission received: 25 March 2022 / Revised: 29 August 2022 / Accepted: 7 September 2022 / Published: 20 September 2022

Abstract

:
A focus on implementing social and emotional (SE) learning into curricula continues to gain popularity in K-12 educational contexts at the policy and practitioner levels. As it continues to be elevated in educational discourse, it becomes increasingly clear that it is important to have reliable, validated measures of students’ SE skills. Here we argue that framework and design are additional important considerations for the development and selection of SE skill assessments. We report the reliability and validity evidence for The Mosaic™ by ACT® Social Emotional Learning Assessment, an assessment designed to measure SE skills in middle and high school students that makes use of a research-based framework (the Big Five) and a multi-method approach (three item types including Likert, forced choice, and situational judgment tests). Here, we provide the results from data collected from more than 33,000 students who completed the assessment and for whom we have data on various outcome measures. We examined the validity evidence for the individual item types and the aggregate scores based on those three. Our findings support the contribution of multi-method assessment and an aggregate score. We discuss the ways the field can benefit from this or similarly designed assessments and discuss how the assessment results can be used by practitioners to promote programs aimed at stimulating students’ personal growth.

1. Introduction

In the past 25 years, social and emotional learning (SEL) has demonstrated tremendous growth and continues to gain popularity. There is growing consensus that a number of factors outside of cognitive ability may be nearly, or just, as important for educational and workplace success (e.g., Aspen Institute 2018; Burrus et al. 2017; Durlak et al. 2015; Kautz et al. 2014). These factors are often termed social and emotional (SE) skills, which can be defined as, “individual capacities that (a) are manifested in consistent patterns of thoughts, feelings, and behaviours, (b) can be developed through formal and informal learning experiences, and (c) influence important socioeconomic outcomes throughout the individual’s life” (OECD 2015, p. 34). SE skills predict a variety of important outcomes, which include, but are not limited to, health (Bogg and Roberts 2004), academic performance (Chernyshenko et al. 2018; Mammadov 2021; Poropat 2009), and job performance (Barrick et al. 2001; Zell and Lesick 2021).
As a result of this growing consensus, the implementation of SEL curricula outpaces the development of assessments through which SE skills can be measured (e.g., Denham 2015). Clearly, reliable and valid measures of students’ SE skills are key for tracking student growth and for studying intervention efficacy. Beyond the evidence for reliability and validity, we encourage developers and users to consider two additional features of the SE skill measures—the framework on which the measure is based and the nature of the items that comprise the measure. Though a vast amount of SE skill frameworks exist (see Jones et al. 2016), some frameworks enjoy more empirical backing than others. Likewise, some item types may be more or less advantageous in certain situations. Here, we discuss one particular SE skills assessment, The Mosaic™ by ACT® Social Emotional Learning Assessment (formerly known as ACT® Tessera® and referred to as Mosaic in the remainder of this paper), as an example of an assessment that makes use of an empirically backed framework and a multimethod approach.

2. The Big Five as an Organizing Framework for SE Skills

The Big Five personality trait model is guided by the lexical hypothesis, which assumes that important individual differences will become encoded into language as single terms (Goldberg 1993). The model was born after researchers turned to an English language dictionary and used factor analytic techniques (e.g., Cattell 1943; Tupes and Christal 1961) to uncover five replicable factors. These factors are commonly labeled Extraversion, Agreeableness, Conscientiousness, Emotional Stability, and Openness to Experience, and are often referred to as the Big Five (de Raad and Mlačić 2015). Subsequently, the same five factors have been found in other languages using various methods (e.g., McCrae et al. 2005; Schmitt et al. 2007).
Conceptually, the Big Five can be considered a “Rosetta Stone” for understanding SE skills (Martin et al. 2019). Using the Big Five, we can take constructs expressed as time management in one framework and grit in another, and understand their connectedness as components of Conscientiousness. Soto and colleagues (2022) provide a comprehensive model of SE skills that resembles the Big Five and have constructed a psychometrically sound SE skill inventory based on this model. Additional recent work speaks to the empirical merit of the Big Five-SE skill links. For example, Walton and colleagues (2021) carried out two studies using different methodologies to evaluate this. In one, they asked subject matter experts from the fields of personality psychology and SEL to rate the degree of overlap between the Big Five traits and popular SE skills. As anticipated, the traits and skills were deemed to overlap to a significant degree. In another study, they factor analyzed items from a Big Five measure along with items from a CASEL (the most prominent SEL framework in the US) competency-based measure, and they determined that the best fitting model was one with five factors that aligned with the Big Five traits.
The Big Five is also an ideal candidate for SE skills assessments because of its consistency with SE skills definitions. Recall that the OECD’s (2015) definition states that SE skills influence important outcomes. There is a vast body of research linking the Big Five with many critical outcomes. For example, a meta-analysis by Poropat (2009) found that, during the primary school years, the cognitive ability’s impact on academic performance exceeds that of any Big Five trait, but by secondary education, Conscientiousness is nearly as important for academic performance as cognitive ability. Another recent meta-analysis also found that Conscientiousness was a strong predictor of academic performance across all levels of schooling and predicted performance incrementally above-and-beyond cognitive ability (Mammadov 2021).
As another example of why the five-factor model is a good choice of framework for SE skills, the OECD’s (2015) definition states that such skills can be developed. In their meta-analysis of mean-level personality change, Roberts et al. (2006) found that individuals exhibit development throughout their lifespan. Furthermore, meta-analyses demonstrate that interventions can alter personality traits significantly even when the interventions last for short periods of time (Hudson et al. 2019; Roberts et al. 2017). Similarly, meta-analyses speak to the effectiveness of SEL interventions. Students who participate in SEL programs see greater gains in SE skills and academic performance versus students who do not participate, and both short- and long-term benefits have since been documented (e.g., Corcoran et al. 2018; Mahoney et al. 2018).
Due to its utility in conceptualizing SE skills and growing empirical support in the field, the Big Five has been adopted as an organizing framework for SE skills in prior literature (Abrahams et al. 2019; Casillas et al. 2015; Soto et al. 2022) and other work. Notably, the OECD’s Study on Social and Emotional Skills, which is a large-scale international study, uses the framework (Chernyshenko et al. 2018; Kankaraš and Suarez-Alvarez 2019).

3. Methodological Approaches to the Assessment of Social and Emotional Skills

Several item types have been used to assess SE skills. For example, Likert items have been used in SEL research for decades and are known for their efficiency, reliability, and validity. Individuals indicate their agreement with several statements (e.g., “I work hard”). This item type is preferred when stakes are low and faking is not expected (Lipnevich et al. 2013). However, respondents may have various motives to fake, such as appearing more attractive to a school admissions officer (Zickar et al. 2004). Furthermore, Likert items might be particularly susceptible to reference effects. That is, people often answer such items by asking the question, “compared with whom?” For example, it could be the case that students from very high achieving schools might rate themselves lower on their SE skills than students from low achieving schools simply because they are using a different reference group, not because they are truly lower on these skills (e.g., Marsh and Hau 2003).
Forced choice (FC) items have been used to assess individuals’ tendencies and capabilities as well. With FC items, statements are grouped in blocks and respondents make selections within each block regarding which statements describe him or her best. There are several variations of the FC methodology (Hontangas et al. 2015). For example, respondents may read multidimensional triads and choose, of the three items (e.g., “I enjoy leading class discussions”, “I work hard in school to achieve my goals”, and “I like working in a team”), which is most like them and which is least like them. Reference bias should be minimized with FC because respondents conduct an internal (self vs. self) rather than an external (self vs. other) comparison on these items. Furthermore, there are compelling studies (Christiansen et al. 2005; Jackson et al. 2000; Walton et al. 2021) and meta-analytic data (Cao and Drasgow 2019) suggesting that FC items cannot be faked as easily as Likert items. For example, Christiansen et al. studied responses across two conditions, one in which participants were instructed to answer honestly and one in which participants were instructed to answer as if they were applying for a job and trying to make a positive impression. The effect size (i.e., standardized mean difference in the responses across the honest and positive impression conditions) was just .24 for the FC items, but the effect reached .96 for the single stimulus items.
A third item type used in this domain is the situational judgment test (SJT). In this approach, students imagine themselves in a scenario and then read various behavioral responses to that scenario. They then indicate either how likely they would be to respond in that manner or are asked to select how a person should respond in that scenario. SJTs have several advantages. First, SJTs may be developed to reflect more complex judgment processes than what is possible with Likert scales. Second, SJTs have the advantage of face validity (Lievens et al. 2008). The situations presented to students have the look and feel of those that would be encountered in real life. Third, there is evidence suggesting that SJTs are less prone to faking than Likert items (Hooper et al. 2006), and some have argued for using them in high-stakes settings for this reason (Whetzel and McDaniel 2016). Fourth, SJTs should be less susceptible to reference effects than Likert items because behavioral responses to scenarios are not comparative in nature. They have been used for decades for employee selection, and now even the Association of American Medical Colleges (2022) has developed an SJT measure to assess medical students’ knowledge about effective behaviors.
Given that all item types have different strengths and weaknesses, an assessment employing multiple methods (i.e., one that conforms to a multi-trait multi-method [MTMM] design) ought to minimize the effects of any biases or shortcomings. According to Kenny and Kashy (1992), “The underlying view of measurement in the MTMM analysis is that to measure a theoretical construct, different measures, each with its own bias, are selected. Bias that is due to method effects is reduced through a triangulation process” (p. 170). To our knowledge, there is no research specifically examining the validity evidence for a MTMM SE skills assessment. It is worth examining whether the inclusion of varied item types contributes to the prediction of important outcomes.

4. Current Study

In this study, we briefly describe Mosaic, a multi-method SE skills assessment based on the Big Five framework. In addition to determining the extent to which the Big Five can account for the variance in important school outcomes, we sought to examine whether Mosaic’s MTMM approach leads to stronger validity evidence than what would be available from the individual item types. We anticipated our findings could help inform framework and methodological considerations for future SE skills assessments.

5. Method

Participants

During the 2018–2019 academic year, 24,400 students from 160 schools completed Mosaic Middle School (MS). The demographic information was provided by the schools. The grade level breakdown was: 6th = 3864 (15.8%), 7th = 17,585 (72.1%), and 8th = 2951 (12.1%). The gender breakdown was: female = 12,273 (50.3%) and male = 12,127 (49.7%). Nine thousand one hundred twelve students from 93 schools completed Mosaic High School (HS). The grade level breakdown was: 9th = 5413 (59.4%), 10th = 1739 (19.1%), 11th = 1002 (11.0%), and 12th = 958 (10.5%). The gender breakdown was: female = 4792 (52.6%) and male = 4320 (47.4%). Race/ethnicity data were not available for most participants, and SES is not available for any students. Some data were missing for some schools or students so not all Ns are identical across the measures.
The students were enrolled in schools that purchased the assessment as part of their SEL programming; that is, the students completed the assessment as part of their regular school activities and were not solicited for participation in the research. Students gave their assent before completing the assessment by checking “yes” to an item worded “By clicking YES, I assent to participate in this activity”. The schools were spread across the U.S. and were a mixture of public, private, and charter schools. Schools allotted one class period for completion and most students were able to complete the assessment during that time but were able to take longer than one class period if necessary.

6. Measures

6.1. Mosaic™ by ACT® Social Emotional Learning Assessment

Constructs. The SE skills assessed by Mosaic (ACT 2021) can be aligned with the Big Five constructs on a one-to-one basis (Table 1). This alignment was conducted rationally by subject matter experts comparing the Big Five factors’ definitions and Mosaic item content and skill definitions. The Mosaic skill labels are different from the Big Five trait labels in accordance with educator preference. For example, focus group feedback indicated that the term Social Connection is preferred over Extraversion. The different labels notwithstanding, the construct breadth is similar to the Big Five. For example, the Social Connection items measure more than that single facet of Extraversion; they also tap into assertiveness, leadership, persuasiveness, etc.
Item Types. Mosaic employs the MTMM approach discussed above with Likert, FC, and SJT items. An example Likert item measuring Getting Along with Others is: I offer to help those who need assistance. Students indicate their level of agreement with this statement on a six-point scale from Strongly Disagree to Strongly Agree. There are 40 Likert items with eight per construct. The descriptive statistics and reliability estimates, all of which exceed .70 for MS and HS forms, can be found in Table 2.
Below is an example of a multidimensional FC triad. Students are instructed to select the statement that is most like them and the one that is least like them. In the example below, the statements measure, respectively, Keeping an Open Mind, Sustaining Effort, and Getting Along with Others.
  • I perform well on assignments that require me to use my imagination to find the answers.
  • If I tell my teacher I will do something, I do it.
  • I ignore classmates who are being left out of class discussions.
There are ( 5 3 ) = 10 FC triads. Each possible triad occurred once and only once, so there are a total of 30 items with six per construct. Cronbach’s alpha values (see Table 2) were lower than what is typically found with Likert items, which is to be expected given their ipsative nature (Meade 2004).
Mosaic includes two SJTs per construct yielding 10 total items. Again, Cronbach’s alpha values (see Table 2) were lower than what is typically found with Likert items, given that they are often multidimensional (Whetzel and McDaniel 2009). Students read each scenario then are presented with five behavioral responses, all of which measure a single construct, and they are asked to indicate how likely they are to enact each response on a five-point scale from Very Unlikely to Very Likely. An example of a partial SJT measuring Sustaining Effort reads:
After studying very hard for a math test, the test results are disappointing, and you have yet to do as well as expected. While you are currently proficient, you would like to move up to the next level.
Two of the behavioral responses read:
  • Look over the test to see what questions you got wrong and work on those.
  • Decide there’s no point to studying so hard if you don’t get the results you want.
Scoring. Negatively worded items across the three item types were reverse scored. The Likert and SJT items were scored in the typical manner by averaging across all responses to yield a total score per construct where higher scores are aligned with higher skill levels. The FC items were scored such that the statement that is selected as ‘most like me’ was scored with a 3, the statement not selected was scored with a 2, and the statement selected as ‘least like me’ was scored with a 1, then the values were averaged across the responses to yield a total score per construct where higher scores were aligned with higher skill levels.
To create aggregate scores across the three item types, the Likert, FC, and SJT scores were standardized then averaged. It is important to note that it is strongly preferred to derive a single score per skill for practical purposes. When reporting SE skill scores to students, parents, and educators, this is a more parsimonious approach rather than reporting three scores per skill. Thus, it is not our goal to examine the incremental validity of the item types; that is, below we compare the validity evidence for the single item types and aggregate scores, rather than building a hierarchical model and examining the incremental validity.

6.2. Test-Criterion Data

Below we describe each of the outcome variables that we obtained from the sub-samples of the 33,512 students completing either Mosaic MS or HS.
School Climate. To measure school climate, we used the Relationships with School Personnel and School Safety Climate (hereafter referred to as Relationships and Safety, respectively) scales from ACT Engage®. The scales are supported by strong evidence of reliability and validity (ACT 2016) and have high internal consistency estimates. There were 12 Relationship climate items (Cronbach’s alpha MS = .86 and HS = .88) and 11 Safety climate items (Cronbach’s alpha MS = .84 and HS = .85), all of which are Likert items rated on a six-point agreement scale. These items were identical across the two forms. All Relationships items make mention of how a student perceives his or her relationships with adults at school (e.g., There are adults at my school who care about me). The Safety items primarily focus on students’ perceptions of their physical safety in school (e.g., I feel safe at school). The climate scores used in the current study reflect each student’s individual perceptions. Prior research has shown that SE skills and favorable perceptions of school climate are positively related to one another (Allen et al. 2019; Osher and Berg 2017).
Academic Performance. GPA was self-reported on a 12-point scale (e.g., A+, 97–100%; A, 93–96%; etc.). Meta-analytic data (Mammadov 2021; Poropat 2009) indicate that the Sustaining Effort skill is expected to have the strongest relationship with GPA.
Attendance. Four school districts provided additional MS student data, including absences and disciplinary infractions, and 12 districts reported these data for HS students. We examined the association between Mosaic scores and absenteeism. We used the sum of unexcused and excused absences as a continuous variable and, when appropriate, also split the sample into three groups representing students with chronic absenteeism (defined as 18 or more missed days), habitual absenteeism (defined as at least 10 missed days but fewer than 18), or acceptable absenteeism (fewer than 10 missed days) and examined group mean differences. Most states describe chronic absenteeism as missing 10 percent or more days within a year (Attendance Works n.d.), which equates to 18 or more days, while some states consider missing 10 or more days within a school year as being habitually truant (Colorado Department of Education n.d.). Based on the findings reported by Lounsbury et al. (2004), we expected Keeping an Open Mind to have the strongest association with absenteeism and that they would be negatively related. However, the opposite effects have been found in college students (Chamorro-Premuzic and Furnham 2003; Credé et al. 2010), so this is somewhat of an exploratory analysis.
Discipline. Finally, we considered the number of disciplinary infractions as recorded in school records. For some schools, we had a count of the actual number of disciplinary infractions, but in some cases, schools only provided a binary response (i.e., no infractions vs. at least one). For some analyses, we dichotomized the continuous responses that some schools had provided and combined them with the binary responses. Research on children and adolescents (Tackett 2006; Tackett et al. 2013) led us to anticipate negative associations between disciplinary infractions and Sustaining Effort, Getting Along with Others, and Maintaining Composure. We anticipated positive, yet smaller, associations between the disciplinary infractions and Keeping an Open Mind and Social Connection.

7. Results

7.1. Convergent and Discriminant Validity

The complete MTMM correlation matrix can be found in Table 2. There is evidence for convergent validity with the Likert scales generally correlating most highly with their respective SJT and FC scales and the SJT scales correlating most highly with their respective FC scales. For example, in MS, the correlation between the Likert Maintaining Composure scores and the SJT Maintaining Composure scores (r = .54) was higher than between the Likert Maintaining Composure scores and any of the other four SJT scores (r = .28–.47). However, evidence for discriminant validity was not as strong, as there were several instances of off-trait correlations being higher than expected. For example, in MS, the correlation between the Likert Sustaining Effort scores and the SJT Maintaining Composure scores reached .59. There were fewer instances of this occurring in HS than in MS.

7.2. Test-Criterion Validity

For the MS and HS forms, we examined the test-criterion validity evidence by correlating the SE skills with the outcome measures. We also regressed the outcomes on the SE skill scales. We did this for each item type individually and compared the findings to those for the aggregate skill scores.

7.2.1. School Climate

As expected, all correlations between SE skills and school climate were positive (Table 3). MS associations were slightly stronger than those observed in HS, and SE skills had slightly higher correlations with Relationships than with Safety. For both MS and HS forms and both climate scales, either the Likert or the aggregate scores had the strongest correlations. The Likert scales (MS Relationships R2 = .38; MS Safety R2 = .24; HS Relationships R2 = .28; HS Safety R2 = .16) accounted for slightly more variance than the aggregate sores (MS Relationships R2 = .36; MS Safety R2 = .22; HS Relationships R2 = .25; HS Safety R2 = .14).

7.2.2. Academic Performance

For both forms, Sustaining Effort had the strongest relationship with GPA (Table 3). However, we should note that all skills were related to academic performance, a finding not typically reported in prior research (Mammadov 2021; Poropat 2009). We suspect this is due to the fact that many of our items are contextualized to a school setting, which may lead to higher correlations than what is found with “generic” Big Five measures. Across skills, the aggregate scores generally had the strongest correlations with GPA. In MS and HS, the Likert (MS R2 = .17; HS R2 = .23) and aggregate scores (MS R2 = .18; HS R2 = .23) were on par with one another in terms of variance in GPA accounted for.

7.2.3. Attendance

We examined correlations between the number of absences and the five SE skills (Table 3). In MS, the correlations were all near zero and there was no clear pattern in terms of which item type exhibited the strongest association. In HS, the correlations were generally small and negative, and the SJTs or aggregate scores had the strongest associations. The individual item types and the aggregates did not count for much variance in absenteeism (i.e., no more than 4%). To evaluate whether the effects varied across item types and aggregate scores, we split the HS sample into three groups—students with chronic, habitual, or acceptable absentee records. All effect sizes were small, though the SJT and aggregate effect sizes were slightly higher overall (Table 4).

7.2.4. Discipline

We examined the correlations between the number of reported disciplinary infractions and the five SE skills (Table 3). In MS, the correlations were in the expected direction. In HS, the Social Connection correlation was higher than expected, relative to those for Sustaining Effort, Getting Along with Others, and Maintaining Composure. The SJT and aggregate associations were generally the strongest. In MS, the SJTs accounted for 12% of the variance in disciplinary infractions, while the aggregate scores accounted for 9%. In HS, however, the individual item types and the aggregate scores were fairly even in terms of variance accounted for, with values ranging from 4% to 6%.
As previously stated, some schools only provided a binary response (i.e., no infractions vs. at least one). We dichotomized the continuous responses that some schools had provided and combined them with the binary responses and carried out independent samples t-tests to compare these students and determine if results differed across item type. Results are reported in Table 5. Any differences in the direction of the effect across the binary and continuous analyses are due to non-identical samples used across the analyses. In MS, the SJT and aggregate scores had the greatest effect sizes. The results were more varied in HS, but generally the SJTs had the greatest effect sizes.

8. Discussion

The MS and HS Mosaic forms were developed using the Big Five as its assessment framework and a multi-method approach to measure SE skills. We primarily focused on the validity evidence for three item types (Likert, FC, and SJT) and an aggregate score based on these three. Overall, the associations between the five SE skills, regardless of item type, and the outcome measures were as expected. For example, all skills were positively associated with school climate dimensions. We found support for the use of an aggregate score and determined that, depending on the particular outcome of interest, certain item types may provide stronger test-criterion validity evidence.

8.1. Methodological Considerations

The Likert item scores and aggregate scores were the best predictors of perceptions of school climate and GPA. The SJT-based and aggregate scores showed the strongest effects in terms of attendance and discipline, although effects overall were smaller than what we observed for climate and GPA. The FC items failed to prevail as the strongest predictor for any of the outcomes we examined, though they were significant predictors of all outcomes except for attendance in MS. Note that unlike HS, there was no evidence of relationships between SE skills and absenteeism in MS, which likely reflects a greater autonomy among HS students relative to MS students.
It is not surprising that Likert items accounted for more variance in perceptions of school climate. The climate items were also Likert items and therefore the shared method variance is likely at play here. We can conjecture as to why SJTs had the strongest link with school attendance and discipline. As discussed above, SJTs have the “look and feel” of situations encountered in real life. Of the five outcomes we assessed, attendance and discipline are the most behavioral. That is, you attend school or you do not, and you get in trouble or you do not. These types of behaviors are written into SJTs. For example, one Mosaic SJT provides the possible behavior response of “Angrily confront your coach…” This is an example of a behavior that would likely lead to some disciplinary action. GPA is not as explicitly behavioral. There are multiple processes involved that lead to an academic measure and it’s a more distal outcome. That is, GPA stems from a combination of behavior, cognitive ability, course taking, and teacher judgment. As such, it’s not as pure as a behavioral measure. This may explain why the Likert items did a better job explaining the variation in student academic performance. As mentioned above, although significant predictors of outcomes, the FC items were not “the best” in terms of predicting the outcomes. This may be due to the low reliability of the scores, which, again, is typical for ipsative FC scores (Meade 2004). Despite this artefact of FC scores, FC items have other advantages, such as making it difficult for respondents to engage in impression management (Cao and Drasgow 2019), which might be particularly attractive in high-stakes settings.
In addition to the possibility of reducing the impression management with the FC items and the strong face validity of SJTs, there are other benefits to using a multi-method approach. As discussed above, reference effects and other response tendencies, such as acquiescent responding, might be minimized by using alternatives to Likert items. Another factor that has gone unmentioned until this point is the increased level of engagement respondents may experience when encountering more than one item type in an assessment. All of these factors, along with the evidence for improved validity from the current study, reinforce the notion that a multi-method approach to SE skills assessments has merit.
Many other assessments rely solely on Likert items to obtain scores (e.g., Davidson et al. 2018; LeBuffe et al. 2018). We believe the review and findings presented here provide compelling evidence supporting the inclusion of multiple item types and aggregate scores. This approach is novel, but it is slowly being adopted by others in the field. For example, the OECD’s Study on Social and Emotional Skills also uses an MTMM approach to generate SE skill scores (Kankaraš et al. 2019). In contrast to our approach, which involves multiple student-report item types, the OECD study combined Likert scales from parents, teachers, and students.

8.2. Framework Considerations

Mosaic differs from many SE skills assessments in terms of its organizing framework. This assessment makes use of the Big Five framework, which is advantageous because it is evidence-based, has cross-cultural generalizability, and consistently replicates across age groups. In addition, it can be used as a framework through which to organize the various terms used to describe SE skills (Walton et al. 2021). One issue plaguing the field of SEL is the abundance of frameworks currently used to organize skills, with a recent study identifying 136 frameworks (Berg et al. 2017). Moving toward a single assessment framework can help to unify the field and integrate research findings across research groups. A unified framework should be comprehensive yet parsimonious, consider developmental implications, and be evidence based and data driven, rather than derived from theory or expert consensus alone. The Big Five stands out as the organizing framework that fits each of these recommendations.
Despite support for using the Big Five framework to organize SE skills, it is imperative to recognize that personality traits and SE skills are not synonymous with one another. Once again, consider the OECD’s (2015) definition; it refers to SE skills as “individual capacities.” Capacities and tendencies are distinct, and herein lies the key differentiator between SE skills and personality traits. Soto and colleagues (e.g., 2021) are careful to make this point. While they use the Big Five framework for their assessment of social, emotional, and behavior skills, the BESSI, they reframe the item stems to ask respondents to report how well they are able to do certain things rather than how likely or how often they do those things (e.g., How well can you perform in a leadership role? vs. How likely are you to take on a leadership role?; Soto et al. 2022).

8.3. Applications for Educational Contexts

Research suggests that SE skills predict an array of desirable outcomes including success in school. Obtaining measures of these skills is key to fostering school-wide SEL initiatives. Practitioners can use data from SE skill assessments as part of a formative assessment system. Assessments can be considered formative if the feedback from the assessment is used to guide practices or make changes that increase student learning (Black and Wiliam 2009). An assessment itself is not necessarily formative on its own, but the process of a formative assessment entails multiple learning components working together to improve student learning. Mosaic was designed as part of a larger assessment system, with the purpose of using assessment results to drive intervention and aim instruction at areas of need. Assessments can help schools drive universal interventions and targeted SEL initiatives. The Mosaic system provides individual student reports, following Hattie and Timperley’s (2007) recommendations for delivering formative feedback to students; reports inform students of their current levels on each skill, provide examples of the target skill level, and include recommendations on how students can improve. Mosaic also provides aggregate school-level reports so educators and administrators can be informed of broad areas of need in their schools and districts.

9. Limitations

This study was limited in a few ways. First, the students were not equally distributed across grades with an oversampling of 7th grade students in MS and 9th grade students in HS. This was an artefact of how schools opted to use Mosaic (i.e., to which students they administered Mosaic) and was not in our control. In addition, school-reported outcomes were limited. We relied on self-reported GPA, and the subsamples with school-reported discipline and attendance data were relatively small. Therefore, generalizations are limited. Additional data could be collected, particularly school-reported outcomes, to provide further validity evidence for the assessments. Moreover, while we focused primarily on the test-criterion validity evidence, additional validity evidence (e.g., construct validity) is also important. Assessment validation should be considered a continually ongoing process, with additional sources of evidence needed to strengthen the validity argument.
Another limitation was that we did not have access to participants’ SES or race/ethnicity status for most students. The exclusion of race/ethnicity data was not intentional, but rather a limitation induced from the assessment’s administration procedure during the 2018–2019 school year. During this administration window, schools were asked to provide students’ race/ethnicity information during the registration process. Reporting student race/ethnicity was optional, not required. We found that most schools chose not to provide students’ race/ethnicity information. Following that school year, we changed the assessment process to have students self-report their race/ethnicity during the test administration rather than relying on schools to do so during the registration process. This change in administration has resulted in more complete student race/ethnicity information beginning in the 2019–2020 school year.
However, we did have race/ethnicity data for 2% (n = 503) of the MS students and 18% (n = 1649) of the HS students, and we compared white students with underrepresented minority students. The groups scored significantly differently from one another on just one SE skill, Keeping an Open Mind, with the underrepresented students scoring higher in both MS and HS. We also note that race/ethnicity information was available during earlier phases of the assessment development, and the parameter estimates for scoring were extracted from a diverse and representative sample. Furthermore, we continue to gather assessment data complete with student-reported race/ethnicity information. With more recent assessment data (N = 3720), we have demonstrated that race/ethnicity subgroup differences are consistently small in magnitude across the three item types, suggesting fairness for all test-takers (Walton and Burrus 2022). DIF analyses across subgroups is also a component of our ongoing test validation process as our body of available assessment data continues to grow.
A final area for future work pertains to the discriminant validity issue raised above; that is, particularly in MS, there were several instances of weak discriminant validity. This could be due to a number of reasons. First, as discussed previously, SJT measures are often multidimensional. In line with this, the correlations involving SJT scores was typically where there were instances of weak discriminant validity. Second, we observed this more among the MS sample than the HS sample, and this is consistent with prior work showing that self-reports become more coherent within a skill domain and better differentiated across domains from late childhood to early adulthood (Soto et al. 2008). Soto and colleagues noted that the domains related to Getting Along with Others and Sustaining Effort showed large gains in the differentiation with age but small gains in coherence, and they conjectured that this differentiation could be the result of complex notions of right and wrong or good and bad. One possible implication of this pertains to impression management; that is, as children age, they should be more adept with impression management and able to present a coherent picture of a “good” respondent. This might suggest that the inclusion of the FC items in particular would be recommended for older students. Finally, this research (Soto et al. 2008) has shown that there is a high variability in the acquiescent responding among younger study participants. This may further explain the lack of differentiation among the younger students in the current sample and further point to the need for item types other than the Likert items for younger students.

10. Conclusions

Reliable and valid assessments are paramount for the growth of the field, as assessments enable practitioners to measure progress and maximize student SE skill development. Here, we argue that the assessment framework and response method are also important considerations for the development or selection of a SE skill assessment. Mosaic is one of a growing number of SE skills assessments designed to measure the skills that crosswalk to the Big Five personality framework. The current study offers additional support for this practice given that the skills accounted for a statistically significant amount of variance in important school-related constructs and outcomes. Moreover, the findings reported here highlight the importance of shying away from automatically using Likert-based assessments and considering more innovative item types either alone or in combination. Our findings suggest that the aggregate scores derived from multiple item types are often superior over single item types in terms of predictive validity, and that the single item type with the best validity evidence varies according to the outcome in question. It is often said that we measure what matters, and it is unequivocal that these five SE skills matter. By purposefully leveraging a multi-method SE skill assessment in schools, we can equip students with the skills they need to thrive in school, in the workforce, and beyond.

Author Contributions

Conceptualization, K.E.W., J.B. and R.D.R.; methodology, K.E.W., J.B., R.D.R. and D.M.; validation, K.E.W., J.B., R.D.R., D.M., C.A.-C. and J.W.; formal analysis, K.E.W. and D.M.; resources, J.B. and R.D.R.; writing—original draft preparation, K.E.W., J.B. and D.M.; writing—review and editing, K.E.W., J.B., R.D.R., D.M., C.A.-C. and J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This research is exempt from IRB review according to section 46.104 of the Office for Human Research Protections (OHRP) regulations for the protection of human subjects.

Informed Consent Statement

Student assent was acquired prior to students completing the assessment.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to contractual and privacy issues.

Conflicts of Interest

The authors declare no conflict of interest. The views, opinions, and interpretation of the findings contained in this article are solely those of the authors and do not purport to represent the views of the researchers’ host institutions.

References

  1. Abrahams, Loes, Gina Pancorbo, Ricardo Primi, Daniel Santos, Patrick Kyllonen, Oliver P. John, and Filip De Fruyt. 2019. Social-emotional skill assessment in children and adolescents: Advances and challenges in personality, clinical, and educational contexts. Psychological Assessment 31: 460–73. [Google Scholar] [CrossRef]
  2. ACT. 2016. Development and Validation of ACT Engage®: Technical Manual. Iowa City: ACT, Inc. [Google Scholar]
  3. ACT. 2021. Mosaic by ACT®: Social Emotional Learning Assessment Technical Manual. Iowa City: ACT, Inc. [Google Scholar]
  4. Allen, Jeff, Jason D. Way, and Alex Casillas. 2019. Relating school context to measures of psychosocial factors for students in grades 6 through 9. Personality and Individual Differences 136: 96–106. [Google Scholar]
  5. Aspen Institute. 2018. From a Nation at Risk to a Nation at Hope: Recommendations from the National Commission on Social, Emotional, and Academic Development. Washington: The Aspen Institute. [Google Scholar]
  6. Association of American Medical Colleges. 2022. AAMC PREviewTM Professional Readiness Exam Now Available for 2023 Admissions Cycle. Available online: https://students-residents.aamc.org/aamc-preview/aamc-preview-professional-readiness-exam-now-available-2023-admissions-cycle (accessed on 1 March 2022).
  7. Attendance Works. n.d. Chronic Absence. Available online: https://www.attendanceworks.org/chronic-absence/the-problem/ (accessed on 1 March 2022).
  8. Barrick, Murray R., Michael K. Mount, and Timothy A. Judge. 2001. Personality and performance at the beginning of the new millennium: What do we know and where do we go next? International Journal of Selection and Assessment 9: 9–30. [Google Scholar]
  9. Berg, Juliette, David Osher, Michelle R. Same, Elizabeth Nolan, Deaweh Benson, and Naomi Jacobs. 2017. Identifying, Defining, and Measuring Social and Emotional Competencies. In American Institutes for Research. Washington: American Institutes for Research. [Google Scholar]
  10. Black, Paul, and Dylan Wiliam. 2009. Developing the theory of formative assessment. Educational Assessment, Evaluation and Accountability 21: 5–31. [Google Scholar]
  11. Bogg, Tim, and Brent W. Roberts. 2004. Conscientiousness and health-related behaviors: A meta-analysis of the leading behavioral contributors to mortality. Psychological Bulletin 130: 887–919. [Google Scholar] [PubMed]
  12. Burrus, Jeremy, Krista Mattern, Bobby D. Naemi, and Richard D. Roberts. 2017. Building Better Students: Preparation for the Workforce. Cambridge: Oxford University Press. [Google Scholar]
  13. Cao, Mengyang, and Fritz Drasgow. 2019. Does forcing reduce faking? A meta-analytic review of forced-choice personality measures in high-stakes situations. Journal of Applied Psychology 104: 1347–68. [Google Scholar]
  14. Casillas, Alex, Jason Way, and Jeremy Burrus. 2015. Behavioral skills. In Beyond Academics: A Holistic Framework for Enhancing Education and Workplace Success. Edited by Ryan O’Connor, Wayne Camara, Krista Mattern and Mary Ann Hanson. Iowa City: ACT, Inc., pp. 25–38. [Google Scholar]
  15. Cattell, Raymond B. 1943. The description of personality: The foundations of trait measurement. Psychological Review 50: 559–94. [Google Scholar]
  16. Chamorro-Premuzic, Tomas, and Adrian Furnham. 2003. Personality predicts academic performance: Evidence from two longitudinal university samples. Journal of Research in Personality 37: 319–38. [Google Scholar]
  17. Chernyshenko, Oleksandr S., Miloš Kankaraš, and Fritz Drasgow. 2018. Social and Emotional Skills for Student Success and Wellbeing: Conceptual Framework for the OECD Study on Social and Emotional Skills. OECD Education Working Papers, No. 17. Paris: OECD Publishing. [Google Scholar] [CrossRef]
  18. Christiansen, Neil D., Gary N. Burns, and George E. Montgomery. 2005. Reconsidering forced-choice item formats for applicant personality assessment. Human Performance 18: 267–307. [Google Scholar]
  19. Colorado Department of Education. n.d. Student Attendance. Available online: http://www.cde.state.co.us/communications/schoolattendance-factsheet-2018 (accessed on 1 March 2022).
  20. Corcoran, Roisin P., Alan C. K. Cheung, Elizabeth Kim, and Chen Xie. 2018. Effective universal school-based social and emotional learning programs for improving academic achievement: A systematic review and meta-analysis of 50 years of research. Educational Research Review 25: 56–72. [Google Scholar]
  21. Credé, Marcus, Sylvia G. Roch, and Urszula M. Kieszczynka. 2010. Class attendance in college: A meta-analytic review of the relationship of class attendance with grades and student characteristics. Review of Educational Research 80: 272–95. [Google Scholar]
  22. Davidson, Laura A., Marisa K. Crowder, Rachel A. Gordon, Celene E. Domitrovich, Randal D. Brown, and Benjamin I. Hayes. 2018. A continuous improvement approach to social and emotional competency measurement. Journal of Applied Developmental Psychology 55: 93–106. [Google Scholar]
  23. de Raad, Boele, and Boris Mlačić. 2015. The lexical foundation of the Big Five-Factor Model. In The Oxford handbook of the Five Factor Model. Edited by Thomas A. Widiger. New York: Oxford University Press, pp. 191–216. [Google Scholar]
  24. Denham, Susanne A. 2015. Assessment of SEL in educational contexts. In Handbook of Social and Emotional Learning: Research and Practice. Edited by Joseph A. Durlak, Celene E. Domitrovich, Roger P. Weissberg and Thomas P. Gullotta. New York: The Guilford Press, pp. 285–300. [Google Scholar]
  25. Durlak, Joseph A., Celene E. Domitrovich, Roger P. Weissberg, and Thomas. P. Gullotta. 2015. Handbook of Social and Emotional Learning: Research and Practice. New York: The Guilford Press. [Google Scholar]
  26. Goldberg, Lewis R. 1993. The structure of phenotypic personality traits. American Psychologist 48: 26–34. [Google Scholar] [CrossRef] [PubMed]
  27. Hattie, John, and Helen Timperley. 2007. The power of feedback. Review of Educational Research 77: 81–112. [Google Scholar] [CrossRef]
  28. Hontangas, Pedro M., Jimmy De La Torre, Vicente Ponsoda, Iwin Leenen, Daniel Morillo, and Francisco J. Abad. 2015. Comparing traditional and IRT scoring of forced-choice tests. Applied Psychological Measurement 39: 598–612. [Google Scholar] [CrossRef] [PubMed]
  29. Hooper, Amy C., Michael J. Cullen, and Paul R. Sackett. 2006. Operational threats to the use of Situational Judgment Tests: Faking, coaching, and retesting issues. In Situational Judgment Tests. Edited by Jeff A. Weekley and Robert E. Ployhart. Mahwah: Lawrence Erlbaum Associates, pp. 205–323. [Google Scholar]
  30. Hudson, Nathan W., Daniel A. Briley, William J. Chopik, and Jaime Derringer. 2019. You have to follow through: Attaining behavioral change goals predicts volitional personality change. Journal of Personality and Social Psychology 117: 839–57. [Google Scholar]
  31. Jackson, Douglas N., Victor R. Wroblewski, and Michael C. Ashton. 2000. The impact of faking on employment tests: Does forced choice offer a solution? Human Performance 13: 371–88. [Google Scholar] [CrossRef]
  32. Jones, Stephanie, Rebecca Bailey, Katharine Brush, Bryan Nelson, and Sophie Barnes. 2016. What Is the Same and What Is Different? Making Sense of the “Noncognitive” Domain: Helping Educators Translate Research into Practice. Harvard Graduate School of Education. Available online: https://easel.gse.harvard.edu/files/gse-easel-lab/files/words_matter_paper.pdf (accessed on 1 March 2022).
  33. Kankaraš, Miloš, and Javier Suarez-Alvarez. 2019. Assessment Framework of the OECD Study on Social and Emotional Skills. Available online: https://www.oecd-ilibrary.org/education/assessment-framework-ofthe-oecd-study-on-social-and-emotional-skills_5007adef-en (accessed on 1 March 2022).
  34. Kankaraš, Miloš, Eva Feron, and Rachel Renbarger. 2019. Assessing Students’ Social and Emotional Skills through Triangulation of Assessment Methods. OECD Education Working Papers, No. 208. Paris: Organization for Economic Cooperation and Development. [Google Scholar]
  35. Kautz, Tim, James J. Heckman, Ron Diris, Bas ter Weel, and Lex Borghans. 2014. Fostering and measuring skills: Improving noncognitive skills to promote lifetime success. In OECD, Educational and Social Progress. Paris: OECD. [Google Scholar]
  36. Kenny, David A., and Deborah A. Kashy. 1992. Analysis of the multitrait-multimethod matrix by confirmatory factor analysis. Psychological Bulletin 112: 165–72. [Google Scholar]
  37. LeBuffe, Paul A., Valerie B. Shapiro, and Jennifer L. Robitaille. 2018. The Devereux Student Strengths Assessment (DESSA) comprehensive system: Screening, assessing, planning, and monitoring. Journal of Applied Developmental Psychology 55: 62–70. [Google Scholar]
  38. Lievens, Filip, Helga Peeters, and Eveline Schollaert. 2008. Situational judgment tests: A review of recent research. Personnel Review 37: 426–41. [Google Scholar] [CrossRef]
  39. Lipnevich, Anastasiya A., Carolyn MacCann, and Richard D. Roberts. 2013. Assessing noncognitive constructs in education: A review of traditional and innovative approaches. In Oxford Handbook of Child Psychological Assessment. Edited by Donald H. Saklofske, Cecil R. Reynolds and Vicki Schwean. Cambridge: Oxford University Press, pp. 750–72. [Google Scholar]
  40. Lounsbury, John W., Robert P. Steel, James M. Loveland, and Lucy W. Gibson. 2004. An investigation of personality traits in relation to adolescent school absenteeism. Journal of Youth and Adolescence 33: 457–66. [Google Scholar] [CrossRef]
  41. Mahoney, Joseph L., Joseph A. Durlak, and Roger P. Weissberg. 2018. An update on social and emotional learning outcome research. Phi Delta Kappan 100: 18–23. [Google Scholar]
  42. Mammadov, Sakhavat. 2021. Big Five personality traits and academic performance: A meta-analysis. Journal of Personality. [Google Scholar] [CrossRef]
  43. Marsh, Herbert W., and Kit-Tai Hau. 2003. Big-fish--little-pond effect on academic self-concept: A cross-cultural (26-country) test of the negative effects of academically selective schools. American Psychologist 58: 364–76. [Google Scholar]
  44. Martin, Jonathan E., Alexis Menten, Kate E. Walton, Jeremy Burrus, Cristina Anguiano-Carrasco, Gabriel Olaru, and Richard D. Roberts. 2019. A Rosetta Stone for Social and Emotional Skills. New York: Asia Society. [Google Scholar]
  45. McCrae, Robert R., Antonio Terracciano, and Members of the Personality Profiles of Cultures Project. 2005. Universal features of personality traits from the observer’s perspective: Data from 50 cultures. Journal of Personality and Social Psychology 88: 547–61. [Google Scholar]
  46. Meade, Adam W. 2004. Psychometric properties and issues involved with creating and using ipsative measures for selection. Journal of Occupational and Organizational Psychology 77: 531–52. [Google Scholar] [CrossRef]
  47. Organisation for Economic Co-Operation and Development. 2015. Skills for Social Progress: The Power of Social and Emotional Skills. Paris: OECD Publishing. [Google Scholar]
  48. Osher, David, and Juliette Berg. 2017. School Climate and Social and Emotional Learning: The Integration of Two Approaches. University Park: Edna Bennet Pierce Prevention Research Center, Pennsylvania State University. [Google Scholar]
  49. Poropat, Arthur E. 2009. A meta-analysis of the five-factor model of personality and academic performance. Psychological Bulletin 135: 322–38. [Google Scholar] [CrossRef]
  50. Roberts, Brent W., Kate E. Walton, and Wolfgang Viechtbauer. 2006. Patterns of mean-level change in personality traits across the life course: A meta-analysis of longitudinal studies. Psychological Bulletin 132: 1–25. [Google Scholar] [CrossRef] [PubMed]
  51. Roberts, Brent W., Jing Luo, Daniel A. Briley, Philip I. Chow, Rong Su, and Patrick L. Hill. 2017. A systematic review of personality trait change through intervention. Psychological Bulletin 143: 117–41. [Google Scholar] [CrossRef] [PubMed]
  52. Schmitt, David P., Jüri Allik, Robert R. McCrae, and Verónica Benet-Martínez. 2007. The geographic distribution of Big Five personality traits: Patterns and profiles of human self–description across 56 nations. Journal of Cross-Cultural Psychology 38: 173–212. [Google Scholar]
  53. Soto, Christopher J., Oliver P. John, Samuel D. Gosling, and Jeff Potter. 2008. The developmental psychometrics of the Big Five self-reports: Acquiescence, factor structure, coherence, and differentiation from ages 10 to 20. Journal of Personality and Social Psychology 94: 718–37. [Google Scholar] [CrossRef]
  54. Soto, Christopher J., Christopher M. Napolitano, Madison N. Sewell, Hee J. Yoon, and Brent W. Roberts. 2022. An integrative framework for conceptualizing and assessing social, emotional, and behavioral skills: The BESSI. Journal of Personality and Social Psychology. [Google Scholar] [CrossRef]
  55. Tackett, Jennifer L. 2006. Evaluating models of the personality-psychopathology relationship in children and adolescents. Clinical Psychology Review 26: 584–99. [Google Scholar] [CrossRef]
  56. Tackett, Jennifer L., Shauna C. Kushner, Filip De Fruyt, and Ivan Mervielde. 2013. Delineating personality traits in childhood and adolescence: Associations among measures, temperament, and behavioral problems. Assessment 20: 738–51. [Google Scholar] [CrossRef]
  57. Tupes, Ernest C., and Raymond E. Christal. 1961. Recurrent Personality Factors Based on Trait Ratings. USAF ASD Technical Report. Lackland Air Force Base: U.S. Air Force, pp. 61–97. [Google Scholar]
  58. Walton, Kate E., and Jeremy Burrus. 2022. Mosaic™ by ACT® Social Emotional Learning Assessment: Evaluating Racial Subgroup Differences Across Item Types. Iowa City: ACT, Inc. [Google Scholar]
  59. Walton, Kate E., Dana Murano, Jeremy Burrus, and Alex Casillas. 2021. Multi-method support for using the Big Five framework to organize social and emotional skills. Assessment. [Google Scholar] [CrossRef]
  60. Whetzel, Deborah L., and Michael A. McDaniel. 2009. Situational judgment tests: An overview of current research. Human Resource Management Review 19: 188–202. [Google Scholar] [CrossRef]
  61. Whetzel, Deborah L., and Michael A. McDaniel. 2016. Are situational judgment tests better assessments of personality than traditional personality tests in high-stakes testing? In The Wiley Handbook of Personality Assessment. Edited by Updesh Kumar. Hoboken: John Wiley & Sons, pp. 205–14. [Google Scholar]
  62. Zell, Ethan, and Tara L. Lesick. 2021. Big five personality traits and performance: A quantitative synthesis of 50+ meta-analyses. Journal of Personality 90: 559–73. [Google Scholar] [CrossRef] [PubMed]
  63. Zickar, Michael J., Robert E. Gibby, and Chet Robie. 2004. Uncovering faking samples in applicant, incumbent, and experimental data sets: An application of mixed-model item response theory. Organizational Research Methods 7: 168–90. [Google Scholar] [CrossRef]
Table 1. Alignment of Mosaic Skills to the Big Five Framework.
Table 1. Alignment of Mosaic Skills to the Big Five Framework.
Big Five FactorMosaic SkillMosaic Skill Example Descriptors
ConscientiousnessSustaining Effortpersistence, goal striving, reliability, attention to detail
AgreeablenessGetting Along with Otherscollaboration, empathy, helpfulness, trustworthiness
Emotional StabilityMaintaining Composurestress management, emotional regulation, poise
OpennessKeeping an Open Mindcreativity, inquisitiveness, open-mindedness, embracing diversity
ExtraversionSocial Connectionassertiveness, influence, optimism, enthusiasm
Table 2. Multi-Trait Multi-Method Correlation Matrix.
Table 2. Multi-Trait Multi-Method Correlation Matrix.
Middle SchoolLikertFCSJT
SEGAMCKOSCSEGAMCKOSCSEGAMCKOSC
Likert
SE
GA.53
MC.58.63
KO.47.60.58
SC.39.63.52.54
FC
SE.43.29.31.13.20
GA.42.49.40.33.28.35
MC.41.31.43.22.29.53.43
KO.24.27.26.36.30.03.42.40
SC.27.32.26.30.48.15.27.38.52
SJT
SE.55.32.31.25.16.32.36.36.23.21
GA.44.38.28.24.12.29.43.33.25.17.63
MC.59.54.54.43.37.37.46.41.29.29.54.52
KO.41.48.47.55.40.14.26.21.30.27.25.23.50
SC.45.45.38.37.39.26.34.35.31.36.49.53.55.45
M (SD)4.41 (.67)4.84 (.72)4.37 (.77)4.28 (.79)4.46 (.71)2.28 (.35)2.46 (.38)2.15 (.43)2.13 (.37)2.20 (.40)3.24 (.63)3.53 (.68)3.54 (.55)3.38 (.55)3.41 (.52)
Cronbach’s Alpha.71.79.74.76.72.37.50.54.35.46.76.79.63.68.59
High SchoolLikertFCSJT
SEGAMCKOSCSEGAMCKOSCSEGAMCKOSC
Likert
SE
GA.55
MC.45.53
KO.51.58.51
SC.45.57.44.56
FC
SE.38.21.23.13.17
GA.30.50.33.32.27.24
MC.28.24.44.27.34.49.35
KO.08.21.20.40.37−.10.38.35
SC.21.23.12.30.53.10.22.39.51
SJT
SE.62.47.31.42.33.28.33.22.13.18
GA.45.57.33.45.36.23.43.24.20.19.58
MC.28.38.30.30.26.25.36.36.22.21.40.53
KO.34.41.29.46.27.15.36.24.29.19.46.59.49
SC.40.40.29.41.43.17.26.25.27.28.42.45.31.42
M (SD)4.50 (.78)4.94 (.61)4.16 (.73)4.54 (.68)4.45 (.76)2.34 (.35)2.52 (.34)2.20 (.43)2.14 (.38)2.23 (.43)3.65 (.51)3.65 (.48)3.61 (.57)3.41 (.52)3.28 (.44)
Cronbach’s Alpha.84.77.77.76.79.40.46.58.43.57.68.63.71.65.44
Note. SE = Sustaining Effort. GA = Getting Along with Others. MC = Maintaining Composure. KO = Keeping an Open Mind. SC = Social Connection.
Table 3. Associations Between Mosaic Skills and Outcome Variables: Correlations and Regression Statistics.
Table 3. Associations Between Mosaic Skills and Outcome Variables: Correlations and Regression Statistics.
Relationships with School Personnel
LikertFCSJTAggregate
Middle School a
Sustaining Effort.47.25.32.44
Getting Along with Others.52.37.28.49
Maintaining Composure.54.31.47.55
Keeping an Open Mind.46.22.41.47
Social Connection.45.28.37.47
F (5, 24,394)2933.42 *1119.52 *1851.83 *2690.51 *
R2.38.19.28.36
High School b
Sustaining Effort.41.20.32.40
Getting Along with Others.44.29.34.44
Maintaining Composure.37.25.28.39
Keeping an Open Mind.37.15.30.36
Social Connection.43.21.25.38
F (5, 9106)690.87 *268.64 *339.57 *615.53 *
R2.28.13.16.25
School Safety Climate
LikertFCSJTAggregate
Middle School a
Sustaining Effort.37.21.24.35
Getting Along with Others.42.32.23.41
Maintaining Composure.45.24.37.44
Keeping an Open Mind.32.13.29.32
Social Connection.28.14.25.28
F (5, 24,394)1548.96 *660.76 *903.88 *1360.81 *
R2.24.12.16.22
High School b
Sustaining Effort.31.16.23.30
Getting Along with Others.34.23.25.33
Maintaining Composure.33.17.22.32
Keeping an Open Mind.22.06.20.21
Social Connection.20.05.12.16
F (5, 24,394)345.68 *134.80 *162.65 *307.34 *
R2.16.07.08.14
GPA
LikertFCSJTAggregate
Middle School c
Sustaining Effort.41.24.33.41
Getting Along with Others.21.24.26.30
Maintaining Composure.21.26.30.32
Keeping an Open Mind.17.15.15.20
Social Connection.17.18.25.26
F (5, 22,777)919.84 *515.13 *706.51 *973.06 *
R2.17.10.13.18
High School d
Sustaining Effort.47.28.36.47
Getting Along with Others.23.22.28.30
Maintaining Composure.15.23.29.29
Keeping an Open Mind.20.06.23.21
Social Connection.12.15.17.19
F (5, 8876)542.49 *232.46 *322.70 *521.24 *
R2.23.12.15.23
Absences
LikertFCSJTAggregate
Middle School e
Sustaining Effort−.05.03−.08−.04
Getting Along with Others−.01−.10−.06−.07
Maintaining Composure.10.01−.03.03
Keeping an Open Mind.02−.02.08.03
Social Connection.07−.03.01.02
F (5, 288)1.78.951.291.01
R2.03.02.02.02
High School f
Sustaining Effort−.14−.11−.14−.17
Getting Along with Others−.11−.10−.12−.14
Maintaining Composure−.06−.07−.12−.11
Keeping an Open Mind−.04−.03−.10−.07
Social Connection.02−.04−.06−.04
F (5, 884)6.66 *3.30 *4.39 *6.14 *
R2.04.02.02.03
Discipline
LikertFCSJTAggregate
Middle School g
Sustaining Effort−.15−.11−.20−.20
Getting Along with Others−.03−.15−.31−.21
Maintaining Composure−.09−.05−.11−.11
Keeping an Open Mind−.02.07.04.04
Social Connection.04.09−.11.01
F (5, 344)2.29 *3.81 *9.20 *7.10 *
R2.03.05.12.09
High School h
Sustaining Effort−.01−.03−.13−.08
Getting Along with Others−.07−.08−.12−.12
Maintaining Composure−.07−.01−.14−.10
Keeping an Open Mind.01.08−.16−.03
Social Connection.10.14.02.12
F (5, 715)5.68 *5.39 *7.00 *9.19 *
R2.04.04.05.06
Note. an = 24,400. b n = 9,112. c n = 22,783. d n = 8,882. e n = 294. f n = 890. g n = 350. h n = 721. * p < .05.
Table 4. Mean Differences on Mosaic Skills Between Students with Acceptable, Habitual, and Chronic Absentee Records in High School Students.
Table 4. Mean Differences on Mosaic Skills Between Students with Acceptable, Habitual, and Chronic Absentee Records in High School Students.
Acceptable aHabitual bChronic c
MSDMSDMSDFη2
Likert
Sustaining Effort (SE)4.50.784.27.814.12.8212.38 *.03
Getting Along with Others (GA)4.92.614.90.714.68.735.26 *.01
Maintaining Composure (MC)4.17.734.08.704.03.832.30.01
Keeping an Open Mind (KO)4.48.714.40.704.42.701.04.00
Social Connection (SC)4.38.764.48.794.37.821.26.00
FC
SE2.34.362.27.352.24.364.93 *.01
GA2.52.332.47.322.41.384.52 *.01
MC2.21.442.14.432.12.412.71.01
KO2.13.412.16.372.10.38.67.00
SC2.25.442.25.442.17.421.32.00
SJT
SE3.63.513.52.463.47.476.71 *.02
GA3.62.483.60.453.41.477.25 *.02
MC3.61.563.47.573.38.579.08 *.02
KO3.38.523.32.483.22.473.91 *.01
SC3.28.453.30.403.15.453.62 *.01
Aggregate
SE.002.36−.742.17−1.082.2413.05 *.03
GA−.122.39−.332.33−1.252.598.40 *.02
MC.042.29−.502.10−.772.127.69 *.02
KO−.202.39−.372.16−.692.041.86.00
SC−.092.40.082.25−.582.312.23.01
Note.an = 636. b n = 168. c n = 86. F df = 2, 887. * p < .05.
Table 5. Mean Differences on Mosaic Skills Between Students with None and Some Discipline Infractions.
Table 5. Mean Differences on Mosaic Skills Between Students with None and Some Discipline Infractions.
Middle School0 Infractions a>0 Infractions b
MSDMSDtd
Likert
Sustaining Effort (SE)4.49.634.25.594.25 *.40
Getting Along with Others (GA)4.87.674.68.673.09 *.29
Maintaining Composure (MC)4.38.684.18.773.09 *.29
Keeping an Open Mind (KO)4.27.784.05.812.93 *.28
Social Connection (SC)4.42.694.35.661.14.11
FC
SE2.34.332.23.323.45 *.33
GA2.53.332.40.384.10 *.39
MC2.18.442.12.431.65 *.16
KO2.12.382.13.41−.20.02
SC2.22.422.22.40−.11.01
SJT
SE3.39.633.07.545.56 *.52
GA3.72.593.41.625.51 *.52
MC3.64.543.42.544.16 *.39
KO3.41.563.26.572.82 *.27
SC3.49.543.32.523.39 *.32
Aggregate
SE.412.31−.822.085.82 *.55
GA.392.28−.782.365.40 *.51
MC.282.26−.552.413.81 *.36
KO.172.35−.352.362.34 *.22
SC.142.40−.272.241.87 *.18
High School0 Infractions c>0 Infractions d
MSDMSDtd
Likert
SE4.56.754.46.782.36 *.13
GA4.97.574.88.662.80 *.15
MC4.21.704.10.772.93 *.16
KO4.59.664.56.701.05.06
SC4.36.754.60.76−5.59 *.31
FC
SE2.34.352.34.35.21.01
GA2.55.332.49.333.81 *.21
MC2.22.442.20.42.79.04
KO2.16.392.18.38−.62.03
SC2.21.452.30.39−3.74 *.20
SJT
SE3.72.493.60.534.62 *.25
GA3.73.473.63.483.78 *.21
MC3.69.533.57.584.02 *.22
KO3.50.513.33.506.05 *.33
SC3.32.473.29.441.25.07
Aggregate
SE.262.27−.122.293.11 *.17
GA.312.30−.252.404.33 *.24
MC.262.25−.162.203.43 *.19
KO.272.33−.082.222.79 *.15
SC−.132.44.312.19−3.39 *.19
Note. an = 335. b n = 169. c n = 982. d n = 508. * p < .05.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Walton, K.E.; Burrus, J.; Murano, D.; Anguiano-Carrasco, C.; Way, J.; Roberts, R.D. A Big Five-Based Multimethod Social and Emotional Skills Assessment: The Mosaic™ by ACT® Social Emotional Learning Assessment. J. Intell. 2022, 10, 72. https://doi.org/10.3390/jintelligence10040072

AMA Style

Walton KE, Burrus J, Murano D, Anguiano-Carrasco C, Way J, Roberts RD. A Big Five-Based Multimethod Social and Emotional Skills Assessment: The Mosaic™ by ACT® Social Emotional Learning Assessment. Journal of Intelligence. 2022; 10(4):72. https://doi.org/10.3390/jintelligence10040072

Chicago/Turabian Style

Walton, Kate E., Jeremy Burrus, Dana Murano, Cristina Anguiano-Carrasco, Jason Way, and Richard D. Roberts. 2022. "A Big Five-Based Multimethod Social and Emotional Skills Assessment: The Mosaic™ by ACT® Social Emotional Learning Assessment" Journal of Intelligence 10, no. 4: 72. https://doi.org/10.3390/jintelligence10040072

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop