Next Article in Journal
Physical Sciences Teachers’ Enactment of Simulations in 5E Inquiry-Based Science Teaching
Previous Article in Journal
Teaching Is a Story Whose First Page Matters—Teacher Counselling as Part of Teacher Growth
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Effect of Peer Influence and Self-Reflection on Scaffolded Out-of-Class Activity Administered Using a Mobile Application

1
Department of Computer Science, Winston-Salem State University, Winston-Salem, NC 27110, USA
2
Department of Computer Science, The University of Texas at El Paso, El Paso, TX 79968, USA
*
Author to whom correspondence should be addressed.
Educ. Sci. 2022, 12(12), 863; https://doi.org/10.3390/educsci12120863
Submission received: 3 November 2022 / Revised: 13 November 2022 / Accepted: 21 November 2022 / Published: 25 November 2022
(This article belongs to the Section Technology Enhanced Education)

Abstract

:
Student engagement with out-of-class activities is becoming more difficult as students spend fewer hours outside the classroom studying the content. This research developed a mobile educational platform, Dysgu, to provide students with an optimal learning experience outside of the classroom. Dysgu includes social networking and gamification features to increase student engagement. The platform offers interactive auto-graded assessments to help students practice concepts and take tests. Students can see their scores and a summary of the performance of the rest of the class. We used Dysgu for multiple out-of-class activities at two universities with different student demographics for two semesters. The data shows that students obtain better grades when using Dysgu. We also saw more on-time or ahead-of-time submissions with Dysgu. Survey responses indicated several Dysgu features which students found helpful. We conclude that digital educational platforms should consider features to support scaffolding to master the concept, peer influence to keep students engaged, self-reflection to foster critical thinking, and easy adaption of the platform to reduce faculty workload and improve students’ acceptance of the system.

1. Introduction

Classroom instructions and interactions are important for student learning. However, a big part of student learning involves practicing the concept outside of the class. In fact, college guidelines suggest that students should spend 2–3 h outside the classroom (per credit hour of class per week) reviewing and practicing the concepts [1,2]. However, students are struggling to keep up with these out-of-class activities due to several reasons, including job commitments and the medium of these out-of-class activities [3,4]. Thus, it is essential to develop methods that will allow the effective usage of the students’ available time for studying outside the classroom and provide students with interactive out-of-class activities that will involve them with the help of engaging in the features of such activities and the medium through which such activities are provided. The objective of this study is to evaluate the role of peer influence, self-reflection, and scaffolding in a mobile platform in order to keep students engaged with out-of-class activities.
In this paper, we present a mobile application that allows students to complete their out-of-activities at their preferred schedule. The application allows students to see their scores in these activities and compare their standing with their peers in the class. Additionally, the application supports the scaffolding of the activities along with contextual notifications for them to be kept involved with those activities. The data suggests that students find it easier to use an app rather than a pen-paper version of out-of-class activities. We also saw better student performance in the activities (e.g., lower D/F/W rates). Our findings suggest that peer influence and gamification can engage students in out-of-class activities while improving their grades and reducing the rate of procrastination.

2. Materials and Methods

This section contains a review of the literature related to some of the major features of Dysgu—the mobile application we developed for engaging students with out-of-class activities. We later provide a description of Dysgu’s significant features and details of the studies carried out at two different universities.

2.1. Related Work

The role of peer influence in non-academic and academic settings has been studied in depth [5,6]. There are many forms of direct peer influence in education, including peer learning, peer-teaching, peer feedback, and peer assessment [7,8]. Multiple research has repeatedly found positive impacts due to peer instructions on student performance and retention [9,10]. Indirect peer influence also comes in various forms, including gamification [11]. One study revealed that one of the non-academic factors influencing university students’ self-regulation of study is peer influence [12]. Multiple research has identified connections between peer influence and improved student engagement, as well as academic performance [13,14].
Self-reflection is linked with improved learning and academic performance [15,16]. Many researchers have studied the role of intervention techniques in self-reflection and academic performance [17,18]. Additionally, researchers showed that being aware of student activities supports self-reflection [19,20].
Researchers have noted that student engagement can be increased through instructional scaffolding [21,22]. Such scaffolding can be achieved through dynamic assessments and providing various forms of support, including executive control (e.g., support for time management) [23].
With the increasing access to mobile (i.e., cell) phones, recommendations are made to include mobile platforms in teaching and learning to support student learning outcomes and increase student engagement [24]. Mobile platforms are frequently used to complement teaching and learning in formal and informal educational settings [25,26]. Mobile games also leverage the mobile platform for education [27]. Others use mobile apps to train or teach concepts [28,29]. At the same time, some have used mobile app development within a course to motivate students and improve student learning [30,31]. Numerous studies have shown that mobile learning is effective in engaging students [32,33] and increasing student achievements [34,35].

2.2. Dysgu and Its Features

Dysgu [36,37] provides features to achieve student learning and engagement outside of the classroom. Although the authors have visualized this system, design decisions for Dysgu were mostly directed by several mobile learning requirement catalogs [38,39] to make the software more responsive and engaging. Dysgu has two different applications: the faculty side and the student side. Both sides communicate with a cloud-based repository which stores all the data. Some of Dysgu’s major features are described next.
(A).
The system’s background operations are transparent to the stakeholder so that there is no need for administration or the day-to-day management of the software. It is achieved through a cloud-based repository system designed explicitly for Dysgu. Neither the faculty nor the students have to worry about managing or paying for such cloud infrastructure and services, as Dysgu only uses the cloud vendors’ free options.
(B).
Dysgu allows activities to be put in learning paths with different degrees of difficulty to support instructional scaffolding. As shown in Figure 1a, a module can have different learning paths with different activities. Some of the modules’ paths can be designated as practice paths to help students practice the concept. Other paths are designated as student learning outcome (SLO) paths, where students are assessed on their learning skills. Additionally, some paths can be assigned as extra credit, which creates problems as this requires additional effort from the students. Students see the same module in the app (Figure 1b) with a different representation where exercises in the SLO path are revealed one at a time, and other paths are shown fully at the beginning.
(C).
Students can earn a score in the SLO path, which is used to calculate their course grades; however, students can only earn points by solving problems in other paths. Points are an internal currency in the software which students can use to extend the deadlines of the modules.
(D).
Each exercise in a path is modeled as an interactive activity. While these activities must be finished within a timeframe, they have interactive elements, such as multiple screens, multiple user interface components to interact with, cause-effect scenarios, and ways to traverse the problem before submitting it for grading. Figure 2a shows a sample interactive activity where answers can have right or wrong hints, or students can select multiple answers or modify their answers by looking at the question’s components. Figure 2b shows another example where students classify entries by dragging them to the corresponding bucket.
(E).
To motivate and engage students, Dysgu provides lightweight gamification and social networking aspects within student privacy regulations. Students can compare their progress (Figure 3a) with their classmates. Dysgu shows the student score, placement (module, path, exercise-specific) compared to the class, and timing information regarding problem-solving activities. Additionally, students are awarded badges depending on different conditions (the fastest in class to answer, highest score, etc.), and they can see what badges they have won and compare them with the rest of the class (Figure 3b). Additionally, it also shows how many other students have checked their progress, which might encourage the students to engage more in the activities.
(F).
Dysgu is developed to support several adaptations for different class situations and personalization to address student needs. For instance, the client app allows students to set (Figure 4a) the number of notifications they will receive from the app each day or if there are any blackout days. This enables students to change and update their study time to match their class, work, and other schedules and instruct the app when to remind them of pending tasks and how many times to remind them during the day. Table 1 shows a list of notifications that Dysgu uses to remind students of their progress, and Figure 4b shows how students can visualize such notifications. Although a student can control the number and time of their notifications, some (neutral ones) cannot be controlled by students and will appear if corresponding conditions are met.

2.3. Study Details

Dysgu was deployed in the CS1301 course at the University of Texas-El Paso (UTEP) and in the CSC/CIT 1311 course at Winston-Salem State University (WSSU). Both of these courses are freshmen-level courses and expose students to Java programming language. Table 2 shows the data that was collected before using the software in the class and during its usage in the class. Note that, at UTEP, in Spring 2022, Author 2 taught two sections of the CS1 course. Dysgu was deployed in Section A, whereas Section B students used Blackboard to submit their responses to the same questions.
UTEP is a majority-minority Hispanic-serving Institution that enrolls over 27,000 undergraduate students, of which over 85% are Hispanic-American [40]. WSSU is a predominately black university where 86% of the students are African American or from other minority groups [41]. Table 3 shows student information on both campuses during the intervention semesters. It is nice to see a gender balance in one of the campuses (WSSU), whereas the other (UTEP) showed regular patterns found in similar studies (i.e., 77% male). UTEP has 70% of students enrolled in 12 or more credit hours, whereas WSSU has more than 93% of students with such a course load. It is eye-opening to see that roughly half of the students work extensively (e.g., more than 10 h) outside the class for financial reasons. Having software such as Dysgu allows us to address such situations that students are in. Out-of-class activities through Dysgu allow students to work on them anytime and anywhere, allowing them to best use their time.
The same assignment was used before and after deploying Dysgu in the class. Each assignment asked students a sequence of questions that students needed to answer by the due date. Before Dysgu, questions from different topics were given to the student through the learning management system (LMS) with a deadline of a week. Students were expected to submit their answers using the LMS. During Dysgu, we used the same set of questions to assess the students’ learning. However, to comply with Dysgu’s scaffolded assignment approach, each assignment was structured as follows:
  • Questions from the same topic were put on a module with the topic’s name.
    Therefore, each assignment had multiple modules in them (having the same questions as before for assessment and scoring).
  • Each module:
    Has one SLO path with the questions enabled one at a time. Once the student answers a question, the next question in that path is turned on. Each question can only be tried once.
    Has one practice path with questions to practice the specific topic.
    Has an optional extra credit path. Questions in the extra credit path only gave points (not scores).
Table 4 shows the properties and comparisons of those assignments.
The same instructors (both authors) taught the courses before Dysgu was used and during the deployment of Dysgu. Similar course settings and teaching practices were consistently maintained before and during the Dysgu interventions. To diminish the impact of confounding variables, such as the size, display, and appearance of mobile devices, each student was given an identical Samsung tablet for the study duration, which the students took home. The software was used two times during the semester at WSSU and three times at UTEP. Table 5 shows the content of each module.
At UTEP, the course was an introductory CS course (CS1), and the Dysgu-based out-of-class activities were deployed throughout the semester. At WSSU, the course was a freshman programming-I course, and the out-of-class activities administered by Dysgu were to assess their knowledge of the content of that course. Therefore, those modules were administered at the beginning of the semester. The study was performed following all Institutional Review Board (IRB) protocols involving human subjects, where student participation was entirely voluntary, and students were not provided with any extra benefits or penalized for non-participation. In addition, all students signed a consent agreement before the start of the study, where they were informed about data collection and other related factors of the study.

3. Results

It is hypothesized that the students’ usage of Dysgu, with its real-time status updates, allows students to self-reflect on their progress and compare it with their peers. Such influence will allow the students to engage with the content more and eventually learn the subject material better. Dysgu’s impact on the different student demographics and students’ open-ended feedback regarding the strengths and limitations of the features of Dysgu are also explored. This research was directed by the following research questions:
  • Is there any improvement in student performance after using this software?
  • What is the effect of the new intervention on student procrastination and participation?
  • Are there any impacts on students with self-reflection and peer influence?
  • What feature of the mobile platform do students find most engaging and helpful?
  • What type of student behavior is noticed when using this software for out-of-class activities?
  • Is there any notable difference between how the intervention impacts two distinct student demographics?
To measure the students’ learning improvement in the topics, we used the assignment grades as evaluation metrics. The graded Dysgu-based out-of-class exercises contributed toward 12% of the total course grade at UTEP and 10% of the total course grade at WSSU. However, the students’ long-term knowledge acquisition was further tested on the same topics during the midterm and final exams, and each of these exams contributed between 15% and 20% toward the total course grade. In addition, assignment and course failure rates are also considered a measure of knowledge acquisition and retention. A pre- and post-survey was conducted before and after deploying Dysgu in the classroom to observe the students’ perceptions about the different facets of out-of-class activities. In addition, to assess student engagement and perception of the software and its effectiveness, an experience survey was offered to the learners only after they completed using Dysgu. The students gave their opinion about the statements in all surveys using a Likert scale of four values (Strongly Agree, Agree, Disagree, and Strongly Disagree), with an agreement scale ranging from strongly agree (4) to strongly disagree (1). To ensure that the students were actually looking at the survey content, not just “clicking it through,” each survey was designed to contain both positive and negative questions. Additionally, the experience survey asked students to select the most helpful features from a list of features provided by the software.

3.1. Student Grade

Research question A seeks to investigate the role of Dysgu on student performance. Figure 5 shows grade comparisons of the student’s work before and after using Dysgu. The results indicate that fewer students were failing, and the average student’s grade was improving. Although for UTEP, the improvements in the highest grade were more pronounced, for WSSU, improvement in the failure rate was more noticeable. It was also observed that, along with reducing course failure and withdrawal rates at WSSU, this intervention was also helpful for average or above-average students, who would otherwise receive a lower grade; however, with this intervention, they were able to engage more and achieve higher grades. On average, across both universities, the D/F/W rate was reduced by 27%.

3.2. Timing Information

Our second research question seeks to understand the effect of the new intervention on student procrastination and participation. A critical aspect of the new intervention (i.e., Dysgu) was to start students working on the out-of-class activity as early as possible and reduce their chance of procrastinating and submitting subpar work. This also provided one way to track when students started their work on the modules. Figure 6 shows when students started working before the deadline. It is obvious that most of the students started working more than 24 h before the deadline. This not only indicates an improvement in the students’ procrastination behavior but also shows that students wanted to start early so that they could see their progress in the activities and compare them to the rest of the class. Such an outcome also sheds light on research question C as it correlates to the student survey responses, as presented in Section 3.3. Without Dysgu, we would never have such timing data from the LMS and could not relate it to student performance and behavior. Additionally, the student’s final submission time is also compared, and it was assumed that if students submitted the activity just before it was due, they might have started working on it late. Since LMS provides us with the student’s submission time, it is possible to compare the two times. As shown in Figure 7, with Dysgu, more than half of the students submitted their work more than a day before the deadline. This improvement was clearly visible in both universities. In both cases, with Dysgu, more students submitted before a day, and fewer submitted just before the deadline compared to when they did not use Dysgu. The benefits of early submission also represented an early start on the assignment and an improved grade, as depicted earlier.

3.3. Student Experience

The fourth research question focuses on finding the most engaging and helpful components of the mobile educational platform. To gather student impressions of the software and to learn through out-of-class exercises, an experience survey was given on both campuses during the intervening semesters. The survey contained 10 questions with a Likert scale of four values (Strongly Agree, Agree, Disagree, and Strongly Disagree) and two open-ended questions. Figure 8 shows the results of some of the questions of this survey. For simplifications, “Strongly Agree” and “Agree” were combined as “Agreeing,” and the two disagreement answer choices were combined as “Disagreeing” in these charts. Figure 8 shows that students, on average, reported very high levels of agreement on the survey questions. On both campuses, the students found it helpful to be able to compare their grades with the rest of the class (Figure 8a) and noted that the ability to compare their progress to the rest of the class encouraged their involvement with the out-of-class activities (Figure 8e). In general, students felt more confident in completing Dysgu out-of-class activities than pen-and-paper-based out-of-class activities (Figure 8b) and found it enjoyable to complete these activities as per their own schedule (Figure 8c). Students also noted that badges and points encouraged them to stay involved in out-of-class activities (Figure 8d). These results show strong support for Dysgu-based out-of-class activities and a preference towards the Dysgu features, which were included to improve student engagement (e.g., badge, grade comparison).
In general, students on both campuses show similar impressions of the intervention using Dysgu. Although students at WSSU agreed less (compared to UTEP students) on feeling more confident completing Dysgu activities, their response to question iii of the pre- and post-survey contradicts this. The WSSU students’ agreement on question (iii) (“I prefer out-of-class activities that can be completed using mobile devices”) increased by 26% after using Dysgu. Table 6 shows the descriptive statistics (µ = mean, σ = standard deviation) for those same questions. The results indicate that the average response represents the students’ uniform attitude towards the intervention, as reflected by a lower standard deviation value.
In the same experience survey, students were asked to pick the app’s most useful features. Figure 9 shows a few of the most chosen features (across both campuses) that students thought were useful in the learning process and for engagement with out-of-class activities. The ability to compare their grades with the rest of the class was chosen as the top feature at both campuses. The second most liked feature was the ability for self-assessment. It is evident from the chart that students liked the features that helped them to succeed in the rest of the course. It also showcases that scaffolded out-of-class activities that provide interactions are mostly preferred. Additionally, a mobile device’s notification features are preferred—most likely for helping students manage time and keep up with their peers.

3.4. Pre and Post-Survey

Pre- and post-surveys were conducted to assess the students’ change in perception of the new intervention. The pre-survey was given before the students were introduced to the app, and the post-survey was conducted after the end of the intervention. As earlier, we combined the two “agrees” into an Agreement and the other as a Disagreement. We then aggregated all the responses for each statement and calculated their percentages. For each statement, we then compared the change in percentage for each of the two (Agreement and Disagreement) choices, which are listed in Table 7.
For each statement in the survey, we listed whether the students’ agreement or disagreement increased and the amount of the increase. For most of the statements, we assumed a response pattern that we anticipated seeing because of the intervention. However, in a few of the statements, the changes were not as expected and are highlighted in orange in the table. For statement I, we anticipated that faster feedback would be helpful for learning; hence, we expected to see an increase in agreement with this statement. However, we can see a bit of an anomaly in the student response in one of the campuses (WSSU), although their grade and submission timeframe improved significantly, as shown in Figure 5 and Figure 7. This could be due to the students not seeing any connection between receiving faster feedback and better learning. The value of different forms of feedback (e.g., formative, immediate) has been studied in depth and shows that different types of feedback affect the students’ performance, motivation, and engagement [42]. This disagreement shows that further investigation is needed to understand student perception of learning and what is the impact of different types on student learning.
We observed a similar anomaly for statement v. This statement asked students if they liked classes with hands-on activities such as out-of-class activities. We expected students to agree with this statement. However, one of the campuses showed a very small disagreement (<0.5%) and needed further investigation. Similarly, with statement vi, we expected an increase in disagreement as any form of practice could be helpful for tests. One of the campuses (i.e., UTEP) showed an agreement with this statement, indicating that the students did not find the out-of-class activities helpful for the exams.
Other than these anomalies, we can see that students are largely in agreement with the statements related to the usefulness of interactive out-of-class activities and the ability to see how their peers are doing in those activities. Overall, the pre- and post-survey data reconfirm the novelty of this research and the usefulness of the presented approach.

3.5. Usage Analytics

To investigate research question E, Dysgu collected student usage analytics while the students were using the app for out-of-class activities. As shown in Figure 10a, the students at both universities checked their status, which provides a comparison of their scores with their peers in the class. This was the most checked feature of the app in both universities. The second most used feature at UTEP was the badge students received to encourage their participation and efforts. At WSSU, the second most used feature was the notifications students received. The third most used feature at both universities was the points students received for the extra-credit pathway. We also noticed that students from both campuses were using the app mostly in the morning (Figure 10b).

3.6. Student Comments

The last question on the experience survey was open-ended. We asked for participants’ suggestions on improving Dysgu or their experience with Dysgu. We received a total of 39 text responses for this question across two campuses over the two semesters. We analyzed the responses to identify emerging themes in the responses. There were three major themes repeated throughout the response (e.g., GUI, more challenging questions). We coded those themes and counted their occurrences in the responses. The three themes in the responses were related to: the Dysgu app (e.g., graphical user interface (GUI), the features of Dysgu, that they liked Dysgu and had no suggestions for its improvement, and suggestions related to the content (e.g., challenging questions). Figure 11 shows the distribution of these themes in the responses. Following is a brief discussion of these themes.
Dysgu-related suggestions: These responses were related to various aspects of the Dysgu app. One of those aspects was the GUI—the user interface of Dysgu. Most responses in this theme suggested making the interface more user-friendly and intuitive to use. One such response was, “I understand that the app is in an early stage, but I feel like having more modern-looking menus, tabs, and that kind of material inside the app would have made the experience even more enjoyable and fun”.
Within Dysgu codes, aside from the GUI sub-theme, another theme emerged that was related to the features of Dysgu. Students suggested additional features in Dysgu to make it more effective or engaging. Most suggestions in this area were related to the time limit of the questions and the ability to pause the timer if the app was turned off. One suggestion was related to peer communication, “another is being able to communicate/text with classmates within the app for help.”
Some responses were related to the performance of the app or the device. Some suggestions were to make the response of the app faster. Dysgu is available for android platforms. One suggestion was to make the app suitable for all phones, “If it was available on my typical device (phone), I would use it much more …”
Content: This theme is the least related to Dysgu. The responses related to this theme discussed how the questions could be more challenging or better articulated. One such response was, “The dysgu should include more challenging questions too”. Responses to this theme also made the suggestion of having more assignments, more questions, or having an adaptive system that would set the difficulty of the questions based on the student’s level. A student suggested, “The app was great. One thing I would suggest is to put a challenge section where the student gets questions one at a time based on the material learned, the next question being harder than the last. The moment they get a question wrong, the challenge stops…
Liked it/nothing to add: 27% of the participants liked the Dysgu app. They did not have any suggestions on improving Dysgu or the experience with Dysgu.

3.7. Self-Reflection, Peer-Influence, and Different Student Demographics

Going back to our research questions, to answer research question C (to study any impacts of Dysgu on students regarding self-reflection and peer influence), we looked at the data related to student grades, submission time, and their feedback on the useful features of Dysgu. Results from the pre-post survey and the experience surveys showed that the students found the ability to see their progress in the app helpful. This shows that students across both campuses value features that enable them to reflect on their standing in the class—thus, self-reflection is essential. Similarly, students noted that knowing their standing in the class compared to the rest of the class was also helpful in keeping them involved with the out-of-class activities. We saw this in both student demographics, indicating that peer influence acts as a strong motivator for students to stay engaged with class activities outside the classroom. These statements can be complemented with the grade and timing data, as we see improved grades and better submission times when Dysgu was deployed. Altogether the data indicates that self-reflection and peer influence positively impact student engagement and performance.
The last research question (question F) seeks to identify any notable difference between how the intervention impacts the two distinct student demographics. In terms of grades, we saw a significant drop in D/F/W rates at WSSU, with an increase in the number of students getting either grades B or C. At UTEP, the D/F/W rate slightly increased with Dysgu; the number of students receiving grade A also increased. Though the grade distribution is different on these campuses (with fewer B and C grades at UTEP compared to WSSU), the number of students with better grades increased on both campuses. In terms of the timeline, students from both campuses started work on the activity more than 24 h before the deadline. Most students also submitted the activity more than 24 h before the deadline at both campuses. Both student groups worked mainly in the morning or in the evening. The afternoon hours had the least activity. Without Dysgu, the students at WSSU were more likely to submit their work within the last five hours of the deadline. We also saw a similar preference for Dysgu features across the campus. One notable difference here was that UTEP students viewed their badges a lot; WSSU students, on the other hand, were checking the app’s notifications. Both student groups used the app in the morning, but the usage percentage was higher at WSSU. Overall, we can see that there are differences in the usage behavior of the app, but the results (in terms of grades and submission time) are somewhat the same across both demographics (i.e., Hispanic and African American).

4. Discussion

We investigated the impact of a mobile platform in engaging students with out-of-class activities. Dysgu allowed us to study the impact of scaffolded out-of-class activities administered through a mobile platform in order to keep students engaged outside the class. Dysgu also includes features that support self-reflection and peer influence. The results show that students perform better when they remain engaged with out-of-class activities and with their peers. We also saw increased on-time or ahead-of-time submissions using Dysgu. The student responses showed that students preferred mobile platforms for out-of-class activities. They also favored the notification feature to remind them of the deadline. The grading and timing data and experience surveys show that peer influence and self-reflection in the mobile platform are good motivators and have a positive impact on performance and on-time submission. The results of this study show the promise of peer influence, self-reflection, and scaffolding in keeping students engaged with out-of-class activities. This result should pave the path for further longitudinal studies to identify the long-term effects of peer influence and self-reflection on student learning and engagement.
We conducted multiple studies across two campuses with a majority-minority population. Our study revealed several Dysgu features which students preferred across both campuses. One of the key findings of our study was that the design of mobile educational platforms should consider four key factors to achieve higher student engagement, student learning, and improved self-efficacy in students, as described next (Figure 12).
  • Scaffolding—digital platforms should support scaffolding in various forms. Content or assessments should be dynamic and adaptive to support the students with different learning paces. For example, an assessment can start with easier questions and build the difficulty levels of subsequent questions based on the student’s responses. Auto-graded practice and assessment questions would make it easier for instructors to provide instant feedback—which also supports student learning and engagement.
  • Peer influence—Peer influence acts as a motivator for many students. Digital platforms should consider the effective integration of peer influence in supporting student learning and increasing student engagement. Chapin [43] identified three factors that impact the self-efficacy of early Computer Science students: engagement with the problem, engagement with others, and engagement with the environment. Scaffolding, when used right, can be leveraged to create engagement with the problem. Similarly, knowing how peers are doing in the class can help students stay engaged with others in the class and the class environment; eventually, this can help to improve self-efficacy. However, attention should be given while designing a system that supports peer influence to limit any potential negative consequences of peer influence.
  • Self-reflection—students should be aware of their learning status. Any educational platform should have support for some form of self-reflection. Self-reflection can be supported through automatic grading in tests or reflective journaling. Self-reflection not only helps with acquiring content knowledge but also with metacognition and developing critical thinking skills.
  • Usability—Digital tools should be at least available on mobile devices to support easier access and ubiquitous learning. These tools should also be available for different platforms (e.g., cell phones, tablets, laptops). Tools should be user-friendly and intuitive to use. The tool’s design should consider usability to keep students engaged with the platform. Gamification (e.g., badge, points) can provide an engaging virtual learning environment.
Although Dysgu utilizes the transformative power of mobile technology for out-of-class activities, student engagement in active learning outside the classroom, and the possibility of providing early intervention, has some limitations. Most significantly, the app is only supported in the Android ecosystem. Although Dysgu is created in a way that supports any interactive activity, the current version of the app only supports two different kinds of activities, and other types of activities need to be developed for broader acceptance. Finally, data from the long-time usage of the system should be collected to see any trends developing that might provide better insights into this approach.
The use of any technology requires an understanding of the potential impacts and unintended consequences. Any new tool or technique thus requires extreme caution and monitoring when deployed. A platform that supports peer influence and promotes self-awareness is no different. In designing such systems, we should strive to create an inclusive environment, keeping many factors, such as collaboration, reflection, scaffolding, and communication, in mind. Systems should ensure that peer influence is positive and that self-reflections support developing self-efficacy. With careful design and deployment, such systems can create an inclusive and equitable environment that is conducive to student engagement and growth.

Author Contributions

Conceptualization, M.F.; methodology, M.F. and M.A.; software, M.F.; validation, M.F. and M.A.; formal analysis, M.F.; investigation, M.F. and M.A.; resources, M.F. and M.A.; data curation, M.F. and M.A.; writing—original draft preparation, M.F. and M.A.; writing—review and editing, M.F. and M.A.; visualization, M.F.; supervision, M.A.; project administration, M.F. and M.A.; funding acquisition, M.F. and M.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Science Foundation, grant numbers 1712030 and 1712073.

Institutional Review Board Statement

The study was conducted in accordance with the Federal Regulation 45CFR46 (at Winston-Salem State University), and 45 CFR 46.101(b)(1) (at the University of Texas-El Paso) and was approved by the Institutional Review Board of Winston-Salem State University (2986-18-0001, 11 July 2017) and the University of Texas-El Paso (1098619-1, 13 July 2017).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Acknowledgments

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Utah State University. “How Many Hours Do I Need To Study?”. Available online: https://www.usu.edu/academic-support/time/estimate_study_hours (accessed on 31 October 2022).
  2. Bennett, J. Hints on How to Succeed in College Classes. Available online: http://www.jeffreybennett.com/pdf/How_to_Succeed_general.pdf (accessed on 31 October 2022).
  3. Gose, B. The Chronicle of Higher Education. 16 January 1998, pp. A37–A39. Available online: https://www.chronicle.com/ (accessed on 31 October 2022).
  4. Nonis, S.A.; Hudson, G.I. Academic Performance of College Students: Influence of Time Spent Studying and Working. J. Educ. Bus. 2006, 81, 151–159. [Google Scholar] [CrossRef]
  5. Ciranka, S.; van den Bos, W. Social Influence in Adolescent Decision-Making: A Formal Framework. Front. Psychol. 2019, 10, 1915. [Google Scholar] [CrossRef]
  6. Li, Y.; Qiu, L.; Sun, B. Two Ways Peer Interactions Affect Academic Performance. In Proceedings of the 4th International Conference on Crowd Science and Engineering, Jinan, China, 18–21 October 2019; ACM: Jinan, China, 2019; pp. 90–94. [Google Scholar]
  7. Biggers, M.; Yilmaz, T.; Sweat, M. Using Collaborative, Modified Peer Led Team Learning to Improve Student Success and Retention in Intro Cs. In Proceedings of the 40th ACM Technical Symposium on Computer Science Education-SIGCSE ’09, Chattanooga, TN, USA, 4–7 March 2009; ACM Press: Chattanooga, TN, USA, 2009; p. 9. [Google Scholar]
  8. Sondergaard, H. Learning from and with Peers: The Different Roles of Student Peer Reviewing. SIGCSE Bull. 2009, 41, 31–35. [Google Scholar] [CrossRef]
  9. Zingaro, D.; Porter, L. Peer Instruction: A Link to the Exam. In Proceedings of the 2014 Conference on Innovation & Technology in Computer Science Education-ITiCSE ’14, Uppsala, Sweden, 21–25 June 2014; ACM Press: Uppsala, Sweden, 2014; pp. 255–260. [Google Scholar]
  10. Pappas, I.O.; Giannakos, M.N.; Jaccheri, L. Investigating Factors Influencing Students’ Intention to Dropout Computer Science Studies. In Proceedings of the 2016 ACM Conference on Innovation and Technology in Computer Science Education, Arequipa, Peru, 11–13 July 2016; ACM: Arequipa, Peru, 2016; pp. 198–203. [Google Scholar]
  11. Indriasari, T.D.; Luxton-Reilly, A.; Denny, P. Gamification of Student Peer Review in Education: A Systematic Literature Review. Educ. Inf. Technol. 2020, 25, 5205–5234. [Google Scholar] [CrossRef]
  12. Balapumi, R.; von Konsky, B.R.; Aitken, A.; McMeekin, D.A. Factors Influencing University Students’ Self-Regulation of Learning: An Exploratory Study. In Proceedings of the Australasian Computer Science Week Multiconference, Canberra, Australia, 1–5 February 2016; ACM: Canberra, Australia, 2016; pp. 1–9. [Google Scholar]
  13. DeLay, D.; Zhang, L.; Hanish, L.D.; Miller, C.F.; Fabes, R.A.; Martin, C.L.; Kochel, K.P.; Updegraff, K.A. Peer Influence on Academic Performance: A Social Network Analysis of Social-Emotional Intervention Effects. Prev. Sci. 2016, 17, 903–913. [Google Scholar] [CrossRef]
  14. Furrer, C.J.; Skinner, E.A.; Pitzer, J.R. The Influence of Teacher and Peer Relationships on Students’ Classroom Engagement and Everyday Motivational Resilience. Teach. Coll. Rec. 2014, 116, 101–123. [Google Scholar] [CrossRef]
  15. May, D.B.; Etkina, E. College Physics Students’ Epistemological Self-Reflection and Its Relationship to Conceptual Learning. Am. J. Phys. 2002, 70, 1249–1258. [Google Scholar] [CrossRef]
  16. Lew, M.D.N.; Schmidt, H.G. Self-Reflection and Academic Performance: Is There a Relationship? Adv. Health Sci. Educ. 2011, 16, 529–545. [Google Scholar] [CrossRef] [Green Version]
  17. Chen, P.; Chavez, O.; Ong, D.C.; Gunderson, B. Strategic Resource Use for Learning: A Self-Administered Intervention That Guides Self-Reflection on Effective Resource Use Enhances Academic Performance. Psychol. Sci. 2017, 28, 774–785. [Google Scholar] [CrossRef] [Green Version]
  18. Kefalidou, G.; Skatova, A.; Brown, M.; Shipp, V.; Pinchin, J.; Kelly, P.; Dix, A.; Sun, X. Enhancing Self-Reflection with Wearable Sensors. In Proceedings of the 16th International Conference on Human-Computer Interaction with Mobile Devices & Services-MobileHCI ’14, Toronto, ON, Canada, 23–26 September 2014; ACM Press: Toronto, ON, Canada, 2014; pp. 577–580. [Google Scholar]
  19. Govaerts, S.; Verbert, K.; Duval, E.; Pardo, A. The Student Activity Meter for Awareness and Self-Reflection. In Proceedings of the CHI ’12 Extended Abstracts on Human Factors in Computing Systems, Austin, TX, USA, 5–10 May 2012; ACM: Austin, TX, USA, 2012; pp. 869–884. [Google Scholar]
  20. Guiffrida, D.A. The Emergence Model: An Alternative Pedagogy for Facilitating Self-Reflection and Theoretical Fit in Counseling Students. Couns. Educ. Superv. 2005, 44, 201–213. [Google Scholar] [CrossRef]
  21. Linn, M.C. Designing Computer Learning Environments for Engineering and Computer Science: The Scaffolded Knowledge Integration Framework. J. Sci. Educ. Technol. 1995, 4, 103–126. [Google Scholar] [CrossRef]
  22. Ambrose, S.A.; Bridges, M.W.; DiPietro, M.; Lovett, M.C.; Norman, M.K. How Learning Works: Seven Research-Based Principles for Smart Teaching; John Wiley & Sons: Hoboken, NJ, USA, 2010. [Google Scholar]
  23. Belland, B.R. Instructional Scaffolding in STEM Education: Strategies and Efficacy Evidence; Springer Nature: Berlin/Heidelberg, Germany, 2017. [Google Scholar]
  24. West, D.M. Mobile Learning: Transforming Education, Engaging Students, and Improving Outcomes. Brook. Policy Rep. 2013, 9, 1–7. [Google Scholar]
  25. Huy, N.P.; vanThanh, D. Evaluation of Mobile App Paradigms. In Proceedings of the the 10th International Conference on Advances in Mobile Computing & Multimedia-MoMM ’12, Bali, Indonesia, 3–5 December 2012; ACM Press: Bali, Indonesia, 2012; p. 25. [Google Scholar]
  26. Kizilcec, R.F.; Chen, M. Student Engagement in Mobile Learning via Text Message. In Proceedings of the Seventh ACM Conference on Learning @ Scale, Virtual Event, 12–14 August 2020; ACM: New York, NY, USA, 2020; pp. 157–166. [Google Scholar]
  27. Zafar, F.; Wong, J.; Khalil, M. Gamifying Higher Education: Enhancing Learning with Mobile Game App. In Proceedings of the Fifth Annual ACM Conference on Learning at Scale, London, UK, 26–28 June 2018; ACM: London, UK, 2018; pp. 1–2. [Google Scholar]
  28. Palomo-Duarte, M.; Berns, A.; Dodero, J.M.; Cejas, A. Foreign Language Learning Using a Gamificated APP to Support Peer-Assessment. In Proceedings of the Second International Conference on Technological Ecosystems for Enhancing Multiculturality-TEEM ’14, Salamanca, Spain, 1–3 October 2014; ACM Press: Salamanca, Spain, 2014; pp. 381–386. [Google Scholar]
  29. Mao, A.; Luo, J. A Light-Weight Mobile Education App for 3D Modelling Course Teaching. In Proceedings of the 2019 the 3rd International Conference on Digital Technology in Education, Yamanashi, Japan, 25–27 October 2019; ACM: Yamanashi, Japan, October 25, 2019; pp. 223–227. [Google Scholar]
  30. Honig, W.L. Teaching and Assessing Programming Fundamentals for Non Majors with Visual Programming. In Proceedings of the 18th ACM Conference on Innovation and Technology in Computer Science Education-ITiCSE ’13, Canterbury, UK, 1–3 July 2013; ACM Press: Canterbury, UK, 2013; p. 40. [Google Scholar]
  31. Kurkovsky, S. Engaging Students through Mobile Game Development. SIGCSE Bull. 2009, 41, 44–48. [Google Scholar] [CrossRef]
  32. Pechenkina, E.; Laurence, D.; Oates, G.; Eldridge, D.; Hunter, D. Using a Gamified Mobile App to Increase Student Engagement, Retention and Academic Achievement. Int. J. Educ. Technol. High Educ. 2017, 14, 31. [Google Scholar] [CrossRef] [Green Version]
  33. Heflin, H.; Shewmaker, J.; Nguyen, J. Impact of Mobile Technology on Student Attitudes, Engagement, and Learning. Comput. Educ. 2017, 107, 91–99. [Google Scholar] [CrossRef]
  34. Crompton, H.; Burke, D. The Use of Mobile Learning in Higher Education: A Systematic Review. Comput. Educ. 2018, 123, 53–64. [Google Scholar] [CrossRef]
  35. Martin, F.; Ertzberger, J. Here and Now Mobile Learning: An Experimental Study on the Use of Mobile Technology. Comput. Educ. 2013, 68, 76–85. [Google Scholar] [CrossRef]
  36. Fuad, M.; Akbar, M.; Zubov, L. Keeping Students Occupied with the Course Contents After Leaving the Classroom. In Proceedings of the 2020 ACM Conference on Innovation and Technology in Computer Science Education, Trondheim, Norway, 15–19 June 2020; ACM: Trondheim, Norway, 2020; pp. 545–546. [Google Scholar]
  37. Fuad, M.; Akbar, M.; Gloster, C.; Aun, N.; Zubov, L. A Mobile Educational Platform Based on Peer Influence and Instructional Scaffolding for Engaging Students in Out-of-Class Activities. In Proceedings of the 2021 International Conference on Advanced Learning Technologies (ICALT), Tartu, Estonia, 12–15 July 2021; IEEE: Tartu, Estonia, 2021; pp. 61–65. [Google Scholar]
  38. Filho, N.F.D.; Barbosa, E.F. A Requirements Catalog for Mobile Learning Environments. In Proceedings of the 28th Annual ACM Symposium on Applied Computing-SAC ’13, Coimbra, Portugal, 18–22 March 2013; ACM Press: Coimbra, Portugal, 2013; p. 1266. [Google Scholar]
  39. Soad, G.W.; Fioravanti, M.L.; Falvo, V.; Marcolino, A.; Duarte Filho, N.F.; Barbosa, E.F. ReqML-Catalog: The Road to a Requirements Catalog for Mobile Learning Applications. In Proceedings of the 2017 IEEE Frontiers in Education Conference (FIE), Indianapolis, IN, USA, 18–21 October 2017; IEEE: Indianapolis, IN, USA, 2017; pp. 1–9. [Google Scholar]
  40. The University of Texas at El Paso UTEP at a Glance. Available online: https://www.utep.edu/initiatives/at-a-glance/ (accessed on 31 October 2022).
  41. UNC. UNC Data Dashboard. Available online: https://myinsight.northcarolina.edu/t/Public/views/db_enroll/BuildYourOwnReport?%3Aembed=y&%3Ais%20GuestRedirectFromVizportal=y&iid=1 (accessed on 31 October 2022).
  42. Wisniewski, B.; Zierer, K.; Hattie, J. The Power of Feedback Revisited: A Meta-Analysis of Educational Feedback Research. Front. Psychol. 2020, 10, 3087. [Google Scholar] [CrossRef] [Green Version]
  43. Chapin, J. The Effect of Whiteboarding on Student Self-Efficacy in the Computer Science Classroom. Doctoral Dissertation, Virginia Tech., Blacksburg, VA, USA, 2022. [Google Scholar]
Figure 1. Exercises in learning paths. (a) Faculty side representation; (b) Student side representation.
Figure 1. Exercises in learning paths. (a) Faculty side representation; (b) Student side representation.
Education 12 00863 g001
Figure 2. Interactive activity in Dysgu. (a) Flashcard-based activity; (b) Drag-and-Drop activity.
Figure 2. Interactive activity in Dysgu. (a) Flashcard-based activity; (b) Drag-and-Drop activity.
Education 12 00863 g002
Figure 3. Self-reflection in Dysgu. (a) Progress comparison; (b) Badges.
Figure 3. Self-reflection in Dysgu. (a) Progress comparison; (b) Badges.
Education 12 00863 g003
Figure 4. Dysgu personalization and notifications. (a) Personalizing notification time and frequency; (b) A sample notification.
Figure 4. Dysgu personalization and notifications. (a) Personalizing notification time and frequency; (b) A sample notification.
Education 12 00863 g004
Figure 5. Students’ grade comparison. (a) Students’ grade comparison in UTEP; (b) Students’ grade comparison in WSSU.
Figure 5. Students’ grade comparison. (a) Students’ grade comparison in UTEP; (b) Students’ grade comparison in WSSU.
Education 12 00863 g005
Figure 6. Students’ start of work time.
Figure 6. Students’ start of work time.
Education 12 00863 g006
Figure 7. Students’ submission timeframe. (a) Students’ submission timeframe in UTEP; (b) Students’ submission timeframe in WSSU.
Figure 7. Students’ submission timeframe. (a) Students’ submission timeframe in UTEP; (b) Students’ submission timeframe in WSSU.
Education 12 00863 g007
Figure 8. Student response through an experience survey. (a) Grade comparison; (b) Pen-and-paper-based activity vs. Dysgu activity; (c) Personalized schedule; (d) Badges and points; (e) Progress comparison.
Figure 8. Student response through an experience survey. (a) Grade comparison; (b) Pen-and-paper-based activity vs. Dysgu activity; (c) Personalized schedule; (d) Badges and points; (e) Progress comparison.
Education 12 00863 g008
Figure 9. Students’ most selected features of the mobile learning system.
Figure 9. Students’ most selected features of the mobile learning system.
Education 12 00863 g009
Figure 10. Usage analytics from the app. (a) Most used features; (b) App usage time.
Figure 10. Usage analytics from the app. (a) Most used features; (b) App usage time.
Education 12 00863 g010
Figure 11. Summary of student comments.
Figure 11. Summary of student comments.
Education 12 00863 g011
Figure 12. Key factors of a digital platform for engaged student learning.
Figure 12. Key factors of a digital platform for engaged student learning.
Education 12 00863 g012
Table 1. Dysgu notifications.
Table 1. Dysgu notifications.
Notification TypeSample Notification
Encouraging
  • Congratulations! You have the highest score in the module so far.
  • Congratulations! You have the most points in the class.
  • Congratulations! You have a higher score than the class average on a module.
  • Congratulations! On average, you answered questions the fastest.
  • Congratulations! On average, you answered questions with fewer tries.
Cautioning
  • Careful! You have the lowest score in the module so far.
  • Careful! You have the lowest point in the module so far.
  • Careful! You have a lower score than the class average on a module.
  • Careful! On average, you answered questions the slowest.
  • Careful! On average, you answered questions with the most tries.
  • You did not work on any single problem today!
Neutral
  • You have module(s) due in less than 12 h!
  • Do not forget to finish your activities.
  • You did not solve any problem in the last 24 h!
Table 2. Details of Dysgu deployment.
Table 2. Details of Dysgu deployment.
UTEPWSSU
Without InterventionWith InterventionWithout InterventionWith Intervention
SemestersSpring 2021, Spring 2022 (Section B)Fall 2021, Spring 2022 (Section A)Spring 2019, Fall 2019Spring 2020, Spring 2022
Number of students66403117
Table 3. Student population information.
Table 3. Student population information.
Gender (in %)Course Load (in %)Out-of-Class Work Hours (in %)
UTEPWSSU UTEPWSSU UTEPWSSU
Male77.2758.82Less than full time (<12 credit hours)29.555.88Less than 10 h5052.94
Female18.1841.18Full-time (12–16 credit hours)61.3670.5910 to 20 h18.1817.65
Prefer not to say4.550Above 16 credit hours9.0923.5321 to 30 h15.9117.65
30 to 40 h11.3611.76
Above 40 h4.550
Table 4. Module comparison for Dysgu and LMS.
Table 4. Module comparison for Dysgu and LMS.
How They are SameHow They are Different
  • At WSSU, each assignment was given on a Monday, and students were asked to submit it by midnight on Friday. At UTEP, each assignment was given a week to complete.
  • At both universities, each assignment asked the same questions on the same topics.
  • At WSSU, students were asked to submit an MS Word document containing answers in Canvas LMS. At UTEP, students were given auto-graded questions in Blackboard LMS.
  • Students had time until the due date to answer the questions, whereas, in Dysgu, once students started to answer a question within a module, they had a set amount of time (around 5–10 min) to answer it. The module was open until the final due date.
Table 5. Modules used in this study.
Table 5. Modules used in this study.
UniversityModuleContent
UTEPFirstVariable
SecondConditional/Selection
ThirdMethod
WSSUFirstProgramming Ethics
Java Basics
Classes and Objects
SecondDecision Structures
Loops
Arrays
Text processing
Table 6. Descriptive statistics for experience survey response.
Table 6. Descriptive statistics for experience survey response.
(a)(b)(c)(d)(e)
µσµσµσµσµσ
UTEP (N = 24 *)3.380.713.420.723.50.723.330.863.420.78
WSSU (N = 17)3.270.702.870.993.20.942.80.863.00.85
* At UTEP, every participant did not complete the experience survey.
Table 7. Pre- and post-survey results.
Table 7. Pre- and post-survey results.
StatementsUTEPWSSU
i.
Faster feedback on assignments helps me to learn better.
Agreement increasesDisagreement increases
2.27%12.5%
ii.
Learning to use technologies for coursework is simply additional work beyond the normal coursework.
Disagreement increasesDisagreement increases
8.97%15.07%
iii.
I prefer out-of-class activities that can be completed using mobile devices.
Agreement increasesAgreement increases
3.59%28.68%
iv.
Using interactive out-of-class activities enhances my learning.
Agreement increasesAgreement increases
1.2%11.03%
v.
I like classes that allow me to practice concepts taught in-class in a hands-on-approach, such as take-home activities.
Agreement increasesDisagreement increases
1.2%0.37%
vi.
Performing out-of-class hands-on-exercises are NOT helpful in preparing for midterm and final exams.
Agreement increasesDisagreement increases
2.15%16.54%
vii.
When I see how my peers have performed in the out-of-class activities, it increases my course engagement.
Agreement increasesAgreement increases
3.47%4.8%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Fuad, M.; Akbar, M. Effect of Peer Influence and Self-Reflection on Scaffolded Out-of-Class Activity Administered Using a Mobile Application. Educ. Sci. 2022, 12, 863. https://doi.org/10.3390/educsci12120863

AMA Style

Fuad M, Akbar M. Effect of Peer Influence and Self-Reflection on Scaffolded Out-of-Class Activity Administered Using a Mobile Application. Education Sciences. 2022; 12(12):863. https://doi.org/10.3390/educsci12120863

Chicago/Turabian Style

Fuad, Muztaba, and Monika Akbar. 2022. "Effect of Peer Influence and Self-Reflection on Scaffolded Out-of-Class Activity Administered Using a Mobile Application" Education Sciences 12, no. 12: 863. https://doi.org/10.3390/educsci12120863

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop