Next Article in Journal
An Improved Cutting Force Model for Ultrasonically Assisted Grinding of Hard and Brittle Materials
Next Article in Special Issue
Toward Adaptability of E-Evaluation: Transformation from Tree-Based to Graph-Based Structure
Previous Article in Journal
Hydrogen Permeability of Epoxy Composites as Liners in Lined Rock Caverns—Experimental Study
Previous Article in Special Issue
Towards a Technological Ecosystem to Provide Information Dashboards as a Service: A Dynamic Proposal for Supplying Dashboards Adapted to Specific Scenarios
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Impact of Transparency in the Teamwork Development through Cloud Computing

by
María Luisa Sein-Echaluce
1,*,
Angel Fidalgo-Blanco
2,
Francisco José García-Peñalvo
3 and
David Fonseca
4
1
EtnoEdu Research Group, Applied Mathematics Department, University of Zaragoza, 50018 Zaragoza, Spain
2
LITI Laboratory, Geology and Mining Department, Technical University of Madrid, 28003 Madrid, Spain
3
GRIAL Research Group, Computer Science Department, Research Institute for Educational Sciences, University of Salamanca, 37008 Salamanca, Spain
4
GRETEL Research Group, La Salle, Architecture Department, Ramon Llull University, 08022 Barcelona, Spain
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(9), 3887; https://doi.org/10.3390/app11093887
Submission received: 30 March 2021 / Revised: 20 April 2021 / Accepted: 22 April 2021 / Published: 25 April 2021
(This article belongs to the Special Issue Innovations in the Field of Cloud Computing and Education)

Abstract

:

Featured Application

This research shows that identifying, managing, and making visible the evidence produced in a Cloud Computing system, while doing collaborative work, produces homogeneity in the work load between teams as well as among team members.

Abstract

Active educational methodologies promote students to take an active role in their own learning, enhance cooperative work, and develop a collective understanding of the subject as a common learning area. Cloud Computing enables the learning space to be supported while also revolutionizing it by allowing it to be used as a link between active methodology and students’ learning activities. A Cloud Computing system is used in conjunction with an active methodology to recognize and manage individual, group, and collective evidence of the students’ work in this research. The key hypothesis shown in this work is that if evidence management is made clear and evidence is consistently and gradually presented to students, their level of involvement will increase, and their learning outcomes will improve. The model was implemented in a university subject of a first academic year using the active Flipped Classroom methodology, and the individual, group and collective evidence is constantly worked with throughout the implementation of a teamwork method.

1. Introduction

The elimination of staffing and computing resource costs for companies that use Cloud Computing is one of the most notable features of this service [1]. Its implementation in education follows the same idea, resulting in the incorporation and establishment of ICT in educational institutions [2].
The three types of services offered by Cloud Computing are classified as SaaS (Software as a Service), PaaS (Platform as a Service), and IaaS (Infrastructure as a Service) [3]. Among them, SaaS, where an example is email, and IaaS, where data storage is an example, are the types that started to be used in education and are now fully implemented [4].
An important milestone in the incorporation of Cloud Computing in education has been social software 2.0, such as social networks, blogs and wikis [5]. Their use in education has several benefits, as students are already accustomed to using these programs, and there are learning theories focused on collaboration, cooperation, and connectivism [6].
The effect of Covid-19 on society is another significant milestone of Cloud Computing, as the use of teleworking, videoconferencing, telecare, and online teaching has increased. Cloud Computing systems have been widely used in education for the development of learning competencies that were traditionally done face-to-face [7,8,9,10,11].
Recent studies have particular relevance in this regard, as they have performed reviews of the existing literature on the use of Cloud Computing in education. These reviews indicate and prove the need to focus on the evaluation of the specific educational benefits of this application, as well as to analyze its impact through active methodologies, such as Flipped Classroom [12]. However, there has been a lack of research on the effect of Cloud Computing on educational processes [13].
On the other hand, in education processes based on Cloud Computing, it acts as a mediating element between teachers and students, as it provides support for activities that enable the acquisition of learning competencies. This same support emerges from student collaboration, interaction, and the outcomes of learning tasks and activities [14]. As a result, Cloud Computing is a place where all learning evidence can be found. The management of this evidence may form the basis of a new model, with Cloud Computing acting as a catalyst for change [15].
This management of evidence generated by Cloud Computing in education can be done in a way that is transparent to the student. As a result, the student is aware of the progress of his or her own learning process. But transparency is a feature that is associated with new ways of teaching, such as open education and open science [16,17,18]. Likewise, data transparency through open knowledge is one of the options that studies of Cloud Computing literature point to as a solution to one of Cloud Computing’s problems of data confidentiality and security [19,20,21,22].
This work applies an educational methodology based on the cooperation and interaction of students and generates evidence through a Cloud Computing system. This evidence is analyzed to determine the learning progress of each student, and this analysis is made transparent to the students through the technologies that make up the Cloud Computing system. As a result, in this research, a Cloud Computing-based training model is used, which can be applied to a variety of educational processes and approaches.
However, one of the challenges of teamwork is the difficulty in identifying evidence that can be used to measure an individual’s workload in relation to the team [23,24]. To be able to identify individual evidence (that affects each student), group evidence (that affects a group of students), and collective evidence (that affects the entire group of students in a class), the proposed model is implemented, along with an active learning methodology, during the development of teamwork, where the aforementioned three types of evidence are generated [25,26]. Finally, the proposal’s premise is that by making evidence management transparent throughout the learning process, it is possible to improve individual, group, and collective learning.
The context of this research was higher education and working specifically with students of the first academic year. To make this study transferable to any assignment, rather than focusing on the assignment in which teamwork is performed, it focuses on the student’s own acquisition of teamwork skills.
The hypothesis of research work is that transparency of individual, group, and collective evidence improves team member involvement, improves the learning of their own competence, and results in a more homogeneous involvement of different work teams.
The following section illustrates the proposed educational model, which is based on Cloud Computing and has been obtained by incorporating learning methods that have been validated in mixed training (face-to-face and distance). Following that, the context in which this research was conducted will be described. The analysis of results will continue through ongoing evaluations of learning progress and effect on student motivation, and the study will be completed by presenting the discussion and conclusions.

2. Proposal for an Educational Model Based on Cloud Computing

The proposed educational model is based on Cloud Computing. This should allow for flexibility when using it in various circumstances and educational environments, as it is structured as a series of interconnected layers that can be adapted when other methodologies are used. As a result, each layer must have a distinct capability, with clearly defined procedures, where different Cloud Computing-based methods can be used.
On the other hand, the model must allow for the following: the use of an active educational method based on Cloud Computing, the execution of processes to produce individual and collective evidence, and the analysis of evidence in real-time.
This necessarily involves a disclosure process on two levels: internally and externally to the team. Each team member can see what is going on within his or her own team on an internal level. Individual evidence (messages in the forums), group evidence (phase results), and follow-up evidence (tables that are filled in during the development) can all be used to achieve this. At the external level, each student can compare and contrast the evidence of other teams with that of his or her own. It is the responsibility of the teaching staff to share the evidence in order for this to happen. In this case, the evidence of each team’s relative participation is used, as well as the sharing of the group evidence of each team.
TRansparent InterAction in TEamwork Competence Acquisition within a cloud computing ecosystem (TRIATECA) is the proposed model’s name, and it is structured in three layers to include all of the features of cooperation, active learning, evidence generation and analysis, and decision-making transparency:
Layer 1: Learning method. It includes the training method using Cloud Computing tools and comprises educational processes such as active learning methodologies (Flipped Classroom, Gamification, and Challenge Based Learning, among others) [27].
Layer 2: Transparency process based on evidence for teamwork. This is the method for organizing teamwork so that all interactions between students, teachers, and activity performance leave analyzable evidence [28].
Layer 3: Evidence analysis. A Learning Analytics system is used to analyze the evidence of the interaction in layer 2 [29].
The TRIATECA model’s layers are outlined in detail below.

2.1. Layer 1- Learning Methodology

The mission of layer 1 is to provide an active learning method based on a fully online training modality. The Hybrid Flipped Classroom method was used, which is a fully online adaptation of the Flipped Classroom method used in face-to-face training [30].
From its inception, the Flipped Classroom method has been based on taking theory teaching out of the classroom and allowing students to obtain theoretical concepts before entering the classroom [31,32]. As a result, during class, students engage in active and cooperative learning based on their prior experience [33,34].
There are several variations of the Flipped Classroom technique. The degree of communication between previous activities and those carried out in the classroom is one way of classifying the various methods [35]. In this sense, the techniques are divided into three categories: (a) those with no interaction between the activities, (b) those in which the students’ learning from the previous activity is tested in class, and (c) those that use the results of the previous activities to work in the face-to-face class. The method that will be used to create the hybrid model is of type (c). It is called MicroFlipTeaching (MFT), and its use has been validated in theoretical classes [36], practical classes [37], and teamwork competency training [38].
The functional model used in face-to-face training and its adaptation to fully online training (layer 1) are shown in Figure 1. In face-to-face training, the MFT method is shown in Figure 1a. The following tasks are carried out remotely prior to the face-to-face class: concept acquisition through videos, doubt resolutions and discussions among students, and a practical task that applies the concepts covered in the video. The results obtained by the students when completing the practical task are used as resources in the face-to-face class to provide feedback on the learning process. The students generate knowledge that they integrate into their work space as a result of this feedback. All of this is combined with micro-lectures, which are used to either extend concepts or explain the following sessions.
The MFT model has been adapted to Cloud Computing, as shown in Figure 1b. We know that communication through a cloud system can be synchronous (when the interlocutors’ temporal coincidence is required) or asynchronous (when the interlocutors’ temporal coincidence is not required). This same feature is used in the MFT model, which divides the learning process into two stages: asynchronous learning (when teachers and students do not need to be on the same time to complete learning activities) and synchronous learning (when teachers and students must be on the same time to complete learning activities).
As a result, the MFT Cloud model adapts distance tasks and turns them into asynchronous learning, while synchronous learning includes face-to-face activities.

2.2. Layer 2- Transparency through the Generation of Evidence

The aim of layer 2 is to generate evidence through the interactions between students and the Cloud Computing system while developing different phases of teamwork.
The method used is Tuckman’s for small teams, in which the teams must go through four phases: Forming, Storming, Norming, Performing, and Adjourning [39]. Tuckman’s method has been extensively used by prestigious universities such as MIT [40] and organizations that use the International Project Management Association (IPMA) model [41,42]. As a result, the Tuckman model’s phases have been modified and expanded; Adjourning being replaced by Delivery and Documentation, and a new phase on Lessons Learned to apply to future projects being added.
This research uses the CTMTC model (Comprehensive Training Model of Teamwork Competence) [43], which provides group evidence that allows for the assessment of each phase’s development (Forming, Storming, Norming, Performing, Delivery, and Documentation).
Each phase has its own set of goals, as well as a set of collective and individual evidence that can be used to verify that these objectives have been met:
  • Forming defines the team’s formation and the coordinator’s election. Individual evidence is focused on messages exchanged to identify the team leader selection process. Group evidence is the team management structure itself (e.g., coordinator and sub-coordinator).
  • Storming is a phase in which the team may have disputes and conflicts. The team determines the theme of the task, the mission, the objectives, the need, and the rationale of the work’s outcome in this phase of the CTMTC method. Individual evidence is messages that the team exchanges in order to accomplish the objectives, while group evidence is a document that contains the mission and objectives, as well as the justification and need for them. Likewise, the roles of each team member are established. A map of responsibilities is created using the CTMTC process, which reflects the sections of the work that each member will be responsible for: expertise to use the Cloud Computing system, monitoring-leadership, and content creation tasks. Each team member should have all three responsibilities, and each member’s role is defined by their collection of responsibilities. The group evidence is a table that shows who is responsible for what in the team. Individual evidence is messages exchanged between team members in order to complete this phase.
  • Norming is a phase where the team is ready to begin working, and rules are defined to ensure the team’s proper functioning. A set of mandatory rules is drawn up in the CTMTC method to regulate the team’s day-to-day work as well as the consequences of not following the rules.
  • Performing is the phase in which the entire team works together to accomplish the project’s objectives and milestones. The CTMTC method establishes evidence for this phase in the form of a work schedule and a follow-up table that represents individual and group activities carried out week by week. The activities in the follow-up table are compared to the map of responsibilities to determine individual involvement. Individual evidence, as with the other phases, includes messages exchanged for phase completion and a work diary in which each team member reflects on the activities carried out together in real-time, including any form of incident or comment.
  • Documentation. These are the temporary documents and lessons learned that were generated during the process. Individual evidence includes video recordings of the lessons learned (all team members participate). The partial resources created and their organization are the group evidence. The resources are stored using Google Drive and a custom knowledge management system.
  • Delivery. The deliverables are as follows. The CTMTC system uses a web page that displays the final work and is incorporated into TRIATECA, as well as a video presentation of the final work’s outcome. The evidence, in this case, is limited to group evidence.
  • Lessons Learnt. These are resources that contain the information gained from a specific experience or activity. These are created for future teamwork, both within the team and for other work teams. Teams make videos with lessons learned about the different phases of the CTMTC method.
The students use three types of communication for each phase: a forum discussion thread, a wiki where progress tables are created, and a cloud storage system to identify, organize, and store the resources obtained in each phase. Similarly, each team member has a personal thread in a team forum that serves as a work diary, as well as a separate set of threads for following up on work, incidents, or thematic meetings. The generation of individual and group evidence by using the CTMTC method, which allows monitoring the acquisition of teamwork competence, has been validated in previous research [44,45].
This layer 2 of the model is represented in Figure 2, and the evidence is divided into two categories: individual and group. Individual ones are extracted from messages that each team member contributes to interactions with other team members, as well as from his or her work diary.
The work team’s members use the Cloud Computing system to exchange information in the following ways: by participating in forums, collaboratively creating tables, and sharing resources. The above-mentioned communication elements assist each phase of teamwork.
The transparency of each team’s learning progress is achieved by the following actions, as shown in Figure 2:
  • Responsibilities map: where each team member’s responsibilities are expressed. Each team member is responsible for three types of activities: designing tasks that contribute to the work’s final outcome, interacting with the Cloud Computing system, and leadership responsibilities. Each team has a leader who is in charge of the team’s activities. However, some of these roles are shared among the group’s members.
  • Individual monitoring maps (each student indicates the activities he/she is carrying out), and group monitoring maps (the group indicates the evolution of the work by linking the individual activities weekly).
  • Collaborative learning outcomes from the completion of the CTMTC phases are exchanged in real-time in the Cloud Computing system.
Each work team can keep track of each member’s activities (tasks, interactions with the Cloud Computing system, and leadership) in real-time (transparent processes for the team). As all of this is mirrored in the Cloud Computing system, the team will track the progress of the work being done and the fulfillment of each member’s commitments.

2.3. Layer 3- Analysis of Evidence

This layer’s aim is to examine the evidence gathered in the previous layer in order to monitor learning and identify the most common errors. The use of the Flipped Classroom method in an online environment produces group evidence, and teamwork in an online environment allows for access to individual evidence in collaborative production [46].
Individual monitoring is achieved with the aid of a Learning Analytics system that searches all of a forum’s threads for messages exchanged between team members [47]. The interaction of each team member in different phases of the development of the teamwork, in the work diary and in the other threads used for coordination, is obtained in this manner.
The Learning Analytics system also compares the number of messages exchanged in various teams and analyzes each team’s global involvement. Individual workloads, group workloads, and job completion dates can all be seen using this analytics [26]. This allows you to look at not just the workload but also how the work is organized (e.g., whether they have left everything to the last minute or have distributed it properly over time).
The monitoring of group evidence, which is obtained by evaluating the monitoring tables and the resources provided in each phase of the development of teamwork, is used to identify common errors (CTMTC). The common mistakes made by students during the production of each phase of teamwork are identified following the study.
Figure 3 shows the working scheme of this layer 3. A few hours before the synchronous session, the teacher analyzes the individual and group evidence and then presents it to all of the students in the teaching group, resulting in the collective evidence.
  • Analysis of individual evidence. The messages and views reflected in each forum’s threads are analyzed using a Learning Analytics system. Since each phase of teamwork is linked to at least one forum thread, each team member’s involvement in the phase’s development can be monitored. Additionally, the start and end dates of each phase and each member’s role in the overall development of the teamwork can be monitored. If the individual evidence has been compiled, the teacher compares the findings (among the members of each team and among the teams) and transmits them to all of the teaching group’s students. The flow (a) of Figure 3 depicts the entire procedure.
  • Analysis of group evidence. The teaching staff is in charge of the evidence analysis. They can be of two types: results from each phase (for example, rules established during the “norming” phase) and monitoring tables that show each team’s progress. The teacher presents the findings to the entire teaching group, highlighting the most common errors in each phase as well as the teams responsible for them. Figure 3 illustrates the entire process.
In addition to error identification, other aspects associated with the responses are identified, such as severity, complexity and creativity.
  • Severity: Errors can have various levels of impact, from mild to severe. In this model, the severity is measured along with the error through the color of the traffic light, with green being slight errors, medium errors being yellow, and serious errors being red.
  • Complexity: Often, the result of a task does not have a single answer, and even the answer may be conditioned by other variables external to the problem. Such as subjectivity, which from one point of view makes the answer correct and from another point of view it is not. Answers of this type are used as a tool to encourage discussion in face-to-face/synchronous sessions.
  • Creativity: Some groups obtain adequate results by following processes to develop an activity that is different from the one expected by the teacher. This degree of creativity is used by teachers to explain different processes that lead to good results, which contributes to deeper learning of knowledge.

3. The TRIATECA Model’s Implementation Context

The model was applied in the first year of the Biotechnology degree program at the Technical University of Madrid in the subject “Fundamentals of Programming”.
The experience involved a total of 91 students: 48 from the morning teaching group and 43 from the afternoon teaching group. The part of the course in which the study was completed (teamwork) accounts for 30% of the overall grade. The aim of this part of the course is for students to enhance teamwork skills by completing a project in which they must improve the course’s content while also adding their experience to the course’s learning process.
Individual competence (involvement of each student in teamwork), group competence (development of teamwork stages), and final outcomes are the three components of teamwork assessment (final result of the teamwork). Each component accounts for a third of the overall teamwork grade. All team members share in the grading of group competency and the final outcome. Each team member has an individual competency grade, and it has an effect on the rest of the assessment. For example, even though the final results are well graded, if a student fails the individual competency, he or she will fail the teamwork. As a consequence, there might be students who pass and those who fail in the same team.
Because of the influence of the situation resulting from COVID-19, the classes were conducted entirely online using Cloud Computing. We used an eco-system consisting of a Moodle e-learning platform, 2.0 tools (wiki and online storage), Microsoft Teams for videoconferencing system, Youtube to manage the videos made, and a knowledge management system of our own design [45] to organize all of the resources generated during the teamwork’s development.
The Hybrid Flipped Classroom was chosen as the teaching method, as described in layer 1 of the TRIATECA model. Microsoft Teams was used to teach eight synchronous two-hour sessions between September and December 2020. Students worked asynchronously between synchronous sessions, using Moodle, a Wiki, Youtube, cloud storage, and the own knowledge management system to organize their partial results.
During the asynchronous sessions, students worked in teams to complete the teamwork phases as well as to build the final work. During the synchronous sessions, the entire teaching group participated in group sharing, and new topics and activities were clarified.
The students used Moodle forums to interact with each other and collaboratively developed both the contents of the final work and the resources associated with each teamwork phase. The forum messages are the individual evidence and the contents, together with the intermediate resources created, are the group evidence.
The teachers reviewed the group evidence shortly before the synchronous session to get a summary of the results of each team’s activities and prepare a document with the errors they made. Individual evidence, on the other hand, was analyzed using the Learning Analytics system to assess the workload of each team member.
A control group and an experimental group were chosen from among all of the students involved in the research. The errors made by the different teams were shown in the control group. Following that, each form of error was solved in collaboration with the students. Finally, micro-lessons were given to clarify the next steps to be carried out, overcoming doubts and energizing debates. The control group was made up of 48 students who were divided into eight teams and were part of the morning teaching group.
The experimental group’s sessions followed the same dynamics as the control group’s sessions. The difference was that, in addition to identifying errors by team (as in the control group), color was added to show the magnitude of the error, and the progress of the work was identified in comparison to previous phases. That is, each team’s involvement was compared (based on the total number of messages exchanged), and the people in each team who had worked the least were listed (by analyzing the messages in each team). The experimental group consisted of 43 students from the afternoon teaching group, who were divided into seven teams.
Since the students in the experimental and control groups cannot be selected at random, the research is quasi-experimental. The reason for this is the experiment must be done with a full teaching group (since it is carried out during the synchronous session, and this must be given to the whole group). The experimental group was selected at random from the morning and afternoon groups, and it fixed the control group.
The same Cloud Computing system was used in both the experimental and control groups, ensuring that all students have the same educational model and the variables that support the research work’s hypothesis can be contrasted: the mechanism of transparency in the management of evidence. The results obtained in the research are shown and contrasted below.

4. Results

The validity of contrasting the experimental and control groups is demonstrated in the first subsection (Section 4.1). The equivalence in the profile of the students and the treatment received by the teachers was investigated for this purpose. The results of group evidence transparency are shown in the second subsection (Section 4.2), collective evidence transparency in the third (Section 4.3), and individual evidence in the fourth (Section 4.4).

4.1. Homogeneity of the Expzerimental and Control Groups

The homogeneity of the control and experimental groups was established using a survey. Homogeneity is investigated from two perspectives: the student profile and teacher treatment.
The survey was completed by 83 students out of a total of 96 students, representing an 86.45% response rate. Out of a total of 44 people, 36 joined in the experimental group, accounting for 81%. In the control group, 47 people out of 51 took part, accounting for 92% of the total.

4.1.1. Homogeneity of The Experimental and Control Groups with Respect to Student Profile and Team Creation

The equivalence is investigated to ensure that both groups are homogeneous by asking the following questions:
  • Q1—university entrance qualification
  • Q2—the total number of times the student has enrolled in the course
  • Q3—gender
  • Q4—the method by which you joined the team. In this case, the possible values are as follows:
    Some of the members of the team were friends of mine
    I knew a few of the team members
    I did not know anyone on the team
    I was the team’s first sign-up
To determine if the distribution is normal, the Shapiro–Wilk test is used first. The p-value of the data series for both the control and experimental groups is shown in Table 1.
Table 1 shows that all of the values have a p-value of less than 0.01; hence, none of the data series are normal.
Equivalence has been investigated using the Wilcoxon p-value for unpaired samples to see if there are any major differences. All samples are categorized by group (experimental and control) and compared for the following data sets: university entrance qualification, gender, and team creation (Table 2).
All of the p-values obtained are greater than 0.01; thus, the student body profile is equivalent.

4.1.2. Homogeneity in the Way the Teachers Treat Their Students

Part of the MUSIC® Model of Motivation Inventory test was included in the survey [48], which is a model described as usable: “in any subject area (a) to design instruction that motivates students, (b) to diagnose strengths and weakness of instruction, and (c) to research relationships among factors critical to student motivation.” The model defines five fundamental principles [49,50]:
  • Feel empowered by having the ability to make decisions about some aspects of their learning.
  • Understand why what they are learning is useful for their short- or long-term goals.
  • Believe that they can succeed if they put forth the effort required.
  • Are interested in the content and instructional activities.
  • Believe that the instructor and others in the learning environment care about their learning and them as a person.
Here, we focus on the findings of the set of questions that measure the part “Caring” (the fifth principle, on the instructor’s care for students), despite the fact that we completed the test in its entirety. The following are the questions that make up the Caring indicator out of the total of 26 that make up the MUSIC®:
  • Q16.4 The teachers are available to answer my questions about the course work.
  • Q16.16 The teachers are willing to help me if I need help in the subject.
  • Q16.20 The teachers care about how I am doing in the subject.
  • Q16.22 The teachers respect me.
  • Q16.24 The teachers are friendly.
  • Q16.27 The teachers care about my feelings.
As in the previous section, we used the Shapiro–Wilk p-value test to determine whether the individual data series have a normal distribution (Table 3). Additionally, we used the Wilcoxon p-value for unpaired samples to determine whether they are equivalent (Table 4).
All of the p-values are less than 0.01; thus, none of the response distributions to the different questions are normal.
The Wilcoxon test is used on two unpaired samples to see if they are equivalent (Table 4). The equivalence between the groups (experimental and control) and the corresponding variable is represented in each row of the matrix.
All of the p-values obtained are greater than 0.01; thus, the treatment of the teachers in the experimental and control groups is equivalent.
Table 5 shows the mean values and standard deviations for each of the “Caring” indicator questions. The “Caring” measurement received a 5.3 out of 6 average.

4.2. Impact of Cloud Computing Transparency on Group Evidence

The errors made by the teams during the preceding asynchronous activities were displayed in each synchronous session. Each error is assigned a code and assigned to the corresponding team.
Figure 4 depicts an example from the Performing phase. The errors found in the teamwork phases were presented with a code, and the team that committed the error in the control group (Figure 4a).
In the experimental group (Figure 4b,c), a traffic light indicating the seriousness of the error was shown for each team in addition to the errors found and the team making the errors (green with a mark between 7 and 10, yellow for passing with a mark between 5 and 6, and red for failing, with less than 5). The traffic light technique is widely used to monitor data because it has been demonstrated to provide a fast and easy view of data processing through a dashboard [51]. The traffic light is used in the following situations:
  • The correction of previous phases’ errors by each team (Figure 4b).
  • The severity of each team’s error in the corresponding phase (Figure 4c).
Each error has a code that corresponds to:
  • E-1: It is very generic, and no details are given.
  • E-2: The resource is unavailable.
  • E-3: A part of the phase is missing.
  • E-4: The technology is not included.
  • E-5: There are repeated tasks.
  • E-6: Does not apply to this phase.
  • E-7: Work log is missing.
Once the collective evidence was exposed to the teaching group, the development of the synchronous session was exactly the same for each group: the errors and their solutions were explained (each error was corrected for a specific team), and the rest of the teams were instructed to correct the error if they had it.
The final evaluation of teamwork took place after the course had ended but before the date of the course’s final exam.
The results of the evolution of the errors in the final evaluation are shown in Figure 5. The results of the control group teams’ errors are shown in Figure 5a. The majority of the teams’ errors have not been corrected, as can be seen. Figure 5b shows that the majority of the teams of the experimental group have corrected their errors.
Everything looks green during the Storming-I and Norming phases for the following reasons:
  • Storming-I: Each team selects the work’s theme from a list of three choices (each option corresponds to a part of the subject). As a result, whatever the team chooses is accepted as correct.
  • Norming: Each team suggests its own norms during this phase, so the one defined by the team is accepted. The norms are shared for all teams during the synchronous session. Any rule from another team can be incorporated by each team. As a result, the final rules selected are deemed correct.
In this case, there is no distinction between the teams of the experimental and control groups during the Storming and Norming phases. There is, nonetheless, a difference between the phases of the responsibility map (Storming-II), planning (Performing-I) and execution (Performing-II).
Table 6 shows the error percentage in each phase according to the color of the traffic light for the experimental and control groups.
When we look at the individual scores for each student in each phase, we see that they all have a non-normal distribution since the Shapiro–Wilk test shows that they all have a value less than 0.01, as shown in Table 7.
The Wilcoxon p-value for unpaired samples was used to see if there were any significant differences in the scores achieved by members of the experimental and control groups. Table 8 displays the results.
As a result, there are significant differences in learning between the experimental and control groups, as all of the p-value are below 0.01.

4.3. Impact of Cloud Computing Transparency on Collective Evidence

No percentage of the workload per team was shown in each synchronous session with the control group. In the experimental group, the relative participation of each team in comparison to the others was shown, as well as the load situation from the start of teamwork to the date of the synchronous session where this exposition was made.
The alarm level is shown using three colors, with M representing the average workload of the experimental group teams.
  • Green: the workload is higher than M-(M*0.19).
  • Yellow: the workload is in the range between M-M*0.20 and M-M*0.39. This is a pre-alert state with the goals of advertising the danger of entering the red alert state and encouraging them to pass to the green alert.
  • Red: The workload is in the (M-M*0.40) range. Actually, this is the most important color because it shows that the teams in this category work considerably less than the rest of the teams, and these are the cases that need to be addressed.
Figure 6a depicts the alertness levels of the seven teams in the experimental group during the last three sessions: EX S6, EX S7, and EX S8. The result of the experimental group’s final evaluation, F EX, is also shown. The final evaluation of the 8 teams of the control group F CO is shown in Figure 6b.
Box plots of the maximum and minimum values, the median, and the workload distribution by quartiles used to visualize the workload distribution. The box plots are depicted in Figure 7.
It can be seen that as the experimental group progresses, the workload of the teams converges, eliminating the extreme loads. Only one final measurement was taken in the control group, and it is more similar to the experimental group’s first measurement in terms of extreme values.
In Figure 6, we observe that 6 of the 7 experimental groups have the green color in the final evaluation F EX (85.7%), 1 of the 7 groups has the yellow color (14.3%), and the red color does not appear. Table 9 shows the percentage of collective evidence depending on the error severity (the color of traffic light) for experimental (7 groups) and control groups (8 groups).

4.4. Impact of Cloud Computing System Transparency on Individual Evidence

The existence of individuals who work considerably less than the rest of the team is one of the issues that arise in the development of teamwork. It is critical to identify these team members so that they can be redirected or expelled by the team.
The number of messages exchanged in the forums to improve each phase of teamwork is one way to measure individual involvement. In this case, it is not so much the total number of messages sent by an individual as it is the percentage of those sent in comparison to the rest of the team.
It is possible to determine the number of messages exchanged by the components of each team using the Learning Analytics tool described in the model, either by phase or in total. The following thresholds were established to determine if a person works less than the rest of the team:
  • M = average number of team messages.
  • p = number of messages from a specific team member.
Table 10 depicts the colors associated with the alert level in relation to the number of messages posted in each team member’s forums. This serves to notify the teachers of each member’s behavior when interacting with the other members of his or her team.
The control group received no warning signs from the teachers. The experimental group was shown the alert levels in the last three sessions, with special attention paid to the students who were at the yellow or red level (Figure 8a). Individual participation was evaluated using the Learning Analytics system. Figure 8b shows that no one with the red alert level was detected in the experimental group, while they were detected in two teams in the control group.
A contingency table with the total percentage of alarm frequencies is shown in Table 11. The proportion of the green level in the control and experimental groups is the same. The experimental group has more yellow alerts, while the control group has considerably more red alerts, which are the most dangerous.
Table 12 shows a summary of the results of Section 4.2, Section 4.3 and Section 4.4 with the percentage of evidence considered as the worst level (red) between the group evidence (Figure 5), collective evidence (Figure 6), and individual evidence (Figure 8).

5. Discussion

In the context of this study, Cloud Computing technologies were of mandatory use because it was conducted in a pandemic situation. Therefore, a discussion could be established as to whether the methodology could work without Cloud Computing.
From a methodological point of view, Cloud Computing is a tool; therefore, the proposed methodology could also be considered to work without it. However, it would be extremely difficult to implement it for two reasons: on the one hand, less evidence would be generated, since much of it can be captured through student interaction with the Cloud Computing tools, and on the other hand, the effort to manage the whole process would be unfeasible without technology in contexts similar to the one in the research work (around 100 students per teaching group), as has been demonstrated in previous research [47].
Transparency is linked to citizen science as a means of broadening and extending the scope of research in research centers [18]. Transparency is also used in this work to move toward a learning process in which the students’ progress is made visible to them, both individually and collectively.
Students expect feedback from their teachers when they build knowledge and share it, according to previous research [44]. Based on this feedback, the learner should either improve his or her understanding or share it with the assurance that it is correct. This work applies this principle to the process of learning. Students consciously (e.g., by solving a problem) and unconsciously (e.g., by completing a task) produce evidence of their progress (e.g., with the messages they exchange with other students). As a result, in this research project, the teacher interprets the evidence and visualizes the individual, team, and collective progress [52].
It can be thought of as a continuous, diagnostic, and reverse-formative assessment process. Teachers use diagnostic and formative assessments in traditional methods to take various corrective actions if necessary [53,54,55,56]. The teacher makes the process transparent in this approach, so the students themselves take corrective actions, as shown in this work.
Students’ participation in the learning process itself [57,58,59,60,61] is one way to improve the learning process, and active methodologies [62,63,64,65] should be used in conjunction with continuous learning assessment [66,67,68,69,70,71].

6. Conclusions

Making the progress and evolution of learning-related aspects, such as errors and workload, transparent has had a significant impact on both learning and participation levels.
Both the type of error and the proportion of errors committed by work teams are very similar in the control and experimental groups at any given phase (see Figure 4a,b). The transparency process, in this case, consists of displaying the seriousness of the error (traffic light in Figure 4b) as well as the progression of errors committed in previous phases.
All of the teams, both the control and experimental groups, are given instructions on how to correct their errors. However, as shown in Figure 5 (evaluation at the end of the course), the experimental group’s teams have corrected the errors (Figure 5b), while the control group’s teams have not (Figure 5a).
Table 7 shows that the scores obtained by the teams in the control group and those in the experimental group are not equivalent in all phases analyzed, as determined by the p-value obtained by applying Wilcoxon. This shows that the transparency process affects the teams’ learning (as they correct their mistakes) and the outcomes of the various phases of teamwork; namely, it produces an improvement of the teamwork competence.
The workload of each team in the experimental group was also shown, which was calculated based on the messages exchanged by team members during the work. Figure 6a shows the workload evolution in the experimental teams. When a team has a red color (meaning its workload is considerably lower than the rest of the teams), the teams in the group become increasingly homogeneous. The evolution of those in red turns to yellow, then to green, and the different teams converge in terms of workload, as shown in Figure 6. As a result, it would take three sessions for the teams that work much less to stabilize.
In terms of the post-course evaluation, Figure 6 (F CO for the control group and F EX for the experimental group) shows that the control group has teams with red alerts, while the experimental group has none.
Figure 7 shows, using box plots, that the experimental groups’ first measurement (EX S6) is similar to the control group’s first measurement (F CO), indicating that both groups have teams with extreme values (both at the top and bottom). The boxes, on the other hand, show that the workload evolution of the experimental group’s teams, as shown in three synchronous sessions (EX S6, EX S7, and EX S8), tends to converge. There are no longer any teams with significant differences, and learning is becoming more homogeneous (the boxes are flattening).
As a result, it has been shown that having the workload visible continuously (in each synchronous session) results in a workload that is similar for all teams, varying between the average to 20% of the average. As a result, this transparency technique is perfect for ensuring that the workloads of different teams working on the same task are similar.
On the other hand, visibility of evidence on each team member’s individual workload has been critical, as team members strive to achieve a homogeneous workload. The same transparency technique was used for team workloads, i.e., a level of warning (traffic light) was displayed based on each member’s individual workload in relation to the other members of his team. There is no standardized transition between the different colors of the traffic light in this case since the red color can stay the same or change from yellow to red, or vice versa, as shown in Figure 8a. This does not necessarily mean that the person who is red for two sessions is the same person, and it may be a different member of the same team. In any case, the objective is for no one in the team to function significantly less than the rest of the team. It can be shown that no team in the experimental group has the red color in the final assessment process, although there are teams in the red color in the control group (Figure 8b).
Table 10 shows that 20% of the teams in the control group had members with a significantly lower workload (yellow and red colors), whereas no team in the experimental group had a red color. As a result, gradually exposing team members’ workloads to a red alertness level causes these members to raise their workload until it is homogeneous with the rest of the team.
The research team intends to carry out future work on the application of the TRIATECO model in types of learning other than teamwork and different training contexts. On the other hand, the research was conducted in a technological university and with first-year students. In these circumstances, cooperative work and technology are commonly used, and the incidence of these factors has not been analyzed in this research work. Therefore, the transfer of this methodology to other areas, such as non-university education, where Cloud Computing technologies and/or cooperative work are not commonly used, should be the subject of a future study.

Author Contributions

Conceptualization, M.L.S.-E. and A.F.-B.; methodology, A.F.-B. and D.F.; formal analysis, M.L.S.-E., A.F.-B., and D.F.; investigation, all authors; resources, F.J.G.-P.; data curation, M.L.S.-E. and A.F.-B.; writing—original draft preparation, all authors; writing—review and editing, all authors; supervision, M.L.S.-E. and F.J.G.-P.; project administration, A.F.-B. and F.J.G.-P.; funding acquisition, all authors. All authors have read and agreed to the published version of the manuscript. Please turn to the CRediT taxonomy for the term explanation. Authorship must be limited to those who have contributed substantially to the work reported.

Funding

This research has been partially funded by the Spanish Government Ministry of Economy and Competitiveness through the DEFINES project (Ref. TIN2016-80172-R), the Educational Innovation Services of the Technical University of Madrid (Project IE IE1920.0601), and the University of Zaragoza (Project PRAUZ_19_326). The authors would like to thank the research groups: EtnoEdu (https://socioconstructivismo.unizar.es), GRETEL La Salle (supported by the Secretaria d’Universitats i Recerca of the Department of Business and Knowledge of the Generalitat de Catalunya, 2017-SGR-934), GRIAL (http://grial.usal.es), and LITI (http://www.liti.es) for their support.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ali, M.B.; Wood-Harper, T.; Mohamad, M. Benefits and Challenges of Cloud Computing Adoption and Usage in Higher Education. Int. J. Enterp. Inf. Syst. 2018, 14, 64–77. [Google Scholar] [CrossRef] [Green Version]
  2. Pardeshi, V.H. Cloud Computing for Higher Education Institutes: Architecture, Strategy and Recommendations for Effective Adaptation. Procedia Econ. Financ. 2014, 11, 589–599. [Google Scholar] [CrossRef] [Green Version]
  3. Mell, P.; Grance, T. The NIST definition of cloud computing: Recommendations of the National Institute of Standards and Technology. In Cloud Computing and Government: Background, Benefits, Risks; National Institute of Standars and Technology. U.S. Departament of Commerce: Gaithersburg, MD, USA, 2011; pp. 171–173. ISBN 9781617617843. [Google Scholar]
  4. Sclater, N. ELearning in the cloud. Int. J. Virtual Pers. Learn. Environ. 2010, 1, 10–19. [Google Scholar] [CrossRef] [Green Version]
  5. Guha, R.; Al-Dabass, D. Impact of Web 2.0 and Cloud Computing platform on software engineering. In Proceedings of the 2010 International Symposium on Electronic System Design, ISED 2010, Bhubaneswar, India, 20–22 December 2010; pp. 213–218. [Google Scholar]
  6. Siemens, G. Connectivism: A learning theory for the digital age. Int. J. Instr. Technol. Distance 2005, 2, 3–10. [Google Scholar]
  7. García-Peñalvo, F.J.; Corell, A. La Covid-19: ¿enzima de la transformación digital de la docencia o reflejo de una crisis metodológica y competencial en la educación superior? Campus Virtuales 2020, 9, 83–98. [Google Scholar]
  8. García-Peñalvo, F.J.; Corell, A.; Abella-García, V.; Grande, M. Online assessment in higher education in the time of COVID-19. Educ. Knowl. Soc. 2020, 21. [Google Scholar] [CrossRef]
  9. García-Peñalvo, F.J.; Corell, A.; Abella-García, V.; Grande-de-Prado, M. Recommendations for Mandatory Online Assessment in Higher Education During the COVID-19 Pandemic. In Lecture Notes in Educational Technology; Springer Science and Business Media Deutschland GmbH: Berlin/Heidelberg, Germany, 2021; pp. 85–98. [Google Scholar]
  10. García-Peñalvo, F.J.; Corell, A.; Rivero-Ortega, R.; Rodríguez-Conde, M.J. Rodríguez-García, N. Impact of the COVID-19 on Higher Education: An Experience-Based Approach. In Information Technology Trends for a Global and Interdisciplinary Research Community; García-Peñalvo, F.J., Ed.; Advances in Human and Social Aspects of Technology (AHSAT) Book Series; IGI Global: Hershey, PA, USA, 2021; pp. 1–18. [Google Scholar]
  11. Fardoun, H.; González, C.; Collazos, C.A.; Yousef, M. Estudio exploratorio en iberoamérica sobre procesos de enseñanza-aprendizaje y propuesta de evaluación en tiempos de pandemia. Educ. Knowl. Soc. 2020, 21, 9. [Google Scholar] [CrossRef]
  12. Scalera, M.; Gentile, E.; Plantamura, P.; Dimauro, G. A Systematic Mapping Study in Cloud for Educational Innovation. Appl. Sci. 2020, 10, 4531. [Google Scholar] [CrossRef]
  13. Baldassarre, M.T.; Caivano, D.; Dimauro, G.; Gentile, E.; Visaggio, G. Cloud Computing for Education: A Systematic Mapping Study. IEEE Trans. Educ. 2018, 61, 234–244. [Google Scholar] [CrossRef]
  14. Rizzardini, R.H. Cloud interoperability service architecture for education environments. J. Univers. Comput. Sci. 2015, 21, 656–678. [Google Scholar]
  15. García-Peñalvo, F.J.; Rivero-Ortega, R.; Rodríguez-Conde, M.J.; Rodríguez-García, N. The institutional decisions to support remote learning and teaching during the COVID-19 pandemic. In Proceedings of the X International Conference on Virtual Campus, Tetouan, Marooc, 3–5 December 2020; IEEE: Piscatway, NJ, USA, 2020. [Google Scholar]
  16. Santos-Hermosa, G. La educación abierta en Europa: Avances, integración con la ciencia abierta y rol bibliotecario. BiD 2019. [Google Scholar] [CrossRef]
  17. Burgos, D.; Tlili, A. Openness as the key factor to support education in times of crisis. In Proceedings of the Eighth International Conference on Technological Ecosystems for Enhancing Multiculturality, TEEM’20, ACM International Conference Proceeding Series, Salamanca, Spain, 21–23 October 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 491–495. [Google Scholar]
  18. Nascimbeni, F.; Burgos, D.; Campbell, L.M.; Tabacco, A. Institutional mapping of open educational practices beyond use of Open Educational Resources. Distance Educ. 2018, 39, 511–527. [Google Scholar] [CrossRef]
  19. Suciu, G.; Anwar, M.; Istrate, C. Mobile Application and Wi-Fi Network Security for e-Learning Platforms. In Proceedings of the Conference »eLearning and Software for Education« (eLSE), Bucharest, Romania, 11–12 April 2019; “Carol I” National Defence University Publishing House: Washington, DC, USA, 2019; Volume 1, pp. 393–399. [Google Scholar]
  20. Amor, A.B.; Abid, M.; Meddeb, A. Secure Fog-Based E-Learning Scheme. IEEE Access 2020, 8, 31920–31933. [Google Scholar] [CrossRef]
  21. Amo, D.; Alier, M.; García-Peñalvo, F.J.; Fonseca, D.; Casañ, M.J. Protected users: A moodle plugin to improve confidentiality and privacy support through user aliases. Sustainability 2020, 12, 2548. [Google Scholar] [CrossRef] [Green Version]
  22. Amo, D.; Alier, M.; García-Peñalvo, F.; Fonseca, D.; Casañ, M.J. Privacidad, seguridad y legalidad en soluciones educativas basadas en Blockchain: Una Revisión Sistemática de la Literatura. RIED Rev. Iberoam. Educ. Distancia 2020, 23, 213. [Google Scholar] [CrossRef]
  23. Friess, W.A.; Goupee, A.J. Using Continuous Peer Evaluation in Team-Based Engineering Capstone Projects: A Case Study. IEEE Trans. Educ. 2020, 63, 82–87. [Google Scholar] [CrossRef]
  24. O’Neill, T.A.; Pezer, L.; Solis, L.; Larson, N.; Maynard, N.; Dolphin, G.R.; Brennan, R.W.; Li, S. Team dynamics feedback for post-secondary student learning teams: Introducing the “Bare CARE” assessment and report. Assess. Eval. High. Educ. 2020, 45, 1121–1135. [Google Scholar] [CrossRef]
  25. Conde, M.A.; Colomo-Palacios, R.; García-Peñalvo, F.J.; Larrucea, X. Teamwork assessment in the educational web of data: A learning analytics approach towards ISO 10018. Telemat. Inform. 2018, 35, 551–563. [Google Scholar] [CrossRef]
  26. Conde, M.Á.; García-Peñalvo, F.J.; Fidalgo-Blanco, Á.; Sein-Echaluce, M.L. Study of the flexibility of a learning analytics tool to evaluate teamwork competence acquisition in different contexts. In Proceedings of the CEUR Workshop, Nanjing, China, 4 December 2017; Volume 1925. [Google Scholar]
  27. García-Peñalvo, F.J.; Alarcón, H.; Dominguez, A. Active learning experiences in engineering education. Int. J. Eng. Educ. 2019, 35, 305–309. [Google Scholar]
  28. Khagram, S.; Thomas, C.W. Toward a platinum standard for evidence-based assessment by 2020. Public Adm. Rev. 2010, 70, S100–S106. [Google Scholar] [CrossRef]
  29. García-Peñalvo, F.J. Learning Analytics as a Breakthrough in Educational Improvement. In Lecture Notes in Educational Technology; Springer: Berlin/Heidelberg, Germany, 2020; pp. 1–15. [Google Scholar]
  30. Fidalgo-Blanco, Á.; Sein-Echaluce, M.L.; García-Peñalvo, F.J. Hybrid Flipped Classroom: Adaptation to the COVID situation. In Proceedings of the TEEM’20. Eighth International Conference on Technological Ecosystems for Enhancing Multiculturality, Salamanca, Spain, 21–23 October 2020; García-Peñalvo, F.J., Ed.; ACM: New York, NY, USA, 2020; pp. 405–409. [Google Scholar]
  31. Lage, M.J.; Platt, G.J.; Treglia, M. Inverting the classroom: A gateway to creating an inclusive learning environment. J. Econ. Educ. 2000, 31, 30–43. [Google Scholar] [CrossRef]
  32. Baker, J.W. The “Classroom Flip”: Using Web Course Management Tools to Become the Guide by the Side. In Selected Papers from the 11th International Conference on College Teaching and Learning; Florida Community College at Jacksonville: Jacksonville, FL, USA, 2000. [Google Scholar]
  33. Bergmann, J.; Sams, A. Flip Your Classroom: Reach Every Student in Every Class Every Day; International Society for Technolo-gy in Education: Washington, DC, USA, 2012; ISBN 978-1-56484-315-9. [Google Scholar]
  34. Väisänen, S.; Hirsto, L. How Can Flipped Classroom Approach Support the Development of University Students’ Working Life Skills?—University Teachers’ Viewpoint. Educ. Sci. 2020, 10, 366. [Google Scholar] [CrossRef]
  35. Sánchez-Canales, M.; García-Aranda, C.; Morillo-Balsera, M.C.; S-de-la-Muela, A.M.; Fernández-GutiérrezdelAlamo, L. Clasificación de los diferentes modelos de Aula invertida y su aplicación en la Universidad Politécnica de Madrid. In Proceedings of the Aprendizaje, Innovación y Cooperación como impulsores del cambio metodológico. Proceedings CINAIC 2019, Zaragoza, España, 9–11 October 2019; Sein-Echaluce, M.L., Fidalgo-Blanco, A., García-Peñalvo, F.J., Eds.; Servicio de Publicaciones Universidad de Zaragoza: Zaragoza, Spain, 2019; pp. 607–611. [Google Scholar]
  36. Fidalgo-Blanco, A.; Martinez-Nuñez, M.; Borrás-Gene, O.; Sanchez-Medina, J.J. Micro flip teaching ‒ An innovative model to promote the active involvement of students. Comput. Human Behav. 2017, 72, 713–723. [Google Scholar] [CrossRef]
  37. Fidalgo Blanco, Á.; Sein-Echaluce Lacleta, M.L.; García-Peñalvo, F.J. Indicadores de participación de los estudiantes en una metodología activa. In Proceedings of the Aprendizaje, Innovación y Cooperación como impulsores del cambio metodológico. Proceedings CINAIC 2019, Zaragoza, España, 9–11 October 2019; Sein-Echaluce, M.L., Fidalgo-Blanco, A., García-Peñalvo, F.J., Eds.; Servicio de Publicaciones de la Universidad de Zaragoza: Zaragoza, Spain, 2019; pp. 596–600. [Google Scholar]
  38. Sein-Echaluce, M.L.; Fidalgo-Blanco, Á.; Peñalvo, F.J.G. A repository of students’ resources to improve the teamwork competence acquisition. In Proceedings of the ACM International Conference Proceeding Series; Association for Computing Machinery: New York, NY, USA, 2015; pp. 173–180. [Google Scholar]
  39. Tuckman, B.W.; Jensen, M.A.C. Stages of Small-Group Development Revisited. Gr. Organ. Stud. 1977, 2, 419–427. [Google Scholar] [CrossRef]
  40. Stein, J. Using the Stages of Team Development|MIT Human Resources. Available online: https://hr.mit.edu/learning-topics/teams/articles/stages-development (accessed on 4 January 2021).
  41. IPMA. ICB-IPMA Competence Baseline Version 3.0; International Project Management Association: Nijkerk, The Netherlands, 2006. [Google Scholar]
  42. IPMA. IPMA Reference Guide ICB4 in an Agile World. Version 1.0; International Project Management Association: Nijkerk, The Netherlands, 2018. [Google Scholar]
  43. Fidalgo-Blanco, Á.; Léris, D.; Sein-Echaluce, M.L.; García-Peñalvo, F.J. Monitoring Indicators for CTMTC: Comprehensive Training Model of the Teamwork Competence in Engineering Domain. Int. J. Eng. Educ. 2015, 31, 829–838. [Google Scholar]
  44. Fidalgo-Blanco, Á.; Sein-Echaluce, M.L.; García-Peñalvo, F.J. Enhancing the main characteristics of active methodologies: A case with Micro Flip Teaching and Teamwork. Int. J. Eng. Educ. 2019, 35, 1–12. [Google Scholar]
  45. Fidalgo Blanco, A.; Sein Echaluce, M.L.; García Peñalvo, F.J.; Sánchez Canales, M. Validation of a semantic search engine for academic resources on engineering teamwork. Int. J. Eng. Educ. 2020, 36, 341–351. [Google Scholar]
  46. Conde, M.; Rodríguez-Sedano, F.J.; Rodríguez-Lera, F.J.; Gutiérrez-Fernández, A.; Guerrero-Higueras, Á.M. Assessing the individual acquisition of teamwork competence by exploring students’ instant messaging tools use: The WhatsApp case study. Univers. Access Inf. Soc. 2020, 1–10. [Google Scholar] [CrossRef]
  47. Fidalgo-Blanco, Á.; Sein-Echaluce, M.L.; García-Peñalvo, F.J.; Conde, M.Á. Using Learning Analytics to improve teamwork assessment. Comput. Human Behav. 2015, 47, 149–156. [Google Scholar] [CrossRef]
  48. Jones, B.D. Motivating Students to Engage in Learning: The MUSIC Model of Academic Motivation. Int. J. Teach. Learn. High. Educ. 2009, 21, 272–285. [Google Scholar]
  49. Jones, B.D. Motivating Students by Design: Practical Strategies for Professors; CreateSpace Independent Publishing Platform: Scotts Valley, CA, USA, 2015; ISBN 1517547504. [Google Scholar]
  50. Jones, B.D. User Guide for Assessing the Components of the MUSIC® Model of Motivation. Available online: https://www.themusicmodel.com (accessed on 24 March 2021).
  51. Zdonek, I. Project indicators visualization using an interactive dashboard. Sci. Pap. Silesian Univ. Technol. Organ. Manag. Ser. 2020, 143, 329–339. [Google Scholar] [CrossRef]
  52. Lázaro-Carrascosa, C.; Hernán-Losada, I.; Palacios-Alonso, D.; Velázquez-Iturbide, J.Á. Flipped Classroom and Aronson’s Puzzle: A combined evaluation in the master´s degree in pre universitary teaching. Educ. Knowl. Soc. 2021, 22. [Google Scholar] [CrossRef]
  53. Coulange, S.; Jouannaud, M.P.; Cervini, C.; Masperi, M. From placement to diagnostic testing: Improving feedback to learners and other stakeholders in SELF (Système d’Evaluation en Langues à visée Formative). Lang. Learn. High. Educ. 2020, 10, 195–205. [Google Scholar] [CrossRef]
  54. Myyry, L.; Karaharju-Suvanto, T.; Virtala, A.-M.K.; R Raekallio, M.; Salminen, O.; Vesalainen, M.; Nevgi, A. How self-efficacy beliefs are related to assessment practices: A study of experienced university teachers. Assess. Eval. High. Educ. 2021, 1–14. [Google Scholar] [CrossRef]
  55. Lucas Delgado, D.G.; Anchundia Delgado, F.E.; Zambrano Quiroz, P.M. Permanent Application of Diagnostic Assessment on Learning Teaching Process. Int. J. Linguist. Lit. Cult. 2019, 15, 34–45. [Google Scholar] [CrossRef]
  56. Saeed, M.; Latif, I. Teachers’ Perceptions about the Use of Classroom Assessment Techniques in Elementary and Secondary Schools. Bull. Educ. Res. 2018, 40, 115–130. [Google Scholar]
  57. Fonseca, D.; Martí, N.; Redondo, E.; Navarro, I.; Sánchez, A. Relationship between student profile, tool use, participation, and academic performance with the use of Augmented Reality technology for visualized architecture models. Comput. Human Behav. 2014, 31, 434–445. [Google Scholar] [CrossRef]
  58. Fonseca, D.; García-Peñalvo, F.J. Interactive and collaborative technological ecosystems for improving academic motivation and engagement. Univers. Access Inf. Soc. 2019, 18, 423–430. [Google Scholar] [CrossRef] [Green Version]
  59. Grijpma, J.W.; de la Croix, A.; Kleinveld, J.H.; Meeter, M.; Kusurkar, R.A. Appreciating small-group active learning: What do medical students want, and why? A Q-methodology study. Med. Teach. 2020, 1–18. [Google Scholar] [CrossRef]
  60. Hyun, J.; Ediger, R.; Lee, D. Students’ Satisfaction on Their Learning Process in Active Learning and Traditional Classrooms. Int. J. Teach. 2017, 29, 108–118. [Google Scholar]
  61. Riswanto, A.; Aryani, S. Learning motivation and student achievement: Description analysis and relationships both. Int. J. Couns. Educ. 2017, 2, 42–47. [Google Scholar] [CrossRef] [Green Version]
  62. Esmaeili, M.; Eydgahi, A. The Relationship of Active Learning Based Courses and Student Motivation for Pursuing STEM Classes. In Proceedings of the ASEE Zone I Conference, Bridgeport, CT, USA, 3–5 April 2014. [Google Scholar]
  63. Fernandez-Rio, J.; Sanz, N.; Fernandez-Cando, J.; Santos, L. Impact of a sustained Cooperative Learning intervention on student motivation. Phys. Educ. Sport Pedagog. 2017, 22, 89–105. [Google Scholar] [CrossRef]
  64. Perez Poch, A.; Sánches Carracedo, F.; Salán Ballesteros, M.N.; López Álvarez, D. Cooperative learning and embedded active learning methodologies for improving students’ motivation and academic results. Int. J. Eng. Educ. 2019, 35, 1851–1858. [Google Scholar]
  65. Vergara, D.; Paredes-Velasco, M.; Chivite, C.; Fernández-Arias, P. The Challenge of Increasing the Effectiveness of Learning by Using Active Methodologies. Sustainability 2020, 12, 8702. [Google Scholar] [CrossRef]
  66. Cole, J.S.; Spence, S.W.T. Using continuous assessment to promote student engagement in a large class. Eur. J. Eng. Educ. 2012, 37, 508–525. [Google Scholar] [CrossRef]
  67. Coll, C.; Rochera, M.J.; Mayordomo, R.M.; Naranjo, M. Continuous assessment and support for learning: An experience in educational innovation with ICT support in higher education. Electron. J. Res. Educ. Psychol. 2007, 13, 783–804. [Google Scholar]
  68. Day, I.N.Z.; van Blankenstein, F.M.; Westenberg, P.M.; Admiraal, W.F. Explaining individual student success using continuous assessment types and student characteristics. High. Educ. Res. Dev. 2018, 37, 937–951. [Google Scholar] [CrossRef]
  69. Hernández, R. Does continuous assessment in higher education support student learning? High. Educ. 2012, 64, 489–502. [Google Scholar] [CrossRef] [Green Version]
  70. Holmes, N. Student perceptions of their learning and engagement in response to the use of a continuous e-assessment in an undergraduate module. Assess. Eval. High. Educ. 2015, 40, 1–14. [Google Scholar] [CrossRef]
  71. Holmes, N. Engaging with assessment: Increasing student engagement through continuous assessment. Act. Learn. High. Educ. 2018, 19, 23–34. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Functional model of Layer 1.
Figure 1. Functional model of Layer 1.
Applsci 11 03887 g001
Figure 2. Functional model of layer 2.
Figure 2. Functional model of layer 2.
Applsci 11 03887 g002
Figure 3. Functional model of layer 3.
Figure 3. Functional model of layer 3.
Applsci 11 03887 g003
Figure 4. Errors of the teams in the control and experimental groups during the Performing Phase (a) Control groups in corresponding phase, (b) Experimental groups in previous phases and (c) Experimental groups in corresponding phase.
Figure 4. Errors of the teams in the control and experimental groups during the Performing Phase (a) Control groups in corresponding phase, (b) Experimental groups in previous phases and (c) Experimental groups in corresponding phase.
Applsci 11 03887 g004
Figure 5. Traffic light for the control and experimental groups during the phases. (a) Control groups and (b) Experimental groups.
Figure 5. Traffic light for the control and experimental groups during the phases. (a) Control groups and (b) Experimental groups.
Applsci 11 03887 g005
Figure 6. Traffic light depicting alert levels of the teams shown in the last three synchronous sessions and the final evaluation (a) Experimental Groups and (b) Control groups (only final evaluation).
Figure 6. Traffic light depicting alert levels of the teams shown in the last three synchronous sessions and the final evaluation (a) Experimental Groups and (b) Control groups (only final evaluation).
Applsci 11 03887 g006
Figure 7. Box plots of the control group and the experimental group in the last three sessions and the final session.
Figure 7. Box plots of the control group and the experimental group in the last three sessions and the final session.
Applsci 11 03887 g007
Figure 8. Alert levels referring to a team member’s low participation (a) during the last three sessions of experimental groups and (b) final evaluation of experimental and control groups.
Figure 8. Alert levels referring to a team member’s low participation (a) during the last three sessions of experimental groups and (b) final evaluation of experimental and control groups.
Applsci 11 03887 g008
Table 1. Shapiro–Wilk test for experimental and control groups.
Table 1. Shapiro–Wilk test for experimental and control groups.
p-ValueExperimentalControl
Q11.409 × 10−104.554 × 10−13
Q24.043 × 10−133.614 × 10−15
Q30.0000000058129.647 × 10−10
Q40.00070280.0001633
Table 2. Homogeneity of the control and experimental groups based on the students’ profile and the team creation.
Table 2. Homogeneity of the control and experimental groups based on the students’ profile and the team creation.
Equivalencep-Value
Q10.8
Q20.8627
Q30.3594
Q40.35
Table 3. Shapiro–Wilk test between experimental and control groups.
Table 3. Shapiro–Wilk test between experimental and control groups.
Questionp-Value Experimentalp-Value Control
Q16.40.00065880.0005686
Q16.160.000000034360.0000002875
Q16.200.0000000064580.000000001403
Q16.220.0000000013151.345 × 10−12
Q16.240.0000055860.00002406
Q16.270.0000019280.00000002615
Table 4. Wilcoxon test between experimental and control groups.
Table 4. Wilcoxon test between experimental and control groups.
Questionp-Value
Q16.40.7382
Q16.160.2007
Q16.200.739
Q16.220.03096
Q16.240.5408
Q16.270.4151
Table 5. Mean values and standard deviation for the “Caring” questions.
Table 5. Mean values and standard deviation for the “Caring” questions.
QuestionMeanSD
Q16.44.571.12
Q16.165.350.91
Q16.205.580.73
Q16.225.700.73
Q16.244.831.24
Q16.275.370.80
Table 6. Error percentages in each phase for teams of experimental and control groups.
Table 6. Error percentages in each phase for teams of experimental and control groups.
Storming-IIExperimentalControl
Green85.7125
Yellow14.2937.5
Red0.0025
Performing-IExperimentalControl
Green85.7150
Yellow14.29 50
Red0.00 0
Performing-IIExperimentalControl
Green10062.5
Yellow012.5
Red025
Table 7. Shapiro–Wilk test for experimental and control groups.
Table 7. Shapiro–Wilk test for experimental and control groups.
p-ValueStorming-IIPerforming-IPerforming-II
Control0.000001370.0000000019680.000000007835
Experimental6.583 × 10−126.583 × 10−126.583 × 10−12
Table 8. Wilcoxon test between experimental and control groups.
Table 8. Wilcoxon test between experimental and control groups.
WilcoxonStorming-IIPerforming-IPerforming-II
p-value0.00000036520.00015360.00144
Table 9. Percentage of collective evidence depending on the error severity.
Table 9. Percentage of collective evidence depending on the error severity.
%GreenYellowRed
Control62.512.525
Experimental85.714.30
Table 10. Colors that correspond to the level of alert in relation to the number of messages in each forum.
Table 10. Colors that correspond to the level of alert in relation to the number of messages in each forum.
Alert LevelColor
Si p > M − (M * 0.19)Green: homogeneous participation
M − (M * 0.20) < = p < = M − (M * 0.39)Yellow: Pre-alert state
p <= M − (M * 0.40)Red: Alert (possible freeloader)
Table 11. Percentage of different alert levels in the experimental and control groups.
Table 11. Percentage of different alert levels in the experimental and control groups.
Group/ColorControlExperimental
Green13.313.3
Yellow20.033.3
Red20.00
Table 12. Percentage of red level in the control and experimental groups for group, collective, and individual evidence.
Table 12. Percentage of red level in the control and experimental groups for group, collective, and individual evidence.
% Red AlertGroup CollectiveIndividual
Control252520
Experimental000
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sein-Echaluce, M.L.; Fidalgo-Blanco, A.; García-Peñalvo, F.J.; Fonseca, D. Impact of Transparency in the Teamwork Development through Cloud Computing. Appl. Sci. 2021, 11, 3887. https://doi.org/10.3390/app11093887

AMA Style

Sein-Echaluce ML, Fidalgo-Blanco A, García-Peñalvo FJ, Fonseca D. Impact of Transparency in the Teamwork Development through Cloud Computing. Applied Sciences. 2021; 11(9):3887. https://doi.org/10.3390/app11093887

Chicago/Turabian Style

Sein-Echaluce, María Luisa, Angel Fidalgo-Blanco, Francisco José García-Peñalvo, and David Fonseca. 2021. "Impact of Transparency in the Teamwork Development through Cloud Computing" Applied Sciences 11, no. 9: 3887. https://doi.org/10.3390/app11093887

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop