Next Article in Journal
Neurocomputing Modelling of Hydrochemical and Physical Properties of Groundwater Coupled with Spatial Clustering, GIS, and Statistical Techniques
Next Article in Special Issue
Qualitative Analysis of Use of ICTs and Necessary Personal Competencies (Self-Efficacy, Creativity and Emotional Intelligence) of Future Teachers: Implications for Education
Previous Article in Journal
Composite Soil Made of Rubber Fibers from Waste Tires, Blended Sugar Cane Molasses, and Kaolin Clay
Previous Article in Special Issue
Advances in Research on Technological, Pedagogical, Didactical, and Social Competencies of Preservice TCFL Teachers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Design, Experimentation and Statistical Validation of a Methodology to Solve Complex Engineering Problems in Higher Education

by
Conrado Rosales-Torres
1,2,
Carlos Gijón-Rivera
1,*,
Claudia Lizette Garay-Rondero
1,3,
Álvaro Castillo-Paz
1 and
Gerardo Domínguez-Ramírez
1
1
Tecnologico de Monterrey, School of Engineering and Sciences, Ave. Eugenio Garza Sada 2501, Monterrey 64849, Mexico
2
Mechanical and Industrial Engineering Department, Universidad de las Américas Puebla, Ex-Hacienda Santa Catarina Mártir, San Andrés Cholula 72810, Mexico
3
Institute for the Future of Education, Tecnologico de Monterrey, Ave. Eugenio Garza Sada 2501, Monterrey 64849, Mexico
*
Author to whom correspondence should be addressed.
Sustainability 2022, 14(4), 2240; https://doi.org/10.3390/su14042240
Submission received: 15 November 2021 / Revised: 23 December 2021 / Accepted: 8 February 2022 / Published: 16 February 2022

Abstract

:
This article proposes a methodology that addresses the problem that many university professors often have with their students when facing complex engineering problems, causing frustration and desertion (abandonment of the problem to be solved). Although there are antecedents of works that emphasize the relevance of the realistic context in engineering problems and the importance of being structured in solving problems, we did not find measured effectiveness from the study of a group of students. This methodology focuses on engineering problems, in such a way that the decomposition of the problems in four steps responds to the solution process of the profiles of the analyzed subjects. The process followed in the preparation, implementation, validation, and reliability of this methodology is detailed. The experiment was designed to test both the effectiveness and reliability of the methodology. Four control groups for three different courses and periods were analyzed before and after the training of the four-step methodology. The observed factor was the variable score (0–100 points). The statistical analysis comprises descriptive statistics; Normality test for each population group; Paired t-Test/Wilcoxon test, and General linear model ANOVA (2 factors). The statistical analysis and tests show how the groups involved in the experiment obtained a significant benefit when the methodology for academic performance evaluations was applied.

1. Introduction

In the academic departments of science and engineering at universities, the phenomenon of students dropping out from the most demanding subjects occurs much more frequently than ever before [1]. In most cases this is triggered when students are asked to show their competence by solving a complex problem in a particular engineering subject [2]. This situation is aggravated when the student must utilize concepts and knowledge from more than one engineering subject. Complex problem solving has topped the list in both the 2015 and 2020 World Economic Forum (WEF) report presented in its five-year report on most desired job skills, focusing on cognitive aspects [3].
In Engineering, it is common to present problems that are related to real situations for training or evaluation. Tasks such as design, optimization or application of different mechanisms (whether they are pulleys or gears in vehicles or industrial machines) are commonly included in highly difficult engineering situations that can represent a challenge for most students [4].
The relevance of practice in engineering education has been documented. Brumm et al. found that 90% of engineering graduates who had cooperative experiences were practicing engineers, compared to only 54% of graduates from 1996 to 2009 who had no experiential education [5,6]. In addition, practice-based activities could enhance students’ job skills, work experience, and teamwork [7]. These promote mentoring and less superficial learning [8]. Among the different tools and techniques used so that students have experiences of solving real cases, when conditions are risky, are their simulations through specialized programs [9].
Contemporary life in the workplace is increasingly demanding. Engineering graduates are under heavy pressure to learn how to solve problems fresh out of college. Although one of the problems in the application of the Project-Based Learning (PBL) strategy within the subjects that addresses real issues, cases and situations, is that they are decontextualized to be an effective training [10]. In addition to the above, studies have been carried out in which it is essential to consider professional training in engineering, considering its scientific basis as a fundamental part [11].
Regarding the mentioned high-difficulty engineering problems, multidisciplinary subjects are needed to provide an accurate solution, such as physics, mechanics, ergonomics or materials science, to name just a few. Taking into consideration that engineering problems of high complexity ultimately lead students to desert before having a complete solution to the problem in question. Desertion often generates frustration that slows down the teaching and learning process, additionally, in critical cases, the desertion concludes in the termination of the student career due to a great loss of motivation [2].
Considering the above, the research objective of this work is to explore the results using a four-step methodology that aims to help students find a well-organized, coherent and efficient solution to highly complex engineering problems.
The general objective of this research triggers to study and validate hypotheses about how the four-step methodology could improve student performance and reflect higher levels in their proficiency assessments. In doing so, the findings will guide the authors in future studies to discover how the method could be correlated with a significant decrease in dropout from higher education STEM courses.
In order to design an automated tool to solve a significant set of problems we focus on the development of the fundamental steps in this task, which is based in the scientific method. These steps, in which is focused the present paper, were practiced by the students of the present study, while performing course contents during regular class sessions, in groups or individually, having access to the educator’s help to guide those who don’t apply it correctly. The path to reach gets the solution of complex problems can now be seen less hidden, that is, in the case it was obtained, the student outcome in this way has been developed. The intervention showed in the past sentence, is one strategy commonly in high school and lower grades, nevertheless has showed good result at university [12,13].
Section 2 presents the review of the literature on the development of competencies in complex problems and how this is a relevant topic, especially in Science, Technology, Engineering and Mathematics (STEM) courses. Furthermore, the relevance of education is discussed, particularly for the competence of solving complex problems in today’s society. Section 3 describes the methodology and experimentation with the student population. In Section 4 the results are presented, before and after the application of the method, using the different statistical tests to validate the analysis. Finally, Section 5 shows the conclusions of the experimental results obtained, in accordance with the hypothesis presented at the beginning of this work.

2. Relevance of Student Outcomes in Engineering Education

In the development of societies, the emphasis on the role of education serves as a critical factor on a worldwide basis. Examples of forums in which different ideas and collaborations on education are: summits of world organizations such as United Nations Educational, Scientific and Cultural Organization (UNESCO) and Economic Commission for Latin America and the Caribbean (ECLAC); conferences specialized in education as Educon from the Institute of Electrical and Electronics Engineers (IEEE), the Congreso Internacional de Innovación Educativa (CIIE), the European Conference on Education (ECE), etc.; and legislative congresses in different countries. These are only a small sample of different forums into which the relevance of education is mentioned, as it is a recurring theme, on the world agenda.
Talking specifically about education on the engineering field, there is a large number of definitions about engineering and STEM education, into which is important to identify the different dimensions that make it up to guide efforts in its study [14,15]. In the context of this article, engineering education is an activity in which a person acquires different skills with the objective of developing the ability to solve complex problems in science and engineering.
Engineering is classified as a region in which singularities are found, concepts from pure areas such as mathematics and physics are combined to solve specific situations in a realistic context related to the fundamentals reviewed in the theory [16]. Furthermore, Bao et al. emphasize the relevance of paying attention not only to the pragmatic competences of a university degree, but also to the formation of high reasoning skills, without conforming to the most applicable part [17]. Some factors internal to the undergraduate problem solver, both cognitive and affective, influence the development of problem-solving skills [10]. Being from this study’s perspective, the different backgrounds with which they start university, time they have to get a correct answer, its interpretation, feedback where mistakes made are corrected, to mention a couple of them. According to Jonassen the three mainly individual internal differences that mediate the ability to solve problems are prior domain knowledge, prior experience in solving similar problems, and cognitive skills [18].
Furthermore, teachers, as an empiric learning along years, have faced the experience where they find theoretical problems without a realistic and applied context, which, when solved by students, only a few are successful, and many of them drop out [19].
Lee et al. [4] report that students, in general, grow different strategies in order to solve problems, which could be resumed in the next steps: 1. to locate the related formulas, materials, and material properties to solve specific problems; 2. to use solved problems which are very similar to the one being studied, that could be situations where the same variable is asked; and 3. a desire to access external sources as a supplement or alternative to the textbook.
A good way for students to acquire required competences to solve complex problems consists in exposing them to real life situations [19,20]. In the last twenty years, students have greater access to information in digital broadcast media, and this amount is continuously growing; most of the information is related to leisure activities and general culture [21]. The previous mentioned motivates the search for strategies in teaching-learning activities, that incorporate elements belonging to these new characteristics of the environment around the university student, refocused on gaining competencies within their scope of professional preparation.
Considering that, just as high-performance athletes require adequate training, an engineer needs to train their scientific logic thinking in order so that their results are significant contributions to knowledge and have the potential to transform their surrounding reality and the social sector [22]. Therefore, giving the student the ability to carry out systematized activities that are part of the complex thinking that leads to the construction of knowledge [23].
Due to the increase of distractors during last years [24] and the number of dropouts in courses considered hard or difficult within the engineering area [1,25], it is of interest to focus special attention on pedagogical alternatives to be implemented within the classroom that would allow the student to reach intermediate goals and develop skills to achieve progress in complex problem solving.

3. Materials and Methods

3.1. The Four-Step Proposal Methodology

The four-step proposal methodology can be used in diverse STEM courses. The main objective is helpful to identify the phase(s) where the student usually fails when solving a complex problem. The process and method can help alumni to be able to improve their effectiveness and develop structure in their solutions, thus raising their performance regarding the resolution time (minutes) and score (0/100 points) and competencies in complex-problem solving.
Due to the findings in the literature review, described in Section 1, a series of steps that the student goes through when approaching a problem are proposed, discretizing its solution. The method arises first from professors’ teaching experience and later on complemented by being collated doing literature review on methodologies focused on engineering problems.
Therefore, the following stages in Section 3.2 and Section 3.3 are summarized as a compilation of the practice of counseling that a teacher typically offers to a student when they ask for an assessment in tasks assigned within a subject, or by their own initiative, based on the search for metacognitive learning. After several iterations, these stages are presented in a sufficiently simple form that can serve as a guide solving a complex problem.

3.2. Stages of the Methodology

Even though reading the problem carefully, it is an inherent step when solving an exercise, thus necessary to show it explicitly. Most teachers find that at the early stages of problem solving a large number of students don’t do it. The foregoing leads to the first obstacle when looking for an answer: faced with an inadequate reading, the student lacks the precise information available to them when starting the task. For this reason, once this has been done, the following stages are considered in the four-step methodology (Figure 1):

3.2.1. Stage 1: Sketch the Actual Situation Representing the Problem and Paraphrase the Requirement(s)

The student resorts too quickly to analytical developments that involve formulas and information at hand since it appears in the text of the problem [4]. Even when the situation has already been understood, the context and application must be specified. Stage 1 proposes developing an outline, no matter how simple, that will clarify how the problem is solved and that the solution presents a correct approach.
Once a problem is understood, solutions are created from different approaches, such solutions may be lost or diluted over time and the essence of the idea is lost. Therefore, sketching is a very effective tool for students to record clues and their mental processes for themselves and others [26]. In addition, a graphic intelligence is being reinforced, making sure that it is carrying out a cognitive process and its understanding from its interaction (visual characteristics), coding (internal interpretation of the image), inference (internal representation of the referent based on envelope a domain of knowledge) [27].

3.2.2. Stage 2: Clearly Identify and Write Theory, Concepts and Formulas Related to the Problem

Students analyze and describe the theory to begin addressing the problem. They register the equations and formulas and another relevant construct.
This stage is, on many occasions, the first stage that students carry out, thinking that it is the one that gives them the most significant advantage in solving the problem as described by Lee et al. [4]. Moreover, is vitally essential to be carried out properly, according with the experiences on good practices in solving complex problems presented in this article. It should be located in the current position within the proposed methodology.

3.2.3. Stage 3: Document the Information Necessary to Resolve the Problem. Make Assumptions If Necessary

Once the equations in the previous point of the methodology have been established, the available information must be reviewed ensuring that the variables involved in the expressions previously obtained can be solved, otherwise, it can be obtained from the different means available according to the work carried out up to the third stage. These sources can be:
  • Technical information tables.
  • Simplification of an equation due to the order of magnitude of some of the variables, or the interaction of two or more variables.
  • The constitutive equation between two or more involved variables that had not been involved.

3.2.4. Stage 4: Solve, Analyze and Verify That There Is No Conflict in the Dimensional Analysis during the Solution Process

The different elements necessary to solve the problem have been calculated, it only remains to obtain the value of the variable(s) of interest for the specific problem. It is important to mention that there must be sensitivity and coherence in the results, so it is proposed to reflect on them to identify any numerical inconsistencies.

3.3. Requirements of the Methodology

The proposal methodology needs three constraints to consider before, during and after the application. The detailed steps and processes delivered by professors are described in the following sections.

3.3.1. Before the Application

The process begins with the design and standardization of the observation tools (See Appendix A and Appendix B), based on each methodology stage to solve complex problems. This design considers relevant quantitative and qualitative variables and factors, such as the time execution during the application, the student’s scores (0–100 points), the number of students per collaborator and the performance and confidence during the experience solving the complex problem (both from the students and professors in charge). Next, the roles of the instructor, collaborators and students are explained below:
  • Instructor/Professor: the academic and intellectual authority figure that validates the correct assessment of the answers proposed by the students, as well as the development of the design, organization and qualification of the exam. It performs a continuous and periodic rotation in each of the groups to resolve doubts and/or attend to certain validations regarding the answers given by the students, as well as to confirm that there is order and transparency throughout the evaluation.
  • Collaborators: instructor’s assistants available to make observations according to the checklist. They carry out continuous monitoring of the performance, development and cooperation of those evaluated at the time of developing the fundamental solution to the proposed problem. They attend to the most fundamental doubts on the subject in case there is one in particular, direct the round of questions to each member of the team in a totally impartial way. They also take evidence and notes of everything done, said or pointed out by the team members during the evaluation, and they communicate with the lead evaluator if her presence is required.
  • Students: the population evaluated to demonstrate the effectiveness of the methodology. They maintain academic integrity and honesty throughout the scope of the assessment. They actively participate in the group portion of the exam, caring of healthy social norms among all team members, and respect the authority of the evaluators at all times.

3.3.2. During the Application

In this step, complex problems of a heuristic nature were developed by professors (and tested on a small group of collaborators present in the application of the test) according to the experience obtained by them in previous courses (See Section 3.4).
The tests are carried out in the same group experimentation, that is a set of dependent observations, that is measurements made on the same student’s population but under different conditions. The first condition is solving a problem before they were trained in four-step methodology, and the second condition is solving another problem with the same complexity level. In this last experimentation students use the four-step proposal methodology, once they were trained in it during regular class lectures and by practicing on exercises they must develop by themselves. Both groups solved the same problem. The observation for the declared variables and factors must be recorded in control sheets, as well as the competences directly registered on the standard rubric and checklist declared as observation tools.
It will be essential to note that the learning experience before and after using the four-step methodology and its outcomes/observations should be obtained from the same students, with the characteristics of being the same course/professor, with the same conditions of the professor, and with the standardized evaluation instruments ( Appendix A and Appendix B).

3.3.3. After the Application

The evaluation is carried out in a group manner between the teacher and the collaborators to identify any behavior that justified the dropping out of the problem and to deliberate on the possible causes. With this, data is collected for statistical analysis.
Ethical aspects must be taken care of given the sensitivity of the students’ information. The legal aspects are formalized through the statutes of the university, the students give their consent in writing to the handling of their data in the analyzes carried out.
It is worth mentioning that before the pandemic, small groups of students were assigned to teachers and collaborators to observe the students’ performance during the resolution of the pre and post-test. During the pandemic, the same procedure is performed, but making a group through the platform of digital delivery of classes.

3.4. Case Study Description

The four-step methodology has been proven in four higher education courses in the manufacturing, materials and mechanical design discipline. This case study helps to demonstrate the method, homologation and definition of the complex problems to solve. Furthermore, the data and quantitative and qualitative observations present information and exploration for the hypothesis declared in this research work. The following sections describe information about the population characteristics and data gathering.

3.4.1. Population Size, Description and Characteristics

The following section will describe the experimentation conducted during the periods: (i) August–December 2019, (ii) January–May 2020, and (iii) August–December 2019. Table 1 shows the experimentation consisted of three courses of the Mechanical Engineering area in the following format Course ID-Name: M2031-Manufacturing Technologies, M2014-Materials Applications, M2023-Mechanics of Materials. See Table 1 for the details about the professors’ assignment for each course and teaching period. The total population size observed during the experimentation from August 2019 to December 2020 was 61 students.
As described in Section 1, aligned to the research objectives, data collection and statistical analysis have obtained the response variable “score” before and after the four-step methodology. The authors recognize the implications of small sample sizes in this exploratory study for the courses observed. Therefore, future studies intend to record massive data through technological tools (apps, machine learning software) for better power in statistical analysis.
For the particular case of the M2023-Manufacturing Technologies course, the experimentation was directed by Professor A during the period (i) August–December 2019 (before the pandemic) and one year later, in the period (ii) August–December 2020 (during the pandemic in online teaching). Professor A also imparted the M2031-Manufacturing Technologies course during (iii) August–December 2020 (during pandemic in online teaching). Finally, the M2023-Mechanics of Materials course observations were gathered during the pandemic and taught by Professor B.
Each teacher selects and/or designs an exercise with the appropriate level of difficulty based on the number of variables [28]. Findings relevant to this case study can be found in the results regarding the student score (0–100 point) performance.
It is important to mention that the observations and assessments of the (i) August–December 2019 period were carried out in a presence and synchronous learning classroom where the professor observed, evaluated, and graded face to face. The particular case of (ii) January–May 2020 period was where the first evaluation (without-steps methodology), was in a face-to-face format, and the second evaluation was in a virtual classroom in a synchronous online teaching.
From March 2020 onwards, all evaluations were online, under the same control system, with the difference that evaluations, observations, and grades were conducted virtually through the Learning Management System Canvas.

3.4.2. Data Gathering Protocols

During the experimentations within the groups defined in Table 1, the following protocols gathered the data used for these analyses.
  • The methodology consisted of two evaluations separated by a reasonable period of time within the boundaries of the entire duration of the course. The assessments were aligned to the course curricula.
  • First evaluation:
    (a)
    The first evaluation (pre-training) lasted no more than 30 min. In such an evaluation, professors were asked to design a particular problem of high difficulty but entirely plausible to be finished during the brief evaluation time.
    (b)
    The students would have to solve the test with any procedure that would be best fitted for the problem in question. Moreover, the student was free to organize the information at their discretion.
    (c)
    In this first round of evaluation, the student was monitored in terms of time execution of the problem and the grade obtained after having solved the complex problem. Professor would grade the student´s performance according to the Rubric and Checklist in Appendix A and Appendix B, correspondingly.
  • Second evaluation:
    (a)
    At this step, once the professors were completely and broadly trained on the methodology. They conducted a training process for their students, an interactive training protocol, which involved live demonstrations of the use of the method in the context of the subject and homework and in-classroom problems that required the assertive use of the methodology.
    (b)
    The second test or evaluation (post-test) was contemplated with a duration of no more than 30 min.
    (c)
    During the second test, the student had to solve a high difficulty problem using the methodology for organizing their information and their procedure of solution.
    (d)
    The test evaluations and the numerical qualification were graded with a standardized four-levels Rubric and the score evaluation Checklist (see Appendixes A and Appendix B).
    (e)
    The procedure listed before, is the same for the problem evaluation without training (pre) and training (post) of the complex problem-solving four-step methodology.
Once the Instructors/Collaborators performed the observation, the numerical qualification (Appendix A) and competence level evaluation (Appendix B) of the students in each group and period, the Professors delivered the collected data to the group of analysts for consolidation, cleaning, and proper treatment towards the next steps of statistical analysis.

3.5. Statistical Analysis Methodology

As mentioned in Section 2, the indicators observed during the experimentation of the four-step methodology were the student’s score obtained in the complex problem, taking into account the technical aspect of the subject (Checklist-Appendix A), their performance while solving the problem (Rubric-Appendix B) and time of resolution.
When analyzing the behavioral data, it was clear that a more in-depth literature review and analysis and validated instruments were necessary to judge the observed behaviors to examine this qualitative variable. For the current scope of our investigation, we will avoid the interpretation of the behavioral data. Thus, only the scores will be taken into consideration, primarily due to its numerical and continuous nature, which are both characteristics that will ease the statistical analysis.
On the other hand, because the data collection methodology regarding the resolution time of each student was not standardized among the groups, unusual observations, biases, and dependence among the data were found, so it was decided to exclude them from the final analysis.
Furthermore, it is important to mention the variable score was obtained from the reported grades by the professors for each evaluation or problem. The context of periods during the pandemic because of COVID-19 produced some trends and bias within the groups’ data observations.
The statistical analysis consists in the following parts: Section 4.1 Descriptive Statistics; Section 4.2 Normality Test (for each group, pre and post-test); Section 4.3 Paired t-Tests and Wilcoxon Test; Section 4.4 General Linear Model ANOVA (2 factors). The software utilized for the mathematical study was Minitab®.
The first part of the analysis has the objective to illustrate the general results of the experiment, by providing general statistics of each group when they were both untrained and trained on their respective evaluations. The following part, normality tests for each group was conducted to verify the plausibility of analysis for each one of the groups involved, taking into consideration their various evaluation’s conditions. Once the normality of each group was determined, we proceeded to conduct paired t-tests to determine which groups presented a significant difference regarding their score’s means, if there was no significant difference, it was therefore stated that the methodology did not positively influence the student’s performance under pretty specific circumstances. On the contrary, if the bilateral paired t-test suggested that there was indeed a significant difference between the means, unilateral paired t-tests were conducted to determine whether the trained students performed better or worse than their untrained counterparts. In a particular case, it was not possible to conduct parametric statistics due to a non-normal distribution of the data, so a nonparametric analysis (Wilcoxon test) was used instead. Ultimately, a general linear model ANOVA of two factors was implemented to determine whether the course or the use of methodology itself was a factor that had a significant influence on the student’s performance or not.

4. Results

4.1. Descriptive Statistics

Table 2 shows the descriptive statistics of the four groups observed. The groups M2031 (i) August–December 2019 and M2014 (ii) January–May 2020 present a higher value of the means and a lower standard deviation for the tests with trained students. Other behavior was for both groups observed during period (iii) August–December 2020 M2031 and M2023. Although these two groups have different syllabus and objectives, their behaviors during the four-step methodology application were very similar, both pre-training and post-training tests. This sighting supports the exploration and findings of the four-step methodology in terms of the standardization it represents when applied to complex problem-solving.
In regard to the group M2031 (i) August–December 2019 (UTND), it is appreciated that its standard deviation and standard error are relatively high, such a behavior can be explained by analyzing the results obtained by the students. During the untrained evaluation, several students obtained a null score, primarily because of the student’s decision to give up on the problem and not presenting any formal solution, on the contrary there was an equal number of perfect scores on such evaluation, which ultimately led to a high value for the standard deviation and standard error. Once the methodology was instructed to the students and the second evaluation took place (TND), it is noticeable the change in the students’ performance, because there are not null scores for any students and the distribution was presented a more uniform behavior.

4.2. Normality Test

The results of the normality tests are displayed in Table 3. It must be clear that the data shown corresponds to the same population of students per course and their scores before and after the training on their respective assessments.
As shown in the table, the vast majority of populations involved in the experiment followed a normal distribution, which helped us use capable and parametric statistical tests to determine whether or not the four-step methodology was effective in terms of the change of the factor “score”. A significant finding is how the three populations of the three groups studied (M2031–Manufacturing Technologies, M2014–Materials Applications and M2023–Mechanics of Materials) before training were non-normally. After training, they reach a normal distribution.
In the M2014–Materials Applications course, when the students did not have any training on the method, they did not reflect a normal distribution. This data behavior can be related to bias concerning this course’s pandemic context and teaching characteristics during the period (ii) January–May 2020, previously described in Section 3.4.1 Once that group was trained in using a methodology, they presented a quite normal distribution, which can be interpreted as an implicit benefit of using a standardized methodology for the solution of complex problems in engineering.

4.3. Paired Test and Wilcoxon Test

The objective of this part of the analysis is to determine whether or not there was a significant difference in the mean between the trained population and the untrained population within each course defined in Table 1. The population always remains unchanged, the only considerable difference is the training process that is between the two evaluations, so the test that will be applied is the paired t-test, due to the dependency that UTND and TND have on each other, same sample, difficulty and professor but different circumstances on the training status about the application of the methodology.
As it can be seen in Table 4, among the four samples there are only two that show a significant difference regarding the means of each population’s sample. However, in this case, the only information that can be concluded is that three of the analyzed samples reflect a different performance when the methodology is used, the other two could ultimately lead to various conclusions, and for thus, further analysis was conducted, primarily to demonstrate whether there was some significance regarding the use of the methodology or not. For that purpose, unilateral paired t-test analyses were conducted, with the aim to determine whether or not there was a significantly higher mean within each population.
First of all, the group M2023 AD 2020 will be analyzed to determine if there was indeed a unilateral difference that could help us to demonstrate the progress achieved when the methodology is used. A unilateral test will be conducted, and in this case, we will always commence by analyzing if the untrained population had a lower mean when compared to the trained population. As it can be seen in the Table 4 for the M2023 AD 2020 group we have a significant difference between the two means, and the population that was trained to use the methodology presented a higher mean, thus demonstrating the potential of the methodology.
By applying the same paired t-test on the other population that presented a significant difference between the means of the untrained and trained population, the M2031 AD 2020 group, we obtained a similar result, in which the population that was trained to use the methodology presented a higher mean, thus demonstrating the potential of the methodology.

4.4. Wilcoxon Test

In regard to the population that did not follow a normal distribution, a Wilcoxon test was applied to determine if there was a significant difference between the median obtained in the non-normal distribution and the theoretical median value. The theoretical median value corresponds to the median value obtained by the normally distributed trained population, which resulted to be 65. Finally, due to the p-value of 0.045 on the bilateral test, it was concluded that there was indeed a significant difference between the hypothesized median of seventy and the median of the untrained populations. Furthermore, in order to obtain more important conclusions with respect to that group, it was decided that a unilateral Wilcoxon test will be performed in order to determine if the untrained group median was indeed lower than the hypothesized median. As it can be seen in the table, the results show that the median of the untrained group is certainly lower than the hypothesized median of seventy, which corresponds to the median of the trained group. With that in mind, the results of this test favor the use of the methodology.
During these tests (Paired t-tests & Wilcoxon test) a confidence level of 90.0% was used, due to the variability of the data.

4.5. General Linear Model ANOVA of Two Factors

Table 5 shows the information of the factors taken into consideration for the students who were involved in this research. For this reason, a proposal for applying the General Linear Model-ANOVA method of two variables aims to obtain the sums of squares of the averages of each population observed in terms of two different factors: training and course.
Factor A is defined as the engineering course that the students were taking and Factor B is defined as the training process on the methodology of solving complex problems. The data is shown as the average of the grades of each of the groups with the characteristic of having been trained or not.
The data observed in each of the populations in the entire experiment were validated in terms of their normal distribution, homoscedasticity and independence. Once this validation was done, we proceeded to obtain Table 6 of ANOVA.
The results of the ANOVA method for the two factors studied, suggest that there is indeed a significant effect of the methodology in which students were trained for systematic resolution in complex problems in regard to their reflected performance during the evaluations. Contrary to that, the course factor does not have a significant effect on the reflected academic performance.
Finally, Figure 2 presents the Confidence Intervals for the means with a Confidence Level of 95%, this reaffirms what was observed within the ANOVA analysis. It is remarkable the fact that the same student population, after being trained on the use of the methodology for solving complex problems in engineering, presented a significantly higher grade than at that previous evaluation when they were not trained on the use of the previously mentioned methodology.
While all statistical tests have been meticulously developed to ensure the maximum reliability possible in regard to our results’ suggestions, it must be denoted that the sample size for the experiment under a general criticism was notably small. In relation to that, the sample size was of 61 students, who were sorted among four different groups in the engineering department, which ultimately led to a small sample size of 13-19 students in each group. For the parametric statistical analysis applied it should be enlightened that a bigger sample size would be desirable if the objective is to give a broad validation to the methodology, and of course, not only the sample size must increase significantly in order to achieve that objective, but also the population involved in the experiment will have to be part of varied branches of the engineering department, to study the possibility of expansion of the methodology and its applicability to other academic areas.

5. Conclusions

As a general discussion about the statistical outcomes obtained, we can appreciate that in the set of observed data between groups and methodologies, the ANOVA model results strongly suggest that in the set of observed data between groups and methodologies a significant difference was observed on the average grades of students when the methodology was employed.
Additionally, the paired t-tests and Wilcoxon tests show that at least three out of four of the groups involved in the experiment obtained a significant benefit when the methodology was applied. The results obtained for the groups that did not present a significant difference when the methodology was used may indicate that there is a need for focalized control during the evaluations and grading process. As an overall description of the population’s performance, we can conclude a moderate increment of the reflected academic performance during the evaluations in which the methodology was used. However, it is essential to note that while an academic score (0–100 points) increase is widely desirable for many educators, the experience obtained with this exploration suggests that many variables can be significant regarding the methodology’s effectiveness. One of the relevant variables to consider is the training process given to students and educators. To establish better-structured conclusions about the methodology’s effectiveness on the experimental populations, deeper and more detailed analysis must be conducted on future studies.
As an overall discussion of the experiences learned throughout this iteration, the validation of the methodology is quite ambitious, primarily due to the objective of applying the methodology in a vast area of academic branches, and the major difficulty relies on the extensive number of human resources that must be included in the experiment. Considering that the only professors involved belong as well to the experimentation team, no further training process was required to teach the students how to use the methodology properly. The applicability of the methodology for branches of engineering that do not envision an explicit use of mathematics for physical theorems is still unbeknownst. Ultimately, the grading process was, in qualitative terms expressed by the professors involved, eased in a considerable manner, due to the step-by-step nature of the alumni’s solutions to the examinations, so not only the methodology can potentially benefit the students in terms of their reflected performance but also the professor’s work to grade evaluations can also benefit from the proper use of the methodology.

Future Work

The results obtained allow us to explore and propose future studies regarding this matter. Protocolized data gathering procedures and standardized grading processes will be ideal to amplify the potential of these studies, moreover, some variables like behavior, time efficiency and general acceptance of the methodology should be considered when designing more sophisticated experiments to obtain a greater understanding about the plausibility of massive usage of the methodology. (It is expected that the use of technological tools that improve behavioral observation on the students subjected to controlled evaluations will lead the statistical analysis to more innovative and conclusive directions.) Further and deeper analysis will have to be conducted in order to have a broader criterion on the general effectiveness and practicality about the usage of this methodology in engineering courses.
In a similar fashion, a strict model of training the educators into the use of the methodology may significantly increase the control of experimentation and the blocking of uncontrollable variables that may have caused noise in the test results presented in the current research work. The results of the statistical analysis open the possibility of applying the methodology in a massive way through some digital applications incorporating elements of artificial intelligence.
The analysis results suggest a promising future for the methodology, but there are still many areas of improvement in regards to the broad standardization of the design in question, the next steps that must be taken in order to have a wider understanding of the effectiveness and applicability of the methodology are the protocolized control over the data involved, such as a redesigned standardized rubric for the grading process, big data approaches in the scope of future studies, and a validated protocol for training in the sue of the methodology for both teachers and students. The field of action is also expanded to test it in different areas of engineering, so this will undoubtedly be the next step where there is a bank of problems based on the four-step methodology, and a quick and time efficient feedback will be provided by the intelligent tutor tool. By having a bigger set of recollected data from different branches of engineering with the optimum control over the potential bias that an experiment of a wide scale could present, the validation of the methodology will be more reliable and applicable to many areas of undergraduate academies. In addition to this, the alumni would have the possibility to learn from the process to improve the provided feedback. Among the areas of opportunity for improvement, we envision the following: Improve the monitoring process in the face of the pandemic (since one of the groups of the periods evaluated in 2020 were carried out remotely), although we only had one group in this modality, we observed that we need to employ a tool that allows us to monitor the behavior of students in a more detailed way. Improving the monitoring process will allow us to make richer and more varied observations.
Talking specifically about the results obtained during this round of experimentation, the data obtained is barely sufficient to consider the methodology promising and a potential success if some critical factors in its application are taken into consideration, apart from the statistical results previously mentioned in this paper, we have appreciated that the students behavior must be deeply analyzed for future iterations in order to have a more profound comprehension in regards to the influence that the methodology might have on the students’ perception of difficulty and their processes of reasoning while solving a complex engineering problem using the methodology, this will enlighten unbeknownst areas in which the methodology could potentially have a positive influence on the students and their reflected performance on academic evaluations, however, while such an information can be interesting to analyze, it may be complex to perform statistical test on the obtained data due to its qualitative nature, hence the method for collecting qualitative data from the observed behavior of students must be meticulously designed.
In the former experiment, any behavioral observation was excluded from the analysis and scope of the current research, so the data collected must only be interpreted as a partial reflection of each student’s performance; it is known that critical information regarding the behavioral influence of the methodology might have been lost in this round of experimentation. Without behavioral data on the current analysis, the numerical data obtained through examination might not be conclusive to many criteria of expansion in the use of the methodology. Also, as every single evaluation had a very strict time limit, the opportunity to freely study about the total impact that the methodology could have on the reduction of time completion of each evaluation was partially ignored, if the future evaluations are of a considerable extension the time effectiveness of the methodology may possibly be thoroughly examined, it ought to be considered for future iterations of the experiment.
In terms of the training process involved in the leisure of the methodology in different branches of engineering, a protocolized standard of training must be established in order to ensure a common homogeneity among the groups that will be trained to adopt the methodology, without such a protocol inacceptable bias can be presented in the collected data and the validation of the methodology will neither be solid nor correct.

Author Contributions

All authors have collaborated and made relevant contributions to the present work and made the following individual contributions to the paper. Conceptualization: C.R.-T., C.G.-R. Á.C.-P., data curation: Á.C.-P., C.G.-R., C.R.-T., C.L.G.-R.; formal analysis: C.R.-T., Á.C.-P., C.L.G.-R., C.G.-R.; funding acquisition: C.R.-T.; investigation: C.R.-T., Á.C.-P., C.L.G.-R., C.G.-R., G.D.-R.; methodology: C.R.-T., Á.C.-P., C.G.-R., C.L.G.-R.; resources: C.G.-R.; supervision: C.R.-T., C.G.-R., Á.C.-P.; visualization: G.D.-R.; writing—original draft: C.R.-T., C.G.-R., Á.C.-P., C.L.G.-R., G.D.-R.; writing—review and editing: G.D.-R., Á.C.-P., C.R.-T., C.G.-R., C.L.G.-R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Novus Grant with PEP number PHHT046-20ZZNV001, TecLabs, Tecnológico de Monterrey, México.

Institutional Review Board Statement

Ethical review and approval were waived for this study, due to no personal identity being involved or reported.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data sets were obtained from sensitive individual grades of students from institutional and author records, then processed losing their individual character for the statistical report. They are partially available upon request to the authors.

Acknowledgments

The authors would like to acknowledge the financial and technical support of the Writing Lab, Institute for the Future of Education, Tecnologico de Monterrey, Mexico, in the production of this paper. The authors would like to acknowledge Alberto Oliart and Verenice Chamu for their participation in the production of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Observation Guide

Table A1. Pre and Post Test. Four-step methodology for complex-problem solving.
Table A1. Pre and Post Test. Four-step methodology for complex-problem solving.
ObservedCriteriaPointsNon Observed
1. Methodology: There is a sequence of stages - beginning and end.25
2. There are diagrams, sketches, situations that give information that the student understands the problem.25
3. The equations used have the correct parameters according to the problem.25
4.- The result # is correct with a +/- (squiggle) +/- (squiggle) 5%.25

Appendix B. Rubric

Table A2. Pre and Post Test. Four-step methodology for complex-problem solving.
Table A2. Pre and Post Test. Four-step methodology for complex-problem solving.
Problem-Solving
Level
Deficient
A
Proficient
B
Satisfactory
C
Very Satisfactory
D
Sketching/ ParaphrasingDoes not finish defining
a good diagram that
allows them to establish
their variables of interest,
the paraphrasing of
statement does not allow
to appreciate that the
student understood what
is being asked. There are
serious inconsistencies
around the design of the
free body diagram and/or
significant errors in
interpretation of the
statement.
Manages to establish a
free body diagram that
vaguely summarizes
characteristics of the
system, likewise, its
interpretation is basic
and excludes important
information. However,
it has a brief paraphrase
that indicates an
understanding of what
is asked to perform.
The diagram is a tool
with very clear elements
that show the correct
physical understanding
of the fundamental
concepts of the subject.
It also presents a valid
geometry and according
to the guidelines of the
bibliography. In addition,
the paraphrase shows a
correct understanding
of what is requested
to solve the problem.
The free body diagram
is precise to the point
of accuracy. It presents
elements of the subject
that comply with the
indications of the
bibliography. In addition,
it is very clear each
element involved in the
system and the geometry
is neat. The paraphrase
demonstrates a deep
understanding of the
problem, and integrates
elements of theoretical
assimilation.
Theoretical concepts
and formulas
The formulas proposed
by the student do not
have a direct relation
with the topic to be
reviewed in the problem.
In addition to that, the
theoretical concepts
explained are deficient
or erroneous, thus
limiting the correct
interpretation of both.
Contains a reduced
number of correct
formulas, but these
present inconsistencies,
or are not complete.
The theoretical concepts
are not always correct,
or present a very basic
structure.
The student presents
enough formulas, and
the mathematical
definition is adequate.
Shows an acceptable
stratification of
theoretical concepts
and does not have
fundamental errors of
the subject.
It expresses all the
necessary formulas for
the correct resolution
of the problem and does
not present mathematical
errors in its definitions.
Likewise, the theoretical
concepts expressed
are accurate, assertive
and complete.
Information and
assumptions
Does not present a
complete informative
synthesis, and does not
establish any substantial
assumptions that could
delimit the application
of the subject matter to
the problem at hand.
Establishes a reduced
amount of information,
and the assumptions
established have no
transcendence for the
resolution.
Proposes an acceptable
amount of information,
while the assumptions
manage to define the
limits of the problem in
a basic way.
It manages to establish
the amount of information
necessary for the correct
resolution of the problem.
Likewise, it includes
assumptions of great
relevance for the
delimitation of the
applicability of the subject,
as a cornerstone of the
analysis.
Problem resolutionThe numerical solution
to the problem varies
by a factor greater than
20% of the expected
value.
The numerical solution
of the problem varies
by a factor between
10% and 15%.
The numerical solution
of the problem varies
by a factor between
5% and 10%.
The numerical solution
of the problem varies
by a factor of less than
5%.

References

  1. Meyer, M.; Marx, S. Engineering Dropouts: A Qualitative Examination of Why Undergraduates Leave Engineering. J. Eng. Educ. 2014, 103, 525–548. [Google Scholar] [CrossRef]
  2. McNeill, N.J.; Douglas, E.P.; Koro-Ljungberg, M.E.; Therriault, D.J.; Krause, I. Undergraduate Students’ Beliefs about Engineering Problem Solving. J. Eng. Educ. 2016, 105, 560–584. [Google Scholar] [CrossRef]
  3. McDermott, T.; Hutchison, N.; Crick, R. The Evolution of HELIX: A Competency Model for Complex Problem Solving. INCOSE Int. Symp. 2021, 31, 907–925. [Google Scholar] [CrossRef]
  4. Lee, C.S.; McNeill, N.J.; Douglas, E.P.; Koro-Ljungberg, M.E.; Therriault, D.J. Indispensable Resource? A Phenomenological Study of Textbook Use in Engineering Problem Solving. J. Eng. Educ. 2013, 102, 269–288. [Google Scholar] [CrossRef]
  5. Brumm, T.; Hanneman, L.; Mickelson, S. The Data Are in: Student Workplace Competencies in the Experiential Workplace. In Proceedings of the 2005 American Society for Engineering Education Annual Conference & Exposition, Portland, OR, USA, 12–15 June 2005; pp. 10.1271.1–10.1271.10. [Google Scholar] [CrossRef]
  6. Fortenberry, N.L. Teaching the Practical Skills. ASME Mech. Eng. 2011, 133, 36–40. [Google Scholar] [CrossRef] [Green Version]
  7. Thaenkaew, S.; Sungkedkit, C.; Ruangsiri, K. Promotion of Advanced Practical Skills Using Disciplinary Integration Teaching and Learning Activity for Electrical Engineering Education. In Proceedings of the 7th International Conference on Technical Education (ICTechEd7), Bangkok, Thailand, 25–26 March 2020; pp. 24–27. [Google Scholar] [CrossRef]
  8. Mir, M. Embedded peer-to-peer coaching for developing practical skills of engineering students. In Proceedings of the 15th International Conference on Interactive Collaborative Learning (ICL) 2012, Villach, Austria, 26–28 September 2012; pp. 1–5. [Google Scholar] [CrossRef]
  9. Lindgren, R.; Tscholl, M.; Wang, S.; Johnson, E. Enhancing learning and engagement through embodied interaction within a mixed reality simulation. Comput. Educ. 2016, 95, 174–187. [Google Scholar] [CrossRef] [Green Version]
  10. Karin, W. Engineering problem-solving knowledge: The impact of context. J. Educ. Work. 2017, 30, 840–853. [Google Scholar] [CrossRef]
  11. Bucciarelli, L.L.; Kuhn, S. Engineering Education and Engineering Practice: Improving the Fit. In Between Craft and Science: Technical Work in the United States; Barley, S.R., Orr, J.E., Eds.; Cornell University Press: Ithaca, NY, USA, 2018; pp. 210–229. [Google Scholar] [CrossRef]
  12. Hu, Y.; Xing, J.; Tu, L. The Effect of a Problem-oriented Teaching Method on University Mathematics Learning. EURASIA J. Math. Sci. Technol. Educ. 2018, 14, 1695–1703. [Google Scholar] [CrossRef]
  13. Giasiranis, S.; Sofos, L. Educational Technology. Int. Forum Educ. Technol. Soc. 2017, 20, 78–88. [Google Scholar]
  14. Van den Beemt, A.; MacLeod, M.; Van der Veen, J.; Van de Ven, A.; van Baalen, S.; Klaassen, R.; Boon, M. Interdisciplinary engineering education: A review of vision, teaching, and support. J. Eng. Educ. 2020, 109, 508–555. [Google Scholar] [CrossRef]
  15. Hasanah, U. Key Definitions of STEM Education: Literature Review. Interdiscip. J. Environ. Sci. Educ. 2020, 16, e2217. [Google Scholar] [CrossRef]
  16. Brown, J.R.; Newton-Smith, W.H. Mathematics, Role in Science. In In a Companion to the Philosophy of Science; Blackwell Publishers Ltd.: Oxford, UK, 2017. [Google Scholar] [CrossRef]
  17. Bao, L.; Koenig, K. Physics education research for 21st century learning. Discip. Interdiscip. Sci. Educ. Res. 2019, 1, 1–12. [Google Scholar] [CrossRef] [Green Version]
  18. Jonassen, D.H. Learning to Solve Complex Scientific Problems, 1st ed.; Routledge: London, UK, 2007. [Google Scholar] [CrossRef]
  19. Kirn, A.; Benson, L. Engineering Students’ Perceptions of Problem Solving and Their Future. J. Eng. Educ. 2018, 107, 87–112. [Google Scholar] [CrossRef]
  20. Khosronejad, M.; Reinmann, P.; Markauskaite, L. ‘We are not going to educate people’: How students negotiate engineering identities during collaborative problem solving. Eur. J. Eng. Educ. 2020, 46, 557–574. [Google Scholar] [CrossRef]
  21. Dogruer, N.; Eyyam, R.; Menevis, I. The use of the internet for educational purposes. Procedia-Soc. Behav. Sci. 2011, 28, 606–611. [Google Scholar] [CrossRef] [Green Version]
  22. Campitelli, G.; Gobet, F. Current Directions in Psychological Science. Curr. Dir. Psychol. Sci. 2011, 20, 280–285. [Google Scholar] [CrossRef]
  23. Lehtinen, E.; Hannula-Sormunen, M.; McMullen, J.; Gruber, H. Cultivating mathematical skills: From drill-and-practice to deliberate practice. ZDM Math. Educ. 2017, 49, 625–636. [Google Scholar] [CrossRef]
  24. May, K.E.; Elder, A.D. Efficient, helpful, or distracting? A literature review of media multitasking in relation to academic performance. Int. J. Educ. Technol. High. Educ. 2018, 15, 1–17. [Google Scholar] [CrossRef]
  25. Tayebi, A.; Gómez, J.; Delgado, C. Analysis on the Lack of Motivation and Dropout in Engineering Students in Spain. IEEE Access 2021, 9, 66253–66265. [Google Scholar] [CrossRef]
  26. Alok, G. Sketching—An Iterative Tool for Engineering Problem Solving. J. Eng. Educ. Transform. 2021, 34, 51–57. [Google Scholar] [CrossRef]
  27. Hegarty, M. The cognitive science of visual-spatial displays: Implications for design. Top. Cogn. Sci. 2011, 3, 446–474. [Google Scholar] [CrossRef] [PubMed]
  28. Bol, L.; Hacker, D.J.; Mattarella-Micke, A.; Beilock, S.L.; Seel, N.M.; Rosenstand, C.A.F.; Ell, S. Complex Problem Solving. Encycl. Sci. Learn. 2012, 37, 682–685. [Google Scholar] [CrossRef]
Figure 1. Four-step methodology’s diagram.
Figure 1. Four-step methodology’s diagram.
Sustainability 14 02240 g001
Figure 2. Interval Plot of Students’ Score-95% Confidence Interval for Means.
Figure 2. Interval Plot of Students’ Score-95% Confidence Interval for Means.
Sustainability 14 02240 g002
Table 1. Course name, ID, sample size and period.
Table 1. Course name, ID, sample size and period.
InstructorIDCourse NamePeriodCOVID-19 Context# Students
Professor 1M2031Manufacturing TechnologiesAugust–December 2019Before pandemic14
Professor 2M2014Materials ApplicationsJanuary–May 2020Before and during pandemic15
Professor 3M2031Manufacturing TechnologiesAugust–December 2020During pandemic13
Professor 4M2023Mechanics of MaterialsAugust–December 2020During pandemic19
Table 2. Descriptive statistical results-Detailed for each group. UTND = Untrained or pre-training test; TND = Trained or post-training test.
Table 2. Descriptive statistical results-Detailed for each group. UTND = Untrained or pre-training test; TND = Trained or post-training test.
Period(i) August–December 2019
M2031
(ii) January–May 2020
M2014
(iii) August–December 2020
M2031
(iii) August–December 2020
M2023
CourseUTNDTNDUTNDTNDUTNDTNDUTNDTND
Population size1414151513131919
Median65.0062.5065.0070.0045.0070.0050.0075.00
Mean55.0061.4359.1763.3346.9266.1550.2667.89
SE Mean11.406.125.204.366.246.284.785.45
Standard Deviation42.5022.9120.1316.8722.5023.2920.8523.60
Skeweness−0.29−0.06−1.78−0.85−0.040.150.59−0.73
Kurtosis−1.76−1.045.03−0.24−1.22−1.54−0.23−1.34
Table 3. Ryan-Joiner Normality test findings applied to each population. UTND = Untrained or pre-training test; TND = Trained or post-training test.
Table 3. Ryan-Joiner Normality test findings applied to each population. UTND = Untrained or pre-training test; TND = Trained or post-training test.
Period/FormatCourse ID-NameTrainingp-Value of
Normality Test
Findings in Normal
Distribution
(i) August–December 2019
Face-to-face
learning
M2031
Manufacturing Technologies.
UTND>0.100The population follows
a normal distribution
TND>0.100The population follows
a normal distribution
(ii) January–May 2020
Hybrid learning
M2014
Materials Applications.
UTND0.100The population does not
follow a normal distribution
TND>0.100The population follows
a normal distribution
(iii) August–December 2020
Virtual learning
M2031
Manufacturing Technologies.
UTND>0.100The population follows
a normal distribution
TND>0.100The population follows
a normal distribution
(iii) August–December 2020
Virtual learning
M2023
Mechanics of Materials.
UTND>0.100The population follows
a normal distribution
TND>0.100The population follows
a normal distribution
Table 4. Results for the paired and Wilcoxon tests for the 4 populations.
Table 4. Results for the paired and Wilcoxon tests for the 4 populations.
Population GroupTestType of Testp-ValueT-ValueFinal Criteria
M2031 AD2019Paired t-testBilateral paired t-test
H 0 : μ difference = 0
H 1 : μ difference 0
0.478−0.730There is no significative
difference between
the population’s means.
M2014 EM2020Wilcoxon textTheoretical median of 70
H 0 : η = 70
H 1 : η 70
0.045-There is indeed a
significative difference
between the untrained
median and the
hypothesized median of 70.
M2014 EM2020Wilcoxon textTheoretical median of 70
H 0 : η = 70
H 1 : η < 70
0.022-The median of the untrained
group is significantly
lower than the hypothesized
median of 70.
M2031 AD2020Paired t-testBilateral paired t-test
H 0 : μ difference = 0
H 1 : μ difference 0
0.052−2.16There is indeed a significative
difference between the
populations means, favoring
the trained population.
M2031 AD2020Paired t-testUnilateral paired t-test
H 0 : μ difference = 0
H 1 : μ difference < 0
0.026−2.16The difference between
the untrained group and
the trained group means
is less than zero, favoring
the trained population.
M2023 AD2020Paired t-testBilateral paired t-test
H 0 : μ difference = 0
H 1 : μ difference 0
0.007−3.04There is indeed a significative
difference between the
populations means, favoring
the trained population.
M2023 AD2020Paired t-testUnilateral paired t-test
H 0 : μ difference = 0
H 1 : μ difference < 0
0.004−3.04The difference between
the untrained group and
the trained group means
is less than zero, favoring
the trained population.
Table 5. Two-Way ANOVA data collection where Factor A: Training and Factor B: Course.
Table 5. Two-Way ANOVA data collection where Factor A: Training and Factor B: Course.
(i) August–December 2019(ii) January–May 2020(iii) August–December 2020(iii) August–December 2020
Factor B: Course
Factor A: TrainingM2031 Manufacturing
Technologies
M2014 Material
Applications
M2031 Manufacturing
Technologies
M2023 Mechanics of
Materials
Untrained55.0059.1752.3350.66
Trained61.4363.3365.1967.29
Table 6. ANOVA model for the experiment.
Table 6. ANOVA model for the experiment.
SourceDFAdj SSAdj MSF-Valuep-ValueAlphaResult
Training1281.44281.4412.370.025<0.05There is significant effect between
the training methodology
Course220.2710.140.450.669>0.050There is no significant effect
between courses
Error491.0122.75
Lack of fit247.2323.621.080.481
Pure error243.7821.89
Total7392.72
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Rosales-Torres, C.; Gijón-Rivera, C.; Garay-Rondero, C.L.; Castillo-Paz, Á.; Domínguez-Ramírez, G. Design, Experimentation and Statistical Validation of a Methodology to Solve Complex Engineering Problems in Higher Education. Sustainability 2022, 14, 2240. https://doi.org/10.3390/su14042240

AMA Style

Rosales-Torres C, Gijón-Rivera C, Garay-Rondero CL, Castillo-Paz Á, Domínguez-Ramírez G. Design, Experimentation and Statistical Validation of a Methodology to Solve Complex Engineering Problems in Higher Education. Sustainability. 2022; 14(4):2240. https://doi.org/10.3390/su14042240

Chicago/Turabian Style

Rosales-Torres, Conrado, Carlos Gijón-Rivera, Claudia Lizette Garay-Rondero, Álvaro Castillo-Paz, and Gerardo Domínguez-Ramírez. 2022. "Design, Experimentation and Statistical Validation of a Methodology to Solve Complex Engineering Problems in Higher Education" Sustainability 14, no. 4: 2240. https://doi.org/10.3390/su14042240

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop