Next Article in Journal
Multi-View Cosine Similarity Learning with Application to Face Verification
Next Article in Special Issue
Redesign of the Statistics Course to Improve Graduates’ Skills
Previous Article in Journal
GATSMOTE: Improving Imbalanced Node Classification on Graphs via Attention and Homophily
Previous Article in Special Issue
A Creative Design Framework for Math Exam Questions Concerning Inequalities and Zeros of Functions with an Unknown Parameter in China National College Entrance Examination
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Implementation of Pedagogical Research into Statistical Courses to Develop Students’ Statistical Literacy

by
Dalibor Gonda
1,*,
Gabriela Pavlovičová
2,
Viliam Ďuriš
2 and
Anna Tirpáková
2,3
1
Department of Mathematical Methods and Operations Research, Faculty of Management Science and Informatics, University of Žilina, Univerzitná 1, 01026 Žilina, Slovakia
2
Faculty of Natural Science, Constantine the Philosopher University in Nitra, Tr.A. Hlinku 1, 94901 Nitra, Slovakia
3
Department of School Education, Faculty of Humanities, Tomas Bata University in Zlin, Stefanikova 5670, 76001 Zlin, Czech Republic
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(11), 1793; https://doi.org/10.3390/math10111793
Submission received: 26 April 2022 / Revised: 12 May 2022 / Accepted: 21 May 2022 / Published: 24 May 2022

Abstract

:
To participate fully in society, a person needs to critically evaluate statistical information. To be able to do this, it is appropriate to start developing the statistical literacy of young people in diverse levels in schools. In our article, we present a modified seminar on statistics for prospective humanities teachers. During the seminar, we planned a pedagogical experiment, the aim of which was to statistically verify the students’ own diagnostic competencies as prospective teachers of humanities. The statistical analysis confirmed that the implementation of real research into the course on statistics has a positive impact on the development of statistical literacy but does not change the attitude of students toward statistics and its use in the future teaching practice.

1. Introduction

Today, thanks to advanced information technologies, almost everyone has access to a large amount of information. Statistical information is largely represented among this information. Probably the most represented currently are statistically evaluated data, mainly in the form of graphs, concerning the COVID-19 pandemic. According to Best [1] “such statistics are not only intended to inform people, but to shape their attitudes and behavior.” Thus, the aim of statistics is not only to inform but, above all, to understand numerical information obtained in each context. At the same time, statistics seek to provide relevant evidence for the claims of experts in various fields as well as politicians. There is a broad consensus that in an increasingly quantitative world, it is essential to understand, interpret, and act on statistical information [2]. Thus, statistics play a key role in fulfilling the Istanbul Declaration [3], which promotes an evidence-based approach to social progress. The OECD recommends involving ordinary citizens in the development of society through appropriate investment in their statistical capacity. As a result of this OECD initiative, the availability of information in the form of statistical files is constantly increasing, which is gradually being made available by government statistical offices as well.
Providers of statistical information are encouraged to use all available technologies to help users understand the data provided [4]. It follows that to participate fully in society, one needs to be able to interpret statistical information. These are presented in graphs, tables, or verbally, and it is necessary to be able to critically evaluate whether certain conclusions can be justified based on the presented statistical results [5]. It has long been known that knowledge of statistics is essential in solving certain types of everyday problems [6] and statistical understanding is an important part of the skills of workers in various sectors e.g., [7,8].
It follows from the above that statistics-based information is ubiquitous today. However, according to experts e.g., [9,10], citizens’ statistical skills are not at a level where accessible data can be used and interpreted appropriately. This fact is also a challenge for teachers who prepare students for their future role in society. The challenge is to find effective ways to use large sets of data that are commonly available in teaching. This challenge applies to both mathematics and non-mathematics teachers, where students need to understand evidence-based arguments. Curriculum areas, such as geography, citizenship, and sociology, deal with complex contexts where the situation is influenced by several factors and where real data are relevant.
These subjects currently make little use of relevant quantitative information because students are expected to have major difficulty understanding multidimensional data [5]. The use of real data and their statistical interpretations is more interesting for students, but also for teachers, than working with data that lack a real context [11]. Based on the results of the research [12], it is necessary to strengthen the use of real data in school teaching, as teachers mostly use textbook data, which are not always up to date. Greater use of up-to-date data would create a suitable motivational environment for the development of students’ statistical literacy, not only in mathematics teaching. An explosion in the availability of information means that the term “statistical literacy” is currently considered a basic life skill of a fully functioning citizen [5]. Our goal was to implement the results of pedagogical research into the content and form of the Statistics Seminar and experimentally verify the effectiveness of this new method of teaching, i.e., whether the implementation of a new method of teaching will increase students’ level of statistical literacy. The partial goal was to find out whether the changed design of the seminar will affect the perception of statistics by students.

2. Statistical Literacy

The term literacy is used to describe people’s skill to achieve targeted behavior in a particular area. Increasingly, this concept suggests a wide range of not only factual knowledge and certain abilities but also adequate critical attitudes [13]. Thus, the definition of the term “statistical literacy” does not look for the minimum set of basic statistical knowledge and skills that a student (person) should master. Rather, it is a matter of defining a set of statistical knowledge and skills that will enable one to use statistical data appropriately in everyday life to shape one’s own attitudes and forms of behavior. At the same time, it should be noted that the concept of statistical literacy is a dynamic concept that changes flexibly with the changing requirements in real life for human statistical competencies. Broers [14] states that experts had to solve two fundamental problems when trying to define statistical literacy. One is to determine exactly what a statistically literate citizen should know about statistics. The second problem was to determine what other than statistical knowledge was needed to achieve statistical literacy.
According to Wallman [15], statistical literacy is the skill to understand and critically evaluate the statistical results that permeate our daily lives—along with the skill to appreciate the benefits that statistical thinking can bring in public, private, and professional lives.
Watson [16] presented a statistical literacy framework consisting of three levels of increasing difficulty:
  • A basic understanding of probabilistic and statistical terminology;
  • An understanding of statistical language and concepts when placed in the context of a wider social debate;
  • The questioning position that can be taken in applying concepts to contradict claims without a proper statistical basis.
Gal [13] extended the description of statistical literacy to cover two interrelated components:
  • People’s skill to interpret and critically evaluate statistical information;
  • Their skill to discuss or communicate their reactions to this statistical information.
According to Ben-Zvi and Garfield [17], statistical literacy includes basic and important skills that can be used to understand statistical information or research results. These skills include the ability to organize data, create and display tables, and work with various data representations. Statistical literacy also includes an understanding of concepts, vocabulary acquisition, and the symbols used in statistics, and it includes the perception of probability as a measure of uncertainty. A person who has achieved statistical literacy has acquired the skill to explain statistical processes and fully interpret statistical results. At the same time, they will develop aspects of statistical thinking, where “statistical thinkers” are able to criticize and evaluate the results of a solved problem or a statistical study [17].
Watson and Callingham’s analyses [18,19] led to the introduction of one basic variable for statistical literacy, which can be described by six levels. The increasing levels reflect more knowledge of statistical procedures, for example, concerning tables, graphs, averages, and probabilities; a better appreciation of the variations and contexts they encounter in tasks; and increased critical thinking in decisions that acknowledge uncertainty. Only at the highest level are students able to apply an understanding of proportional reasoning in performing statistical tasks.
The above attempts to define statistical literacy involve several important ideas. Wallman [15] emphasized the ability to appreciate, understand, and evaluate the statistical evidence that affects our daily lives. Gal [13] emphasized the ability to interpret, evaluate, and communicate statistical evidence and provided an initial model that assumes that statistically literate behavior includes both a knowledge component and a dispositional component. Ben-Zvi and Garfield [17] sought to distinguish statistical literacy from statistical reasoning and statistical thinking in terms of the types of understanding or cognitive outcomes. Watson and Callingham [18,19] established a hierarchy within statistical literacy. This indicated the direction in which the gradual development of statistical literacy should go.
The findings of Callingham and Watson [20] suggest that this hierarchy provides a sound basis for school-level curriculum development. At the same time, their research confirmed that the statistical literacy hierarchy can be used to monitor progress and that students can improve their performance through targeted and appropriate teaching. Overall, a suitable definition of statistical literacy has been found. It is clear from the above that the concept of statistical literacy includes basic and important skills that can be used to understand statistical information or research results. These skills include the ability to organize data, create and display tables, and work with various data representations. Statistical literacy also includes an understanding of concepts, vocabulary, and symbols and includes an understanding of probability as a measure of uncertainty.

3. Statistics for “Non-Mathematicians”

The natural environment for the development of mathematical and thus statistical literacy is the school. During their education, students encounter statistics, especially in mathematics lessons. At the same time, almost every university study program includes a statistics course. Participants in university statistics courses are often students who do not primarily study mathematics or statistics, and therefore it is relatively difficult to motivate them to study statistics and thus develop their statistical literacy [21]. The same is true of students in teacher training programs who are preparing to teach subjects other than mathematics. However, the substantial number of different statistics and their interpretations in everyday life leads to the fact that teachers are aware of the need for their own statistical literacy to effectively use quantitative data in the teaching of their subjects [13,17,22]. Many studies highlight the role of the teacher as one of the critical factors in relation to students’ progress in education e.g., [23,24].
The research by Callingham et al. [25] showed that higher levels of teachers’ statistical skills lead to a better performance of students in education. Recent research has identified teachers’ interest in understanding statistics in a context where it is necessary to interpret data obtained primarily from student testing [26,27]. It could also be classified as statistical literacy in the workplace and not as knowledge needed to teach students statistical literacy. This requirement of teachers can be used as the basic context for teaching statistics in teacher education programs. Within the courses, they would learn to statistically evaluate “data” that are obtained from students’ written work. Data from differently oriented questionnaires would also be evaluated, as the teacher can use questionnaires to find out the opinions and attitudes of his students. This would make the tasks in the statistics course authentic, i.e., tasks that naturally simulate tasks from normal teaching practice [28]. Hahn [29] points out the importance of using real contexts to teach statistics. For students, there are then numbers that statistically process the numbers in context [30] and then they are better able to link the data in a real research environment. Based on the above, we have created a new concept of a statistics course for future humanities teachers to strengthen their statistical literacy.

4. Description of a New Way of Teaching a Statistics Course

As we have already mentioned, the problem in traditional statistics courses is usually the lack of motivation among students. It is important for students and their performance to be motivated and engaged in learning the subject [31,32]. Students’ motivation can also be strengthened by changing the focus of the course [33]. Based on the above findings, we tried to motivate future humanities teachers to actively develop their own statistical literacy in our research by changing the focus of the seminar on statistics. We focused the statistics course on the application of statistical methods in pedagogical diagnostics. Diagnostic competence is generally considered to be a teacher’s skill to assess the learning and behavior of their students according to predetermined criteria. In addition, diagnostic competence includes the skill to support the learning processes of individual students [34,35]. Diagnostic competence can be considered as a “cross-cutting” competence and its level fundamentally affects the level of other competencies of the teacher. The skill to diagnose usually goes hand in hand with the quality of the teacher: they can adapt their teaching to the individual needs of their students [36,37]. Diagnostic competence is basically the teacher’s skill to recognize students’ abilities, assumptions, and specific characteristics and, subsequently, to positively influence students’ learning processes and performance [38,39].
A diagnostically competent teacher can match the assigned tasks with the knowledge and skills of their students [40]. This change of focus means focusing the course on the real problems of the teacher, which is important in teaching statistics for non-mathematicians [41]. According to Roboack, the main goal of the statistics course is for students to develop statistical literacy and understand the basic elements of statistics that can help them critically evaluate data outcomes in their area of interest and in their professional lives. Cobb [42] noted that most courses can be improved by emphasizing the use of empirical data and the concepts to which the data relate, at the expense of students having fewer statistical theories and fewer “recipes” for solving statistical problems. Thus, students should develop their statistical literacy by doing statistics and not statistical calculations [43]. Therefore, in our course, we did all the statistical calculations using the generally available Excel program, and thus we tried to eliminate students’ problems with mastering mathematical calculations [44].
In agreement with the students of the teacher’s study program, we planned a pedagogical experiment within the subject “Statistics Seminar”, which was voluntarily attended by 23 students—future teachers of humanities. The aim of the experiment was a statistical verification of the students’ own diagnostic competencies using statistics methods as future teachers of humanities. In the previous semester, these students had attended a seminar called Pedagogical Diagnostics, in which they acquired the basic diagnostic competencies of a teacher. During the seminar Pedagogical Diagnostics, the students became acquainted with, among other things, the classification sheet of talents [45], which we used in our planned pedagogical experiment.
At the first meeting of the Statistics Seminar, we asked students to fill in an anonymous questionnaire of their own design—pre-test. The questionnaire contained twelve questions. In creating the concept of the self-assessment questionnaire, we used the concept of Soto-Andrade [46], who conducted research focused on self-assessment. In his research, Soto-Andrade concluded that self-assessment is one of the possible ways to obtain feedback and thus a basis for improving the learning process. The aim of the questionnaire in our case was to self-assess students’ own statistical literacy and their perception of statistics before completing our proposed model of teaching statistics. Subsequently, according to Wallman [15], we divided the items (questions) included in the questionnaire into three dimensions:
  • Dimension A: the ability to interpret and critically evaluate statistical information;
  • Dimension B: the ability to appreciate the benefits that statistical thinking can bring in public, private, and professional life;
  • Dimension C: the perception of statistics. Within dimension C, we surveyed the perception of statistics by students.
Questions no. 1, 3, 4, and 10 were included in dimension A, questions no. 2, 5, 6, and 8 were included in dimension B, and questions no. 7, 9, 11, and 12 in dimension C. It is recalled that the answers to the questions were scaled, so students had the opportunity to choose one of the options from a five-point scale from 1 (disagree) to 5 (agree) when answering each question.
Before starting a new, experimental way of teaching a Statistics Seminar, we made the following working hypothesis:
Hypothesis 1 (H1).
The proposed innovative way of teaching statistics will increase students’ statistical literacy.
To verify the validity of the research hypothesis, we conducted a pedagogical experiment with students—prospective teachers—which was the basis of the proposed method of teaching statistics. Subsequently, after completing our proposed innovative model of teaching a Statistics Seminar, i.e., with an interval of 13 weeks, we asked the same students to fill out a questionnaire (post-test), which contained the same questions as the pre-test. Subsequently, we verified the validity of the research hypothesis H1 by statistical analysis and comparison of the results of the pre-test and post-test.
A new way of teaching statistics—the procedure.
To acquaint students with the use of several statistical methods and tools in the field of pedagogical diagnostics, we held Statistics Seminar in three basic phases.
In the first phase of the seminar, we focused with students (future teachers) on the implementation of data collection. Data collection took place within the monthly pedagogical practice of those students in the teaching study program who also attended the seminar on statistics. Students had to identify gifted pupils in the form of observation during classes and during breaks within their monthly pedagogical practice. The students, based on personal judgment, selected those pupils they considered gifted.
In the second phase of the seminar, the students again evaluated the students, but this time the evaluation was carried out using the Talent Classification Sheet (TCS). We remind you that TCS allows an indicative assessment of the level of students’ talent in the classroom, regardless of school success. To maintain the anonymity of the pupils, their names were replaced by codes. Students (future teachers) evaluated TCS in the class in which they performed each student’s internship, using twenty-one items, through a five-point scale, with one being the best grade and five the worst. According to this criterion, each student in the class created a ranking of pupils according to TCS, which is also freely available at https://vudpap.sk/wp-content/uploads/2019/01/vudpap-informacie-Klasifikacny-harok-talentov-v-ZS.pdf (accessed on 27 April 2022).
In the third phase of the seminar, we worked with students to evaluate the data obtained using selected statistical methods. At the same time, we placed great emphasis on the interpretation of the obtained results of statistical data evaluation, because one of the goals of this part of the seminar was to make students aware that developed statistical literacy contributes to increasing their diagnostic competencies, i.e., the use of statistical methods allows them to verify working hypotheses.
In this part of the seminar, we therefore sought an answer to the following research hypothesis:
Hypothesis 2 (H2).
Are the diagnostic competencies of all twenty-three students (prospective teachers) developed to the extent that they can objectively assess students’ abilities in two ways: through observation and through the “Talent Classification Sheet”?
In search of the answer to the given question, we challenged the students to, first, graphically represent the obtained evaluations of pupils, which were carried out by students—future teachers in pedagogical practice in 13 classes (Figure 1). In search of the answer to the given question, we first graphically represented the obtained evaluations of pupils, which were carried out by students—future teachers within the pedagogical practice in 13 classes (Figure 1). Results—we compared the arithmetic average of the point score of the selected pupils with the average score of other pupils in each class (Figure 1).
We pointed out to the students that, in Figure 1, we can see that the “selected pupils”, as expected, achieved a lower score on average in the overall score, calculated according to the Talent Classification Sheet. These results show that future teachers (students, internships) were able to identify more capable pupils in the class based on their observations. However, their assessment was not comprehensive for all selected pupils. We can also see from the graph that in classes C, E, I, and M, the difference between the compared groups of pupils (“selected” and others) is the least significant. Probably, the pupils in these classes did not manage to identify the group of so-called “Selected” pupils.
In the next part of the experiment, in cooperation with the schools where the students’ internships took place, the teachers gave the pupils of the examined classes selected tasks from the intelligence tests. In this case, the higher the number of points obtained meant a better result of the pupil in the intelligence test. Data from intelligence tests were processed by students (trainees), each for “their” class. This is because they calculated, in each class, the average scores obtained by the selected pupils as well as the average scores of the other pupils. Subsequently, the students presented both results graphically. Subsequently, both results were plotted. We remind you that pupils with higher intellectual abilities achieved higher scores in the intelligence test (Figure 2).
In Figure 2, we see that in all classes, except those already mentioned in classes C, E, I, and M, as expected, the selected students achieved significantly better scores from intelligence tests than other students.
In the next part, we informed students that to assess the level of their own diagnostic competencies, it is necessary to compare the compiled order of students according to TCS in each class and the order based on the results of intelligence tests. In other words, we compared with the students (future teachers) the rankings they assigned to the pupils in the individual classes based on TCS with the rankings of the pupils according to their score obtained in the tasks from the intelligence tests. Thus, we wondered if there was a relationship between students’ scores in TCS and students’ scores in intelligence tests.
We pointed out to the students that in case of connection verification, we would like to use parametric t-tests and Pearson’s correlation coefficient, while we emphasized the conditions for the use of these statistical methods. We explained to the students that because the above methods can be used only if the assumption of a normal distribution of the statistical set is met, we first verified the hypothesis of a normal distribution of the observed features by the Shapiro–Wilk test. Given that, based on the results obtained by the Shapiro–Wilk test, the assumption of a normal distribution of the observed traits was not met, to verify the relationship between the two traits, we used nonparametric methods between the pupils’ scores in TCS and the scores obtained in the intelligence tests, namely the Spearman rank correlation coefficient R, which, like the Pearson correlation coefficient, takes values from the interval ⟨−1, 1⟩. The students calculated the correlation coefficient in Excel (Real Statistics).
To express the degree of connection between the two features, i.e., between the pupils’ score in TCS and the score obtained in the intelligence tests, we used Spearman’s rank correlation coefficient R, which takes values from the interval ⟨−1, 1⟩. The students calculated the correlation coefficient in Excel. The students clearly listed the calculated values of the Spearman rank correlation coefficient in Table 1.
Based on the results shown in Table 1, we, together with students, can state that between the score given to pupils in TCS and the score obtained by students in intelligence tests in individual classes, there is a significant degree of binding in 9 classes (A, B, D, F, G, H, J, K, L). This means that in these classes, students (future teachers) have identified gifted pupils well. There is a moderate or weak correlation dependence in four cases, namely in classes C, E, I, and M. This low value of the rank correlation coefficient R in classes C, E, I, and M confirmed our assumption that not all future teachers are able to reliably assess the above-average level of development of the pupils’ mental abilities and thus identify the gifted or the most capable pupils. Therefore, diagnostic competencies are not sufficiently developed in some future teachers. If we generalize the results in individual classes, then we can state that most future teachers are able to use the Classification of Talents to identify gifted and the most capable pupils in the class.
Finally, together with the students (future teachers), using Spearman’s correlation coefficient, we calculated the degree of agreement between the performance of the observed pupils and their results in the intelligence test (R1) and the degree of agreement between the performance of the observed pupils and their score achieved in TCS (R2). The results are shown in Table 2. We proceeded in the same way as in the previous case when calculating the coefficient of rank correlation between the mentioned features.
Based on the results shown in Table 2, students who conducted the above pedagogical experiment in statistics can assess how closely the pupils’ results in the intelligence test are related to their learning outcomes, which are presented by their average achievement (R1). For example, in classes E, F, and H, there is an extremely high degree of agreement between intelligence and benefit, and in the other classes, there is a high degree of agreement. Based on these findings, we discussed with students the possible aspects that arise from this finding for evaluating the quality of work of teachers teaching in the given classes. We also discussed with the students in the seminar how it would be possible to evaluate the pupils’ approach to learning and fulfilling their school responsibilities in other ways. Based on the calculated values of the coefficient R2, it is also possible to compare the degree of agreement between the evaluation of students through TCS and their benefit. The results show that it is in classes C, E, I, and M where students—interns—least reliably identified gifted pupils, and there is also the lowest (medium) degree of agreement between the evaluation of students’ giftedness using TCS and their average benefit. In conclusion, the students were pointed out to the fact that this finding also supports the demonstrable reliability and argumentative power of statistical data processing.

5. Statistical Analysis of Research Results (Comparison of Pre-Test and Post-Test Results)

As we have already mentioned, we were also interested in whether the implementation of a new method of teaching statistics changed students’ self-esteem, not only in dimensions but also in individual issues. The main research hypothesis (H1: The proposed innovative method of teaching statistics will increase students’ statistical literacy…), for this reason, was divided into two working hypotheses H1a and H1b as follows:
Hypothesis 1 (H1a).
Students’ self-assessment in the pre-test and post-test in individual dimensions will be different.
Hypothesis 1 (H1b).
Students’ answers to the same questions in the pre-test and in the post-test will be different.
  • Verification of the validity of the working hypothesis H1a (Self-assessment in individual dimensions).
As we have already mentioned, we were interested in whether, by conducting pedagogical research within the Statistics Seminar, students changed their attitudes in individual dimensions—and if so, in which dimensions the respondents (students) had statistically significant changes. We first calculated the average scores obtained by the respondents in the individual dimensions in both the pre-test and the post-test (Table 3).
In Table 3, we can see that the values of the achieved scores in the individual dimensions in the pre-test and in the post-test are different. We wondered if these differences were also statistically significant. Because the assumption of a normal distribution of observed traits was not met, we used the nonparametric Wilcoxon signed-rank test, which is a nonparametric analogy of a paired parametric t-test, to verify the statistical significance of differences in the level of observed traits.
The observed features in our case were X and Y, where X is the average values of the answers to questions in dim A (i.e., dim B, dim C) before the start of the pedagogical experiment (in the pre-test) and Y is the average values of the answers after its completion (in the post-test).
We tested hypothesis H0: the distributions of the characters X and Y are the same as the alternative hypothesis H1, that they are not the same. We implemented the test in the STATISTICA program. After entering the input data, we received the following results in the computer output for each dimension: the value of the test criterion Z of the Wilcoxon one-sample test and the value of the probability p (Table 4).
We evaluate the test using a p-value. Because the calculated p-value of probability is a small number in dimensions A and B, we reject the tested hypothesis H0 at the significance level α = 0.01. This means that the implementation of the pedagogical experiment significantly changed the answers of students in dimensions A and B. However, in the case of dimension C, hypothesis H0 cannot be rejected (p = 0.226). This means that in dimension C, the students’ responses did not change significantly, i.e., there was no statistically significant change in self-assessment.
II.
Validation of the working hypothesis H1b (comparison of answers to individual questions in the pre-test and post-test).
We first calculated the average values of the respondents’ answers to individual questions in both tests (Figure 3).
From Figure 3, we can see that some average values of respondents’ answers to individual questions in both tests are different. Again, we were interested in whether these differences between the two tests in the answers to the individual questions are also statistically significant. To verify the validity of the working hypothesis H1a, we used nonparametric test methods, namely the χ 2 —independence test for a contingency table of type k × m . χ 2 —using the independence test for the contingency table k × m, we verified whether the answers to each of the 12 questions of the respondent are related to whether the respondent answered the given question in the pre-test or in the post-test. In this case, too, we performed the test using the STATISTICA program. After entering the input data in the computer’s output report, we received a contingency table for each question, the value of the test criterion χ 2 test and the value of p (Table 5).
Based on the results shown in Table 5, we can see that within dimension A, there were statistically significant changes in the answers to questions 1, 4, and 10. A statistically significant difference in the answers to question 1 (I can correctly interpret the pre-test statistical information as in the post-test) is illustrated in Figure 4.
In Figure 4, we can see that in the pre-test, up to 75% of students evaluated their ability to interpret statistical information as low, but in the post-test, already 78% of students evaluated this ability positively. There was a similar change in self-assessment in the answers to questions 4 and 10. In the case of comparing the answers of the respondents to question no. 3 (controlling the calculation procedures will increase my ability to evaluate statistical data), in the pre-test and post-test (Table 5), we state that the difference in the answers is not significant. This means that the students answered the question in both the pre-test and the post-test.
In Table 5, we can also see that, in dimension B, there were statistically significant changes only in the answers to questions 2 and 8. To compare the results more clearly, we expressed the individual levels of the responses of the respondents from the given scale to question 2 in the pre-test and post-test in percent (Figure 5).
Based on the results obtained using the χ 2 test and from Figure 5, it follows that the respondents had a notable change in self-evaluation, from a negative to a positive evaluation. While 43% of the respondents answered positively in the pre-test, in the post-test, it was up to 100% who agreed with the statement: I consider the contribution of statistics to everyday life to be significant.
Moreover, based on the results shown in Table 5, we can state that the answers of the respondents to questions 5 (I am interested in a deeper understanding of statistical procedures) and 6 (I will use the knowledge of statistics in my teaching practice) in the pre-test and post-test did not change significantly statistically—they even remained largely negative. The situation in the case of choosing the answers to question 6 is illustrated in Figure 6.
Based on the results of the statistical analysis (Table 5), we can also state that in dimension C there was no statistically significant change in the answers of the respondents in the pre-test and in the post-test to questions 7, 9, and 11. A statistically significant difference in the answers of the respondents to the questions in the pre-test and in the post-test was only found for question 12 (Figure 7).
In Figure 7, we can see that in the pre-test, 83% of respondents disagreed with question 12 (I was motivated to study statistics by attending the Statistics course), so in the post-test, only 26% of respondents were negative about this question.

6. Discussion

Based on the results obtained by the statistical analysis of the pre-test and post-test, we can state that based on the average values of the pre-test (Table 3), respondents in the pre-test rated their statistical skills as low—they answered the individual questions of the pre-test mostly negative (Figure 3). If we start from the formulation of the questions in the pre-test, we assume that one of the main reasons may be the lack of quality mathematical training. According to Cladera et al. [47], the lack of a mathematical basis for students is one of the main failures of students in statistical courses.
By comparing the average values of the post-test with the pre-test (Table 3) and the subsequent statistical analysis, we can see (Table 4) that there was a significant positive change in students in dimensions A and B, which followed students’ self-assessment in their statistical literacy. The statistical analysis confirmed the increase in students’ level of perception of their own statistical literacy, based on which it can be stated that the statistics course was set in such a way that it demonstrably supported the increase in the level of students’ statistical literacy.
Within dimension A, a statistically significant increased ability to interpret and critically evaluate statistical information after completing the proposed course was also found. The statistical analysis confirmed a positive change in the ability to interpret statistical data, as well as the ability to critically evaluate the submitted statistical information. Both competencies are an essential element of statistical literacy [48]. After passing the post-test, up to 69% of students said that they understood the basic concept and principles of statistics (question 10). The implementation of mathematical calculations using computer software also contributed to the observed improvement in statistics. This eliminated the barrier for students caused by a negative attitude toward mathematics [49]. The students themselves expressed both pre-test and post-test that mastering computational procedures has a gentle impact on their statistical skills, which was confirmed by the results of statistical analysis.
Within dimension B, after completing the proposed innovative way of teaching statistics, students evaluated the positive contribution that statistical thinking can bring into one’s public, private, and professional life. Despite the recognition of the importance and need for real-life statistics and their own ability to interpret statistical data (dim A results), students did not show interest in further statistical education (results of the analysis of the answers to question 5, average 2.61). Moreover, the pre-test did not increase their belief in the use of statistics in their teaching practice (answers to question 6). We believe this may be due to their previous failures in mathematics which, according to several studies, have influenced students’ attitudes toward mathematics [50,51,52,53].
This fact was also reflected in the results of the statistical analysis of the answers to the questions in dimension C, where even after completing the innovative way of teaching statistics, the student’s negative perception of statistics persists. There was no improvement in the perception of statistics, even though our students considered the proposed way of teaching statistics to be very motivating to study statistics (answers to question 12). They continue to believe statistics are only for gifted students, which corresponds to the findings of several studies e.g., [54,55].
By analyzing the students’ answers to the individual questions in both tests, we came to several important and interesting conclusions, which are in line with the conclusions of several experts. For example, the use of real data recommended by several researchers (see [56]) has increased students’ motivation to study different statistical methods. It is experimentally verified that most students tend to believe that statistics are complicated due to complex calculations and irrelevant to their lives [57,58,59]. The implementation of numerical calculations by statistical software reduced the risk of failure in the proposed course. At the same time, they have more time to understand the basic principles of statistics and to interpret the results obtained. On the other hand, the implementation of the experiment has shown that it is not easy, if not directly impossible, to change students’ perception of statistics in general in a brief period. According to [60], a negative attitude toward statistics reduces students’ motivation to use statistics in their professional lives, which has been confirmed in our research.
Because, as already mentioned, attitudes are, according to Comas et al. [61], an integral part of learning any educational content, when teaching statistics, we would recommend looking for forms of teaching that would, in addition to teaching a specific statistical curriculum, develop students’ positive attitudes to statistics. In accordance with Garfield, Hogg, Schau, and Whittinghill [62], we can conclude that the required results in the introductory course in statistics are closely related to the learning, perseverance, and attitudes and beliefs of students.

7. Conclusions

During the implementation of this pedagogical experiment in the seminar on statistics, we obtained results that correspond to the conclusions [12], which means that working with real data, which in our case directly concerned the course participants, can have a strong motivational potential. According to several research results [63,64,65], students have a predominantly negative attitude toward statistics, which is often reflected in seminar passivity. In our concept of the Statistics Seminar, we have replaced the solution of examples from pre-prepared data by carrying out the entire real research. In doing so, we wanted students to know that statistics are not just numbers but numbers in context [13]. By implementing this innovative way of teaching statistics, we came to the same conclusions as Galiazzi [66], i.e., that this method of teaching allows the student to create a link between theory and practice, which can strengthen the student’s motivation to acquire new knowledge, which they evaluate as useful or necessary for their professional and personal life. It is needed or necessary to acquire some theoretical knowledge [67] and it is equally important that this finding corresponds to the claim of Piaget [68] that with such activities for students, they undertake an active approach to learning. Based on the observation of students’ attitudes and their activities within the Seminar of Statistics implemented in the proposed form, we can state that the change in the design of the seminar had an incredibly positive effect on students’ attitudes.

Author Contributions

Conceptualization, D.G.; Data curation, A.T.; Formal analysis, D.G.; Investigation, V.Ď.; Methodology, G.P.; Resources, V.Ď.; Software, A.T.; Validation, G.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the projects of KEGA 015UKF-4/2021. This work was supported by the Slovak Research and Development Agency under the contract No. APVV-14-0446.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Best, J. Beyond calculation: Quantitative literacy and critical thinking about public issues. Calc. Context Quant. Lit. Implic. Teach. Educ. 2008, 125–135. [Google Scholar]
  2. Jolly, R. Myrdal and the Economic Commission for Europe. UN Chronicle 2012, 46, 70. [Google Scholar] [CrossRef]
  3. OECD. Istanbul Declaration. 2007. Available online: www.oecd.org/dataoecd/14/46/38883774.pdf (accessed on 30 June 2007).
  4. Giovannini, E. Dynamic Graphics: Turning Key Indicators into Knowledge. In Proceedings of the 56th Session of the International Statistical Institute, Lisbon, Portugal; International Statistical Institute: Voorburg, The Netherlands, 2007; Available online: www.stat.auckland.ac.nz/~iase/publications (accessed on 7 December 2007).
  5. Ridgway, J.; Nicholson, J.; McCusker, S. Developing statistical literacy in students and teachers. In Teaching Statistics in School Mathematics-Challenges for Teaching and Teacher Education; Springer: Dordrecht, The Netherlands, 2011; pp. 311–322. [Google Scholar]
  6. Kosonen, P.; Winne, P.H. Effects of teaching statistical laws on reasoning about everyday problems. J. Educ. Psychol. 1995, 87, 33. [Google Scholar] [CrossRef]
  7. Camevale, A.P.; Gainer, L.J.; Meltzer, A.S. Workplace Basics: The Essential Skills Employers Want; Jossey-Bass: San Francisco, CA, USA, 1990. [Google Scholar]
  8. Packer, A. Mathematical Competencies that employers expect. In Why Numbers Count: Quantitative Literacy for Tomorrow’s America; Steen, L.A., Ed.; The College Board: New York, NY, USA, 1997; pp. 137–154. [Google Scholar]
  9. Chance, B.L. Components of statistical thinking and implications for instruction and assessment. J. Stat. Educ. 2002, 10. [Google Scholar] [CrossRef]
  10. Nicholson, J.; Ridgway, J.; McCusker, S. Reasoning with data–time for a rethink? Teach. Stat. 2006, 28, 2–9. [Google Scholar] [CrossRef]
  11. Ridgway, J.; McCusker, S.; Nicholson, J. Reasoning with evidence–Development of a scale. In Proceedings of the 32nd Annual Conference of the International Association for Educational Assessment Conference, Singapore, 21–26 May 2006. [Google Scholar]
  12. Muñiz-Rodríguez, L.; Rodríguez-Muñiz, L.J.; Alsina, Á. Deficits in the statistical and probabilistic literacy of citizens: Effects in a world in crisis. Mathematics 2020, 8, 1872. [Google Scholar] [CrossRef]
  13. Gal, I. Adults’ statistical literacy: Meanings, components, responsibilities. Int. Stat. Rev. 2002, 70, 1–25. [Google Scholar] [CrossRef]
  14. Broers, N.J. Learning goals: The primacy of statistical knowledge. In Proceedings of the 7th Annual Meeting of ICOTS, Salvador, Brazil, 2–9 July 2006; Volume 7. [Google Scholar]
  15. Wallman, K.K. Enhancing statistical literacy: Enriching our society. J. Am. Stat. Assoc. 1993, 88, 1–8. [Google Scholar]
  16. Watson, J. Assessing statistical literacy through the use of media surveys. In The Msessment Chullenge in Statistics Education; Gal, I., Garfield, J., Eds.; International Statistical Institutel IOS Press: Amsterdam, The Netherlands, 1997; pp. 107–121. [Google Scholar]
  17. Ben-Zvi, D.; Garfield, J. (Eds.) Statistical literacy, reasoning, and thinking: Goals, definitions, and challenges. In The Challenge of Developing Statistical Literacy, Reasoning and Thinking; Kluwer: Dordrecht, The Netherlands, 2004; pp. 3–25. [Google Scholar]
  18. Watson, J.M.; Callingham, R. Statistical literacy: A complex hierarchical construct. Stat. Educ. Res. J. 2003, 2, 3–46. [Google Scholar] [CrossRef]
  19. Watson, J.M.; Callingham, R. Statistical literacy: From idiosyncratic to critical thinking. In Curricular Development in Statistics Education. International Association for Statistical Education (IASE) Roundtable, Lund, Sweden, 2004; Burrill, G., Camden, M., Eds.; International Statistical Institute: Voorburg, The Netherlands, 2005; pp. 116–162. [Google Scholar]
  20. Callingham, R.; Watson, J.M. The development of statistical literacy at school. Stat. Educ. Res. J. 2017, 16, 181–201. [Google Scholar] [CrossRef]
  21. Wiberg, M. Teaching statistics in integration with psychology. J. Stat. Educ. 2009, 17. [Google Scholar] [CrossRef]
  22. Watson, J.M. Statistical Literacy at School: Growth and Goals; Lawrence Erlbaum: Mahwah, NJ, USA, 2006. [Google Scholar]
  23. McKinsey & Company. How the World´s Best Performing School Systems Come Out on Top. London. Retrieved September 16 2013. 2007. Available online: http://mckinseyonsociety.com/downloads/reports/Education/Worlds_School_Systems_Final.pdf (accessed on 1 September 2007).
  24. Hattie, J.; Yates, G.C. Visible Learning and the Science of How We Learn; Routledge: Abingdon, UK, 2013. [Google Scholar]
  25. Callingham, R.; Carmichael, C.; Watson, J. Explaining student achievement: The influence of teachers’ pedagogical content knowledge in statistics. Int. J. Sci. Math. Educ. 2016, 14, 1339–1357. [Google Scholar] [CrossRef] [Green Version]
  26. Chick, H.L.; Pierce, R.; Wander, R. Sufficiently assessing teachers’ statistical literacy. In Proceedings of the Ninth International Conference on Teaching Statistics: Sustainability in Statistics Education; Makar, K., deSousa, B., Gould, R., Eds.; International Statistical Institute: Voorburg, The Netherlands, 2014. [Google Scholar]
  27. Pierce, R.; Chick, H.; Watson, J.; Les, M.; Dalton, M. A statistical literacy hierarchy for interpreting educational system data. Aust. J. Educ. 2014, 58, 195–217. [Google Scholar] [CrossRef]
  28. Palm, T. Impact of authenticity on sense making in word problem solving. Educ. Stud. Math. 2007, 67, 37–58. [Google Scholar] [CrossRef]
  29. Hahn, C. La recherche internationale en éducation statistique: État des lieux et questions vives. Stat. Enseign. 2015, 6, 25–39. [Google Scholar]
  30. Cobb, G.W.; Moore, D.S. Mathematics, statistics, and teaching. Am. Math. Mon. 1997, 104, 801–823. [Google Scholar] [CrossRef]
  31. Gal, I.; Garfield, J. Curricular goals and assessment challenges in statistics education. Assess. Chall. Stat. Educ. 1997, 1–13. [Google Scholar]
  32. Gal, I.; Ginsburg, L.; Schau, C. Monitoring attitudes and beliefs in statistics education. Assess. Chall. Stat. Educ. 1997, 12, 37–51. [Google Scholar]
  33. Heikkil, A.; Lanka, K. Studying in Higher Education: Students’ approaches to Self-learning, Self-regulation and Cognitive Strategies. Stud. High. Educ. 2006, 31, 99–117. [Google Scholar] [CrossRef]
  34. Artelt, C.; Gräsel, C. Diagnostic Competence of Teachers. Z. Pädagogische Psychol. 2009, 23, 157–160. [Google Scholar] [CrossRef]
  35. Schwarz, B.; Wissmach, B.; Kaiser, G. “Last Curves Not Quite Correct”: Diagnostic Competences of the Future Teachers with Regard to Modelling and Graphical Representations. ZDM Math. Educ. 2008, 40, 777–790. [Google Scholar] [CrossRef]
  36. Helmke, A.; Hosenfeld, I.; Schrader, F.-W. Comparative Tests as an Instrument to Improve the Diagnostic Skills of Teachers. In Schulleitung und Schulentwicklung; Arnold, R., Griese, C., Eds.; Schneider: Hohengehren, Germany, 2004; pp. 119–143. [Google Scholar]
  37. Vogt, F.; Rogalla, M. Developing Adaptive Teaching Competency Through Coaching. Teach. Teach. Educ. 2009, 25, 1051–1060. [Google Scholar] [CrossRef] [Green Version]
  38. Kultusministerkonferenz (KMK). Competencies and Standards for Teacher Training. Beschluss der Kultusministerkonferenz vom 16.12.2004. Retrieved 15.03.2012. 2004. Available online: http://www.kmk.org/fileadmin/veroeffentlichungen_beschluesse/2004/2004_12_16-Standards-Lehrerbildung.pdf (accessed on 15 March 2012).
  39. Bruder, S.; Klug, J.; Hertel, S.; Schmitz, B. Measuring, Modeling and Advancement of Teachers’ Advisory Skill and Diagnostic Competence. In Themenheft, Lehrerbildung auf dem Prüfstand; Beck, K., Zlatkin Troischanskaia, O., Eds.; Empirische Pädagogik: Landau, Germany, 2010; pp. 173–195. [Google Scholar]
  40. Hollenweger, J. Teachers’ Ability to Assess Students for Teaching and Supporting Learning. Prospects 2011, 41, 445–457. [Google Scholar] [CrossRef]
  41. Roback, P.J. Teaching an Advanced Methods Course to a Mixed Audience. J. Stat. Educ. 2003, 11, 2. [Google Scholar] [CrossRef] [Green Version]
  42. Cobb, G.W. Teaching Statistics: More Data, Less Lecturing. Amstat News, December 1991; pp. 1–4. [Google Scholar]
  43. Smith, G.; Colleger, P. Learning Statistics by doing Statistics. J. Stat. Educ. 1998, 6, 3. [Google Scholar] [CrossRef]
  44. Emanovský, P.; Gonda, D. Mathematical Calculations within Physics Lessons and Their Popularity among Learners. J. Effic. Responsib. Educ. Sci. 2020, 13, 204–211. [Google Scholar] [CrossRef]
  45. Dočkal, V.; Kováč, T. On talent development. Studia Psychol. 1993, 35. [Google Scholar]
  46. Soto-Andrade, J. Enactive metaphorising in the learning of mathematics. In Invited Lectures from the 13th International Congress on Mathematical Education; Springer: Cham, Switzerland, 2018; pp. 619–637. [Google Scholar]
  47. Cladera, M.; Rejón-Guardia, F.; Vich-i-Martorell, G.À.; Juaneda, C. Tourism students’ attitudes toward statistics. J. Hosp. Leis. Sport Tour. Educ. 2019, 24, 202–210. [Google Scholar] [CrossRef]
  48. Mclaughlin, J.E.; Kang, I. A flipped classroom model for a biostatistics short course. Stat. Educ. Res. J. 2017, 16, 441–453. [Google Scholar] [CrossRef]
  49. Boaler, J. Mathematical Mindsets: Unleashing Students’ Potential through Creative Mathematics, Inspiring Messages and Innovative Teaching; John Wiley & Sons: Hoboken, NJ, USA, 2022. [Google Scholar]
  50. Primi, C.; Bacherini, A.; Beccari, C.; Donati, M.A. Assessing Math Attitude through the Attitude toward Mathematics Inventory–Short form in Introductory Statistics Course Students. Stud. Educ. Eval. 2020, 64, 100838. [Google Scholar] [CrossRef]
  51. García, T.; Rodríguez, C.; Betts, L.; Areces, D.; González-Castro, P. How affective-motivational variables and approaches to learning predict mathematics achievement in upper elementary levels. Learn. Individ. Differ. 2016, 49, 25–31. [Google Scholar] [CrossRef] [Green Version]
  52. Lipnevich, A.A.; MacCann, C.; Krumm, S.; Burrus, J.; Roberts, R.D. Mathematics attitudes and mathematics outcomes of US and Belarusian middle school students. J. Educ. Psychol. 2011, 103, 105. [Google Scholar] [CrossRef]
  53. Moenikia, M.; Zahed-Babelan, A. A study of simple and multiple relations between mathematics attitude, academic motivation and intelligence quotient with mathematics achievement. Procedia-Soc. Behav. Sci. 2010, 2, 1537–1542. [Google Scholar] [CrossRef] [Green Version]
  54. Witteman, H.O.; Hendricks, M.; Straus, S.; Tannenbaum, C. Are gender gaps due to evaluations of the applicant or the science? A natural experiment at a national funding agency. Lancet 2019, 393, 531–540. [Google Scholar] [CrossRef] [Green Version]
  55. Leslie, S.J.; Cimpian, A.; Meyer, M.; Freeland, E. Expectations of brilliance underlie gender distributions across academic disciplines. Science 2015, 347, 262–265. [Google Scholar] [CrossRef] [Green Version]
  56. Aziz, A.M.; Rosli, R. A systematic literature review on developing students’ statistical literacy skills. J. Phys.: Conf. Ser. 2021, 1806, 12102. [Google Scholar] [CrossRef]
  57. Oslington, G.; Mulligan, J.; Van Bergen, P. Third-graders’ predictive reasoning strategies. Educ. Stud. Math. 2020, 104, 5–24. [Google Scholar] [CrossRef]
  58. Ben-Zvi, D.; Gravemeijer, K.; Ainley, J. Design of statistics learning environments. In International Handbook of Research in Statistics Education; Springer: Cham, The Netherlands, 2018; pp. 473–502. [Google Scholar]
  59. Slootmaeckers, K.; Kerremans, B.; Adriaensen, J. Too afraid to learn: Attitudes towards statistics as a barrier to learning statistics and to acquiring quantitative skills. Politics 2014, 34, 191–200. [Google Scholar] [CrossRef]
  60. Peiró-Signes, Á.; Trull, O.; Segarra-Oña, M.; García-Díaz, J.C. Anxiety towards Statistics and Its Relationship with Students’ Attitudes and Learning Approach. Behav. Sci. 2021, 11, 32. [Google Scholar] [CrossRef]
  61. Comas, C.; Martins, J.A.; Nascimento, M.M.; Estrada, A. Estudio de las Actitudes hacia la Estadística en Estudiantes de Psicología. Bolema Bol. Educ. Matemática 2017, 31, 479–496. [Google Scholar]
  62. Garfield, J.; Hogg, B.; Schau, C.; Whittinghill, D. First courses in statistical science: The status of educational reform efforts. J. Stat. Educ. 2002, 10. [Google Scholar] [CrossRef]
  63. Hogg, R.V. Statistical education: Improvements are badly needed. Am. Stat. 1991, 45, 342–343. [Google Scholar]
  64. Gal, I.; Ginsburg, L. The Role of Beliefs and Attitudes in Learning Statistics: Towards an Assessment Framework. J. Stat. Educ. 1994, 2, 2. [Google Scholar] [CrossRef] [Green Version]
  65. Peterson, D.R. Connection and disconnection of research and practice in the education of professional psychologists. Am. Psychol. 1991, 46, 422. [Google Scholar] [CrossRef]
  66. Galiazzi, M.D.C.; Moraes, R.; Ramos, M.G. Educar pela pesquisa: As resistências sinalizando o processo de profissionalização de professores. Educ. Rev. 2003, 21, 1–15. [Google Scholar] [CrossRef] [Green Version]
  67. da Silva, M.P.M.; Pinto, S.S. Teaching statistics through learning projects. Stat. Educ. Res. J. 2014, 13, 177–186. [Google Scholar] [CrossRef]
  68. Piaget, J. Piaget’s theory. In Piaget and His School; Springer: Berlin/Heidelberg, Germany, 1976; pp. 11–23. [Google Scholar]
Figure 1. Average score according to the Talent Classification Sheet in thirteen classes.
Figure 1. Average score according to the Talent Classification Sheet in thirteen classes.
Mathematics 10 01793 g001
Figure 2. Average score in the intelligence test.
Figure 2. Average score in the intelligence test.
Mathematics 10 01793 g002
Figure 3. Average values of answers to questions in both tests.
Figure 3. Average values of answers to questions in both tests.
Mathematics 10 01793 g003
Figure 4. Respondents’ answers to question no. 1 (in %) in pre-test and post-test.
Figure 4. Respondents’ answers to question no. 1 (in %) in pre-test and post-test.
Mathematics 10 01793 g004
Figure 5. Respondents’ answers to question no. 2 (in %) in pre-test and post-test.
Figure 5. Respondents’ answers to question no. 2 (in %) in pre-test and post-test.
Mathematics 10 01793 g005
Figure 6. Respondents’ answers to question no. 6 (in %) in pre-test and post-test.
Figure 6. Respondents’ answers to question no. 6 (in %) in pre-test and post-test.
Mathematics 10 01793 g006
Figure 7. Respondents’ answers to question no. 12 (in %) in pre-test and post-test.
Figure 7. Respondents’ answers to question no. 12 (in %) in pre-test and post-test.
Mathematics 10 01793 g007
Table 1. Spearman’s rank correlation coefficient between students’ scores in KHT and scores in intelligence tests in individual classes.
Table 1. Spearman’s rank correlation coefficient between students’ scores in KHT and scores in intelligence tests in individual classes.
ClassR
A−0.5389 *
B−0.5405 *
C−0.2618
D−0.4967 *
E−0.2149
F−0.4939 *
G−0.6218 *
H−0.5726 *
I−0.2401
J−0.6087 *
K−0.5881 *
L−0.5972 *
M−0.2784
* Statistically significant results.
Table 2. Spearman coefficient of rank correlation between students’ performance and their results in the intelligence test and the performance and overall score achieved in TCS.
Table 2. Spearman coefficient of rank correlation between students’ performance and their results in the intelligence test and the performance and overall score achieved in TCS.
ClassR1R2
A−0.70 *0.63 *
B−0.59 *0.48 *
C−0.63 *0.38 *
D−0.75 *0.82 *
E−0.83 *0.37 *
F−0.89 *0.88 *
G−0.76 *0.56 *
H−0.82 *0.52 *
I−0.67 *0.41 *
J−0.71 *0.82 *
K−0.64 *0.64 *
L−0.59 *0.63 *
M−0.60 *0.37 *
* Statistically significant results.
Table 3. Average scores in individual dimensions.
Table 3. Average scores in individual dimensions.
Dim ADim BDim C
pre-test2.282.821.86
post-test3.823.412.37
Table 4. Results of the Wilcoxon single-sample test.
Table 4. Results of the Wilcoxon single-sample test.
DimZp
A4.1970.000 *
B3.2480.001 *
C1.8510.226
* Statistically significant results.
Table 5. Test results (pre-test and post-test).
Table 5. Test results (pre-test and post-test).
DimensionQuestion χ 2 p
Dim A1. question28.6510.000 *
3. question1.2630.867
4. question30.0550.000 *
10. question21.6450.000 *
Dim B2. question18.8870.000 *
5. question0.8100.937
6. question0.6220.960
8. question12.1420.016 *
Dim C7. question5.2380.155
9. question1.4380.487
11. question5.1520.160
12. question18.3820.000 *
Note: Values exceeding the critical value are indicated * in the table ( α = 0.05 ).
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gonda, D.; Pavlovičová, G.; Ďuriš, V.; Tirpáková, A. Implementation of Pedagogical Research into Statistical Courses to Develop Students’ Statistical Literacy. Mathematics 2022, 10, 1793. https://doi.org/10.3390/math10111793

AMA Style

Gonda D, Pavlovičová G, Ďuriš V, Tirpáková A. Implementation of Pedagogical Research into Statistical Courses to Develop Students’ Statistical Literacy. Mathematics. 2022; 10(11):1793. https://doi.org/10.3390/math10111793

Chicago/Turabian Style

Gonda, Dalibor, Gabriela Pavlovičová, Viliam Ďuriš, and Anna Tirpáková. 2022. "Implementation of Pedagogical Research into Statistical Courses to Develop Students’ Statistical Literacy" Mathematics 10, no. 11: 1793. https://doi.org/10.3390/math10111793

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop