Next Article in Journal
An Entropy Analysis of Classroom Conditions Based on Mathematical Social Science
Next Article in Special Issue
How do Online Learning Networks Emerge? A Review Study of Self-Organizing Network Effects in the Field of Networked Learning
Previous Article in Journal
Conclusion: Perspectives on Language, Literacy, and Deafness
Previous Article in Special Issue
To Be or Not to Be: Social Justice in Networked Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Exploring Conditions for Enhancing Critical Thinking in Networked Learning: Findings from a Secondary School Learning Analytics Environment

by
Elizabeth Koh
1,
Christin Jonathan
1,* and
Jennifer Pei-Ling Tan
2
1
Office of Education Research, National Institute of Education, Nanyang Technological University, Singapore 639798, Singapore
2
Group Human Resources, OCBC Bank, Singapore 049513, Singapore
*
Author to whom correspondence should be addressed.
Educ. Sci. 2019, 9(4), 287; https://doi.org/10.3390/educsci9040287
Submission received: 7 September 2019 / Revised: 6 November 2019 / Accepted: 8 November 2019 / Published: 4 December 2019

Abstract

:
Networked learning provides opportunities for learners to develop their critical thinking, an important 21st century competency, through dialogue with fellow learners to consider other perspectives and negotiate and critique ideas and arguments. However, much extant literature has not examined networked learning environments among younger learners nor the optimal conditions for enhancing critical thinking. Therefore, a study was carried out to investigate these conditions. A learning analytics networked learning environment was designed and 264 secondary three students participated in the 10-week long intervention as part of their English curriculum. Individual and collective social network metrics, critical reading scores, and self-reported survey data were used to quantitatively evaluate students’ critical reading performance in relation to their participation in networked learning. Results highlight several optimal conditions, notably that it is not just participation of the learner that enhances critical thinking but the learners’ reciprocity in replying and the distance of those posts in the network. Discussions and implications of the findings follow to provide insightful understanding of how the rich and complex settings of networked learning can enhance critical thinking capacities in secondary schooling.

1. Introduction

Critical thinking is one of the core competencies for learners in the 21st century and beyond. Networked learning provides opportunities for learners to develop their critical capacities through dialogue with fellow learners, to consider other perspectives, and negotiate and critique ideas and arguments [1]. While online learning forums are important predecessors of networked learning, technological and pedagogical developments in learning analytics point to the importance of dynamic visual learning analytics in computer-mediated learning environments [2], as well as multimodal social dialogic learning spaces [1,3]. Participation in such social dialogic spaces can dispose networked learners toward higher criticality [4].
Many existing networked learning environments have been implemented in post-secondary educational settings, such as distance education and higher education. Besides originating from the higher education context, it may be relatively easier to enact such networked learning with older learners as compared to younger learners due to the learner agency and self-regulation required [5]. A related issue is that existing pedagogical models for younger learners tend to focus on information or knowledge accumulation rather than productive interactions for learning [6]. Consequently, the research-informed empirical understandings of younger learners’ (e.g., adolescents) participation in social dialogic learning and the extent to which this affects their learning outcomes is limited [7,8]. As networked learning environments and pedagogy are increasingly being extended to K-12 educational settings [9], it is important to explore and examine the optimal application of such learning environments with younger learners.
Since social and intellectual interactions are key aspects of networked learning [10], examining interaction patterns to examine participants’ activities and relationships in the learning network have the potential to uncover insights into the interactions in networked learning and their association with learning outcomes [11]. Exploring collaborative and technology-dependent processes in networked learning through social network analysis (SNA) can potentially uncover the conditions necessary for effective learning and characterize productive interaction patterns for optimising networked learning environments [12].
Toward identifying the optimal conditions for developing critical thinking in secondary schooling, a 10-week intervention was carried out using WiREAD, a networked learning environment with affordances of multimodal social dialogic learning and dynamic visual learning analytics. A total of 264 secondary three (grade 9) students from a mainstream secondary school in Singapore participated in the intervention as part of their English language (EL) curriculum.
Our research aims to contribute to addressing some of the above-mentioned issues by: (1) presenting a web-based collaborative critical reading and learning analytics environment that was designed and implemented in schools; and (2) reporting individual and collective social network metrics, critical reading scores, and self-reported survey data to quantitatively evaluate students’ critical reading performance in relation to their participation in networked learning.
This paper starts with a literature review on networked learning environments and critical thinking, including advanced techno-pedagogical designs of learning network environments; this is followed by a review of quantitative methods for networked learning, primarily SNA for analyzing interactions. We then present the networked learning environment that was implemented. Next, the methods and materials used in the study are described. SNA was used together with multiple regression analysis (MRA) and t-tests to analyse the data. The results are shown and discussed with implications for theory, pedagogy, and design. Lastly, we end with a conclusion with contributions of the study and future ways forward.

2. Literature Review

2.1. Networked Learning Environments and Critical Thinking

The design of networked learning and computer-supported collaborative learning (CSCL) environments is often informed by socio-constructivist theories whereby knowledge is seen as socially negotiated and co-constructed [13,14]. Social and intellectual interactions between people are an integral process of learning [15,16,17] and key to networked learning [10]. These interactions are primarily carried out through dialogue, highlighting the crucial role of dialogic interaction for knowledge co-construction and learning in networked learning environments [1,18]. Accordingly, new theorizations of literacy, such as multiliteracies [19], have reconceptualized literacy as a social practice that is “increasingly multiple, multimodal and mediated through new technology” [20] (p. 1). These new literacies emphasise higher-order learning outcomes, such as the critical and collaborative evaluation and understanding of multiple forms of texts [19,21]. Students’ interactions with peers and instructors have been positively associated with student learning in general [22,23] and with higher-order learning outcomes in online collaboration [24,25]. In the context of networked learning, online discussions have been shown to enhance critical thinking capacities by engaging students in more constructive interactions [26,27].
Although critical thinking can be embedded in the curriculum and taught in various ways, the lesson materials and task criteria will influence students’ critical thinking performance [28]. Correspondingly, Johnson et al. [29] emphasize the importance of intentionally designing interaction into the learning context for high-quality interactions to occur. Aviv et al. [30] found that a higher level of critical thinking was achieved during knowledge construction in a structured learning network than in an unstructured learning network. To provide more structure, scripts can be built into the networked learning environment to scaffold students’ interaction. Schellens et al. [4] found that students using thinking or knowledge types scripts to tag their contributions displayed a greater depth of critical thinking as compared to the control group. These findings stress the importance of structuring interactions in networked learning environments to foster students’ criticality thinking.
Another key aspect of the design of networked learning and CSCL environments is the integration of learning analytics dashboards to support teaching and learning. Learning analytics dashboards can be used as a formative feedback tool to provide opportunities for learners to gain awareness of and reflect on their learning, and subsequently modify their learning behaviours [2]. In networked learning, social learning analytics can also be harnessed to generate insights on learning patterns and behaviors, based on learners’ participation and interactions in the learning environment [18]. The use of learning analytics dashboards has been associated with more opportunities for students to reflect on their learning and to provide greater motivation [2,31] and improved academic achievement [32]. In other words, well-designed dashboards can create a networked learning environment conducive to building students’ critical thinking capacities.

2.2. Quantitative Networked Learning Metrics

To understand learning in a network, there are many methods and measures. While acknowledging the importance of qualitative and mixed method approaches, this research uses a quantitative method with measures and metrics that help to account and control for the scale of networks. Individual measures of learners’ characteristics, such as their participation, achievement, and perceptions, together with collective measures that provide insight into the specific context the learner is situated in, can be useful for understanding networked learning. Understanding the social aspects of learning is gaining focus in education research [33], particularly with the growing use and applicability of SNA in educational settings [30,34]. SNA seeks to identify underlying relationships between actors in a social network based on their interactions with one another [35]. In networked learning environments, system log data can be used to map out interaction patterns to examine participants’ activities and relationships in the learning network [11]. SNA is an effective technique for understanding collaborative and technology-dependent processes [11,12] that has been applied to identify well-connected learners and learners who require more support and to design more effective learning environments [34,36]. SNA has also been used to monitor and assess the effectiveness of virtual learning groups [37,38].
SNA can examine network-level metrics (e.g., density, reciprocity) and individual-level metrics of networks (e.g., degree centrality) [39]. At the network level, SNA metrics provide more information about the network structure and network-level participation patterns. Network structure in terms of density and tie strength was found to influence social learning [40], and network centrality and reciprocity are correlated with grades [41]. Instructor’s presence has been found to influence network density [42], and information flow was highly influenced by instructors [43]. However, in networks with no instructor present, learners took on central roles [44]. Central actors are also influential as research shows that participants are more likely to follow central actors’ recommendations compared to less central actors in the network [45,46].
At the individual level of network metrics, “centrality” measures the extent of an individual’s interactions with other participants in the network [35]. In higher education, low participation in a learning network was identified as a possible predictor of low achievement [47] and low in-degree centrality was predictive of underachievement [48]. Out-degree centrality and interacting with tutors was associated with better grades in a problem-solving course [49], although this effect of out-degree was not significant for students in a medical course [48]. The inconclusive findings for out-degree centrality indicate that further examination of out-degree centrality is needed to understand its relationship to learning.

3. Study Context

The study was carried out using WiREAD, a web-based collaborative critical reading and learning analytics environment that was implemented in a secondary three (grade 9) EL subject in a Singapore secondary school. WiREAD was designed with the aim of fostering critical thinking in the EL domain as a socially generative practice by motivating and scaffolding students’ collaborative knowledge construction around multimodal texts. The techno-pedagogical design of WiREAD focused on two key learning affordances: (i) multimodal social dialogic learning (Figure 1), and (ii) dynamic visual learning analytics (Figure 2). These key affordances are in turn underpinned by four embedded learning and pedagogical frameworks: (i) multiliteracies pedagogical framework [19], (ii) dialogic learning [50], (iii) assessment for learning (AfL) [51], and (iv) CSCL [52]. The key principles of these four interconnected and nested learning and pedagogical frameworks are visually represented in Figure 3.
The multimodal social dialogic learning space allowed teachers to embed multimodal textual resources around social, moral, and/or ethical dilemmas for students to read and collaboratively critique with their peers during and beyond formal class time (see [53]). Critical reading pedagogical scaffolds based on the multiliteracies pedagogical framework [21,54], Paul–Elder’s “wheel of critical reasoning” [55], as well as dialogic indicators of collective creativity and criticality [56], guided students’ reading and discussions of the texts. The pedagogical scaffolds were comprised of seven critical lenses and five critical talk types, which are listed in Table 1. Students had to select one critical lens and one critical talk type for every comment or reply that they posted on WiREAD. Students’ participation on WiREAD was captured and visualized to them through a learning analytics dashboard, enabling them to track their critical reading engagement and progress in terms of: (i) the number of comments and replies posted, (ii) the number of critical lenses and critical talk types used, and (iii) sociograms visualizing their involvement and interactions within their WiREAD learning network.
Typically, a new teacher-nominated text would be uploaded to WiREAD every 1–2 weeks. Students would read and respond to texts on WiREAD using one-to-one devices for one EL lesson (around one hour) every 1–2 weeks in their classrooms or in computer labs, as well as in their own time out of class. During the lesson, teachers would set aside some time for students to read the text before posting comments and replies. The critical lenses acted as questions for students to respond to. Students were encouraged to use a range of critical lenses and critical talk types, as well as engage with their classmates’ comments. To this end, teachers often provided guidelines on the minimum number of comments and replies students should aim to post (e.g., at least three comments and three replies), to ensure that they are engaging sufficiently with the text and classmates’ responses. Although most of the lesson would be spent engaging online on WiREAD, teachers may pause the lesson at critical junctures to provide instructions and feedback to the class or highlight key points.
Previous research on WiREAD documented its impact on students’ critical reading fluency, cognitive reading engagement, and EL self-efficacy [57], and students’ and teachers’ qualitative accounts of its usefulness [53]. In this study, we focus on the conditions for promoting critical reading as generated by the network metrics. To examine and identify conditions for effective critical thinking in WiREAD, a networked learning environment, this paper asks, what are the conditions for promoting critical thinking of students in a networked learning environment?

4. Methods and Materials

4.1. Research Design and Procedure

The paper is situated in a larger study that employs design-based research with multiple trials to develop a techno-pedagogical networked learning environment with multimodal social dialogic learning and dynamic visual learning analytics in the secondary school context. The trial concerned was conducted in 2016 at the participating school, a mainstream secondary school. Seven secondary three classes participated in the 10-week long WiREAD intervention. During the intervention, WiREAD was used as part of the secondary three EL curriculum, with almost weekly WiREAD sessions during curriculum time, as well as out-of-class usage. Seven multimodal texts adapted from newspaper or magazine articles and online posts were embedded in WiREAD for students’ collaborative critique. The topics covered included education systems, abortion, filial piety, and euthanasia.
For this study, quantitative data at the start, during, and at the end of the trial were gathered. Quantitative data were in terms of the pre- and post-tests and log data on WiREAD. Informed consent was obtained from the students and their parents in accordance with the institutional review board’s research ethics requirements. One student opted not to participate in the study and was thus excluded from the data collection and analysis, although the student used WiREAD as part of the curriculum.

4.2. Participants

Consenting participants consisted of 263 students in the seven classes. These classes were taught by a total of five EL teachers. A breakdown of the participants is provided in Table 2 below. All the classes were of similar size, except for Class C, which had 23 participants. There were slightly more female students in the study (53.2%) as compared to male students (46.8%).

4.3. Measures

4.3.1. Social Network Metrics

Social network measures were computed to identify individual (node) and collective (network) characteristics that are optimal for fostering critical reading. On WiREAD, log data on whose posts a student had replied to and who had replied to a student’s posts was captured. This data was then fed into UCINET 6.0 [58] as seven separate matrices, one matrix per class. Node-level and network-level metrics were generated separately for each class of students since each class was a closed network on WiREAD and interactions were restricted to students and teachers within each class.
Several network metrics were calculated in relation to the context and research question:
  • Out-degree centrality: As students were encouraged to post replies to their classmates’ posts, out-degree centrality was computed to measure how many participants a student had out-going ties with (i.e., how many people they sent replies to).
  • Out k-step reach centrality: To measure how closely connected a student was to other participants in the class, out k-step reach centrality was used. Out k-step reach centrality refers to “the proportion of actors that a given actor can reach in k steps or less” [59] (p. 178). It was used instead of closeness centrality as it is more suited to directed networks.
  • Arc reciprocity: This was used as a network- or class-level metric to measure the extent to which replies from a student to another participant was matched by replies from that participant to the student. By default, UCINET dichotomizes the data in order to generate these metrics, meaning that the analysis focused on whether ties/interactions existed between two participants rather than how many ties there were between two participants.
Additionally, the metrics generated for each class of students were normalized so as to allow between-class comparisons.

4.3.2. Engagement in Collaborative or Networked Learning

At the start of the intervention, a self-report questionnaire was used to measure students’ perception of their engagement in collaboration using a five-item scale adapted from Tan and Nie’s [60] group engagement scale. The items are on a seven-point Likert-type scale ranging from “strongly disagree” to “strongly agree.” The scale demonstrated excellent internal reliability, with a Cronbach’s alpha coefficient of 0.94. Items from the scale are listed below:
  • I try my best to contribute to group discussions in EL;
  • I share my ideas during group work in EL;
  • I try my best to get involved in discussions during EL;
  • I try my best to contribute to group work in EL;
  • I enjoy discussions with my classmates in EL.
During the intervention, students’ posting behaviour on WiREAD was captured through the log data. The total number of words each student posted on WiREAD was used to quantify the extent of their participation. The word count of posts is a more meaningful measure as it reflects the depth of students’ participation and contribution compared to a frequency count of the number of posts, which can be inflated by a high frequency of very short posts.

4.3.3. Critical Reading Score

Students’ critical reading competency was measured using an objective critical reading test that was co-designed with the EL teachers. The test aimed to assess students on two dimensions: (i) critical reading quality and (ii) critical reading fluency. Critical reading quality pertains to the accuracy of students’ answers to the test questions. The questions were drawn from the critical lenses (e.g., Message: “What is the text telling you?”). Answers were scored on their accuracy, with 0 for irrelevant responses and up to 3 for responses with deeper inference/insight. Critical reading fluency pertains to the breadth of critical reading sub-skills demonstrated in each answer (idea/claim, agreement/validation, justification, etc.). These sub-skill dimensions were drawn from the critical talk types. Critical reading quality and fluency scores were combined to obtain an overall critical reading competency score for each student. Students completed the critical reading test at the start of the intervention, and again at the end of the intervention. The tests were graded by the EL teachers.

4.4. Data Analysis

SNA data and pre- and post-test variables were collated and descriptive statistics were generated to determine the characteristics of the sample and the variables examined. To identify the optimal conditions for fostering effective critical reading, a standard multiple regression analysis (MRA) was conducted using the enter method. The independent variables were the students’ pre-test critical reading score, self-reported engagement in collaboration, word count of posts, out-degree centrality, out 2-step reach centrality, and class’ arc reciprocity. The dependent variable was the students’ post-test critical reading score. The results are reported in the next section of this paper.

5. Results

5.1. Multiple Regression Analysis

Prior to interpreting the results of the MRA, several assumptions were evaluated. An analysis of standard residuals was carried out, which indicated that the data contained no outliers (std. residual min. = −2.39, std. residual max. = 3.10). To check for influential cases that could bias the model, Cook’s distance and the Mahalanobis distance were examined. Although Cook’s distance was less than 1 for all cases, the Mahalanobis distance exceeded the critical χ2 for df = 6 (at α = 0.001) of 22.458 for four cases in the sample, indicating that these cases could be exerting undue influence over the parameters of the model. A total of 11 potentially influential cases were examined and removed iteratively until a maximum Mahalanobis distance of less than 22.458 was achieved in the final model. Descriptive statistics for the original model (model 1) and the final model (model 2, after the 11 cases were removed) are summarized in Table 3.
Multicollinearity was not a concern in either model as the variance inflation factor (VIF) was below 4 and the tolerance level was above 0.2 for all the independent variables [61,62]. The data met the assumption of independent errors (Durbin-Watson valuemodel1 = 2.04; Durbin-Watson valuemodel2 = 2.04). Inspection of the normal probability plots of standardized residuals and scatterplots of standardized residuals against standardized predicted values indicated that the assumptions of normality, linearity, and homoscedasticity of residuals were met for both models.
In model 1, the six predictor variables explained 38.8% of the variance in the post-test critical reading score, F (6, 256) = 28.652, p < 0.001, R2 = 0.402, adjusted R2 = 0.388. By Cohen’s [63] conventions, an effect of this magnitude can be considered “large” (ƒ2 = 0.67). The pre-test critical reading score (β = 0.30, p < 0.001), word count of posts (β = 0.16, p = 0.011), out-degree centrality (β = −0.17, p = 0.013), out 2-step reach centrality (β = 0.17, p = 0.007), and class arc reciprocity (β = 0.32, p < 0.000) were statistically significant predictors of post-test critical reading score, while engagement in collaboration was not a significant predictor (β = 0.06, p = 0.252).
This pattern was also seen in model 2, where the six predictor variables explained 38.8% of the variance in post-test critical reading score, F (6, 245) = 27.505, p < 0.001, R2 = 0.402, adjusted R2 = 0.388, with a “large” effect size (ƒ2 = 0.67). Pre-test critical reading score (β = 0.27, p < 0.001), word count of posts (β = 0.21, p = 0.003), out-degree centrality (β = −0.20, p = 0.009), out 2-step reach centrality (β = 0.17, p = 0.008), and class arc reciprocity (β = 0.33, p < 0.001) were statistically significant predictors of post-test critical reading score, while engagement in collaboration was not a significant predictor (β = 0.05, p = 0.371). The regression coefficients, standard errors, and t-values are reported in Table 4.
Comparing the two models, the predictors explained 38.8% of the variance in both models. Only engagement in collaboration was not a statistically significant predictor of post-test critical reading score in both models, while pre-test critical reading score, word count of posts, out 2-step reach centrality, and class arc reciprocity were positively-related significant predictors, and out-degree centrality was a negatively-related statistically significant predictor. Given that neither the presence nor absence of the influential cases changes the conclusions that can be drawn from the results, we focus our discussion on the original model, model 1.

5.2. Further Analysis

Out k-Step Reach

To further examine what level of connections in the network out k-step reach centrality were optimal for fostering effective critical reading, the post-test critical reading scores of students with different levels of out k-step reach were examined. To perform this test, we computed a median split for four measures of out k-step reach, where k = 1, k = 2, k = 3, and k = 4, and used a series of independent samples t-tests with the out k-step reach variable as the independent variable and post-test critical reading score as the dependent variable. The results are reported in Table 5.
For out-degree centrality (equivalent to out 1-step reach), preliminary assumption testing indicated that scores for both groups were relatively normally distributed, but Levene’s test was significant as there was substantially more variance in the below-median-group’s scores. Consequently, Welch’s t-test was used to compare the two groups’ post-test critical reading scores. The results indicated that there was no significant difference in scores for students with below median out-degree centrality (n = 136, M = 18.77, SD = 6.54) and students with above median out-degree centrality (n = 127, M = 18.88, SD = 5.11), t (253.20) = −0.15, p = 0.877.
For out 2-step reach centrality, preliminary assumption testing indicated that scores for both groups were relatively normally distributed, and Levene’s test was also non-significant. Thus, equal variances can be assumed. The results indicated that there was a significant difference in scores for students with below median out 2-step reach centrality (n = 136, M = 17.58, SD = 5.80) and students with above median out 2-step reach centrality (n = 127, M = 20.15, SD = 5.71), t (261) = −3.62, p < 0.001.
For out 3-step reach centrality, preliminary assumption testing indicated that scores for both groups were relatively normally distributed, but Levene’s test was significant as there was substantially more variance in the above-median-group’s scores. Consequently, Welch’s t-test was used to compare the two groups’ post-test critical reading scores. The results indicated that there was a significant difference in scores for students with below median out 3-step reach centrality (n = 135, M = 17.21, SD = 4.67) and students with above median out 3-step reach centrality (n = 128, M = 20.52, SD = 6.54), t (228.88) = −4.69, p < 0.001.
For out 4-step reach centrality, preliminary assumption testing indicated that scores for both groups were relatively normally distributed, but Levene’s test was significant as there was substantially more variance in the below-median-group’s scores. Consequently, Welch’s t-test was used to compare the two groups’ post-test critical reading scores. The results indicated that there was a significant difference in scores for students with below median out 4-step reach centrality (n = 167, M = 19.40, SD = 6.36) and students with above median out 4-step reach centrality (n = 96, M = 17.82, SD = 4.83), t (241.53) = 2.27, p = 0.024.

6. Discussion

This paper examines the conditions for promoting critical thinking among grade 9 students in a techno-pedagogical designed networked learning environment. The MRA results show that class arc reciprocity, out 2-step reach centrality, pre-test critical reading score, and word count of posts were significantly positive predictors of higher post-test critical reading score, while out-degree centrality was a significantly negative predictor of post-test critical reading score. Pre-test engagement in collaboration was not found to significantly predict post-test critical reading score. The results are discussed in the following paragraphs and provide some insight into optimal conditions for enhancing the critical thinking capacities of learners.
Starting with students’ prior achievement in terms of pre-test critical reading scores, this was found to be a strong significant predictor of post-test critical reading scores in the networked learning environment. Past studies have shown that individual factors, such as prior achievement and knowledge, can influence knowledge comprehension and construction, with high-achieving students performing better on various thinking skills [64], as well as participating more actively in online discussion tasks [65]. The result implies that students with lower prior critical reading ability levels still require more teacher support and guidance in order to participate adequately in networked learning environments [66,67].
Examining the SNA metrics, the results strongly show that class arc reciprocity predicts students’ post-test critical reading scores. This was the highest unstandardized regression coefficient in the MRA and shows a clear alignment with existing research indicating the positive association between reciprocity and students’ grades [41]. Facilitating a culture of reciprocity at the network-level could contribute toward students’ critical thinking by ensuring that they are interacting sufficiently with other students’ responses, as well as contributing ideas of adequate quantity and quality for others to respond to their contributions, as shown by other studies, e.g., Saqr et al. [48,68].
Although out-degree centrality and out 2-step reach centrality are both measures of active outward interactions, findings from our MRA revealed that out 2-step reach centrality was a significant predictor of higher post-test critical reading scores, whereas out-degree centrality was predictive of lower post-test critical reading scores. These contrasting findings reflect the inconclusive findings around out-degree centrality in the literature [48,49]. One possible interpretation of these results is that maintaining a high out-degree centrality, i.e., interacting with a large number of participants in the network, requires considerable time and effort. Larger discussion threads are important for continuous discussion to develop [69], but a high volume of interactions can result in information overload [70]. Being engaged in a high number of discussions might result in a trade-off between breadth and depth of connections in the learning network. Students with high out-degree centrality could be engaged in many short and disconnected discussions without developing ideas with sufficient depth [70], thus not developing their critical thinking capacities despite their high participation.
In this case, the finding that out 2-step reach centrality was a significant predictor of critical reading scores suggests that forming indirect connections with the entire network through deeper interactions with a smaller number of participants might have been more beneficial for developing students’ critical reading compared with forming direct but shallow connections with all participants in the learning network. As mentioned, we performed further analysis to probe this finding. Further analysis comparing the scores of students with above median and below median out k-step reach centrality revealed that students of above median out 2-step reach centrality and out 3-step reach centrality had significantly higher post-test critical reading scores as compared to their below median peers. For out-degree centrality (out 1-step reach centrality), there was no significant difference in post-test critical reading scores for students above and below the median.
The findings from the further analysis provided more evidence that being able to reach the rest of the learning network through two or three other indirect or intermediary connections is optimal for fostering critical thinking capacities, such that students can easily be connected to the ideas of many different participants in the network without having to directly interact (i.e., reply) with all of them. This implies that effective strategies for fostering interactions and guiding students toward achieving a conducive balance between breadth and depth of interactions in order to develop their critical thinking capacities are needed. For instance, in the design or pedagogical instruction of networked learning environments, learners should be guided to steer away from replying to many posts, but instead to concentrate on certain ones that pique their interest and that allow them to engage deeply in the discussion.
We note that the results were reversed for the further analysis of out 4-step reach centrality; students below the median had significantly higher post-test critical reading scores than students above the median. In this case, the results for out 4-step reach centrality should be interpreted with caution as students in the above-median group were exclusively from three of the seven classes. Teachers in these three classes were highly involved and connected in the network, with extremely high normalized out 2-, 3-, and 4-step reach centralities of 1. Correspondingly, it is likely that these students had high out 4-step reach centrality because of their connection to the teacher rather than their own interactions in the network. Additionally, there was at most only a difference of 0.05 between the out 4-step reach centrality of 98% of the sample whereas there was a greater variation among students for out-degree, out 2-step reach, and out 3-step reach centralities. This suggests that out 4-step reach centrality and beyond might not be useful metrics for analyzing small networks such as the classes in our study.
Lastly, the study also examined conditions related to individual participation. While greater participation in terms of the word count of posts was a significant predictor of students’ post-test critical reading scores, students’ self-reported pre-test engagement in collaboration in EL had no influence over their critical reading scores. Quantity of words can be indicative of the depth of students’ contributions and interactions [71], and could be a proxy for whether students are sufficiently engaged in the process of critical and collaborative evaluation and knowledge co-construction to develop their critical thinking capacities. Greater participation levels in networked learning are associated with achievement [47]. This implies that due to the potential inaccuracy of self-reported measures of participation or collaboration, it is likely more useful to focus on word count of posts and other behavioral measures in examining participation in networked learning environments [72].
Looking at the findings overall, it does suggest that while it is important to have learners participate individually in the network to develop their critical thinking, the extent of that participation matters. Specifically, what seems to matter is the reciprocity between learners’ replies and the distance or nearness between learners’ replies. This means that optimal conditions for learner’s critical thinking in the networked learning environment include learners’ ability to respond to a reply and going deeply into a few posts rather than many. A pedagogical implication, therefore, is that facilitating critical thinking in such environments requires several instructional guidelines for students’ interactions. It should not be “random” replies or posts but thoughtful dialogues between learners clarifying and exchanging perspectives, as well as intentional discussions on various topics/posts that learners can deeply engage in.

7. Conclusions

In this paper, an empirical investigation of the optimal networked learning conditions for enhancing critical thinking capacities among grade 9 students as part of the EL curriculum was conducted. The networked learning environment WiREAD integrated multimodal social dialogic learning and dynamic visual learning analytics to foster students’ 21st century competencies including critical thinking. We found that higher prior achievement in critical reading and greater participation in terms of word count of posts were optimal factors. More importantly, the findings utilized network metrics to highlight that it was not just participation in the network, but also the reciprocity between learners’ replies and the distance between learners’ replies that enhanced students’ critical thinking. There is a need to balance the breadth with the depth of interactions to cultivate students’ critical thinking capacities.
Implications of these results are that the design of the networked learning environment and pedagogical instruction should provide more structure and support for students in developing their critical thinking in terms of both the quantity and quality of interactions. Students should be guided toward greater participation in the learning network. However, this should be done in combination with strategies to help students engage deeply and critically with others in the network so as to encourage higher quality interactions [70,73].
Students with lower prior achievement can also benefit from more teacher support and guidance to optimize their meaningful participation in networked learning environments [66,67]. To meet diverse students’ needs, teachers should be adaptive in their teaching strategies [74] while leveraging contemporary technologies, such as networked learning environments, to provide support and structure for students’ interactions while allowing sufficient agency for students to develop their critical thinking capacities. We acknowledge that the role of the teacher is important in networked learning, as also seen in other learning networks [43] and will be examined in future studies.
This study was not without its limitations and we highlight them and suggest ways forward. First, overall network metrics were used and there was no breakdown of students’ participation for different teacher-nominated texts or by the discussion thread. This resulted in slightly coarse-grained findings with a broad overview that cannot resolve participation levels that can differ across texts and threads. Future studies can look into examining finer-grained metrics to provide more information about interaction behaviors for each text and thread in the network. Second, although this study examined the quality of students’ critical reading through their critical reading scores, our analysis of students’ participation in networked learning involved metrics related to the quantity of their participation. Combining content analysis with SNA can provide a more comprehensive understanding of the type and quality of interactions in networked learning that might be optimal for enhancing students’ critical thinking [11,34] and are future steps for the research team. Another limitation is related to the log data available on the technical system. There was no indicator for whether a post was read, which limited the kind of inferences we could draw. This lack of data is partly due to the current technological limitations that although post clicks can be detected, it cannot accurately inform if a post was read. Examining reading in networks, e.g., requiring eye trackers for each student, is practically and technically challenging at this point in time. Lastly, we acknowledge that the study can be further strengthened through the use of a quasi-experimental design with a control group.
In conclusion, this paper has identified several key conditions for enhancing critical thinking in a collaborative critical reading and learning analytics networked learning environment. While several of the study’s findings support existing research, this is one of a few studies on networked learning that focus on grade 9 learners. In that sense, it helps to widen and extend the optimal conditions. Besides contributing to the broader application of networked learning, the study also showcases a networked learning environment designed with advanced affordances, and provides design, pedagogical, and theoretical implications. It is our hope that the findings and discussion presented here contribute toward identifying optimal networked learning conditions for enhancing critical thinking capacities. By presenting how conditions such as students’ prior achievement, and their participation reciprocity and distance in networked learning affects critical reading scores, we provide a more insightful understanding of how the rich and complex settings of networked learning can enhance critical thinking capacities in secondary schooling.

Author Contributions

Conceptualization, E.K., C.J. and J.P.-L.T.; data curation, C.J.; formal analysis, C.J.; funding acquisition, J.P.-L.T.; investigation, J.P.-L.T.; methodology, J.P.-L.T.; writing—original draft, E.K. and C.J.; writing—review and editing, E.K., C. J. and J.P.-L.T.

Funding

This research was funded by National Research Foundation, grant number NRF2013-EDU001-EL019.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ravenscroft, A. Dialogue and connectivism: A new approach to understanding and promoting dialogue-rich networked learning. Int. Rev. Res. Open. Dis. 2011, 12, 139–160. [Google Scholar] [CrossRef] [Green Version]
  2. Verbert, K.; Duval, E.; Klerkx, J.; Govaerts, S.; Santos, J.L. Learning analytics dashboard applications. Am. Behav. Sci. 2013, 57, 1500–1509. [Google Scholar] [CrossRef] [Green Version]
  3. Cukurova, M.; Luckin, R.; Millán, E.; Mavrikis, M. The NISPI framework: Analysing collaborative problem-solving from students’ physical interactions. Comput. Educ. 2018, 116, 93–109. [Google Scholar] [CrossRef]
  4. Schellens, T.; Van Keer, H.; De Wever, B.; Valcke, M. Tagging thinking types in asynchronous discussion groups: Effects on critical thinking. Interact. Learn. Envir. 2009, 17, 77–94. [Google Scholar] [CrossRef]
  5. Smidt, H.; Thornton, M.; Abhari, K. The future of social learning: A novel approach to connectivism. In Proceedings of the 50th Hawaii International Conference on System Sciences, Big Island, HI, USA, 3–7 January 2017; HICSS Conference Office: Honolulu, HI, USA, 2017. [Google Scholar]
  6. Williams, R.; Karousou, R.; Mackness, J. Emergent learning and learning ecologies in Web 2.0. Int. Rev. Res. Open. Dis. 2011, 12, 39–59. [Google Scholar] [CrossRef]
  7. Bell, F. Network theories for technology-enabled learning and social change: Connectivism and actor network theory. In Proceedings of the 7th International Conference on Networked Learning 2010, Aalborg, Denmark, 3–4 May 2010; Lancaster University: Lancaster, UK, 2010. [Google Scholar]
  8. Redecker, C.; Ala-Mutka, K.; Punie, Y. Learning 2.0-The Impact of Social Media on Learning in Europe. Policy Brief. JRC Scientific and Technical Report. EUR JRC56958 EN. 2010. Available online: http://ftp.jrc.es/EURdoc/JRC56958.pdf (accessed on 6 September 2019).
  9. Hossain, M.M.; Wiest, L.R. Collaborative middle school geometry through blogs and other web 2.0 technologies. J. Comput. Math. Sci. Teach. 2013, 32, 337–352. [Google Scholar]
  10. Jones, C.; Ryberg, T.; de Laat, M. Networked Learning. In Encyclopedia of Educational Philosophy and Theory; Peters, M., Ed.; Springer Singapore: Singapore, 2017; pp. 1–6. [Google Scholar]
  11. De Laat, M.; Lally, V.; Lipponen, L.; Simons, R.-J. Investigating patterns of interaction in networked learning and computer-supported collaborative learning: A role for Social Network Analysis. Int. J. Comp-Supp. Coll. 2007, 2, 87–103. [Google Scholar] [CrossRef]
  12. Carolan, B.V. Social Network Analysis and Education: Theory, Methods & Applications; Sage Publications: Thousand Oaks, CA, USA, 2013. [Google Scholar]
  13. Jonassen, D.H. Thinking technology: Toward a constructivist design model. Educ. Technol. 1994, 34, 34–37. [Google Scholar]
  14. Kent, C.; Rechavi, A.; Rafaeli, S. Networked Learning Analytics: A Theoretically Informed Methodology for Analytics of Collaborative Learning. In Learning In a Networked Society: Spontaneous and Designed Technology Enhanced Learning Communities; Kali, Y., Baram-Tsabari, A., Schejter, A.M., Eds.; Springer International Publishing: Cham, Switzerland, 2019; pp. 145–175. [Google Scholar]
  15. Dede, C. The evolution of distance education: Emerging technologies and distributed learning. Am. J. Dis. Educ. 1996, 10, 4–36. [Google Scholar] [CrossRef]
  16. Wenger, E. Communities of practice: Learning as a social system. Syst. Thinker 1998, 9, 2–3. [Google Scholar] [CrossRef]
  17. Richardson, J.C.; Maeda, Y.; Lv, J.; Caskurlu, S. Social presence in relation to students’ satisfaction and learning in the online environment: A meta-analysis. Comput. Hum. Behav. 2017, 71, 402–417. [Google Scholar] [CrossRef]
  18. Shum, S.B.; Ferguson, R. Social learning analytics. J. Educ. Tech. Soc. 2012, 15, 3–26. [Google Scholar]
  19. The New London Group. A pedagogy of multiliteracies: Designing social futures. Harvard. Educ. Rev. 1996, 66, 60–92. [Google Scholar] [CrossRef]
  20. Burnett, C.; Davies, J.; Merchant, G.; Rowsell, J. Changing Contexts for 21st-Century Literacies. In New Literacies around the Globe; Routledge: New York, NY, USA, 2014; pp. 21–33. [Google Scholar]
  21. Cope, B.; Kalantzis, M. The things you do to know: An introduction to the pedagogy of multiliteracies. In A Pedagogy of Multiliteracies; Cope, B., Kalantzis, M., Eds.; Palgrave Macmillan: London, UK, 2015; pp. 1–36. [Google Scholar]
  22. Bernard, R.M.; Abrami, P.C.; Borokhovski, E.; Wade, C.A.; Tamim, R.M.; Surkes, M.A.; Bethel, E.C. A meta-analysis of three types of interaction treatments in distance education. Rev. Educ. Res. 2009, 79, 1243–1289. [Google Scholar] [CrossRef]
  23. Jaggars, S.S.; Xu, D. How do online course design features influence student performance? Comput. Educ. 2016, 95, 270–284. [Google Scholar] [CrossRef] [Green Version]
  24. Akyol, Z.; Garrison, D.R. Understanding cognitive presence in an online and blended community of inquiry: Assessing outcomes and processes for deep approaches to learning. Brit. J. Educ. Technol. 2011, 42, 233–2500. [Google Scholar] [CrossRef]
  25. Gold, S. A constructivist approach to online training for online teachers. J. Asynchron. Learn. Netw. 2001, 5, 35–57. [Google Scholar] [CrossRef]
  26. Klisc, C.; McGill, T.; Hobbs, V. Use of a post-asynchronous online discussion assessment to enhance student critical thinking. Australas. J. Educ. Tec. 2017, 33, 63–76. [Google Scholar] [CrossRef] [Green Version]
  27. Garrison, D.R.; Akyol, Z. Toward the development of a metacognition construct for communities of inquiry. Internet. High. Educ. 2015, 24, 66–71. [Google Scholar] [CrossRef]
  28. Thompson, C. Critical thinking across the curriculum: Process over output. Int. J. Hum. Soc. Sci. 2011, 1, 1–7. [Google Scholar]
  29. Johnson, R.D.; Hornik, S.; Salas, E. An empirical examination of factors contributing to the creation of successful e-learning environments. Int. J. Hum.-Comput. St. 2008, 66, 356–369. [Google Scholar] [CrossRef]
  30. Aviv, R.; Erlich, Z.; Ravid, G.; Geva, A. Network analysis of knowledge construction in asynchronous learning networks. J. Asynchron. Learn. Netw. 2003, 7, 1–23. [Google Scholar] [CrossRef] [Green Version]
  31. Wise, A.; Zhao, Y.; Hausknecht, S. Learning analytics for online discussions: Embedded and extracted approaches. J. Learn. Anal. 2014, 1, 48–71. [Google Scholar] [CrossRef]
  32. Arnold, K.E.; Pistilli, M.D. Course signals at Purdue: Using learning analytics to increase student success. In Proceedings of the 2nd International Conference on Learning Analytics and Knowledge, Vancouver, BC, Canada, 29 April–02 May 2012; ACM: New York, NY, USA, 2012; pp. 267–270. [Google Scholar]
  33. Dawson, S. ‘Seeing’ the learning community: An exploration of the development of a resource for monitoring online student networking. Brit. J. Educ. Technol. 2010, 41, 736–752. [Google Scholar] [CrossRef]
  34. Cela, K.L.; Sicilia, M.Á.; Sánchez, S. Social network analysis in e-learning environments: A preliminary systematic review. Educ. Psychol. Rev. 2015, 27, 219–246. [Google Scholar] [CrossRef]
  35. Wasserman, S.; Faust, K. Social Network Analysis: Methods and Applications; Cambridge University Press: Cambridge, UK, 1994; Volume 8. [Google Scholar]
  36. Siemens, G.; Long, P. Penetrating the fog: Analytics in learning and education. Educ. Rev. 2011, 46, 30. [Google Scholar]
  37. Daradoumis, T.; Martínez-Monés, A.; Xhafa, F. An integrated approach for analysing and assessing the performance of virtual learning groups. In Proceedings of the International Conference on Collaboration and Technology Workshop on Groupware (CRIWG 2004), San Carlos, Costa Rica, 5–9 September 2004; Springer: Berlin/Heidelberg, Germany, 2004; pp. 289–304. [Google Scholar]
  38. Rakic, S.; Softic, S.; Vilkas, M.; Lalic, B.; Marjanovic, U. Key Indicators for Student Performance at the E-Learning Platform: An SNA Approach. In Proceedings of the 2018 16th International Conference on Emerging eLearning Technologies and Applications (ICETA), Stary Smokovec, Slovakia, 15–16 November 2018; IEEE: New York, NY, USA, 2018; pp. 463–468. [Google Scholar]
  39. Kellogg, S.; Booth, S.; Oliver, K. A social network perspective on peer supported learning in MOOCs for educators. Int. Rev. Res. Open. Dis. 2014, 15, 263–289. [Google Scholar] [CrossRef]
  40. Paredes, W.C.; Chung, K.S.K. Modelling learning & performance: A social networks perspective. In Proceedings of the 2nd International Conference on Learning Analytics and Knowledge, Vancouver, BC, Canada, 29 April–2 May 2012; ACM: New York, NY, USA, 2012; pp. 34–42. [Google Scholar]
  41. Sousa-Vieira, M.; López-Ardao, J.; Fernández-Veiga, M. The Network Structure of Interactions in Online Social Learning Environments. In Proceedings of the 9th International Conference on Computer Supported Education (CSEDU 2017), Porto, Portugal, 21–23 April 2017; Springer: Cham, Switzerland, 2018; pp. 407–431. [Google Scholar]
  42. Martınez, A.; Dimitriadis, Y.; Rubia, B.; Gómez, E.; De la Fuente, P. Combining qualitative evaluation and social network analysis for the study of classroom social interactions. Comput. Educ. 2003, 41, 353–368. [Google Scholar] [CrossRef] [Green Version]
  43. Skrypnyk, O.; Joksimović, S.k.; Kovanović, V.; Gašević, D.; Dawson, S. Roles of course facilitators, learners, and technology in the flow of information of a cMOOC. Int. Rev. Res. Open. Dis. 2015, 16, 188–217. [Google Scholar] [CrossRef] [Green Version]
  44. Laghos, A.; Zaphiris, P. Sociology of student-centred e-learning communities: A network analysis. In Proceedings of the IADIS International Conference on e-Society, Dublin, Ireland, 13–16 July 2006; IADIS Press: Dublin, Ireland, 2006. [Google Scholar]
  45. Beck, R.J.; Fitzgerald, W.J.; Pauksztat, B. Individual behaviors and social structure in the development of communication networks of self-organizing online discussion groups. In Designing for Change in Networked Learning Environments; Springer: Dordrecht, The Nederlands, 2003; pp. 313–322. [Google Scholar]
  46. Cho, H.; Stefanone, M.; Gay, G. Social information sharing in a CSCL community. In Proceedings of the Conference on Computer Support for Collaborative Learning: Foundations for a CSCL Community, Boulder, CO, USA, 7–11 January 2002; Lawrence Erlbaum Associates: Hillsdale, NJ, USA, 2002; pp. 43–50. [Google Scholar]
  47. Sousa-Vieira, M.E.; López-Ardao, J.C.; Fernández-Veiga, M.; Ferreira-Pires, O.; Rodríguez-Pérez, M.; Rodríguez-Rubio, R.F. Prediction of learning success/failure via pace of events in a social learning network platform. Comput. Appl. Eng. Educ. 2018, 26, 2047–2057. [Google Scholar] [CrossRef]
  48. Saqr, M.; Fors, U.; Tedre, M. How the study of online collaborative learning can guide teachers and predict students’ performance in a medical course. BMC Med. Educ. 2018, 18, 24. [Google Scholar] [CrossRef] [PubMed]
  49. Saqr, M.; Alamro, A. The role of social network analysis as a learning analytics tool in online problem based learning. BMC Med. Educ. 2019, 19, 160. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Alexander, R. Essays on Pedagogy; Routledge: New York, NY, USA, 2008. [Google Scholar]
  51. Black, P.; Wiliam, D. Assessment and classroom learning. Assess. Educ. Prin. Policy Pract. 1998, 5, 7–74. [Google Scholar] [CrossRef]
  52. Stahl, G.; Koschmann, T.D.; Suthers, D.D. Computer-supported Collaborative Learning. In Cambridge Handbook of the Learning Science; Cambridge University Press: Cambridge, UK, 2006; pp. 409–426. [Google Scholar]
  53. Tan, J.P.-L.; Koh, E.; Jonathan, C.; Yang, S. Learner dashboards a double-edged sword? Students’ sense-making of a collaborative critical reading and learning analytics environment for fostering 21st-century literacies. J. Learn. Anal. 2017, 4, 117–140. [Google Scholar] [CrossRef] [Green Version]
  54. Tan, J.P.-L.; McWilliam, E. From literacy to multiliteracies: Diverse learners and pedagogical practice. Pedag. Int. J. 2009, 4, 213–225. [Google Scholar] [CrossRef] [Green Version]
  55. Paul, R.; Elder, L. Critical Thinking: Tools for Taking Charge of Your Learning and Your Life; Prentice Hall: Upper Saddle River, NJ, USA, 2001. [Google Scholar]
  56. Tan, J.P.-L.; Caleon, I.S.; Jonathan, C.R.; Koh, E. A Dialogic Framework for Assessing Collective Creativity in Computer-Supported Collaborative Problem-Solving Tasks. Res. Pract. Tech. Enhanc. Learn. 2014, 9, 411–437. [Google Scholar]
  57. Jonathan, C.; Tan, J.P.-L.; Koh, E.; Caleon, I.S.; Tay, S.H. Enhancing students’ critical reading fluency, engagement and self-efficacy using self-referenced learning analytics dashboard visualizations. In Proceedings of the 25th International Conference on Computers in Education, Christchurch, New Zealand, 4–8 December 2017; Asia-Pacific Society for Computers in Education: Jhongli, Taiwan, 2017; pp. 457–462. [Google Scholar]
  58. Borgatti, S.; Everett, M.; Freeman, L. Ucinet 6.0 Version 1.00. Natick: Analytic Technologies. Sofware for Social Network Analysis. Version para Windows. Se puede conseguir en la dirección. 1999. Available online: http://www.analytictech.com/ (accessed on 6 February 2011).
  59. Borgatti, S.P.; Everett, M.G.; Johnson, J.C. Analyzing Social Networks; Sage Publications Ltd: London, UK, 2013. [Google Scholar]
  60. Tan, J.P.-L.; Nie, Y. The role of authentic tasks in promoting twenty-first century learning dispositions. In Authentic Problem Solving and Learning in the 21st Century; Springer: Singapore, 2015; pp. 19–39. [Google Scholar]
  61. Cohen, P.; West, S.G.; Aiken, L.S. Applied Multiple Regression/Correlation Analysis for the Behavioral Sciences; Lawrence Erlbaum Associates: Mahwah, NJ, USA, 2014. [Google Scholar]
  62. Menard, S. Applied Logistic Regression Analysis; Sage Publications: Thousand Oaks, CA, USA, 2002; Volume 106. [Google Scholar]
  63. Cohen, J. Statistical Power Analysis for the Behavioural Sciences, 2nd ed.; Cohen, J., Ed.; Lawrence Erlbaum Associates: Hillsdale, NJ, USA, 1988. [Google Scholar]
  64. Zohar, A.; Dori, Y.J. Higher order thinking skills and low-achieving students: Are they mutually exclusive? J. Learn. Sci. 2003, 12, 145–181. [Google Scholar] [CrossRef]
  65. Prinsen, F.; Volman, M.L.; Terwel, J. The influence of learner characteristics on degree and type of participation in a CSCL environment. Brit. J. Educ. Technol. 2007, 38, 1037–1055. [Google Scholar] [CrossRef]
  66. Azevedo, R. Using hypermedia to learn about complex systems: A self-regulation model. In Proceedings of the 10th International on Artificial Intelligence in Education, San Antonio, TX, USA, 19–23 May 2001; IOS Press: Amsterdam, The Netherlands, 2001. [Google Scholar]
  67. Rahikainen, M.; Lallimo, J.; Hakkarainen, K. Progressive inquiry in CSILE environment: Teacher guidance and students’ engagement. In Proceedings of the 1st European Conference on Computer-Supported Collaborative Learning (Euro-CSCL 2001), Maastricht, The Netherlands, 22–24 March 2001; Maastricht McLuhan Institute: Maastricht, The Netherlands, 2001; pp. 520–528. [Google Scholar]
  68. Philip, D. Social network analysis to examine interaction patterns in knowledge building communities. Can. J. Learn. Technol. 2010, 36, 1–20. [Google Scholar] [CrossRef] [Green Version]
  69. Hewitt, J.; Teplovs, C. An analysis of growth patterns in computer conferencing threads. In Proceedings of the 1999 Conference on Computer Support for Collaborative Learning, Palo Alto, CA, USA, 12–15 December 1999; Lawrence Erlbaum Associates: Mahwah, NJ, USA, 1999; pp. 232–241. [Google Scholar]
  70. Lipponen, L.; Rahikainen, M.; Hakkarainen, K.; Palonen, T. Effective participation and discourse through a computer network: Investigating elementary students’ computer supported interaction. J. Educ. Comput. Res. 2002, 27, 355–384. [Google Scholar] [CrossRef]
  71. Hara, N.; Bonk, C.J.; Angeli, C. Content analysis of online discussion in an applied educational psychology course. Instr. Sci. 2000, 28, 115–152. [Google Scholar] [CrossRef]
  72. Goggins, S.; Xing, W. Building models explaining student participation behavior in asynchronous online discussion. Comput. Educ. 2016, 94, 241–251. [Google Scholar] [CrossRef]
  73. Heo, H.; Lim, K.Y.; Kim, Y. Exploratory study on the patterns of online interaction and knowledge co-construction in project-based learning. Comput. Educ. 2010, 55, 1383–1392. [Google Scholar] [CrossRef]
  74. Martinez-Maldonado, R.; Clayphan, A.; Yacef, K.; Kay, J. MTFeedback: Providing notifications to enhance teacher awareness of small group work in the classroom. IEEE T. Learn. Technol. 2014, 8, 187–200. [Google Scholar] [CrossRef]
Figure 1. Multimodal social dialogic learning space of WiREAD.
Figure 1. Multimodal social dialogic learning space of WiREAD.
Education 09 00287 g001
Figure 2. Dynamic learning analytics visualizations of WiREAD (selected only).
Figure 2. Dynamic learning analytics visualizations of WiREAD (selected only).
Education 09 00287 g002
Figure 3. Learning and pedagogical principles underlying techno-pedagogical design of WiREAD.
Figure 3. Learning and pedagogical principles underlying techno-pedagogical design of WiREAD.
Education 09 00287 g003
Table 1. Critical reading pedagogical scaffolds.
Table 1. Critical reading pedagogical scaffolds.
Critical LensesCritical Talk Types
Message: What is the text/author telling the reader?Ideate: I think that…
Purpose: What is the objective of the text/author?Justify: I think so because…
Audience: Who are the target readers?Validate: I agree…
Assumption: What presuppositions does the author make?Challenge: I disagree…
Viewpoint: Whose perspective is the text written from?Clarify: I need to ask…
Inference: What conclusions did I draw from the text and why?
Impact: How effective are the language/visuals used?
Table 2. Breakdown of participants by class, gender, and English language (EL) teacher.
Table 2. Breakdown of participants by class, gender, and English language (EL) teacher.
Classn% Males% FemalesEL Teacher
Class A4067.5%32.5%Teacher 1
Class B4042.5%57.5%Teacher 2
Class C2334.8%65.2%Teacher 3
Class D3930.8%69.2%Teacher 1
Class E3951.3%48.7%Teacher 4
Class F4042.5%57.5%Teacher 5
Class G4252.4%47.6%Teacher 2
Total26346.8%53.2%5 Teachers
Table 3. Descriptive statistics for students’ critical reading scores, self-reported engagement in collaboration, and online participation on WiREAD.
Table 3. Descriptive statistics for students’ critical reading scores, self-reported engagement in collaboration, and online participation on WiREAD.
Model 1Model 2
VariablenMSDnMSD
Post-test critical reading score26318.825.8825218.945.91
Pre-test critical reading score26317.494.9125217.554.93
Pre-test engagement in collaboration2635.301.072525.360.96
Word count of posts2631012674.79252979.52583.22
Out-degree centrality2630.240.152520.230.13
Out 2-step reach centrality2630.830.222520.830.22
Class arc reciprocity2630.520.072520.520.07
Table 4. Summary of Multiple Regression Analysis.
Table 4. Summary of Multiple Regression Analysis.
Model 1Model 2
PredictorBSEBΒtBSEBβt
Pre-test critical reading score0.3620.0670.3025.431 ***0.3290.0690.2754.751 ***
Pre-test engagement in collaboration0.3190.2780.0581.1480.2840.3170.0460.896
Word count of posts0.0010.0010.1582.560 *0.0020.0010.2082.966 **
Out-degree centrality−6.6702.679−0.166−2.490 *−9.2513.495−0.202−2.647 **
Out 2-step reach centrality4.4261.6380.1682.702 **4.6171.7360.1702.660 **
Class arc reciprocity27.1744.4580.3246.096 ***28.0454.6090.3356.085 ***
Note: B—unstandardized regression coefficient; SEB—Standard error of the coefficient; β—standardized coefficient; * p < 0.05; ** p < 0.01; *** p < 0.001.
Table 5. Out k-step reach centrality: descriptive data, independent samples t-test results and effect sizes.
Table 5. Out k-step reach centrality: descriptive data, independent samples t-test results and effect sizes.
Below-Median GroupAbove-Median Group
VariablenMean (SD)nMean (SD)tCohen’s d (Effect size)
Out-degree centrality13618.77 (6.54)12718.88 (5.11)−0.15−0.02 (Small)
Out 2-step reach centrality13617.58 (5.80)12720.15 (5.71)−3.62 ***−0.45 (Medium)
Out 3-step reach centrality13517.21 (4.67)12820.52 (6.54)−4.69 ***−0.58 (Medium)
Out 4-step reach centrality16719.40 (6.36)9617.82 (4.83)2.27 *0.28 (Medium)
Note: * p < 0.05; ** p < 0.01; *** p < 0.001. SD—standard deviation.

Share and Cite

MDPI and ACS Style

Koh, E.; Jonathan, C.; Tan, J.P.-L. Exploring Conditions for Enhancing Critical Thinking in Networked Learning: Findings from a Secondary School Learning Analytics Environment. Educ. Sci. 2019, 9, 287. https://doi.org/10.3390/educsci9040287

AMA Style

Koh E, Jonathan C, Tan JP-L. Exploring Conditions for Enhancing Critical Thinking in Networked Learning: Findings from a Secondary School Learning Analytics Environment. Education Sciences. 2019; 9(4):287. https://doi.org/10.3390/educsci9040287

Chicago/Turabian Style

Koh, Elizabeth, Christin Jonathan, and Jennifer Pei-Ling Tan. 2019. "Exploring Conditions for Enhancing Critical Thinking in Networked Learning: Findings from a Secondary School Learning Analytics Environment" Education Sciences 9, no. 4: 287. https://doi.org/10.3390/educsci9040287

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop