Next Article in Journal
Is the Education System Prepared for the Irruption of Artificial Intelligence? A Study on the Perceptions of Students of Primary Education Degree from a Dual Perspective: Current Pupils and Future Teachers
Previous Article in Journal
Supervisory Feedback on Chinese Doctoral Students’ Drafts of a Research Article
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Technology Implementation in Pre-Service Science Teacher Education Based on the Transformative View of TPACK: Effects on Pre-Service Teachers’ TPACK, Behavioral Orientations and Actions in Practice

1
Department of Physics Education, Bielefeld Universtiy, 33615 Bielefeld, Germany
2
Department of Science Edcuation, Leuphana Universtity, 21335 Lüneburg, Germany
*
Author to whom correspondence should be addressed.
Educ. Sci. 2023, 13(7), 732; https://doi.org/10.3390/educsci13070732
Submission received: 2 June 2023 / Revised: 6 July 2023 / Accepted: 13 July 2023 / Published: 18 July 2023
(This article belongs to the Section Teacher Education)

Abstract

:
Teaching with and about technology is part of science teachers’ 21st century skills. To foster technology-enhanced practice, teachers need to acquire both technological pedagogical content knowledge (TPACK on action) and positive behavioral orientations toward technology exploitation. However, it remains unclear if the gained knowledge is applied in practice (TPACK in action). Therefore, studies are required to investigate the interplay of programs promoting TPACK on action, behavioral orientations, and resulting TPACK in action. This paper presents an approach that explicitly links pre-service science teachers’ pedagogical content knowledge (PCK) with TPACK development in two undergraduate modules, following the transformative view of TPACK. TPACK on action and behavioral orientations are captured through a questionnaire at three points in time. Additionally, lesson plans are analyzed to evaluate the quality of technology use and cognitive engagement, approximating TPACK in action. The results show a significant increase in pre-service science teachers’ (N = 133) self-rated TPACK on action and behavioral orientations between pre- and post-test, with moderate to large effects. Moreover, the analyses of lesson plans reveal a high quality of technology exploitation in the planned lessons, indicating distinctive TPACK in action after attending the modules. This theory-based approach is supported by empirical data, and highly regarded by participants, making it a successful model for course redesign at other universities.

1. Introduction

Digital transformation has progressed rapidly in everyday and working life in recent decades and has also found its way into the classroom. Especially science education has a great potential for the multifaceted use of technology, not only due to the proximity of science to technology itself but also because of its advantages for teaching and learning. For example, internal and external measurement sensors and devices can be used as alternatives for traditional lab equipment; interactive simulations and remote or video experiments can enable students to explore experimental setups that cannot be realized in the classroom; data analysis and presentation software can contribute to minimizing barriers in dealing with measurements; and modeling software and virtual and augmented reality applications can be applied to model and visualize interrelationships between quantities.
In order to be able to integrate technology purposefully into science education, pre-service science teachers need to develop and interconnect content knowledge (CK) and pedagogical knowledge (PK), as well as technological knowledge (TK), forming technological pedagogical content knowledge (TPACK) [1,2]. Accordingly, universities face the challenge of redesigning existing study programs to meet these requirements. Today, technology is integrated into teacher education in multiple ways, from (elective) add-on modules to the implementation of technology into existing modules [3]. Considering the transformative view of TPACK [4,5,6,7,8,9], especially the implementation of technology into science education modules seems to be most promising, as this allows the development of PCK and TPACK on action simultaneously [8,9,10,11,12,13,14]. Additionally, pre-service science teachers are offered the opportunity to gain positive (theoretical and/or practical) experiences with technology, which, in turn, promote advantageous positive behavioral orientations toward technology exploitation in science education [15,16,17]. In this context, several authors emphasize the theory of planned behavior (ToPB [18]) as a valid framework to assess (pre-service) teachers’ intentions to implement technology into their classroom practice [19,20,21,22,23,24]. Nevertheless, none of these studies go beyond the intended use of technology and address how teachers implement technology into science education (TPACK in action). This is of particular relevance, however, as it is known that teachers’ TPACK on action is often not consistent with their (future) actions in the classroom [25,26]. Therefore, studies combining measures of (pre-service) teachers’ TPACK on action, their behavioral orientations and resulting TPACK in action represent a research desideratum [3,25,26,27]. Additionally, mixed-methods approaches that collect and combine quantitative and qualitative data in order to compensate for the respective limitations of the individual instruments are needed [3,27].
To address this challenge, the current study aims to investigate how the implementation of technology according to the transformative view of TPACK in the science teacher education program at the Leuphana University Lüneburg influences the TPACK of pre-service science teachers both on action and in action, as well as their behavioral orientations.

2. Theoretical Framework

In the following sections, we briefly outline the TPACK framework and provide an overview of TPACK assessment in empirical research. In so doing, we distinguish between TPACK on action and TPACK in action and discuss methodological approaches to investigate both.

2.1. TPACK

Based on Shulman’s PCK model [28], teachers’ professional knowledge about the purposeful integration of technology in class can be described by the TPACK framework [1,2], which distinguishes between pedagogical knowledge (PK), content knowledge (CK), technological knowledge (TK) and the resulting hybrids first (PCK: pedagogical content knowledge, TCK: technological content knowledge, TPK: technological pedagogical knowledge) and second order (TPACK: technological pedagogical content knowledge), as shown in Figure 1.
Building on PCK research by Grossman [29], Niess [30] conceptualized TPACK as consisting of four elements: (1) an overarching conception of teaching a subject with technology, (2) instructional strategies and representations for teaching with technologies, (3) students’ understandings, thinking and learning in a subject with technology, as well as (4) curriculum and curricular materials. Other authors such as Canbazoğlu Bilici et al. [31] added a fifth element: (5) knowledge of assessment with technology, in accordance with Magnusson et al.’s [32] PCK elements.
Initial research on TPACK focused on the framework itself, e.g., the internal structure or interpretation. In particular, two contrasting perspectives on the relation between TPACK dimensions have emerged among researchers [33,34]: the integrative and the transformative view. Following the integrative view, TPACK arises from the integration of all TPACK dimensions (CK, PK, TK, PCK, TCK and TPK). In contrast, the transformative view assumes that TPACK is formed by the first hybrids PCK, TCK and TPK, which themselves depend on CK, PK and TK. Current empirical studies favor the transformative view of TPACK already described theoretically by Mishra and Koehler [2,4,5,6,7,8,35,36,37,38]. Nevertheless, only in a study by Pamuk et al. [6], all first-order hybrids were found to have a significant influence on TPACK. All other studies only found two of the three first-order hybrids to be positive predictors for TPACK, mainly PCK and TPK [4,5,7,8,33].
In recent years, TPACK research has shifted from investigating the framework itself to the application of the framework as well as potential extensions [39,40,41]. Various methods and approaches have been developed to capture TPACK, especially in the context of skill development of (pre-service) science teachers, as discussed in the following section.

2.2. TPACK Assessment in Empirical Research

One goal of current science teacher education programs is for pre-service teachers to develop the necessary skills to integrate technology purposefully into their practice. Corresponding empirical research on TPACK (development) uses various measures, such as self-assessment questionnaires [42,43,44,45,46,47,48,49,50,51,52], performance assessment [25,31,52,53,54], analysis of lesson plans [14,31,55,56] and lesson observations [57,58,59].
These measures can be distinguished into two groups: TPACK self- and performance assessments capture teachers’ knowledge on action. Analysis of lesson plans and lesson observations are used to describe teachers’ practice and, therefore, can be understood as knowledge in action [60]. TPACK on and TPACK in action are linked in that teachers’ knowledge on action forms the basis for planning and teaching topic-specific lessons [61], which integrate technology and is, thus, a prerequisite for TPACK in action.

2.2.1. TPACK on Action

Teacher self-assessments are widely used instruments to capture TPACK on action. Not only are they of high practical applicability, but they also allow for assessing teachers’ confidence in implementing technology and drawing conclusions about their technology-related self-efficacy beliefs [62,63]. However, self-assessment questionnaires often face inherent methodological limitations and present constraints related either to validity or reliability [7]. Some of the instruments only capture selected TPACK dimensions [64,65] or considerably vary in the number of items per TPACK dimension [47,49]. Moreover, subject-independent item formulations, such as “in my content area” [47] or “in my teaching subject” [45], can distort the results, as in some countries (pre-service) teachers study/teach more than one subject.
As several studies have shown, TPACK self-assessments often lack a satisfactory correlation with more objective measurements [25,31,53] and crucially differ from teachers’ actions observed in the classroom [25,26]. In order to obtain a meaningful assessment of (pre-service) science teachers’ TPACK, on the one hand, and their technology-related self-efficacy beliefs, on the other hand, it is, therefore, necessary to combine TPACK on action self-assessments with more objective measures that take into account classroom practice and, thus, teachers’ TPACK in action. Following this, mixed-methods approaches, which combine self-reported TPACK with more objective measurements, have become prominent in recent studies [66].

2.2.2. TPACK in Action

One approach to capture (pre-service) teachers’ TPACK in action is the use of TPACK rubrics to analyze authentic data sources such as lesson plans or observations. However, these rubrics are subject to limitations. For example, existing TPACK rubrics for lesson plans, which follow descriptions of PCK elements [32], adapted for TPACK, sometimes fail to cover all TPACK elements and determine teachers’ TPACK in action level, e.g., by the lowest score across all TPACK elements [56]. Other existing TPACK rubrics assign an overall value to TCK, TPK and TPACK [55] while neglecting PCK. In both cases, potentially varying quality in technology exploitation is neglected and, as in the second case, the related instructional quality is not captured.
It becomes evident that TPACK in action is a rather difficult construct to grasp. Instead, the quality of technology exploitation, which can be understood as a manifestation of teachers’ TPACK in action, can be assessed from lesson plans or observations by the SAMR [67] and the ICAP [68,69] framework, respectively (see Table 1).
The SAMR framework by Puentedura [67] describes the quality of technology exploitation in comparison to traditional teaching and learning media on four ascending levels (substitution, augmentation, modification and redefinition; see Table 1, left column) and has been used successfully in various subjects to analyze teachers’ abilities to plan and conduct lessons containing technology [62,70]. However, the subject-independent applicability of the SAMR framework, at the same time, limits its usability to adequately describe teachers’ underlying content knowledge in one specific subject. Rather, the SAMR framework can be used to describe the relationship and interactions between technology and specific pedagogical practices, particularly how tasks can be enhanced by the use of technology.
The ICAP framework by Chi [68] and Chi and Wylie [69] describes instructional quality in terms of students’ cognitive engagement on four descending levels (interactive, constructive, active and passive; see Table 1, right column) and is not limited to materials or activities involving technology. Nevertheless, it can be used to assess students’ cognitive engagement achieved through technology exploitation, as conducted, for example, by Deepika et al. [71], Kramer et al. [70] and Wekerle et al. [72].
Chi and Wylie [69] argue that only constructive and interactive cognitive engagement can lead to a deep understanding of subject-specific content. However, as learning materials, activities and objectives vary considerably across subjects, cognitive engagement is inextricably linked to the respective subject or content. Especially science education has very unique ways of thinking and working, which are not common in other subjects. This is reflected in the relationship and interactions between learning objectives in science education and corresponding pedagogical practices.
Considering this, we understand the quality of technology exploitation, as conceptualized by the SAMR framework, as an indicator for (pre-service) science teachers’ TPK and students’ cognitive engagement with the scientific content via technology, as conceptualized by the ICAP framework, as an indicator for PCK.
In accordance with the current empirical research on the internal structure of TPACK, as discussed in Section 2.1, PCK and TPK are predictors for TPACK, while TCK has no significant influence [4,5,7,8,33]. Accordingly, a combination of distinctive PCK and TPK can be understood as an indicator for high TPACK. Based on this, we combine the SAMR framework [67] as a measure of TPK in action and the ICAP framework [68,69] as a measure of PCK in action, as an alternative to existing TPACK rubrics. This approach has several advantages. Both frameworks are easily applicable to lesson plans or observations and allow for distinguishing possibly varying qualities of multiple technology exploitations in a single lesson, which, in a second step, can be summarized to an overall rating for the respective lesson. Additionally, an approximation of TPACK in action based on the SAMR (for TPK) and ICAP (for PCK) framework is in line with current empirical findings.

2.3. Moderating Factors between TPACK on and TPACK in Action

In order to gain an in-depth understanding of the interplay between TPACK on and 187 in action mixed-method approaches which combine self-assessments with the analysis of 188 related classroom practice are necessary. This implies that inconsistencies between (pre-service) science teachers’ TPACK on and in action, as observed in previous studies [25,26], are taken into account and potential moderating factors need to be identified. For example, more recent studies consider, according to the theory of planned behavior (ToPB; [18]), that the intention to perform a specific behavior depends on one’s attitudes toward the behavior, subjective norms and perceived behavioral control [20,23,24,25]. Attitudes toward a specific behavior can either be positive or negative. Subjective norms refer to social norms or pressure, such as colleagues’ or superiors’ opinions toward the behavior. Perceived behavioral control describes the availability of resources as well as one’s self-efficacy toward the behavior.
Related to the use of technology in (science) education, Ertmer and Ottenbreit-Leftwich [73] identified four factors that positively influence (pre-service) teachers’ intentions to integrate technology into practice: knowledge, self-efficacy and pedagogical beliefs, as well as subject and school culture. While knowledge refers to teachers’ TPACK on action [73], the remaining three factors can be located in the theory of planned behavior. The authors of [21] argue that subject and school culture determine corresponding subjective norms on the expected use of technology in the respective context (subject/school), that pedagogical beliefs are expressed by teachers’ attitudes toward the use of technology and that self-efficacy (in combination with constraints) forms the perceived behavioral control. Therefore, all factors influencing teachers’ intentions toward the use of technology in education and, ultimately, practice, as identified by Ertmer and Ottenbreit-Leftwich [73], can be related to the ToPB.
Also, several other authors have outlined the relationship between TPACK and ToPB, such that ToPB is a valid framework to study (pre-service) teachers’ behavioral orientations toward the use of technology in practice [19,20,25]. However, studies rarely go beyond teachers’ planned behavior and relate TPACK on action to their actual classroom practice. Studies investigating teachers’ TPACK in action by lesson plans or observations only occasionally address TPACK on action or moderating behavioral orientations [35,48,74,75,76]. To gain an in-depth understanding of the interplay between pre-service teachers’ professional knowledge, intentions, and resulting actions in practice studies, combining various methods, are necessary.

3. Context of the Study

Universities face the challenge of integrating technology into their teacher education programs without omitting other essential content. One possible solution is to offer (elective) add-on modules, which supplement the existing study program by focusing on the use of technology in a class or lab [3]. However, this leads to an additional workload for pre-service teachers, which cannot always be credited. Consequently, these modules are often attended by just a few participants, and in the case of elective modules, pre-service teachers have to make a choice between different courses. One way or another, only some pre-service science teachers can attend and benefit from add-ons addressing technology in science education. However, as we believe that it is crucial that all pre-service teachers develop the necessary knowledge to enhance students’ scientific literacy through technology, we decided to implement technology into modules that are mandatory for all pre-service science teachers. There are two main reasons for this: First, we can assure that all pre-service science teachers graduating from Leuphana University will have attended the modules. Secondly, following the transformative view of TPACK, we want to link the development of PCK and, building on this, TPACK among pre-service teachers more closely. Therefore, teaching and learning technology was implemented systematically into the existing science education modules in the form of digital supplements and digital reminders, as described in the following section. Up to then, science education modules focused “merely” on pre-service teachers’ PCK development.

Module Design

The study is embedded in the three-year (six semesters) bachelor’s program at the Leuphana University. A special feature of the science teacher education program at Leuphana University is that pre-service primary and secondary science teachers jointly attend the science education modules ‘Teaching and Learning Science’ and ‘Science in Everyday life’ in the fourth and fifth semesters of the bachelor’s program. First, pre-service teachers attend the mandatory module ‘Teaching and Learning Science’, followed by ‘Science in Everyday Life’, consisting of a lecture and a complementary seminar in each. Each module consists of about 13 weeks, with 2 h each lecture and seminar. In the first module, fundamental science education topics are discussed in theory and complemented by digital supplements (app. 15–20 min of each lecture), which include exploitation possibilities of technology in class related to the respective topic (see Table 2). In the complementary seminar, pre-service teachers additionally learn to design diagnostic tasks with quiz apps, create short explanatory videos on science topics and use simulations and remote labs as well as digital measuring sensors.
The subsequent module ‘Science in Everyday Life! focuses on the practical design, teaching and reflection of science lessons. Additionally, digital reminders (brief, explicit verbal and/or written reference to the potential use of certain tools for a specific purpose) relating to the digital supplements from the previous module are incorporated through the lecture to demonstrate how technology can be implemented and support science teaching and learning in manifold ways. In the complementary seminar, students deepen their theoretical knowledge in practice by designing a technology-enhanced lesson. In groups of three to four, they present this lesson to the other participants of the seminar by teaching selected parts of the unit and providing an overview of the remaining lesson.

4. The Study: Aims and Research Questions

The aim of the study is to implement technology following the transformative view of TPACK [2,4,5,6,7,8,35,36,37,38] into the science education program at Leuphana University in order to increase pre-service teachers’ professional knowledge, particularly TPACK on and in action, as well as to positively influence their behavioral orientations toward the use of technology (according to ToPB). The study design presented here is a novelty, as the development of TPACK on action, ToPB and TPACK in action are explicitly linked to each other and captured using a mixed-methods approach. In addition, the findings contribute to a better understanding of the impact of integrating technology into existing modules (which focus on PCK) in tertiary science teacher education and are, therefore, useful for redesigning curricula at other universities in the future.
This leads to the following research questions:
  • How do pre-service science teachers’ self-reported professional knowledge, particularly TPACK, and their behavioral orientations develop during the modules?
  • How distinctive are technology exploitations in pre-service science teachers’ lesson plans in terms of quantity and quality after attending the modules?
  • How do pre-service science teachers rate the quality of the redesigned modules?

5. Methods

This section is divided into a characterization of the sample (see Section 5.1), followed by a detailed description of the data collection process and used instruments. We combine measures (by questionnaire) of self-rated TPACK on action and ToPB (see Section 5.2.1) with TPACK in action derived from lesson plans (see Section 5.2.2) to obtain an in-depth understanding of pre-service science teachers’ TPACK development. This approach has three advantages: first, the whole decision-making process for the integration of technology, from underlying TPACK on action and moderating behavioral orientations to resulting TPACK in action, is captured; second, it allows for combining the advantages of various TPACK on/in action measurements (questionnaire/analysis of lesson plans) while compensating limitations; and third, it allows for a higher resolution in terms of individual technology exploitation in practice.

5.1. Participants

In total, the sample consisted of 133 participants separated into 2 cohorts, 74 pre-service teachers in the first year and 59 in the second (Table 3). Sample sizes decreased in-between measurement points (pre/re/post) from 102 down to 55 students. This high dropout rate was due, among other things, to the fact that the pre- and post-tests were two semesters apart. Some students did not attend modules in directly consecutive semesters, and others did not pass the exam in module ‘Teaching and Learning Science’, which was why they did not attend the following module. The study design and data collection process were identical for both cohorts.

5.2. Data Collection

The accompanying research followed a mixed-methods approach. At three measuring points (pre/re/post—before, in-between and after attending the two modules) pre-service science teachers reported their professional knowledge (TPACK on action) and behavioral orientations according to the ToPB through a questionnaire. Lesson plans were analyzed by qualitative content analysis [78] in terms of quantity and quality of technology exploitation as well as pre-service teachers underlying TPACK in action. Additionally, the quality of the modules was assessed through the standardized teaching evaluation form (TEF) of the Leuphana University, supplemented by module-specific items. The study design and accompanying data collection are presented in Figure 2.
Due to the preservation of anonymity and the design of technology-enhanced lessons by groups of three to four pre-service science teachers, it was not possible to combine the quantitative and qualitative data for individual participants. Nevertheless, the analyzed lesson plans serve as authentic data sources, which supplement pre-service science teachers’ self-reported TPACK on action and help to gain an in-depth understanding of pre-service science teachers’ TPACK on and in action after attending the modules.

5.2.1. Questionnaire

At three measuring points (pre/re/post) students’ TPACK on action as well as their behavioral orientations toward the use of digital media in science education according to the ToPB were captured by self-reports. All items were rated on a 5-point Likert scale (from 1 = “strongly disagree” to 5 = “strongly agree”).
Pre-service science teachers’ behavioral orientations were assessed by an established test instrument, which has been extensively reviewed by the authors with regard to validity and reliability [21]. It consists of five scales addressing attitudes toward the use of technology in science education (eight items, α = 0.88), perceived behavioral control (seven items for self-efficacy, α = 0.73, and four items addressing constraints, α = 0.69), subjective norm (four items, α = 0.58) and motivational orientation (six items, α = 0.87) as a measure for the resulting intention to use technology in practice [21]. Sample items for each scale are shown in Table 4.
Within the scope of the development of a TPACK instrument, in the first step, various national and international TPACK instruments were reviewed. Review criteria were that all seven TPACK dimensions are covered by the instrument, that the number of items per TPACK dimension should be comparable and relatively low (due to the study design, completing the questionnaire should take a maximum of 15 min) while good scale reliabilities are ensured. Only the TPACK instrument by Chai et al. [45] fulfilled all of the given criteria. In the second step, the original items were translated to German and adapted for science teaching. A subsequent review by two experts led to a consensus regarding individual item formulations. A comprehensive overview of all items can be found in [8], and sample items for each scale are shown in Table 4. Evidence of reliability is given by internal consistency in terms of Cronbach’s α (αCK = 0.80; αPK = 0.87; αTK = 0.80; αPCK = 0.86; αTCK = 0.71; αTPK = 0.78; αTPACK = 0.86) and confirmatory factor analysis (CFA). In CFA results, the model fit indices were acceptable (RMSEA = 0.07) or slightly less than good fit values (CFI = 0.89, TLI = 0.87).
Note that no (language) instrument, which fulfilled all given criteria, existed at the time of the project launch. In the meantime, instruments comparable in terms of validity and reliability to the one herein presented have been developed, for example, by Schmid et al. [7].

5.2.2. Lesson Plans

To assess the use of technology, 31 lesson plans were analyzed [78]. Pre-service teachers were asked to design and teach a technology-enhanced science lesson for a topic of their choice in groups of three to four students. In doing so, they can benefit from collaborating with peers when implementing technology in the classroom for the first time [10]. Lesson plans included the school type, grade, rationales for their topics (e.g., anchoring in the national curriculum of the respective grade), learning goals, scientific background, students’ conceptions and use of technology, as well as a tabular display of the chronological lesson outline (instructional activities, educational reasoning, social form and materials and technology use). A detailed template for the lesson plans is available as open educational resource in [79]. All lesson plans were coded independently by two trained coders in terms of quantity and quality of technology exploitation, showing a substantial initial inter-coder reliability of κ = 0.66 [80]. Quantity of technology exploitation in lesson plans was coded in terms of the type of technology and user (teacher/learner), with quality in accordance with the SAMR and ICAP frameworks [67,68,69]). If necessary, subsequent consensus finding was conducted [81]. Thus, all technology implementations described in lesson plans could be unambiguously classified.
Technology exploitation was coded multiple times when the same technology was used multiple times during the lesson or by varying users. Types of technologies were aggregated inductively [82] to provide a better overview of the dominant types of technologies used in lesson plans. In accordance with the hierarchical structure of the SAMR and ICAP frameworks, individual technology exploitations in lesson plans were coded with 1 point (SAMR: substitution; ICAP: passive) to 4 points (SAMR: redefinition; ICAP: interactive). Coding rubrics for the quality of technology exploitation in lesson plans are shown in Table 5.

5.2.3. Module Evaluation

The quality of the courses was assessed through the standardized teaching evaluation form (TEF) of Leuphana University, supplemented by module-specific items. Each module was evaluated individually by participants after attending the respective module. Additional items focused either on digital supplements and practical applications of technology in the accompanying seminar or on digital reminders and the planning and teaching of a technology-enhanced science lesson (Table 6). Where possible, analogous item formulations were used. Participants rated all items on a 5-point Likert scale (1 = “strongly disagree” to 5 = “strongly agree”).
The TEFs were organized by Leuphana University as an online questionnaire, which was sent to all pre-service science teachers enrolled in the module at the end of each semester. Participation was voluntary and completely anonymous. Therefore, conclusions about individual participants are not possible. Nevertheless, the TEF gave participants the opportunity to assess the quality of the modules independently of their own participation in the accompanying research. In total, 50 of the participants gave feedback to the first module ‘Teching and Learning Science’ and 24 to the second module ‘Science in Everyday Life’.

6. Results

The results are presented in three parts according to the research questions.

6.1. Development of TPACK on Action and Behavioral Orientations (ToPB)

6.1.1. TPACK on Action Development

The development of pre-service teachers’ professional knowledge before, during and after attending the modules is summarized in Table 7 and Figure 3. Comparisons between pre- and re-test and re- and post-test, as well as pre- and post-test, were conducted by paired t-tests or Wilcoxon test, depending on the respective data distribution.
The results show that both modules have a positive influence on pre-service science teachers’ self-reported TPACK on action with moderate to large effect sizes (Table 7). Especially during the second module, where participants deepen their theoretical knowledge in practice by designing and teaching a technology-enhanced science lesson, a significant change in their self-reported TPACK on action can be observed. Therefore, by adding digital supplements and digital reminders, as well as giving pre-service science teachers the opportunity to apply their knowledge in practice, existing modules focusing on the PCK development can be successfully extended in order to foster pre-service science teachers’ TPACK on action.

6.1.2. Development of Behavioral Orientations (ToPB)

As TPACK is addressed explicitly in the modules, pre-service teachers gain positive experiences with technology, which presumably influence their behavioral orientations toward the use of technology in science education as well [14,15,16]. In accordance with the ToPB, pre-service science teachers’ behavioral orientations were captured before, during and after attending the modules (Table 8; Figure 4). High values indicate strong manifestations of the respective variable, for example, a positive attitude toward the use of technology in science education, a strongly perceived norm to use it, severe constraints in using it, a high self-efficacy expectation in using it and a strong motivational orientation to using it in practice.
Comparisons between measuring points were conducted by paired t-tests or Wilcoxon tests, depending on the respective data distribution.
As can be seen from Table 8, pre-service science teachers already had positive attitudes (M = 3.69; SD = 0.66) and motivational orientations (M = 3.27; SD = 0.80) toward the use of technology in science education before participating in the study. In contrast, their self-efficacy was rather low (M = 2.69; SD = 0.65). Over the course of the study, attitudes, self-efficacy and motivational orientations increased consistently, leading to significant differences between pre- and post-test, as shown in Table 8 and Figure 4. Simultaneously, participants’ constraints decreased over the course of the study, leading to significant differences between pre- and post-test. No significant differences occurred for subjective norms between the beginning and the end of the study.
Again, effect sizes are bigger between re- and post-test compared to pre- and re-test, with the exception of self-efficacy. This highlights the relevance of incorporating practical experiences in study programs in order to foster not only pre-service science teachers’ TPACK on action but also their behavioral orientations toward the use of technology in science education.

6.1.3. Correlations between TPACK on Action and ToPB Components

Assuming TPACK on action as a predictor for pre-service science teachers’ behavioral orientations, positive (negative for constraints) correlations between TPACK on action and ToPB components were expected and verified by the data. Low significant correlations exist between participants’ self-rated TPACK on action and their attitudes (r(129) = 0.261, p < 0.01), moderate between participants’ self-rated TPACK on action and their motivational orientation (r(129) = 0.354, p < 0.01) and high significant (negative) correlations between participants’ self-rated TPACK on action and their perceived constraints (r(129) = −0.551, p < 0.01) as well as their self-efficacy (r(129) = −0.561, p < 0.01).

6.2. Technology Exploitation in Lesson Plans

Lesson plans were analyzed in terms of quantity and quality of technology exploitation according to the SAMR and ICAP frameworks. Based on this, pre-service science teachers’ TPACK in action was derived subsequently and compared to their self-rated TPACK on action.

6.2.1. Quantity and Quality of Technology Exploitation

Overall, participants described 183 technology exploitations in lesson plans (N = 31), ranging from 2 to 14 per lesson plan. In 31%, technology was used by teachers (pre-service teachers), in 66% by students (fictional learning group) and in 3% of cases, no information was given about the user. An overview of technology exploitation in lesson plans distinguished by the type of technology and user is shown in Table 9.
Figure 5 shows the quality of these cases of technology exploitation as assessed based on conceptualizations by Puentedura [67] for the level of technology integration (SAMR) and by Chi and Wylie [69] for the cognitive engagement of students (ICAP). Based on the lesson plans, 139 of the 183 cases of technology exploitation could be unambiguously assigned to the SAMR and ICAP frameworks according to the rubrics shown in Table 5.
The observed technology exploitation in science lessons follows a typical pattern: First, the respective topic and research question are presented by the (pre-service) science teacher with the help of presentation soft- and hardware, whereby technology is used as a substitute (with functional improvement) for traditional media, leading to a predominantly passive cognitive engagement. The following collection of hypotheses on a virtual bulletin board or with response systems varies across all levels according to the SAMR framework, addressing higher levels of cognitive engagement, namely, constructive and interactive. During the experimental phase, students (in the course of the seminar embodied by other pre-service teachers) are asked to plan and conduct their own experiments. Both real-world experiments including digital measurement acquisition as well as the use of interactive simulations and experimental videos take place. Typically, related documentation is created on virtual bulletin boards or with various office applications. During this phase, technology is almost entirely used by students. When testing their hypotheses, students mostly work in groups, which fosters interactive cognitive engagement with the subject-specific content, while in most lesson plans, technology is used as a substitute for functional improvements or allows for a significant redesign of the task. Additionally, several technology-enhanced scaffolding offers are provided, such as explanatory videos or QR tip cards, leading to online resources with additional information, which students are free to use. For the presentation of results, students use mainly the same technology as for documentation. Thus, students’ cognitive engagement decreases from interactive to constructive or passive, depending on the amount of control by the teacher, while the technology use in terms of the SAMR framework is quite similar to the previous phase.
Overall, the quality of technology exploitation and students’ cognitive engagement is higher when students are the actual users of technology.

6.2.2. TPACK in Action Derived from Lesson Plans

To assess pre-service teachers’ TPACK in action after attending the course, arithmetic means and standard deviations for the quality of technology exploitation in terms of the SAMR and ICAP frameworks were calculated for the observed technology exploitation in each lesson plan. Since SAMR and ICAP values are, by definition, in a range of 1 to 4, no extreme outliers were expected. The arithmetic mean is more sensitive to variation within this data range than other measures and, therefore, can more accurately capture the varying quality of technology use within a lesson plan. The results for each of the 31 groups of pre-service science teachers are presented in Figure 6.
Overall, the quality of technology exploitation can be categorized as augmentation ( M = 2.19 ;   S D = 0.80 ) , while students’ corresponding cognitive engagement is mainly constructive ( M = 2.78 ; S D = 0.95 ) .
Converting these results to a five-point scale, TPACK in action derived from lesson plans can be compared to pre-service teachers’ self-rated TPACK on action after participating in the course (Table 10). Note that due to the preservation of anonymity and the design of technology-enhanced lessons by groups of three to four pre-service teachers, it is not possible to combine the quantitative and qualitative data for individual participants. Nevertheless, the data help to gain an understanding of the interplay of pre-service teachers’ TPACK on action and their resulting TPACK in action.
After attending the course, pre-service teachers’ TPK, PCK and resulting TPACK in action were consistently lower than corresponding self-rated TPACK on action. Mann–Whitney tests show that these differences are significant for TPK and TPACK ( TPK :   U = 117.5 ,   Z = 6.64 ,   p < 0.001 ;   TPACK :   U = 475.0 ,   Z = 3.40 ,   p < 0.001 ) , while no significant differences in pre-service teachers’ PCK on and in action were identified ( PCK :   U = 668.5 ,   Z = 1.66 ,   p = 0.097 ) .
The comparably lower TPK in action (and resulting TPACK in action) can have several reasons, e.g., pre-service science teachers overestimate their TPK on action, their behavioral orientations limit technology exploitation in practice, they make a reasoned choice against the use of technology (which cannot be derived from lesson plans) or the comparison of varying samples in terms of number and composition.

6.3. Pre-Service Teachers’ Assessment of the Course Design

The results of the course-specific TEF items are shown in Figure 7 for the module ‘Teaching and Learning Sciene’ (left) and ‘Science in Everyday Life’ (right). Analogously formulated items are presented next to each other.
Pre-service science teachers assessed the number of incorporated technologies in digital supplements (item X.1) as well as their fit to science education (item X.2) as good to very good (number: M = 4.52, SD = 0.71; fit: M = 4.35, SD = 0.85). Practical applications of technology in the seminar (item X.3) as well as planning and teaching a technology-enhanced lesson (item Y.3) were rated as good opportunities to deepen theoretical knowledge in practice (seminar: M = 4.16, SD = 0.86; technology-enhanced lesson: M = 4.13, SD = 0.90). In addition, participants agreed that technology exploitation (as presented in the digital supplements/reminders) can be implemented into school practice (item X.6: M = 3.88, SD = 1.05; item Y.6: M = 4.08, SD = 0.91) and enhances the quality of science teaching and learning (item X.7: M = 4.02, SD = 0.96; item Y.7: M = 4.38, SD = 0.75). Overall, pre-service teachers stated that the incorporation of digital supplements as well as planning and teaching a technology-enhanced science lesson promoted their professional knowledge (item X.8: M = 4.06, SD = 1.04; item Y.8: M = 4.21, SD = 0.82).
Comparing both modules, the quality of the second module is rated slightly better by the attendees. Participants assessed the quality of the approach, which combines fundamental science education topics (PCK) with technology applications for teaching and learning science (TPACK), as good to very good. In particular, practical deepening in the accompanying seminar and by planning and teaching a technology-enhanced science lesson give pre-service teachers the necessary opportunities to apply and deepen their professional knowledge.

7. Discussion

In the presented study, the transformative view of TPACK [2,4,5,6,7,8,35,36,37,38] was used as theoretical background to implement technology systematically into the science teacher education program at the Leuphana University Lüneburg. Participation in the modules led to a significant increase in pre-service teachers’ self-reported professional knowledge in all TPACK on action dimensions with moderate to large effect sizes. These findings are in line with other studies examining technology-specific courses [13]. In addition, Zimmermann et al. [14] stated that “[…] technology integration courses should not be detached from subject didactics since not only the technology related components of the TPACK framework improved but also PCK” (p. 1868). Thus, courses at the university level should not treat technology as an add-on [10] but rather link the PCK and TPACK development of pre-service science teachers systematically. The presented approach not only fulfills this claim by integrating technology into the existing PCK modules, but it also leads to a significant increase in self-rated PCK and TPACK on action after attending the modules, as evidenced by the data.
TPACK is not only addressed theoretically in the form of digital supplements and reminders, but participants were also given various opportunities to deepen their professional knowledge and apply it to practice, e.g., by designing and teaching a technology-enhanced science lesson.
In accordance with previous studies [14,15,16], our results show that these positive experiences with technology influence pre-service teachers’ behavioral orientations toward the use of technology in science education according to the ToPB [18]. During participation, pre-service teachers’ attitudes, self-efficacy and motivational orientations toward the use of technology in science education increased consistently, leading to significant differences between pre- and post-test. Several authors have outlined that the ToPB is a valid framework to study pre-service teachers’ behavioral orientations toward the use of technology in practice [19,20,21,25]. Positive behavioral orientations can be seen as a prerequisite for technology use, with higher values suggesting a greater willingness to integrate technology, which is confirmed by the results of this study. However, to our knowledge, no study combines measures of TPACK on action and behavioral orientations, as well as TPACK in action, to gain an in-depth understanding of pre-service science teachers’ TPACK development in the course of a university program. Therefore, the presented study design provides a novelty.
Pre-service science teachers’ TPACK in action was derived from lesson plans. In contrast to previous studies that applied TPACK rubrics to lesson plans [31,55,56], TPACK in action was approximated following conceptualizations by Puentedura [67] for the quality of technology exploitation (SAMR) and by Chi [68] and Chi and Wylie [69] for the corresponding instructional quality (ICAP). Commonly, the SAMR as well as the ICAP framework are used to rate the quality of (technology-enhanced) activities/tasks [46,62,70,71,72,82,83], not teachers’ underlying professional knowledge. We argue that the SAMR framework can be applied as a measure of pre-service teachers’ TPK in action and the ICAP framework as a measure of PCK in action; thus, the resulting combination of both leads to an approximation of pre-service teachers’ TPACK in action. This assumption is in line with current empirical findings on the transformative view of TPACK, which indicate that only TPK and PCK have a significant influence on TPACK [4,5,7,8,33].
Analysis of lesson plans revealed a medium to high quality of technology exploitations (27% substitution, 38% augmentation, 27% modification and 8% redefinition), corresponding instructional quality ranged from passive (17%), overactive (24%) and constructive (20%) to an interactive cognitive engagement (39%) of students. In comparison, the authors of [78] analyzed 85 lessons from experienced biology teachers in terms of quantity and quality of technology exploitation. A total of 71% of technology exploitation was categorized as substitution, 11% as augmentation and 19% could not be clearly assigned to the SAMR framework. The high quality of technology exploitations, as can be seen in the lesson plans, indicates a distinctive TPACK in action after attending the course. Nevertheless, matching the scales reveals that pre-service teachers’ TPK, PCK and resulting TPACK in action are consistently lower than corresponding self-assessments, with significant differences for TPK and consequently TPACK as well. While the ICAP framework seems to provide a solid approximation of PCK in action that is consistent with pre-service teachers’ self-assessments, it remains to be clarified whether the SAMR framework underestimates TPK in action or if participants overestimate their TPK/TPACK on action. Since TPACK on and in action profiles cannot be matched for individual participants in the present study, it is neither possible to prove nor to disprove both assumptions based on the data. In addition, since pre-service teachers’ behavioral orientations toward the use of technology in science education increased in the course of the study, it seems unlikely that differences between their TPK on and in action are caused by intrinsic beliefs. Nevertheless, social desirability in the ToPB data cannot be ruled out since the redesigned modules convey a positive attitude toward technology exploitation in science education. It is needless to say that in digital supplements/reminders, the advantages, as well as possible disadvantages, of technology exploitation were discussed. A further organizational factor that might cause the differences in TPK on and in action is the planning and teaching of a science lesson in groups consisting of pre-service science teachers with possibly varying professional knowledge or behavioral orientations.
Following the transformative view of TPACK, teaching and learning technology was implemented systematically into the existing science education modules at Leuphana University instead of creating additional offers. The quality of the chosen approach was rated positively throughout by pre-service teachers. Participants especially highlighted practical approaches in terms of seminars (module ‘Teaching and Learning Science’) and the opportunity to plan and teach a science lesson (module ‘Science in Everyday Life’) where they could apply their newly gained knowledge in practice.

8. Limitations

The presented study has several limitations such as the number of participants or the chosen empirical approach.
The first limitation is the number of pre-service science teachers who attended the course as well as the decreasing participation in the accompanying questionnaire.
Second, due to the design of technology-enhanced lessons by groups of three to four pre-service teachers and the preservation of anonymity in the TEF, it is not possible to combine self-rated TPACK on action with TPACK in action for individual participants or student groups. In addition, the quality of the modules rated by participants cannot be related to the actual development of their professional knowledge through attending the modules.
Third, another limitation of the study is the choice to use the SAMR and ICAP frameworks as measures of TPK and PCK. While this approach is in line with current empirical findings on the transformative view of TPACK, it nevertheless neglects possible influences of TCK and is, thus, only an approximation of pre-service teachers’ TPACK in action. Moreover, analyses of lesson plans always fail to identify the reasoned choice against technology, which also is an indicator of a high TPACK on action level.
Fourth, due to economic reasons in the course design and data collection process, pre-service science teachers’ TPACK in action was only measured once at the end of the study. An analysis of the TPACK in action development during the course is, thus, not possible.
Some of the above limitations can only be influenced and minimized to a certain extent, while others can serve as a basis for planning future research, as explained in the following section.

9. Conclusions and Implications

Several implications for curriculum designers and future research can be derived from the results presented. The results show that the chosen course design is highly promising. In the sense of “from PCK to TPACK”, technology should not be taught additionally but systematically integrated into existing science education programs focusing on PCK [8,9]. By doing so, pre-service science teachers can develop the necessary professional knowledge to implement purpose-oriented technology in practice. In addition to the empirically shown increase in pre-service teachers’ self-rated TPACK on action and the positive development of their behavioral orientations toward the use of technology in science education during their participation in the presented study, participants rated the quality of the modules as good to very good, highlighting the practical application of technology in the accompanying seminars. Thus, the presented course design can serve as a successful example for curriculum designers to redesign courses at their universities in a similar way.
A first implication for future research is the successful development of a reliable and valid German instrument to assess TPACK on action [8], which did not exist at the time of the project launch.
In addition, the presented study poses a new approach to accessing (pre-service) science teachers’ TPACK in action by analyzing lesson plans, which can be easily adapted to lesson observations. In contrast to existing TPACK rubrics, TPACK in action is derived by measures of PCK and TPK according to recent empirical findings on the transformative view of TPACK [4,5,7,8,33]. The SAMR framework by Puentedura [67] is thereby used as an approximation for teachers’ TPK. It describes the quality of technology exploitation in comparison to traditional teaching and learning media and has been successfully applied to analyze teachers’ abilities to plan and deliver technology-enriched lessons in various (science) subjects [62,70,83,84]. In addition, PCK is measured in terms of students’ cognitive engagement achieved through the use of technology according to the ICAP framework [68,69]. Both frameworks allow for a detailed resolution of individual technology exploitation as well as an overall judgment of a lesson. Furthermore, their combination results in a reasoned approximation of teachers’ TPACK in action, which takes both teachers’ PCK and TPK, equally into account. Scholars are encouraged to adapt this approach for future research, not only in science education.
Yet, more studies combining measures of TPACK on action, behavioral intentions and TPACK in action are necessary to gain an in-depth understanding of (pre-service) science teachers’ TPACK (development). The presented study can serve as an example for the design of future investigations. Nevertheless, due to limitations in the data collection process, measures of TPACK on action and behavioral orientations could not be combined with TPACK in action for individual participants, which is, thus, a future research desideratum.
The presented results as well as previous studies make evident that teachers’ self-rated TPACK on action and behavioral intentions are sometimes inconsistent with their actions in the classroom [25,26]. To understand this discrepancy, the analysis of lesson plans or observations needs to be extended, e.g., through interviews that focus on the actual planning process. Exclusively looking at the final product, not all indicators for a high TPACK level, such as the reasoned choice against technology exploitation, can be identified. Therefore, approaches that give teachers the opportunity to explain their choice to (not) implement technology in practice are a further research desideratum.
Last but not least, current research focuses on pre-service science teachers’ TPACK development in the course of various study programs. All of these studies (including the presented) assume that participating pre-service science teachers will apply their gained professional knowledge in the future, but the evidence is still missing. Long-term studies that follow pre-service science teachers in their transition from university to school practice are needed to fully understand the impact of university courses that promote pre-service science teachers’ TPACK.

Author Contributions

Conceptualization, L.S.-R., E.H., A.R. and S.A.; methodology, E.H. and A.R.; software, L.S.-R.; validation, L.S.-R. and E.H.; formal analysis, L.S.-R.; investigation, L.S.-R., E.H., A.R. and S.A.; resources, L.S.-R. and S.A.; data curation, L.S.-R.; writing—original draft preparation, L.S.-R.; writing—review and editing, L.S.-R., E.H., A.R. and S.A.; visualization, L.S.-R. and E.H.; supervision, S.A.; project administration, L.S.-R.; funding acquisition, L.S.-R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Joachim Herz Foundation within the funding program ‘Kolleg Didaktik:digital’.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Acknowledgments

The authors would like to thank all participating pre-service teachers from the Leuphana University Lüneburg.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Koehler, M.J.; Mishra, P.; Kereluik, K.; Shin, T.S.; Graham, C.R. The Technological Pedagogical Content Knowledge Framework. In Handbook of Research on Educational Communications and Technology, 4th ed.; Springer: Heidelberg/Berlin, Germany, 2014; pp. 101–111. [Google Scholar] [CrossRef]
  2. Mishra, P.; Koehler, M.J. Technological Pedagogical Content Knowledge: A Framework for Teacher Knowledge. Teach. Coll. Rec. 2006, 108, 1017–1054. [Google Scholar] [CrossRef]
  3. Kay, R.H. Evaluating Strategies Used to Incorporate Technology into Preservice Education. J. Res. Technol. Educ. 2006, 38, 383–408. [Google Scholar] [CrossRef]
  4. Jang, S.-J.; Chen, K.-C. From PCK to TPACK: Developing a Transformative Model for Pre-Service Science Teachers. J. Sci. Educ. Technol. 2010, 19, 553–564. [Google Scholar] [CrossRef]
  5. Jin, Y. The Nature of TPACK: Is TPACK Distinctive, Integrative or Transformative? In Proceedings of the Society for Information Technology & Teacher Education International Conference, Waynesville, NC, USA, 18 March 2019; pp. 2199–2204. [Google Scholar]
  6. Pamuk, S.; Ergun, M.; Cakir, R.; Yilmaz, H.B.; Ayas, C. Exploring Relationships among TPACK Components and Development of the TPACK Instrument. Educ. Inf. Technol. 2015, 20, 241–263. [Google Scholar] [CrossRef]
  7. Schmid, M.; Brianza, E.; Petko, D. Developing a Short Assessment Instrument for Technological Pedagogical Content Knowledge (TPACK.Xs) and Comparing the Factor Structure of an Integrative and a Transformative Model. Comput. Educ. 2020, 157, 103967. [Google Scholar] [CrossRef]
  8. Stinken-Rösner, L. Digitale Medien in Der Naturwissenschaftlichen Lehrkräftebildung. PhyDid B - Didaktik der Physik - Beiträge zur DPG-Frühjahrstagung. 2021. Available online: https://ojs.dpg-physik.de/index.php/phydid-b/article/view/1114 (accessed on 1 June 2023).
  9. Stinken-Rösner, L.; Abels, S. Forschendes Lernen Mit Digitalen Medien Ein Beitrag Zur “Diklusion”. In Unsicherheit als Element von naturwissenschaftsbezogenen Bildungsprozessen. Gesellschaft für Didaktik der Chemie und Physik Virtuelle Jahrestagung 2021; Habig, S., van Vorst, H., Eds.; Duisburg-Essen: Duisburg, Germany, 2022; pp. 72–75. [Google Scholar]
  10. Bell, R.L.; Maeng, J.L.; Binns, I.C. Learning in Context: Technology Integration in a Teacher Preparation Program Informed by Situated Learning Theory. J. Res. Sci. Teach. 2013, 50, 348–379. [Google Scholar] [CrossRef]
  11. Kartal, T.; DïLek, İ. Preservice Science Teachers’ TPACK Development in a Technology-Enhanced Science Teaching Method Course. J. Educ. Sci. Environ. Health 2021, 4, 339–353. [Google Scholar] [CrossRef]
  12. Voet, M.; De Wever, B. Towards a Differentiated and Domain-Specific View of Educational Technology: An Exploratory Study of History Teachers’ Technology Use: Exploring History Teachers’ Technology Use. Br. J. Educ. Technol. 2017, 48, 1402–1413. [Google Scholar] [CrossRef] [Green Version]
  13. Wilson, M.L.; Ritzhaupt, A.D.; Cheng, L. The Impact of Teacher Education Courses for Technology Integration on Pre-Service Teacher Knowledge: A Meta-Analysis Study. Comput. Educ. 2020, 156, 103941. [Google Scholar] [CrossRef]
  14. Zimmermann, F.; Melle, I.; Huwer, J. Developing Prospective Chemistry Teachers’ TPACK—A Comparison between Students of Two Different Universities and Expertise Levels Regarding Their TPACK Self-Efficacy, Attitude, and Lesson Planning Competence. J. Chem. Educ. 2021, 98, 1863–1874. [Google Scholar] [CrossRef]
  15. Bastian, J.; Riplinger, T. Tablets for a Redefinition of Learning? An Analysis of Video Observations to Determine the Integration of Tablets in the Classroom. In EdMedia+ Innovate Learning; Association for the Advancement of Computing in Education (AACE): Asheville, NC, USA, 2016; pp. 143–149. [Google Scholar]
  16. Instefjord, E.J.; Munthe, E. Educating Digitally Competent Teachers: A Study of Integration of Professional Digital Competence in Teacher Education. Teach. Teach. Educ. 2017, 67, 37–45. [Google Scholar] [CrossRef]
  17. Tondeur, J.; Van Braak, J.; Ertmer, P.A.; Ottenbreit-Leftwich, A. Understanding the Relationship between Teachers’ Pedagogical Beliefs and Technology Use in Education: A Systematic Review of Qualitative Evidence. Educ. Tech Res. Dev. 2017, 65, 555–575. [Google Scholar] [CrossRef]
  18. Ajzen, I. The Theory of Planned Behavior. Organ. Behav. Hum. Decis. Process. 1991, 50, 179–211. [Google Scholar] [CrossRef]
  19. Teo, T.; Tan, L. The Theory of Planned Behavior (TPB) and Pre-Service Teachers’ Technology Acceptance: A Validation Study Using Structural Equation Modeling. J. Technol. Teach. Educ. 2012, 20, 89–104. [Google Scholar]
  20. Valtonen, T.; Kukkonen, J.; Kontkanen, S.; Mäkitalo-Siegl, K.; Sointu, E. Differences in Pre-Service Teachers’ Knowledge and Readiness to Use ICT in Education. J. Comput. Assist. Learn. 2018, 34, 174–182. [Google Scholar] [CrossRef]
  21. Vogelsang, C.; Finger, A.; Laumann, D.; Thyssen, C. Vorerfahrungen, Einstellungen und motivationale Orientierungen als mögliche Einflussfaktoren auf den Einsatz digitaler Werkzeuge im naturwissenschaftlichen Unterricht [Experience, Attitudes and Motivational Orientations as Potential Factors Influencing the Use of Digital Tools in Science Teaching]. ZfDN 2019, 25, 115–129. [Google Scholar] [CrossRef]
  22. Jung, L.; Cerreto, F.A.; Lee, J. Theory of Planned Behavior and Teachers’ Decisions Regarding Use of Educational Technology. J. Educ. Technol. Soc. 2010, 13, 152–164. [Google Scholar]
  23. Lindsey, L.; Buss, R.; Foulger, T.; Wetzel, K.; Pasquel, S. The Technology Infusion ITeach Experience: Preparing Student Teachers to Integrate Technology. In Society for Information Technology & Teacher Education International Conference; Association for the Advancement of Computing in Education (AACE): Nashville, TN, USA, 2016; pp. 2923–2930. [Google Scholar]
  24. Sointu, E.; Valtonen, T.; Cutucache, C.E.; Kukkonen, J.; Lambert, M.C.; Mäkitalo-Siegl, K. Differences in Preservice Teachers’ Readiness to Use ICT in Education and Development of TPACK. In Society for Information Technology & Teacher Education International Conference; Association for the Advancement of Computing in Education (AACE): Nashville, TN, USA, 2017; p. 10. [Google Scholar]
  25. Gonzalez, M.J.; Ruiz, I.G. Behavioural Intention and Pre-Service Mathematics Teachers’ Technological Pedagogical Content Knowledge. EURASIA J. Math. Sci. Technol. Ed. 2016, 13, 601–620. [Google Scholar] [CrossRef]
  26. Van Der Ross, D.; Tsibolane, P. The Influence of Teacher Attitudes and Beliefs on Information and Communications Technology Integration Behavior in South African High Schools. CONF-IRM 2017 PROCEEDINGS. Available online: https://core.ac.uk/download/pdf/301372265.pdf (accessed on 1 June 2023).
  27. Kilty, T.J.; Burrows, A.C. Secondary Science Preservice Teachers: Technology Integration in Methods and Residency. J. Sci. Teach. Educ. 2021, 32, 578–600. [Google Scholar] [CrossRef]
  28. Shulman, L.S. Those Who Understand: Knowledge Growth in Teaching. Educ. Res. 1986, 15, 4–14. [Google Scholar] [CrossRef]
  29. Grossman, P.L. A Study in Contrast: Sources of Pedagogical Content Knowledge for Secondary English. J. Teach. Educ. 1989, 40, 24–31. [Google Scholar] [CrossRef] [Green Version]
  30. Niess, M.L. Preparing Teachers to Teach Science and Mathematics with Technology: Developing a Technology Pedagogical Content Knowledge. Teach. Teach. Educ. 2005, 21, 509–523. [Google Scholar] [CrossRef]
  31. Canbazoğlu Bilici, S.; Guzey, S.; Donna, J.; Roehrig, G.; Karahan, E.; Yamak, H.; Kavak, N. A Technological Pedagogical Content Knowledge (TPACK)-Based Lesson Plan Assessment Instrument. In Proceedings of the Association for Science Teacher Education (ASTE), Charleston, SC, USA, 9–12 January 2013. [Google Scholar]
  32. Magnusson, S.; Krajcik, J.; Borko, H. Nature, Sources, and Development of Pedagogical Content Knowledge for Science Teaching. In Examining Pedagogical Content Knowledge: The Construct and its Implications for Science Education; Gess-Newsome, J., Lederman, N.G., Eds.; Science & Technology Education Library; Springer: Dordrecht, The Netherlands, 1999; pp. 95–132. ISBN 978-0-306-47217-6. [Google Scholar]
  33. Angeli, C.; Valanides, N. Epistemological and Methodological Issues for the Conceptualization, Development, and Assessment of ICT–TPCK: Advances in Technological Pedagogical Content Knowledge (TPCK). Comput. Educ. 2009, 52, 154–168. [Google Scholar] [CrossRef]
  34. Graham, C.R. Theoretical Considerations for Understanding Technological Pedagogical Content Knowledge (TPACK). Comput. Educ. 2011, 57, 1953–1960. [Google Scholar] [CrossRef]
  35. Angeli, C.; Valanides, N. Preservice Elementary Teachers as Information and Communication Technology Designers: An Instructional Systems Design Model Based on an Expanded View of Pedagogical Content Knowledge. J. Comput. Assist. Learn. 2005, 21, 292–302. [Google Scholar] [CrossRef]
  36. Celik, I.; Sahin, I.; Akturk, A.O. Analysis of the Relations among the Components of Technological Pedagogical and Content Knowledge (Tpack): A Structural Equation Model. J. Educ. Comput. Res. 2014, 51, 1–22. [Google Scholar] [CrossRef]
  37. Dong, Y.; Chai, C.S.; Sang, G.-Y.; Koh, J.H.L.; Tsai, C.-C. Exploring the Profiles and Interplays of Pre-Service and In-Service Teachers’ Technological Pedagogical Content Knowledge (TPACK) in China. J. Educ. Technol. Soc. 2015, 18, 158–169. [Google Scholar]
  38. Koh, J.H.L.; Chai, C.S.; Tsai, C.-C. Examining Practicing Teachers’ Perceptions of Technological Pedagogical Content Knowledge (TPACK) Pathways: A Structural Equation Modeling Approach. Instr. Sci. 2013, 41, 793–809. [Google Scholar] [CrossRef]
  39. Benton-Borghi, B.H. Intersection and Impact of Universal Design for Learning (UDL) and Technological, Pedagogical, and Content Knowledge (TPACK) on Twenty-First Century Teacher Preparation: UDL-Infused TPACK Practitioner’s Model. In Technological Pedagogical Content Knowledge: Exploring, Developing, and Assessing TPCK; Angeli, C., Valanides, N., Eds.; Springer: Boston, MA, USA, 2015; pp. 287–304. ISBN 978-1-4899-8080-9. [Google Scholar]
  40. Huwer, J.; Irion, T.; Knuntze, S.; Schaal, S.; Thyssen, C. Digitality-Related Paedagogical Content Knowledge (DPaCK)—A Framework for Teacher Education in the Digital Age. In Education Research Highlights in Mathematics, Science and Technology 2019; Shelley, M., Kiray, A., Eds.; IRES Publishing: San Francisco, CA, USA, 2019; pp. 289–309. [Google Scholar]
  41. Rodríguez-Becerra, J.; Cáceres-Jensen, L.; Díaz, T.; Druker, S.; Padilla, V.B.; Pernaa, J.; Aksela, M. Developing Technological Pedagogical Science Knowledge through Educational Computational Chemistry: A Case Study of Pre-Service Chemistry Teachers’ Perceptions. Chem. Educ. Res. Pract. 2020, 21, 638–654. [Google Scholar] [CrossRef]
  42. Archambault, L.; Crippen, K. Examining TPACK among K-12 Online Distance Educators in the United States. Contemp. Issues Technol. Teach. Educ. CITE J. 2009, 9, 71–88. [Google Scholar]
  43. Çetin, I.; Erdogan, A. Development, Validity and Reliability Study of Technological Pedagogical Content Knowledge (TPACK) Efficiency Scale for Mathematics Teacher Candidates. Int. J. Contemp. Educ. Res. 2018, 5, 50–62. [Google Scholar]
  44. Chai, C.S.; Koh, J.H.L.; Tsai, C.C. Exploring the Factor Structure of the Constructs of Technological, Pedagogical, Content Knowledge (TPACK). Asia Pac. Educ. Res. 2011, 20, 595–603. [Google Scholar]
  45. Chai, C.S.; Ng, E.M.; Li, W.; Hong, H.-Y.; Koh, J.H.L. Validating and Modelling Technological Pedagogical Content Knowledge Framework among Asian Preservice Teachers. Australas. J. Educ. Technol. 2013, 29, 41–53. [Google Scholar] [CrossRef] [Green Version]
  46. Kabakci Yurdakul, I.; Odabasi, H.F.; Kilicer, K.; Coklar, A.N.; Birinci, G.; Kurt, A.A. The Development, Validity and Reliability of TPACK-Deep: A Technological Pedagogical Content Knowledge Scale. Comput. Educ. 2012, 58, 964–977. [Google Scholar] [CrossRef]
  47. Sahin, I. Development of Survey of Technological Pedagogical and Content Knowledge (TPACK). Turk. Online J. Educ. Technol. TOJET 2011, 10, 97–105. [Google Scholar]
  48. Schmid, M.; Brianza, E.; Petko, D. Self-Reported Technological Pedagogical Content Knowledge (TPACK) of Pre-Service Teachers in Relation to Digital Technology Use in Lesson Plans. Comput. Hum. Behav. 2021, 115, 106586. [Google Scholar] [CrossRef]
  49. Schmidt, D.A.; Baran, E.; Thompson, A.D.; Mishra, P.; Koehler, M.J.; Shin, T.S. Technological Pedagogical Content Knowledge (TPACK). J. Res. Technol. Educ. 2009, 42, 123–149. [Google Scholar] [CrossRef]
  50. Valtonen, T.; Sointu, E.; Kukkonen, J.; Kontkanen, S.; Lambert, M.C.; Mäkitalo-Siegl, K. TPACK Updated to Measure Pre-Service Teachers’ Twenty-First Century Skills. Australas. J. Educ. Technol. 2017, 33, 15–31. [Google Scholar] [CrossRef] [Green Version]
  51. von Kotzebue, L. Beliefs, Self-reported or Performance-Assessed TPACK: What Can Predict the Quality of Technology-Enhanced Biology Lesson Plans? J. Sci. Educ. Technol. 2022, 31, 570–582. [Google Scholar] [CrossRef]
  52. Drummond, A.; Sweeney, T. Can an Objective Measure of Technological Pedagogical Content Knowledge (TPACK) Supplement Existing TPACK Measures? Br. J. Educ. Technol. 2017, 48, 928–939. [Google Scholar] [CrossRef]
  53. Maderick, J.A.; Zhang, S.; Hartley, K.; Marchand, G. Preservice Teachers and Self-Assessing Digital Competence. J. Educ. Comput. Res. 2016, 54, 326–351. [Google Scholar] [CrossRef]
  54. Große-Heilmann, R.; Riese, J.; Burde, J.P.; Schubatzky, T.; Weiler, D. Fostering Pre-Service Physics Teachers’ Pedagogical Content Knowledge Regarding Digital Media. Educ. Sci. 2022, 12, 440. [Google Scholar] [CrossRef]
  55. Harris, J.B.; Grandgenett, N.; Hofer, M. Testing a TPACK-Based Technology Integration Assessment Rubric. In Society for Information Technology & Teacher Education International Conference; Association for the Advancement of Computing in Education (AACE): Asheville, NC, USA, 2010; Volume 18. [Google Scholar]
  56. Tournaki, N.; Lyublinskaya, I. Preparing Special Education Teachers for Teaching Mathematics and Science with Technology by Integrating TPACK Framework into the Curriculum: A Study of Teachers’ Perceptions. J. Technol. Teach. Educ. 2014, 22, 243–259. [Google Scholar]
  57. Agyei, D.; Voogt, J. Determining Teachers’ TPACK through Observations and Self-Report Data. In Proceedings of the SITE 2011—Society for Information Technology & Teacher Education International Conference, Nashville, TN, USA, 7–11 March 2011; Koehler Matthew, J.M., Mishra, P., Eds.; Association for the Advancement of Computing in Education (AACE): Nashville, TN, USA, 2011; pp. 2314–2319. [Google Scholar]
  58. Hofer, M.; Grandgenett, N.; Harris, J.B.; Swan, K. Testing a TPACK-Based Technology Integration Observation Instrument. In Proceedings of the Society for Information Technology & Teacher Education International Conference, Nashville, TN, USA, 7 March 2011. [Google Scholar]
  59. Schmidt-Crawford, D.A.; Tai, S.-J.D.; Wang, W.; Jin, Y. Understanding Teachers’ TPACK Through Observation. In Handbook of Technological Pedagogical Content Knowledge (TPACK) for Educators; Routledge: Milton Park, UK, 2016; pp. 117–128. [Google Scholar]
  60. Park, S.; Oliver, J.S. Revisiting the Conceptualisation of Pedagogical Content Knowledge (PCK): PCK as a Conceptual Tool to Understand Teachers as Professionals. Res. Sci. Educ. 2008, 38, 261–284. [Google Scholar] [CrossRef]
  61. Gess-Newsome, J. A Model of Teacher Professional Knowledge and Skill Including PCK: Results of the Thinking from the PCK Summit. In Re-Examining Pedagogical Content Knowledge in Science Education; Berry, A., Friedrichsen, P., Loughran, J., Eds.; Routledge: New York, NY, USA, 2015; pp. 28–42. ISBN 978-1-315-73566-5. [Google Scholar]
  62. Backfisch, I.; Lachner, A.; Hische, C.; Loose, F.; Scheiter, K. Professional knowledge or motivation? Investigating the role of teachers’ expertise on the quality of technology-enhanced lesson plans. Learn. Instr. 2020, 66, 101300. [Google Scholar] [CrossRef]
  63. Willermark, S. Technological pedagogical and content knowledge: A review of empirical studies published from 2011 to 2016. J. Educ. Comput. Res. 2018, 56, 315–343. [Google Scholar] [CrossRef]
  64. Brandhofer, M.G. Die Kompetenzen der Lehrenden an Schulen im Umgang mit ardenti Medien und die Wechselwirkungen zwischen Lehrtheorien und mediendidaktischem Handeln. Ph.D. Thesis, Technische Universität Dresden, Dresden, Germany, 2015. [Google Scholar]
  65. Endberg, M. Professionswissen von Lehrpersonen der Sekundarstufe I zum Einsatz Digitaler Medien im Unterricht. Eine Untersuchung auf Basis einer repräsentativen Lehrerbefragung; Empirische Erziehungswissenschaft; Waxmann Verlag: Münster, Germany; New York, NY, USA, 2019; Volume 71, ISBN 978-3-8309-4004-3. [Google Scholar]
  66. Wang, W.; Schmidt-Crawford, D.; Jin, Y. Preservice Teachers’ TPACK Development: A Review of Literature. J. Digit. Learn. Teach. Educ. 2018, 34, 234–258. [Google Scholar] [CrossRef]
  67. Puentedura, R. Transformation, Technology, and Education; Hippasus: Nicosia, Cyprus, 2006. [Google Scholar]
  68. Chi, M.T.H. Active-Constructive-Interactive: A Conceptual Framework for Differentiating Learning Activities. Top. Cogn. Sci. 2009, 1, 73–105. [Google Scholar] [CrossRef] [Green Version]
  69. Chi, M.T.H.; Wylie, R. The ICAP Framework: Linking Cognitive Engagement to Active Learning Outcomes. Educ. Psychol. 2014, 49, 219–243. [Google Scholar] [CrossRef]
  70. Kramer, M.; Förtsch, C.; Aufleger, M.; Neuhaus, B.J. Der Einsatz digitaler Medien im gymnasialen Biologieunterricht. Eine descriptive Auswertung einer quantitativen Videostudie. ZfDN 2019, 25, 131–160. [Google Scholar] [CrossRef]
  71. Deepika, A.; Kandakatla, R.; Saida, A.; Reddy, V.B. Implementation of ICAP Principles through Technology Tools: Exploring the Alignment between Pedagogy and Technology. J. Eng. Educ. Transform. 2021, 34, 542–549. [Google Scholar] [CrossRef]
  72. Wekerle, C.; Daumiller, M.; Kollar, I. Using Digital Technology to Promote Higher Education Learning: The Importance of Different Learning Activities and Their Relations to Learning Outcomes. J. Res. Technol. Educ. 2020, 1–17. [Google Scholar] [CrossRef]
  73. Ertmer, P.A.; Ottenbreit-Leftwich, A.T. Teacher Technology Change. J. Res. Technol. Educ. 2010, 42, 255–284. [Google Scholar] [CrossRef]
  74. Harris, J.B.; Hofer, M.J. Technological Pedagogical Content Knowledge (TPACK) in Action. J. Res. Technol. Educ. 2011, 43, 211–229. [Google Scholar] [CrossRef]
  75. Ocak, C.; Baran, E. Observing the Indicators of Technological Pedagogical Content Knowledge in Science Classrooms: Video-Based Research. J. Res. Technol. Educ. 2019, 51, 43–62. [Google Scholar] [CrossRef]
  76. Valtonen, T.; Leppänen, U.; Hyypiä, M.; Sointu, E.; Smits, A.; Tondeur, J. Fresh Perspectives on TPACK: Pre-Service Teachers’ Own Appraisal of Their Challenging and Confident TPACK Areas. Educ. Inf. Technol. 2020, 25, 2823–2842. [Google Scholar] [CrossRef] [Green Version]
  77. Höttecke, D.; Allchin, D. Reconceptualizing Nature-of-science Education in the Age of Social Media. Sci. Educ. 2020, 104, 641–666. [Google Scholar] [CrossRef] [Green Version]
  78. Kuckartz, U. Qualitative Text Analysis: A Guide to Methods, Practice & Using Software; SAGE Publications Ltd.: Thousand Oaks, CA, USA, 2014. [Google Scholar]
  79. Abels, S.; Hofer, E.; Hollstein, S.; Rodenhauser, A.; Stinken-Rösner, L.; Hüfner, S. Kontextorientierte Unterrichtseinheit Zum Forschenden Lernen Im Inklusiven Naturwissenschaftlichen Unterricht. twillo. 2022. Available online: https://www.twillo.de/edu-sharing/components/render/45f2f9c0-c2f0-4999-bc20-fddfb673dd81 (accessed on 1 June 2023).
  80. Brennan, R.L.; Prediger, D.J. Coefficient Kappa: Some Uses, Misuses, and Alternatives. Educ. Psychol. Meas. 1981, 41, 687–699. [Google Scholar] [CrossRef]
  81. Bortz, J.; Döring, N. Qualitative Methoden. In Forschungsmethoden und Evaluation; Bortz, J., Döring, N., Eds.; Springer: Berlin/Heidelberg, Germany, 2006; pp. 295–350. [Google Scholar] [CrossRef]
  82. Mayring, P. Qualitative Content Analysis: Theoretical Background and Procedures. In Approaches to Qualitative Research in Mathematics Education: Examples of Methodology and Methods; Bikner-Ahsbahs, A., Knipping, C., Presmeg, N., Eds.; Advances in Mathematics Education; Springer: Dordrecht, The Netherlands, 2015; pp. 365–380. ISBN 978-94-017-9181-6. [Google Scholar]
  83. Romrell, D.; Kidder, L.C.; Wood, E. The SAMR Model as a Framework for Evaluating mLearning. J. Asynchronous Learn. Netw. 2014, 18, 1–15. [Google Scholar] [CrossRef] [Green Version]
  84. Schaal, S. Man sieht den Wald vor lauter Bäumen nicht. Wie Digitale Medien für den Biologieunterricht ausgewählt werden. Unterr. Biol. 2017, 429, 46–47. [Google Scholar]
Figure 1. TPACK framework (reproduced by permission of the publisher, © 2012 by tpack.org).
Figure 1. TPACK framework (reproduced by permission of the publisher, © 2012 by tpack.org).
Education 13 00732 g001
Figure 2. Design of the study and data collection.
Figure 2. Design of the study and data collection.
Education 13 00732 g002
Figure 3. TPACK on action development for pre-, re- and post-test. Significant differences are marked with asterisks (*: p < 0.050; **: p < 0.010; ***: p < 0.001).
Figure 3. TPACK on action development for pre-, re- and post-test. Significant differences are marked with asterisks (*: p < 0.050; **: p < 0.010; ***: p < 0.001).
Education 13 00732 g003
Figure 4. Boxplots for pre-, re- and post-test. Significant differences are marked with asterisks (*: p < 0.050; **: p < 0.010; ***: p < 0.001).
Figure 4. Boxplots for pre-, re- and post-test. Significant differences are marked with asterisks (*: p < 0.050; **: p < 0.010; ***: p < 0.001).
Education 13 00732 g004
Figure 5. Quality of technology exploitation related based on the SAMR and ICAP frameworks. The size of the respective bubble indicates the number of categorized technology exploitations for each combination of SAMR and ICAP, ranging from 1 to 26 scenarios.
Figure 5. Quality of technology exploitation related based on the SAMR and ICAP frameworks. The size of the respective bubble indicates the number of categorized technology exploitations for each combination of SAMR and ICAP, ranging from 1 to 26 scenarios.
Education 13 00732 g005
Figure 6. TPACK in action (M ± SD) derived by lesson plans for each group. Combinations indicating low TPACK levels are located in the lower left, high levels in the upper right corner.
Figure 6. TPACK in action (M ± SD) derived by lesson plans for each group. Combinations indicating low TPACK levels are located in the lower left, high levels in the upper right corner.
Education 13 00732 g006
Figure 7. Results from standardized teaching evaluation form (TEF) of the modules ‘Teaching and Learning Science’ (left, N = 50) and ‘Science in Everyday Life’ (right, N = 24). Individual item formulations are given in Table 6.
Figure 7. Results from standardized teaching evaluation form (TEF) of the modules ‘Teaching and Learning Science’ (left, N = 50) and ‘Science in Everyday Life’ (right, N = 24). Individual item formulations are given in Table 6.
Education 13 00732 g007
Table 1. Levels of the SAMR and ICAP frameworks to describe the quality of technology exploitation in practice.
Table 1. Levels of the SAMR and ICAP frameworks to describe the quality of technology exploitation in practice.
SAMR Framework [67]ICAP Framework [69]
RedefinitionTechnology allows for design of novel learning tasks, previously inconceivable, e.g., students use interactive simulations to explore the movement of tectonic plates and resulting earthquakes.InteractiveStudents engage constructively in an equal community of learners, e.g., students explore Avogadro’s law, Boyl’s law, and Gay-Lussac’s law, in groups, (use technology to) document their results (in a shared file) and combine their findings to derive the ideal gas law collaboratively.
ModificationTechnology allows for a significant redesign of the task, e.g., the use of a virtual bulletin board to create a multimodal experimental protocol including pictures, videos or audios.ConstructiveStudents generate new inferences or information, which go beyond the presented materials, e.g., students create their own explanatory videos or e-books to summarize experimental findings in virtual lab reports.
AugmentationTechnology is used as a substitute for traditional media with functional improvement, e.g., students read a text on a tablet computer with embedded hyperlinks to retain more information.ActiveStudents manipulate (technology enriched) learning materials or activities, e.g., students use the play/pause/zoom function to observe video scenes of a pre-recorded experiment in more detail.
SubstitutionTechnology is used as a substitute for traditional media with no functional improvement, e.g., students read a plain text on a tablet computer.PassiveStudents receive information without showing any indicator of interaction, e.g., they read a scientific text or watch a video.
Table 2. Content and digital supplements of the module ‘Teaching and Learning Science’.
Table 2. Content and digital supplements of the module ‘Teaching and Learning Science’.
SessionContentDigital Supplement with Practical Deepening/Application in the Complementary Seminar
1IntroductionPre-test
2Digital mediaGlossary: digital media for science education
3InclusionUniversal design for learning in digital media and assistive operating aids
4DiagnosticsQuiz apps
5DifferentiationQR tip cards
6Materials and tasksE-books and learning apps
7LanguageMultimodal design of interactive work sheets
8ExplanationsExplanatory videos
9Nature of scienceNature of science in society [77]
10ModelsModeling software and virtual and augmented reality
11Students’ beliefsConcept cartoons, mind maps and concept maps
12ExperimentsInternal and external sensors, interactive simulations and remote labs
13Feedback and assessmentClassroom/student response systems
Table 3. Sample of this study.
Table 3. Sample of this study.
N
(Pre/Re/Post)
133
(102/76/55)
Gender
(%)
MaleFemaleDiverseNo answer
12.064.70.023.3
Age
(years)
MSD
23.44.68
Subjects
(%)
Science (primary)Biology (2nd)Chemistry (2nd)No answer
34.635.312.823.3
Teaching experience
(%)
None1–10 h11–30 h>30 hNo answer
28.832.67.67.623.5
Table 4. Sample items for self-rated TPACK on action and behavioral orientations toward the use of technology in science education.
Table 4. Sample items for self-rated TPACK on action and behavioral orientations toward the use of technology in science education.
ScaleSample Item
AttitudesWith technology, I can plan and adapt science lessons more appropriately for my students.
Subjective NormsThe science curriculum requires the use of technology.
ConstraintsThe long preparation time often prevents me from using technology in my science lessons.
Self-EfficacyI can use technology to get feedback from my students on science lessons.
Motivational OrientationI enjoy thinking about how to use technology in the science classroom.
CKI have sufficient knowledge about my science subject.
PKI am able to stretch my students’ thinking by creating challenging tasks for them.
TKI have the technical skills to use technology effectively.
PCKI can help my students to understand my science subject in various ways.
TCKI can use software that is created specifically for my science subject.
TPKI am able to use technology to introduce my students to real-world scenarios.
TPACKI can formulate in-depth discussion topics about my science subject and facilitate students’ online collaboration with appropriate technology.
Table 5. Rubrics for quality of technology use in lesson plans [70].
Table 5. Rubrics for quality of technology use in lesson plans [70].
ICAP frameworkInteractiveStudents engage constructively in an equal community of learners
AND
Technology as substitute without functional improvement
Students engage constructively in an equal community of learners
AND
Technology as substitute with functional improvement
Students engage constructively in an equal community of learners
AND
Technology allows for a significant redesign of the task
Students engage constructively in an equal community of learners
AND
Technology allows for design of novel learning tasks, previously inconceivable
ConstructiveStudents generate new inferences or information, which go beyond the presented materials AND
Technology as substitute without functional improvement
Students generate new inferences or information, which go beyond the presented materials AND
Technology as substitute with functional improvement
Students generate new inferences or information, which go beyond the presented materials AND
Technology allows for a significant redesign of the task
Students generate new inferences or information, which go beyond the presented materials AND
Technology allows for design of novel learning tasks, previously inconceivable
ActiveStudents manipulate learning materials or activities
AND
Technology as substitute without functional improvement
Students manipulate learning materials or activities
AND
Technology as substitute with functional improvement
Students manipulate learning materials or activities
AND
Technology allows for a significant redesign of the task
Students manipulate learning materials or activities
AND
Technology allows for design of novel learning tasks, previously inconceivable
PassiveStudents receive information; no indicator of interaction
AND
Technology as substitute without functional improvement
Students receive information; no indicator of interaction
AND
Technology as substitute with functional improvement
Students receive information; no indicator of interaction
AND
Technology allows for a significant redesign of the task
Students receive information; no indicator of interaction
AND
Technology allows for design of novel learning tasks, previously inconceivable
SubstitutionAugmentationModificationRedefinition
SAMR framework
Table 6. Extract from standardized teaching evaluation form (TEF) of the modules. Only course-specific items addressing digital supplements/reminders and practical application of technology are presented.
Table 6. Extract from standardized teaching evaluation form (TEF) of the modules. Only course-specific items addressing digital supplements/reminders and practical application of technology are presented.
Module ‘Teaching and Learning Science’ incl. Digital SupplementsModule ‘Science in Everyday Life’ incl.
Digital Reminders
1The number of technologies presented in the digital supplements was sufficient.-
2The digital supplements were sufficiently focused on science education.-
3The seminar offers me sufficient opportunities to deepen my knowledge on technology exploitation into practice.Planning and teaching an IBL lesson offers me sufficient opportunities to deepen my knowledge on technology exploitation into practice.
4-The digital reminders helped me to choose technology while planning and teaching a science lesson.
5-It was easy for me to choose technology while planning and teaching a science lesson.
6Technology exploitation (as presented in the digital supplements) can be implemented into school practice.Technology exploitation (as presented in the digital reminders) can be implemented into school practice.
7Technology exploitation (as presented in the digital supplements) enhances the quality of science teaching and learning.Technology exploitation (as presented in the digital reminders) enhances the quality of science lessons.
8Through the digital supplements, I was able to expand my professional knowledge in the area of technology exploitation.By planning and teaching a science lesson, I was able to expand my professional knowledge in the area of technology exploitation.
Table 7. Self-rated TPACK on action before, during and after attending the course and comparison of first and second, second and third, as well as first and third measurement points (last row), respectively. Abbreviations in order of appearance: number N, mean M, standard deviation SD, t-value t, p-value p, effect size d, z-value z, correlation coefficient r, and degrees of freedom df.
Table 7. Self-rated TPACK on action before, during and after attending the course and comparison of first and second, second and third, as well as first and third measurement points (last row), respectively. Abbreviations in order of appearance: number N, mean M, standard deviation SD, t-value t, p-value p, effect size d, z-value z, correlation coefficient r, and degrees of freedom df.
NMSDt-TestWilcoxon Test
tpdzprdf
CKPre1023.470.70
Re763.560.67 3.020.0030.4250
Post553.950.52−2.390.0220.37 40
4.20<0.0010.6345
PKPre1023.530.73
Re763.680.61 2.730.0060.3850
Post553.940.58 3.69<0.0010.5840
5.05<0.0010.7545
TKPre1023.290.82
Re763.300.79 2.020.0430.2850
Post553.770.59 3.100.0020.4840
4.92<0.0010.7345
PCKPre1023.430.67
Re763.710.58 4.92<0.0010.6950
Post554.030.53 3.48<0.0010.5440
5.01<0.0010.7645
TCKpre1023.220.68
re763.490.62 2.480.0130.3550
post543.910.57 3.49<0.0010.5540
5.19<0.0010.7745
TPKPre1023.690.72
Re763.710.58 1.190.2350.1750
Post554.240.56 4.44<0.0010.6940
4.30<0.0010.6445
TPACKPre1013.200.84
Re763.430.60 2.430.0150.3449
Post553.960.52 3.92<0.0010.6140
5.29<0.0010.8044
Table 8. Self-rated behavioral orientations according to the ToPB before, during and after attending the course and comparison of first and second, second and third, as well as first and third, measurement points (last row), respectively. Abbreviations in order of appearance: number, N; mean, M; standard deviation, SD; t-value, t; p-value, p; effect size, d; z-value, z; correlation coefficient, r; and degrees of freedom, df.
Table 8. Self-rated behavioral orientations according to the ToPB before, during and after attending the course and comparison of first and second, second and third, as well as first and third, measurement points (last row), respectively. Abbreviations in order of appearance: number, N; mean, M; standard deviation, SD; t-value, t; p-value, p; effect size, d; z-value, z; correlation coefficient, r; and degrees of freedom, df.
NMSDt-TestWilcoxon Test
tpdzprdf
AttitudesPre1023.690.66
Re763.840.57 1.420.1540.2050
Post554.150.58 3.77<0.0010.5940
3.54<0.0010.5345
Subjective NormsPre1023.430.61
Re763.410.47 −0.270.7900.0450
Post553.660.59 2.060.0390.3240
1.590.1130.2445
ConstraintsPre1022.930.74
Re763.030.71 0.410.6840.0650
Post552.510.70 −3.70<0.0010.5840
−3.040.0020.4545
Self-efficacyPre1022.690.65
Re763.020.57−5.18<0.0010.73
Post553.310.50−3.060.0040.48
−9.13<0.0011.36 45
Motivational OrientationPre1023.270.80
Re763.270.75−1.160.2520.16 50
Post543.660.72 2.430.0150.3840
3.52<0.0010.5245
Table 9. Quantity of technology exploitation in lesson plans distinguished by type of technology, and user.
Table 9. Quantity of technology exploitation in lesson plans distinguished by type of technology, and user.
Technology TypeUser = TeacherUser = StudentsUser = UnknownTotal
Virtual Bulletin Boards1638155
(Explanatory) Videos168024
Presentation Soft- and Hardware119323
Office Applications (Text and Spreadsheet) 415221
Response Systems and Quiz Apps612119
QR Tip Cards011011
Graphic Soft- and Hardware0808
eBooks0707
Virtual Reality Applications1506
Measurement Acquisition0404
Other2305
TOTAL561207183
Table 10. Comparison of pre-service teachers’ self-rated TPACK on action and TPACK in action derived from lesson plans after attending both modules. TPACK in action is approximated by the arithmetic mean of TPK and PCK for each coded technology exploitation.
Table 10. Comparison of pre-service teachers’ self-rated TPACK on action and TPACK in action derived from lesson plans after attending both modules. TPACK in action is approximated by the arithmetic mean of TPK and PCK for each coded technology exploitation.
On Action
(Questionnaire, 5-Point Likert Scale)
In Action
(Lesson Plans, Converted to 5-Point Scale)
NMSDNMSD M
PCK554.030.53313.481.190.55
TPK554.240.56312.741.001.50
TPACK553.960.52313.111.100.85
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Stinken-Rösner, L.; Hofer, E.; Rodenhauser, A.; Abels, S. Technology Implementation in Pre-Service Science Teacher Education Based on the Transformative View of TPACK: Effects on Pre-Service Teachers’ TPACK, Behavioral Orientations and Actions in Practice. Educ. Sci. 2023, 13, 732. https://doi.org/10.3390/educsci13070732

AMA Style

Stinken-Rösner L, Hofer E, Rodenhauser A, Abels S. Technology Implementation in Pre-Service Science Teacher Education Based on the Transformative View of TPACK: Effects on Pre-Service Teachers’ TPACK, Behavioral Orientations and Actions in Practice. Education Sciences. 2023; 13(7):732. https://doi.org/10.3390/educsci13070732

Chicago/Turabian Style

Stinken-Rösner, Lisa, Elisabeth Hofer, Annika Rodenhauser, and Simone Abels. 2023. "Technology Implementation in Pre-Service Science Teacher Education Based on the Transformative View of TPACK: Effects on Pre-Service Teachers’ TPACK, Behavioral Orientations and Actions in Practice" Education Sciences 13, no. 7: 732. https://doi.org/10.3390/educsci13070732

APA Style

Stinken-Rösner, L., Hofer, E., Rodenhauser, A., & Abels, S. (2023). Technology Implementation in Pre-Service Science Teacher Education Based on the Transformative View of TPACK: Effects on Pre-Service Teachers’ TPACK, Behavioral Orientations and Actions in Practice. Education Sciences, 13(7), 732. https://doi.org/10.3390/educsci13070732

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop