Next Article in Journal
Romantic Relationship Quality and Eating Disorder Symptoms in Late Pregnancy: The Serial Mediating Role of Depression and Body Dissatisfaction
Previous Article in Journal
Evolution of Life Satisfaction Throughout the Gestation Process and at Different Postpartum Stages
Previous Article in Special Issue
Teacher Cognition and Practices in Using Generative AI Tools to Support Student Engagement in EFL Higher-Education Contexts
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Effects of Interaction Scenarios on EFL Learners’ Technology Acceptance and Willingness to Communicate with AI

1
School of Foreign Studies, Northeastern University, Shenyang 110819, China
2
School of English for Specific Purposes, Beijing Foreign Studies University, Beijing 100089, China
3
National Research Centre for Foreign Language Education, Beijing Foreign Studies University, Beijing 100089, China
*
Author to whom correspondence should be addressed.
Behav. Sci. 2025, 15(10), 1391; https://doi.org/10.3390/bs15101391
Submission received: 23 August 2025 / Revised: 9 October 2025 / Accepted: 11 October 2025 / Published: 14 October 2025

Abstract

Grounded in a sociocultural theory, this study investigates how distinct interaction scenarios influence Chinese English as a Foreign Language (EFL) learners’ technology acceptance: perceived usefulness (PU) and perceived ease of use (PEU), and their willingness to communicate with AI (AI-WTC). A total of 367 university students completed a scenario-based questionnaire measuring PU, PEU, and AI-WTC across four empirically derived scenarios: advisory interaction, language skills support, academic knowledge inquiry, and factual information retrieval. Repeated-measures ANOVA with Bonferroni tests revealed significant scenario effects on all three constructs, though effect sizes were small to moderate. Factual Information Retrieval Scenario consistently received the highest ratings, whereas Academic Knowledge Inquiry and Language Skills Support Scenario scored lowest. A salient divergence emerged in complex scenarios: Advisory Interaction Scenario was rated more useful than Language Skills Support Scenario, yet both elicited equally low willingness to communicate, indicating that perceived usefulness alone may not sustain engagement under high interactional demands. These findings suggest that the effectiveness of AI as a communicative scaffold is not inherent but co-constructed through scenario-specific affordances and constraints. The study contributes a scenario-sensitive framework to TAM and WTC research, providing pedagogical guidance for designing differentiated AI-mediated language tasks.

1. Introduction

The rapid integration of generative artificial intelligence (GenAI) into educational landscapes has marked a pivotal moment for foreign language pedagogy (Feng et al., 2025; Kohnke et al., 2023; Chiu et al., 2023). Advanced AI tools, such as ChatGPT and other large language models, are no longer mere supplementary instruments but are increasingly positioned as active participants in the learning process. This paradigm shift prompts a reconceptualization of learner–computer interaction (M. Liu et al., 2024, 2025; Rahimi, 2025), wherein, from a sociocultural perspective, GenAI transcends its role as a simple tool to become a “communicative agent” (Fauzi et al., 2023; Wang et al., 2024). This emerging dynamic presents both unprecedented opportunities and notable challenges for second language (L2) education (AlAfnan et al., 2023; Karataş et al., 2024; Apriani et al., 2024). While GenAI offers personalized, on-demand interaction that can mitigate foreign language anxiety and provide extensive output practice, its pedagogical effectiveness is not guaranteed, as empirical findings remain mixed (Fathi et al., 2024; Huang & Zou, 2024; Zhang et al., 2024). The variability in outcomes underscores that AI’s potential is not self-actualizing but is contingent upon learners’ psychological and behavioral engagement. Consequently, the extent to which learners accept this technology (Technology Acceptance) and their willingness to communicate with a non-human agent (WTC) emerge as critical factors that will ultimately determine the instructional value of GenAI (Davis, 1989; Zawacki-Richter et al., 2019).
The Technology Acceptance Model (TAM) has been the cornerstone of research on user adoption of new technologies for over three decades (Al-Emran et al., 2018). Developed by Davis (1989), the original TAM proposed that a user’s intention to adopt a technology is primarily determined by two key beliefs: perceived usefulness (PU) and perceived ease of use (PEU). As Davis (1989, p. 320) defines them, PU refers to “the degree to which a person believes that using a particular system would enhance his or her job performance,” while PEU is “the degree to which a person believes that using a particular system would be free of effort.” These two factors influence a user’s attitude toward using the system, which in turn predicts their behavioral intention and actual use. When AI functions as a communicative agent, PU and PEU are no longer confined to the tool’s interface or functional output; they become intricately linked to the perceived quality of the interaction itself (Strzelecki, 2023). Recent studies applying TAM to ChatGPT in educational contexts have confirmed the continued importance of PU and PEU in shaping learners’ acceptance of AI-based tools (M. Liu et al., 2024; G. Liu & Ma, 2024; Kim & Su, 2024; Ouyang et al., 2024; Wang et al., 2024).
Beyond technology adoption, learners’ communicative engagement with AI also hinges on affective and psychological readiness, as explained by Willingness to Communicate (WTC) theory. WTC describes an individual’s readiness to engage in discourse (McCroskey & Baer, 1985; MacIntyre et al., 1998) and has been widely applied in second language acquisition research. In the L2 context, WTC is recognized not merely as a stable trait but as a dynamic, situation-specific construct that is susceptible to a complex interplay of linguistic, affective, and situational variables (Nematizadeh & Wood, 2021). With the rise of digital technologies, L2 WTC research has expanded into computer-mediated communication (Lee, 2019). Early findings not only validate the applicability of the WTC framework to AI contexts but are also refining its contours, revealing critical distinctions; while AI can offer a low-anxiety space (Lee & Drajati, 2019, 2020; Xie et al., 2025), learners’ willingness to engage is profoundly shaped by their perception of the AI’s competence, reliability, and the specific purpose of the interaction (Fathi et al., 2024; Wang et al., 2024). This evolving line of inquiry not only demonstrates the adaptability of the WTC framework to AI-mediated ecologies but also highlights that AI-WTC is a unique construct sensitive to technological and contextual affordances.
Although TA and AI-WTC have been explored in related strands of research, studies that examine both constructs within the same AI interaction scenario remain limited. This gap is compounded by a tendency in prior research to treat AI as a monolithic entity, thereby overlooking the fundamental ways in which specific interaction scenarios shape learner perceptions and behaviors. This limitation points to the need for a more differentiated analytical framework that can capture the variability of AI interaction scenarios, a perspective reinforced by emerging studies showing that learners engage with AI in distinct, role-based capacities (Peng & Liang, 2025). In this light, a scenario-sensitive perspective offers a promising direction for understanding how PU, PEU and AI-WTC interact in more nuanced ways.
In summary, the existing literature reveals three critical gaps that this study aims to fill. First, while WTC theory is being extended to AI, research has not yet systematically examined how different types of interaction with AI influence learners’ communicative volition. Second, while the TA model is being applied to GenAI, studies have largely overlooked how the purpose of use (i.e., the scenario) dynamically shapes learners’ perceptions of usefulness and ease of use. Third, and most importantly, existing research on learner–AI interaction, though expanding, tends to conceptualize it as a uniform activity. Most studies examine “the effect of ChatGPT” or “the impact of AI” (e.g., Fathi et al., 2024; G. Liu & Ma, 2024) without differentiating how distinct purposes or scenarios of interaction may shape learning outcomes (Wu & Zhang, 2017). This assumption of uniformity obscures the contextual nature of AI-mediated communication and hinders the formulation of targeted pedagogical strategies.
Drawing on sociocultural theory, this study focuses on Chinese EFL learners, examining how scenarios shape their technology acceptance and willingness to communicate with AI. Scenarios define the communicative event by shaping the roles, expectations, and goals assigned to the learner and the AI, thereby influencing both technology acceptance and communicative engagement. Within this framework, learners’ technology acceptance, operationalized as PU and PEU, is expected to vary across four pedagogically relevant interaction scenarios: advisory interaction (AIS), language skills support (LSSS), academic knowledge inquiry (AKIS), and factual information retrieval (FIRS). Similarly, learners’ AI-WTC is theorized to be scenario-sensitive, reflecting the extent to which contextual affordances lower interactional barriers and foster readiness for dialogue. Finally, we examine the extent to which scenario effects on TA correspond to scenario effects on AI-WTC, thereby revealing potential interconnections between learners’ cognitive evaluations of AI and their communicative volition (See Figure 1 for the conceptual model).
Guided by this framework, the study addresses three research questions:
(1) To what extent do interaction scenarios affect learners’ technology acceptance (PU and PEU)?
(2) To what extent do interaction scenarios affect learners’ willingness to communicate with AI (AI-WTC)?
(3) What are the potential interconnections between the effects of interaction scenarios on technology acceptance and WTC?
Based on this, we propose the following hypotheses:
H1. 
Learners’ technology acceptance, measured as PU and PEU, will significantly differ across the four interaction scenarios: Advisory Interaction Scenario, Language Skills Support Scenario, Academic Knowledge Inquiry Scenario, and Factual Information Retrieval Scenario.
H2. 
Learners’ willingness to communicate with AI (AI-WTC) will significantly vary across the four scenarios.
H3. 
Scenario-based differences in technology acceptance (PU and PEU) will correspond to scenario-based differences in AI-WTC.

2. Materials and Methods

2.1. Participants and Procedures

The participants were 367 Chinese university students recruited from two key universities in northern China. The sample comprised both undergraduate and graduate students, with a mean age of 19.93 years. In terms of academic background, 71 students (19.35%) majored in English, 79 (21.53%) in other foreign languages, and 217 (59.13%) in non-language disciplines. Regarding academic level, 233 (63.49%) were first-year undergraduates, 61 (16.62%) were second-year, 33 (8.99%) were third-year, 10 (2.72%) were in their final undergraduate year, 17 (4.63%) were master’s students, and 1 participant (0.27%) was a doctoral student. All participants had prior experience using generative AI tools (e.g., ChatGPT and Wenxin Yiyan) for auxiliary language learning purposes, as the use of such technology is increasingly integrated into the general academic environment at their institutions. The study specifically focused on the context of learning English as a foreign language, and the diverse background ensured a wide range of English proficiency levels and learning needs, thereby enhancing the representativeness of the findings for the broader EFL learner population in Chinese higher education.
Given the lack of existing instruments measuring technology acceptance and willingness to communicate with AI across context-specific interaction scenarios, the present study developed a purpose-designed questionnaire through a multi-phase process. For the TA scale, we adopted Davis’s (1989) Technology Acceptance Model (TAM) as the theoretical framework, adapting items from the original TAM questionnaire (Davis, 1989) to fit AI-mediated language learning contexts. The TA scale comprised two subscales: Perceived Usefulness, with six items assessing learners’ beliefs about AI’s effectiveness in enhancing learning outcomes, and Perceived Ease of Use (PEU), with six items evaluating the accessibility of AI interactions. All items were designed to be scenario-adaptive, allowing for later contextualization within specific interaction scenarios. The AI-WTC scale was adapted from Peng and Woodrow’s (2010) classroom-based WTC questionnaire, which highlights the interplay between individual and contextual variables. Building on this foundation, the present study extended the individual dimension to include learners’ technology acceptance alongside their AI-WTC. With communicative engagement, AI was understood as co-constructed through the interaction between learners’ perceptions and scenario-specific affordances. To capture this, the AI-WTC scale employed an “I am willing to …” structure across four scenarios, with six items per scenario developed in parallel with the TA scale to ensure measurement consistency.
With the theoretical space for TA and WTC specified, we adopted a theory-driven, empirically calibrated approach to scenario construction. Drawing a priori on Peng and Liang’s (2025) taxonomy of AI roles in L2 learning, we operationalized four interaction scenarios: Advisory Interaction Scenario, Language Skills Support Scenario, Academic Knowledge Inquiry Scenario, and Factual Information Retrieval Scenario, to delineate the content domain in which AI assumes distinct social-interactional roles. To develop ecologically valid scenarios for the subsequent scale, we first administered an open-ended questionnaire to a separate, independent sample of 67 university-level EFL learners (Hinkin, 1998), eliciting detailed accounts of their actual and envisioned AI-supported learning activities (938 words). The qualitative data underwent a three-phase thematic analysis involving initial open coding to identify instances of AI use, categorization of codes into broader themes, and consolidation into four overarching scenarios. This process yielded 130 concrete interaction instances. Six representative tasks were selected per scenario based on frequency and pedagogical salience (DeVellis & Thorpe, 2021), with each task either reflecting experiences frequently reported in the open-ended responses or clearly described with sufficient context to ensure participants’ familiarity, so that students could meaningfully evaluate them even if they had not previously performed a specific task.
Building upon the empirically validated scenario framework, we systematically integrated the 24 interaction tasks into the measurement instrument by contextualizing all questionnaire items for technology acceptance (PU/PEU) and AI-WTC within specific scenario parameters. Each of the six tasks per scenario type was carefully embedded into parallel item sets for both constructs, ensuring that respondents evaluated identical interaction contexts across all measures. This dual contextualization process yielded a 72-item instrument (24 scenarios × 3 constructs), enabling direct comparisons of how specific scenario characteristics influence both technology acceptance and communicative willingness.
To ensure that participants could respond without language barriers, the original English questionnaire was collaboratively translated into Chinese by the research team. Semantic and cultural equivalence was maintained through iterative reviews (Brislin, 1970). The preliminary Chinese version was then administered to 40 first-year undergraduate students, who provided feedback on item clarity and relevance. Subsequently, two experts in applied linguistics and educational technology reviewed the revised instrument, offering suggestions to further enhance construct validity and alignment with the intended theoretical dimensions (DeVellis, 2017). Based on this feedback, the final version of the questionnaire was produced.
The finalized survey was distributed online via Wenjuanxing (https://www.wjx.cn/), a widely used data collection platform in China. A simple random sampling method was employed, and data were collected from 13 July to 24 July 2025. Ethical considerations were strictly observed: participants were informed of the study’s purpose, assured of voluntary participation, and notified of their right to withdraw at any time. No personally identifiable information was collected, and submission of the completed survey was considered informed consent. On average, respondents completed the questionnaire in approximately 6 min.

2.2. Instrument

The instrument of this study was an online survey comprising two main sections. The first section collected participants’ demographic information, including gender, age, academic major, and year or level of study (based on the 2024–2025 academic calendar). This information was used to describe the sample and control for potential background variability.
The second section consisted of two scenario-based scales designed to assess learners’ perceptions of AI and their communicative willingness with it. Four distinct AI interaction scenarios were identified to reflect the range of communicative events in which second language learners may engage with AI tools. The four empirically derived interaction scenarios are defined as follows: the Advisory Interaction Scenario involves seeking personalized advice or strategic feedback; the Language Skills Support Scenario focuses on practicing or refining specific linguistic abilities; the Academic Knowledge Inquiry Scenario entails accessing explanatory theoretical or disciplinary knowledge; and the Factual Information Retrieval Scenario concerns retrieving concise, referential information. These scenarios form the contextual foundation for measuring TA and AI-WTC. The first scale targeted TA, measuring two constructs across four interaction scenarios: Perceived Usefulness, referring to the extent to which learners believe that using AI tools would enhance their foreign language learning outcomes. Perceived Ease of Use (PEU) refers to the degree to which learners perceive AI tools as easy and accessible for learning-related tasks. The second scale measured learners’ AI-WTC across four scenarios. In this study, AI-WTC was reconceptualized to capture learners’ context-sensitive readiness to initiate or engage in communicative interactions with AI tools in the context of L2 learning.
The structure and item distribution of the scenario-based questionnaire are presented in Table 1. All items were measured using a five-point Likert scale, ranging from 1 (strongly disagree) to 5 (strongly agree). The full list of items, including construct-scenario mappings, is provided in Appendix A.
Measurement reliability and construct validity were assessed using AMOS 26, where all standardized factor loadings exceeded 0.6 and were statistically significant (p < 0.05), indicating strong associations between each item and its respective construct and a well-fitting measurement model. Composite reliability (CR) values ranged from 0.897 to 0.955, exceeding the recommended threshold of 0.70 and demonstrating strong internal consistency. Average variance extracted (AVE) values ranged from 0.593 to 0.779, surpassing the minimum requirement of 0.50 and supporting convergent validity (see Appendix B, Table A1). Inter-factor correlations, calculated using SPSS 26, ranged from 0.70 to 0.91 between PU and PEU, and from 0.72 to 0.88 between TA and AI-WTC (see Appendix B, Table A2). Although correlations were relatively high, this pattern aligns with theoretical expectations given the close conceptual relationships among constructs. Overall, these results indicate that the instrument provides a reliable and valid basis (Nunnally & Bernstein, 1994) for examining technology acceptance and willingness to communicate with AI.

2.3. Data Analysis

A quantitative research approach was adopted to examine how learners’ technology acceptance and willingness to communicate with AI tools vary across distinct interaction scenarios in L2 learning. All statistical analyses were conducted using IBM SPSS Statistics (version 26.0). The data analysis proceeded in three stages, each aligned with one or more of the study’s research questions.
To address the first and second research questions, namely, whether learners’ PU, PEU, and AI-WTC differ significantly across distinct interaction scenarios, a repeated measures analysis of variance (RM-ANOVA) was employed (Field, 2018; Maxwell et al., 2018). Separate RM-ANOVAs were conducted for each of the three constructs, as they were all measured repeatedly across the same four within-subject conditions. This sequential analytical approach enabled the detection of within-person variability while controlling for individual differences in baseline responses, which is a key strength of within-subject designs (Gravetter & Wallnau, 2016).
Following this, to answer the third research question, whether scenario-based differences in TA align with variations in AI-WTC, post hoc pairwise comparisons were performed using the Bonferroni adjustment method (Abdi, 2007). This conservative correction was used to minimize the risk of inflated Type I error associated with multiple comparisons. The aim was to identify potential convergence or divergence between learners’ evaluations of AI and their willingness to engage with AI across different interactional contexts.

3. Results

3.1. Scenario-Based Differences in Technology Acceptance (PU and PEU)

To examine whether learners’ technology acceptance, operationalized as Perceived Usefulness and Perceived Ease of Use, varied significantly across the four AI interaction scenarios, a within-subjects design was employed, as the same group of participants provided ratings for each construct under all four scenario conditions: Advisory Interaction Scenario, Language Skills Support Scenario, Academic Knowledge Inquiry Scenario, and Factual Information Retrieval Scenario. Repeated Measures Analysis of Variance (RM-ANOVA) was conducted using IBM SPSS Statistics (version 26.0) to account for the correlated nature of the data and to assess within-subject variability while controlling for individual differences in baseline perceptions.
The results, presented in Table 2, revealed a statistically significant main effect of interaction scenario on PU, F(3, 1098) = 16.096, p < 0.001, η2 = 0.042, and on PEU, F(3, 1098) = 24.641, p < 0.001, η2 = 0.063. According to Cohen’s (1988) benchmarks, these effect sizes indicate small-to-moderate effects, suggesting that while the differences are statistically robust, scenario-related differences in technology acceptance account for only a modest proportion of the total variance. This indicates that learners’ perceptions of AI’s usefulness and ease of use varied across scenarios.

3.2. Scenario-Based Differences in Willingness to Communicate with AI

To examine potential differences in learners’ willingness to communicate with AI across the four interaction scenarios, a repeated measures analysis of variance (RM-ANOVA) was performed. This approach enabled comparison of WTC scores for the same participants in the Advisory Interaction Scenario, Language Skills Support Scenario, Academic Knowledge Inquiry Scenario, and Factual Information Retrieval Scenario, while accounting for individual variability.
As shown in Table 3, the analysis identified a significant main effect of interaction scenario on AI-WTC, F(3, 1098) = 9.600, p < 0.001, η2 = 0.026. Interpreted against Cohen’s (1988) guidelines, this effect size is small, indicating that scenario type exerts a measurable yet limited influence on learners’ communicative willingness with AI. In other words, although learners’ AI-WTC scores varied across scenarios, the extent of these differences represented only a modest share of the overall variance.

3.3. Interconnections Between Technology Acceptance and AI-WTC Across Scenarios

To further identify how TA and AI-WTC were interrelated across the four interaction scenarios, Bonferroni-adjusted post hoc pairwise comparisons were conducted following the RM-ANOVA results. This procedure allowed for the identification of specific scenario pairs that differed significantly, while controlling for the inflated Type I error rate associated with multiple comparisons (Abdi, 2007).
As displayed in Table 4, the post hoc analyses revealed variations in scenario effects across the constructs. For PU, the overall effect of scenario type was significant, F(1) = 18.823, p < 0.001, η2 = 0.049, indicating that a small proportion of the variance in perceptions of usefulness was accounted for by the scenario type. Pairwise comparisons suggested a pattern where the Factual Information Retrieval Scenario received the highest ratings, significantly surpassing all others. The Academic Knowledge Inquiry Scenario followed, while the Language Skills Support Scenario received the lowest ratings. Notably, the Advisory Interaction Scenario was perceived as more useful than the Language Skills Support Scenario but less useful than the Factual Information Retrieval Scenario. For PEU, a significant main effect was observed with a moderate-to-large effect size, F(1) = 51.774, p < 0.001, η2 = 0.124, indicating that scenario type accounted for a more substantial portion of the variance in ease-of-use perceptions. Bonferroni comparisons indicated that the Factual Information Retrieval Scenario again received the highest ratings, significantly exceeding all other scenarios, while the Advisory Interaction Scenario consistently ranked lowest. The relative positioning of the Language Skills Support Scenario and the Academic Knowledge Inquiry Scenario (LSSS < AKIS < FIRS) suggested a potential gradient related to task complexity. For AI-WTC, a statistically significant main effect was observed, F(1) = 22.903, p < 0.001, η2 = 0.059, representing a small-to-moderate effect size where scenario type explained a limited portion of the variance in willingness to communicate. Post hoc comparisons indicated a pattern: the Factual Information Retrieval Scenario elicited the highest levels of AI-WTC, significantly surpassing other scenarios. The Academic Knowledge Inquiry Scenario occupied an intermediate position, generating significantly greater AI-WTC than both the Advisory Interaction Scenario and the Language Skills Support Scenario, between which no significant difference was found.
Collectively, these patterns indicate scenario-dependent variation, with the Factual Information Retrieval Scenario consistently rated highest and the Advisory Interaction Scenario (AIS) or Language Skills Support Scenario lowest. The nuanced divergences reveal several insights. First, the effect size for PEU (η2 = 0.124) was notably larger than for PU (η2 = 0.049) and AI-WTC (η2 = 0.059), suggesting that judgments of ease of use were more sensitive to contextual differences than those of usefulness or willingness to communicate. Second, the disparity between the Advisory Interaction Scenario and the Language Skills Support Scenario in PU (AIS > LSSS) versus their equivalence in AI-WTC (AIS ≈ LSSS) might tentatively suggest that perceived utility alone does not necessarily translate to a higher AI-WTC in complex interaction scenarios. Finally, the varying gap between the Factual Information Retrieval Scenario and the Academic Knowledge Inquiry Scenario across constructs highlights differential sensitivity to task structure. These findings collectively showcase that AI-mediated communication is shaped by, but not reducible to, traditional technology acceptance mechanisms.

4. Discussion

The present study aimed to investigate how distinct AI interaction scenarios influence EFL learners’ technology acceptance and their willingness to communicate. Regarding the first research question, the findings demonstrated that learners’ perceptions of usefulness and ease of use differed significantly across the four scenarios, thereby supporting Hypothesis 1. However, the effect sizes were small to moderate (Cohen, 1988), indicating that although scenario type exerted a reliable influence, it accounted for only a limited proportion of the overall variance. From a sociocultural perspective, this pattern suggests that learners’ evaluation of AI technology is not determined solely by the inherent affordances of the tool but also shaped by the situated communicative activity in which AI is embedded (Lantolf & Thorne, 2006; Vygotsky, 1978). The modest yet significant scenario effects point to the role of cultural factors in the Chinese learning context, where the inherent complexity of different tasks creates varying cognitive burdens that subsequently shape learners’ perceptions of the technology’s usefulness and ease of use.
With respect to the second research question, the results showcased that learners’ AI-WTC also varied across the four scenarios, thereby confirming Hypothesis 2. Similar to the findings for technology acceptance, the effect sizes were small to moderate, suggesting that although interaction scenarios exert a consistent influence, the variance they explain remains modest. This finding aligns with sociocultural perspectives that conceptualize WTC not as a fixed trait but as a context-sensitive construct, co-constructed through situated interaction and shaped by the cultural background embedded in each scenario (MacIntyre et al., 1998; Lantolf & Poehner, 2014).
The third research question examined the interconnections between scenario effects on technology acceptance and willingness to communicate with AI. The results showcase a clear hierarchy: the Factual Information Retrieval Scenario consistently secured the highest ratings across all constructs, closely followed by the Academic Knowledge Inquiry Scenario. This stable alignment indicates that for tasks with low complexity and a clear goal, akin to the structured exercises prevalent in Chinese foreign language education, AI’s scaffolding as a “more capable peer” (Vygotsky, 1978) operates effectively within the learner’s ZPD. The predictability and minimal communicative risk of these scenarios lower cognitive load, thereby supporting both positive technology acceptance and communicative engagement. In contrast, a more complex pattern emerges at the lower end of the spectrum. Although both the Advisory Interaction Scenario and the Language Skills Support Scenario received the poorest ratings, their internal profiles diverged: the former was perceived as more useful yet harder to use, while both ultimately led to equally low willingness to communicate. The modest effect sizes affirm that this dissociation, while systematic, is one of several factors at play. This pattern can be interpreted by examining the distinct challenges each scenario poses. The Language Skills Support Scenario demands high linguistic precision, creating a communicative risk of evaluation. Unlike a human teacher who can offer adaptive support, the AI’s standardized feedback fails to provide the nuanced scaffolding needed to mitigate this risk, thus falling short as an accessible “more capable peer” within the learner’s ZPD (Swain et al., 2015). The Advisory Interaction Scenario, conversely, imposes a high cognitive load from strategic judgment and open-ended information processing (Sweller & van Merrienboer, 2005). In both cases, the AI’s scaffolding is either perceived as inadequate for the task or too demanding to integrate effectively. Consequently, compared to the highly accessible support in the Factual Information Retrieval Scenario, these complex scenarios render the AI a less effective peer, suppressing willingness to communicate despite potential utility.
This study makes threefold contributions. At the theoretical level, it moves beyond a monolithic view of AI by introducing a scenario-sensitive perspective into the Technology Acceptance Model and Willingness to Communicate framework. It shows that the relationship between technology acceptance and communication willingness is not fixed but shifts according to the alignment between AI’s affordances and the demands of particular interaction scenarios. This reframing shifts attention from evaluating the technology in isolation to understanding how usage contexts shape learner perceptions and behaviors. By drawing on sociocultural theory, the study further conceptualizes AI as a “more capable peer” whose effectiveness depends on scenario-mediated scaffolding within the learner’s ZPD. Methodologically, the study develops and validates a scenario-based instrument that captures the nuances of human–AI interaction across diverse contexts, providing a resource for future empirical investigations. Practically, the findings offer educators an evidence-based framework for differentiated AI task design and sequencing, while also encouraging designers to build adaptive scaffolding mechanisms that lower interactional friction and promote learner engagement. Together, these contributions advance the understanding of AI-mediated communication by foregrounding the critical, yet underexplored, role of interaction scenarios.

5. Conclusions

Grounded in the sociocultural theory of viewing AI as a “more capable peer” (Vygotsky, 1978), this study quantitatively investigated how four distinct AI interaction scenarios shape Chinese EFL learners’ technology acceptance and their AI-WTC. The results demonstrated that scenario type exerts a significant, albeit modest, influence on both constructs. A generally aligned yet nuanced pattern was observed: while the Factual Information Retrieval Scenario was consistently rated highest across all measures, the relative positioning of the Advisory Interaction and Language Skills Support scenarios differed between constructs. A key finding was the critical divergence in complex scenarios: despite being rated as more useful than language practice, the advisory scenario triggered an equally low level of communicative willingness, suggesting that high perceived usefulness alone is insufficient to foster engagement when interactional demands are high.
The findings highlight that effective AI integration in language education requires intentional scenario design aligned with both technological capabilities and pedagogical goals. Importantly, since most AI systems were not originally designed as language teachers but rather as general-purpose conversational agents, expectations need to be recalibrated by positioning AI primarily as a scaffolding tool or a “more capable peer” rather than a comprehensive teacher substitute. This reframing acknowledges that AI’s core strengths lie in efficient information processing and structured support, whereas nuanced guidance and empathy remain the domain of human educators, especially in complex advisory roles. Furthermore, while the observed effects of scenario type were small to moderate, they nonetheless indicate that sequencing tasks from high to low AI–compatibility may be a prudent pedagogical strategy. Specifically, structured scenarios such as Factual Information Retrieval and Academic Knowledge Inquiry can serve as accessible entry points, leveraging AI’s informational strengths to build learner confidence through low-risk, high-success interactions. As learners gain proficiency, more open-ended scenarios such as Language Skills Support and Advisory Interaction can be gradually introduced. At this stage, explicit scaffolding becomes crucial: curated prompt templates (e.g., “Act as a critical friend and evaluate the thesis statement of my essay…”) and metacognitive training can guide learners to engage productively with AI while critically evaluating its feedback. For highly complex scenarios like Advisory Interaction, strategically permitting the use of the native language can serve as a transitional scaffold that reduces cognitive load, allowing learners to focus on complex advisory tasks before transitioning to the target language. Finally, hybrid human–AI models offer an effective balance, employing AI for initial support while reserving human oversight for higher-order tasks such as assessing argument coherence and cultural appropriateness. Framed through Vygotsky’s (1978) notion of scaffolding, this developmental sequencing illustrates how AI’s role can evolve from an authoritative knowledge resource to a collaborative partner, enabling educators to maximize its utility while mindfully addressing its limitations.
Several limitations in this study warrant consideration. First, the reliance on self-reported data without behavioral triangulation may not fully capture actual engagement, as responses could be influenced by social desirability bias. Second, the generalizability is limited by the Chinese student sample, as culturally specific factors such as the pronounced novelty effect and a tendency toward social desirability in educational settings may have uniquely shaped the responses. Third, the static measurement of technology acceptance and AI-WTC at a single time point overlooks how these perceptions might evolve with prolonged AI use. Future research could address these limitations through several avenues. First, behavioral data such as chat logs or AI interaction transcripts could be analyzed to triangulate learners’ self-reported perceptions with actual communicative behavior, offering richer evidence of engagement patterns. Second, cross-cultural comparisons involving learners from different sociocultural contexts would help clarify whether the observed scenario effects are culturally bound or generalizable across settings. Third, longitudinal studies could trace how learners’ technology acceptance and AI-WTC evolve over extended periods of interaction, revealing potential developmental trajectories of AI-mediated communication in L2 learning.

Author Contributions

Conceptualization, H.X. and H.Y.; methodology, Z.C.; formal analysis, H.X.; investigation, Z.C. and H.Y.; writing–original draft preparation, Z.C.; writing—review and editing, H.Y.; supervision, H.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Ethics Committee of the School of English for Specific Purposes, Beijing Foreign Studies University (protocol code 2025-0015 and date of approval 3 March 2025).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

A publicly available dataset was analyzed in this study. This data can be found here: https://pan.baidu.com/s/19g82JL2jYQDqc0ArEFip9g?pwd=data (accessed on 22 August 2025) (password: data).

Acknowledgments

During the preparation of this manuscript, the authors used ChatGPT 4.0 for the purpose of language polishing. The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
AI-WTCWillingness to Communicate with AI
AKISAcademic Knowledge Inquiry Scenario
GenAIGenerative Artificial Intelligence
L1First Language
L2Second Language
LSSSLanguage Skills Support Scenario
MKOMore Knowledgeable Other
PEUPerceived Ease of Use
PUPerceived Usefulness
RM-ANOVARepeated Measures Analysis of Variance
TATechnology Acceptance
TAMTechnology Acceptance Model
WTCWillingness to Communicate
ZPDZone of Proximal Development

Appendix A. Scenario-Based AI Interaction Questionnaire

All items were measured on a 5-point Likert scale, ranging from 1 (strongly disagree) to 5 (strongly agree).

Appendix A.1. Part 1—Scenario-Based AI Technology Acceptance Questionnaire

Appendix A.1.1. Advisory Interaction Scenario

Perceived Usefulness (PU)
  • Communicating with AI to obtain suggestions on selecting academic research topics would enable me to complete academic tasks more quickly.
  • Communicating with AI to obtain feedback on classroom presentations would improve my performance.
  • Communicating with AI to obtain writing-related advice would increase my productivity.
  • Communicating with AI to discuss research directions or ideas would enhance my effectiveness in academic writing.
  • Communicating with AI would make it easier to understand viewpoints or arguments.
  • I would find communicating with AI useful in personalizing learning strategies.
Perceived Ease of Use (PEU)
  • Learning to operate AI for selecting academic research topics would be easy for me.
  • I would find it easy to get AI to help me generate ideas for academic research.
  • My interaction with AI when seeking writing-related advice would be clear and understandable.
  • I would find AI to be flexible to interact with when asking for feedback on classroom presentations.
  • It would be easy for me to become skillful at communicating with AI to explore personalized learning strategies.
  • I would find AI easy to communicate viewpoints or arguments.

Appendix A.1.2. Language Skills Support Scenario

Perceived Usefulness (PU)
  • Communicating with AI in translation practice would enable me to accomplish language learning tasks more quickly.
  • Communicating with AI in speaking practice would improve my speaking performance.
  • Working with AI to revise and polish my writing would increase my productivity in writing tasks.
  • Interacting with AI to access and understand listening materials would enhance my effectiveness in listening comprehension.
  • Interacting with AI to explore the structure and meaning of texts would make it easier to complete reading tasks.
  • I would find communicating with AI useful in my grammar rules learning.
Perceived Ease of Use (PEU)
  • Learning to operate AI for translation practice would be easy for me.
  • I would find it easy to get AI to help me access and understand listening materials.
  • My interaction with AI for revising and polishing my writing would be clear and understandable.
  • I would find AI to be flexible to interact with in speaking practice.
  • It would be easy for me to become skillful at using AI to understand English texts and their structure.
  • I would find AI easy to interact with in learning grammar rules and explanations.

Appendix A.1.3. Academic Knowledge Inquiry Scenario

Perceived Usefulness (PU)
  • Interacting with AI in academic tasks would help me translate complex academic materials more quickly.
  • Interacting with AI to clarify difficult concepts or theories in academic courses would improve my class performance.
  • Interacting with AI during academic writing tasks would increase my productivity.
  • Working with AI would enhance my effectiveness in understanding academic theories or concepts.
  • Working with AI would make it easier to understand the academic arguments.
  • I would find communicating with AI to explore scholars’ key ideas and theories useful in my academic studies.
Perceived Ease of Use (PEU)
  • Learning to operate AI tools to translate complex materials would be easy for me.
  • I would find it easy to interact with AI to clarify difficult concepts or theories in academic courses.
  • My interaction with AI in understanding complex concepts and theories would be clear and understandable.
  • I would find AI to be flexible to interact with in completing academic writing tasks.
  • It would be easy for me to become skillful at interacting with AI to understand academic arguments.
  • I would find AI easy to interact with to explore scholars’ key ideas and theories.

Appendix A.1.4. Factual Information Retrieval Scenario

Perceived Usefulness (PU)
  • Using AI to quickly look up the meanings of unfamiliar English words would help me complete learning tasks more efficiently.
  • Using AI to confirm grammar structures or usage rules would help me use English more accurately.
  • Using AI to check standard definitions of academic terms would improve my learning efficiency.
  • Using AI to retrieve basic facts about well-known people would support my understanding of related materials.
  • Using AI to find out historical or cultural background information would make it easier to understand texts.
  • I would find it useful to use AI to check daily life knowledge.
Perceived Ease of Use (PEU)
  • Learning to operate AI tools to quickly look up the meaning of English words would be easy for me.
  • I would find it easy to get AI to confirm grammar structures or rules when I need help.
  • My interaction with AI to check definitions of academic concepts would be clear and understandable.
  • I would find AI to be flexible to interact with in retrieving basic information about famous people.
  • It would be easy for me to become skillful at using AI to find out historical or cultural background information.
  • I would find AI easy to use to check daily life knowledge.

Appendix A.2. Part 2—Scenario-Based AI-WTC Questionnaire

Appendix A.2.1. Advisory Interaction Scenario

  • I am willing to communicate with AI to obtain suggestions on choosing academic research topics.
  • I am willing to communicate with AI to improve the content or delivery of my class presentations.
  • I am willing to communicate with AI to receive feedback on my academic writing.
  • I am willing to communicate with AI to explore ideas or directions for academic research.
  • I am willing to communicate with AI to understand certain arguments or topics.
  • I am willing to communicate with AI to explore effective learning strategies.

Appendix A.2.2. Language Skills Support Scenario

  • I am willing to communicate with AI to improve my translation skills through practice.
  • I am willing to communicate with AI to practice English speaking in different situations.
  • I am willing to communicate with AI to revise and polish my writing.
  • I am willing to communicate with AI to understand and practice English listening materials.
  • I am willing to communicate with AI to understand the content and structure of English texts.
  • I am willing to communicate with AI to learn and clarify English grammar rules.

Appendix A.2.3. Academic Knowledge Inquiry Scenario

  • I am willing to communicate with AI to translate complex academic texts.
  • I am willing to communicate with AI to better understand difficult concepts or theories in academic courses.
  • I am willing to communicate with AI to complete academic writing tasks.
  • I am willing to communicate with AI to clarify difficult theories or abstract ideas in academic readings.
  • I am willing to communicate with AI to understand viewpoints, positions, or reasoning in academic texts.
  • I am willing to communicate with AI to explore key ideas and theories of specific scholars.

Appendix A.2.4. Factual Information Retrieval Scenario

  • I am willing to communicate with AI to look up the meanings or usage of unfamiliar English words.
  • I am willing to communicate with AI to confirm grammar rules or sentence structures.
  • I am willing to communicate with AI to check the standard definitions of academic concepts.
  • I am willing to communicate with AI to retrieve basic facts about well-known people.
  • I am willing to communicate with AI to find out historical or cultural background knowledge.
  • I am willing to communicate with AI to ask about daily life knowledge.

Appendix B

Table A1. CR and AVE for the Measurement Model.
Table A1. CR and AVE for the Measurement Model.
FactorCRAVE
AIS PU0.9260.677
AIS PEU0.9170.647
AIS AI-WTC0.9510.764
LSSS PU0.8970.593
LSSS PEU0.9210.66
LSSS WTC0.930.688
AKIS PU0.9470.749
AKIS PEU0.950.761
AKIS AI-WTC0.9550.779
FIRS PU0.940.724
FIRS PEU0.9440.736
FIRS WTC0.940.723
Note. CR = Composite Reliability; AVE = Average Variance Extracted; PU = Perceived Usefulness; PEU = Perceived Ease of Use; AI-WTC = Willingness to Communicate with AI; AIS = Advisory Interaction Scenario; LSSS = Language Skills Support Scenario; AKIS = Academic Knowledge Inquiry Scenario; FIRS = Factual Information Retrieval Scenario.
Table A2. Correlation Matrix of PU, PEU, TA, and AI-WTC.
Table A2. Correlation Matrix of PU, PEU, TA, and AI-WTC.
AIS_PUAIS_PEULSSS_PULSSS_PEUAKIS_PUAKIS_PEUFIRS_PUFIRS_PEUAIS_WTCLSSS_WTCAKIS_WTCFIRS_WTC
AIS_PU1
AIS_PEU0.8491
LSSS_PU0.8160.8201
LSSS_PEU0.7980.8320.9111
AKIS_PU0.7860.7950.8710.8751
AKIS_PEU0.7560.8080.8380.8680.9131
FIRS_PU0.7230.7070.7830.7680.7970.7801
FIRS_PEU0.7230.7170.8040.7920.8150.7830.9261
AIS_WTC0.8050.7830.8270.8240.8450.8210.8120.8431
LSSS_WTC0.7650.7670.8390.8200.8180.8070.8490.8380.8781
AKIS_WTC0.7420.7170.7810.7650.8330.7910.8400.8550.8580.8691
FIRS_WTC0.7360.7160.7930.7690.8080.7580.8630.8460.8630.8520.8791
Note. PU = Perceived Usefulness; PEU = Perceived Ease of Use; AI-WTC = Willingness to Communicate with AI; AIS = Advisory Interaction Scenario; LSSS = Language Skills Support Scenario; AKIS = Academic Knowledge Inquiry Scenario; FIRS = Factual Information Retrieval Scenario.

References

  1. Abdi, H. (2007). The Bonferroni and Šidák corrections for multiple comparisons. In N. J. Salkind (Ed.), Encyclopedia of measurement and statistics (Vol. 1, pp. 103–107). Sage. [Google Scholar]
  2. AlAfnan, M., Dishari, S., Jovic, M., & Lomidze, K. (2023). ChatGPT as an educational tool: Opportunities, challenges, and recommendations for communication, business writing, and composition courses. Journal of Artificial Intelligence and Technology, 3(2), 60–68. [Google Scholar] [CrossRef]
  3. Al-Emran, M., Mezhuyev, V., & Kamaludin, A. (2018). Technology acceptance model in m-learning context: A systematic review. Computers & Education, 125, 389–412. [Google Scholar] [CrossRef]
  4. Apriani, E., Cardoso, L., Obaid, A. J., Muthmainnah, Wijayanti, E., Esmianti, F., & Supardan, D. (2024). Impact of AI-powered chatbots on EFL students’ writing skills, self-efficacy, and self-regulation: A mixed-methods study. Global Educational Research Review, 1(2), 57–72. [Google Scholar] [CrossRef]
  5. Brislin, R. W. (1970). Back-translation for cross-cultural research. Journal of Cross-Cultural Psychology, 1(3), 185–216. [Google Scholar] [CrossRef]
  6. Chiu, T. K. F., Xia, Q., Zhou, X., Chai, C. S., & Cheng, M. (2023). Systematic literature review on opportunities, challenges, and future research recommendations of artificial intelligence in education. Computers and Education: Artificial Intelligence, 4, 100118. [Google Scholar] [CrossRef]
  7. Cohen, J. (1988). Statistical power analysis for the behavioral sciences (2nd ed.). Lawrence Erlbaum Associates. [Google Scholar]
  8. Davis, F. D. (1989). Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Quarterly, 13(3), 319–340. [Google Scholar] [CrossRef]
  9. DeVellis, R. F. (2017). Scale development: Theory and applications (4th ed.). Sage Publications. [Google Scholar]
  10. DeVellis, R. F., & Thorpe, C. T. (2021). Scale development: Theory and applications (5th ed.). Sage Publications. [Google Scholar]
  11. Fathi, J., Rahimi, M., & Derakhshan, A. (2024). Improving EFL learners’ speaking skills and willingness to communicate via artificial intelligence-mediated interactions. System, 121, 103254. [Google Scholar] [CrossRef]
  12. Fauzi, F., Tuhuteru, L., Sampe, F., Ausat, A. M. A., & Hatta, H. R. (2023). Analysing the role of CHAT-GPT in improving student productivity in higher education. Journal on Education, 5(4), 14886–14891. [Google Scholar] [CrossRef]
  13. Feng, H., Li, K., & Zhang, L. J. (2025). What does AI bring to second language writing? A systematic review (2014–2024). Language Learning & Technology, 29(1), 1–27. [Google Scholar] [CrossRef]
  14. Field, A. (2018). Discovering statistics using IBM SPSS statistics (5th ed.). Sage Publications. [Google Scholar]
  15. Gravetter, F. J., & Wallnau, L. B. (2016). Statistics for the behavioral sciences (10th ed.). Cengage Learning. [Google Scholar]
  16. Hinkin, T. R. (1998). A brief tutorial on the development of measures for use in survey questionnaires. Organizational Research Methods, 1(1), 104–121. [Google Scholar] [CrossRef]
  17. Huang, F., & Zou, B. (2024). English speaking with artificial intelligence (AI): The roles of enjoyment, willingness to communicate with AI, and innovativeness. Computers in Human Behavior, 159, 108355. [Google Scholar] [CrossRef]
  18. Karataş, F., Abedi, F. Y., Gunyel, F. O., Karadeniz, D., & Kuzgun, Y. (2024). Incorporating AI in foreign language education: An investigation into ChatGPT’s effect on foreign language learners. Education and Information Technologies, 29, 19343–19366. [Google Scholar] [CrossRef]
  19. Kim, A., & Su, Y. (2024). How implementing an AI chatbot impacts Korean as a foreign language learners’ willingness to communicate in Korean. System, 122, 103256. [Google Scholar] [CrossRef]
  20. Kohnke, L., Moorhouse, B. L., & Zou, D. (2023). ChatGPT for language teaching and learning. RELC Journal, 54(2), 537–550. [Google Scholar] [CrossRef]
  21. Lantolf, J. P., & Poehner, M. E. (2014). Sociocultural theory and the pedagogical imperative in L2 education: Vygotskian praxis and the research/practice divide. Routledge. [Google Scholar] [CrossRef]
  22. Lantolf, J. P., & Thorne, S. L. (2006). Sociocultural theory and the genesis of second language development. Oxford University Press. [Google Scholar]
  23. Lee, J. S. (2019). EFL students’ views of willingness to communicate in the extramural digital context. Computer Assisted Language Learning, 32(7), 692–712. [Google Scholar] [CrossRef]
  24. Lee, J. S., & Drajati, N. (2019). Affective variables and informal digital learning of English: Keys to willingness to communicate in a second language. Australasian Journal of Educational Technology, 35(4), 168–182. [Google Scholar] [CrossRef]
  25. Lee, J. S., & Drajati, N. (2020). Willingness to communicate in digital and non-digital EFL contexts: Scale development and psychometric testing. Computer Assisted Language Learning, 33(5–6), 560–583. [Google Scholar] [CrossRef]
  26. Liu, G., & Ma, C. (2024). Measuring EFL learners’ use of ChatGPT in informal digital learning of English based on the technology acceptance model. Innovation in Language Learning and Teaching, 18(2), 125–138. [Google Scholar] [CrossRef]
  27. Liu, M., Zhang, L. J., & Biebricher, C. (2024). Investigating students’ cognitive processes in generative AI-assisted digital multimodal composing and traditional writing. Computers & Education, 211, 104977. [Google Scholar] [CrossRef]
  28. Liu, M., Zhang, L. J., & Zhang, D. (2025). Enhancing student GAI literacy in digital multimodal composing through development and validation of a scale. Computers in Human Behavior, 166, 108569. [Google Scholar] [CrossRef]
  29. MacIntyre, P. D., Clément, R., Dörnyei, Z., & Noels, K. A. (1998). Conceptualizing willingness to communicate in a L2: A situational model of L2 confidence and affiliation. The Modern Language Journal, 82(4), 545–562. [Google Scholar] [CrossRef]
  30. Maxwell, S. E., Delaney, H. D., & Kelley, K. (2018). Designing experiments and analyzing data: A model comparison perspective (3rd ed.). Routledge. [Google Scholar]
  31. McCroskey, J. C., & Baer, J. E. (1985, November 7–10). Willingness to communicate: The construct and its measurement. Annual Convention of the Speech Communication Association, Denver, CO, USA. [Google Scholar]
  32. Nematizadeh, S., & Wood, D. (2021). Second language willingness to communicate as a complex dynamic system. In N. Zarrinabadi, & M. Pawlak (Eds.), New perspectives on willingness to communicate in a second language (pp. 19–38). Springer. [Google Scholar] [CrossRef]
  33. Nunnally, J. C., & Bernstein, I. H. (1994). Psychometric theory (3rd ed.). McGraw-Hill. [Google Scholar]
  34. Ouyang, Z., Jiang, Y., & Liu, H. (2024). The effects of Duolingo, an AI-integrated technology, on EFL learners’ willingness to communicate and engagement in online classes. International Review of Research in Open and Distributed Learning, 25(3), 97–115. [Google Scholar] [CrossRef]
  35. Peng, J.-E., & Liang, W. (2025). Willingness to communicate with artificial intelligence (AI)? insights from tracking EFL learners’ perceived acceptance and chat output. Computer Assisted Language Learning, 1–27. [Google Scholar] [CrossRef]
  36. Peng, J.-E., & Woodrow, L. (2010). Willingness to communicate in English: A model in the Chinese EFL classroom context. Language Learning, 60(4), 834–876. [Google Scholar] [CrossRef]
  37. Rahimi, M. (2025). Synergising dialogic teaching with competencies-trained GenAI dialoguing for critical thinking and communication competencies. Journal of University Teaching and Learning Practice, 22(2), 1–22. [Google Scholar] [CrossRef]
  38. Strzelecki, A. (2023). To use or not to use ChatGPT in higher education? A study of students’ acceptance and use of technology. Interactive Learning Environments, 32(9), 5142–5155. [Google Scholar] [CrossRef]
  39. Swain, M., Kinnear, P., & Steinman, L. (2015). Sociocultural theory in second language education: An introduction through narratives (2nd ed.). Multilingual Matters. [Google Scholar] [CrossRef]
  40. Sweller, J., & van Merrienboer, J. J. G. (2005). Cognitive load theory and complex learning: Recent developments and future directions. Educational Psychology Review, 53(3), 147–177. [Google Scholar] [CrossRef]
  41. Vygotsky, L. S. (1978). Mind in society: The development of higher psychological processes. Harvard University Press. [Google Scholar]
  42. Wang, C., Zou, B., Du, Y., & Wang, Z. (2024). The impact of different conversational generative AI chatbots on EFL learners: An analysis of willingness to communicate, foreign language speaking anxiety, and self-perceived communicative competence. System, 127, 103533. [Google Scholar] [CrossRef]
  43. Wu, H., & Zhang, L. J. (2017). Effects of different language environments on Chinese graduate students’ perceptions of English writing and their writing performance. System, 65, 164–173. [Google Scholar] [CrossRef]
  44. Xie, X., Teng, M., Zhang, L. J., & Alamer, A. (2025). Exploring AI literacy and AI-induced emotions among Chinese university English language teachers: The partial least squares structural equation modeling (PLS-SEM) approach. International Journal of Applied Linguistics. Advance online publication. [Google Scholar] [CrossRef]
  45. Zawacki-Richter, O., Marín, V. I., Bond, M., & Gouverneur, F. (2019). Systematic review of research on artificial intelligence applications in higher education—Where are the educators? International Journal of Educational Technology in Higher Education, 16(1), 39. [Google Scholar] [CrossRef]
  46. Zhang, D., Wu, J. G., & Fu, Z. (2024). From shy to fly: Facilitating EFL learners’ willingness to communicate with an AI chatbot and an intelligent tutoring system. System, 127, 103501. [Google Scholar] [CrossRef]
Figure 1. Conceptual Model.
Figure 1. Conceptual Model.
Behavsci 15 01391 g001
Table 1. Structure and item distribution of the scenario-based questionnaire.
Table 1. Structure and item distribution of the scenario-based questionnaire.
SectionContent/ConstructScenariosNo. of Items
1. Background InformationGender, Age, Major, Academic Year 4
2. A. TAPU, PEUAIS, LSSS, AKIS, FIRS48
 B. AI-WTCAI-WTCAIS, LSSS, AKIS, FIRS24
Total72
Note. PU = Perceived Usefulness; PEU = Perceived Ease of Use; AI-WTC = Willingness to Communicate with AI. AIS = Advisory Interaction Scenario; LSSS = Language Skills Support Scenario; AKIS = Academic Knowledge Inquiry Scenario; FIRS = Factual Information Retrieval Scenario.
Table 2. Results of repeated measures ANOVA for interaction scenario effects on PU, PEU.
Table 2. Results of repeated measures ANOVA for interaction scenario effects on PU, PEU.
Dependent VariableFdfpη2
PU16.0963, 1098<0.0010.042
PEU24.6413, 1098<0.0010.063
Note. PU = Perceived Usefulness; PEU = Perceived Ease of Use.
Table 3. Results of Repeated Measures ANOVA for Interaction Scenario Effects AI-WTC.
Table 3. Results of Repeated Measures ANOVA for Interaction Scenario Effects AI-WTC.
Dependent VariableFdfpη2
AI-WTC9.6003, 1098<0.0010.026
Note. AI-WTC = Willingness to Communicate with AI.
Table 4. Bonferroni Post Hoc Comparisons across Interaction Scenarios for PU, PEU, and WTC.
Table 4. Bonferroni Post Hoc Comparisons across Interaction Scenarios for PU, PEU, and WTC.
VariableFdfpη2Comparisons
PU18.82310.0000.049FIRS > AKIS > AIS > LSSS
PEU51.77410.0000.124FIRS > AKIS > LSSS > AIS
AI-WTC22.90310.0000.059FIRS > AKIS > LSSS ≈ AIS
Note. PU = Perceived Usefulness; PEU = Perceived Ease of Use; AI-WTC = Willingness to Communicate with AI; AIS = Advisory Interaction Scenario; LSSS = Language Skills Support Scenario; AKIS = Academic Knowledge Inquiry Scenario; FIRS = Factual Information Retrieval Scenario.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cui, Z.; Yang, H.; Xu, H. The Effects of Interaction Scenarios on EFL Learners’ Technology Acceptance and Willingness to Communicate with AI. Behav. Sci. 2025, 15, 1391. https://doi.org/10.3390/bs15101391

AMA Style

Cui Z, Yang H, Xu H. The Effects of Interaction Scenarios on EFL Learners’ Technology Acceptance and Willingness to Communicate with AI. Behavioral Sciences. 2025; 15(10):1391. https://doi.org/10.3390/bs15101391

Chicago/Turabian Style

Cui, Zheng, Hua Yang, and Hao Xu. 2025. "The Effects of Interaction Scenarios on EFL Learners’ Technology Acceptance and Willingness to Communicate with AI" Behavioral Sciences 15, no. 10: 1391. https://doi.org/10.3390/bs15101391

APA Style

Cui, Z., Yang, H., & Xu, H. (2025). The Effects of Interaction Scenarios on EFL Learners’ Technology Acceptance and Willingness to Communicate with AI. Behavioral Sciences, 15(10), 1391. https://doi.org/10.3390/bs15101391

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop