Next Article in Journal
Platform-Dependent Entrepreneurship: A Systematic Review
Previous Article in Journal
How Much Is Too Much? The Impact of Update Frequency on Crowdfunding Success
Previous Article in Special Issue
Leveraging ICT and Generative AI in Higher Education for Sustainable Development: The Case of a Lebanese Private University
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unleashing ChatGPT: Redefining Technology Acceptance and Digital Transformation in Higher Education

by
Loubna Mourtajji
1,2,3 and
Nathalie Arts-Chiss
1,2,3,*
1
LEFMI (Laboratory of Economics, Finance, Management, and Innovation), University of Picardy Jules Verne, 80000 Amiens, France
2
Marketing Department, University Institute of Technology of Oise, 60100 Creil, France
3
Business and Administration Management Department, University Institute of Technology of Oise, 60000 Beauvais, France
*
Author to whom correspondence should be addressed.
Adm. Sci. 2024, 14(12), 325; https://doi.org/10.3390/admsci14120325
Submission received: 22 October 2024 / Revised: 18 November 2024 / Accepted: 23 November 2024 / Published: 2 December 2024
(This article belongs to the Special Issue ChatGPT, a Stormy Innovation for a Sustainable Business)

Abstract

:
This article examines the effects of integrating ChatGPT, a generative language model developed by OpenAI, into educational and training contexts in higher education. The research takes as its conceptual framework models of technology acceptance and questions the relevance of these models to the acceptance and adoption of ChatGPT. A qualitative study carried out with teachers from various higher education establishments in France enables us to propose a model adapted to the specific features of generative AI. The ethical dimension and the controllability of the tools by users, made possible by a progressive training program, are two constructions that are essential to a proper understanding of whether or not these new tools are adopted. Additionally, this research contributes to the growing discourse on how generative AI innovations can be leveraged to enhance digital transformation in the academic sector, with a particular focus on business schools’ stakeholders and strategies. Finally, the contributions and prospects for future research are discussed.

1. Introduction

Since 2022, the introduction of artificial intelligence (AI) technologies in the field of education has marked a significant turning point in teaching and learning methods. Among these technologies, ChatGPT, developed by OpenAI, stands out for its ability to generate text in a contextual manner and to explore different domains, thus offering new pedagogical perspectives (Moussavou 2023; Lo 2023; Rahman and Watanobe 2023; Mogavi et al. 2023). The aim of this research is to examine the factors influencing the perceptions, acceptance and use of ChatGPT, and its influence on the teaching practices of higher education practitioners. Through the application of Technology Acceptance Theory (TAM) as a theoretical framework (Davis 1986), this study aims to propose a TAM model specific to ChatGPT as advocated by Venkatesh (2022), considering the specificities of the application domain, namely education and university teaching. This will make it possible to understand the contexts in which the adoption of ChatGPT could modify teachers’ pedagogical approach, in terms of course design, pedagogical experience, assessment methods and interaction with students (Klyshbekova and Abbott 2024; Karthikeyan 2023; Montenegro-Rueda et al. 2023).
While some teachers proclaim the effectiveness and usefulness of ChatGPT, others remain highly sceptical and fear the excesses of such a tool. Based on TAM theory, the aim of this research is to empirically examine teachers’ perceptions and practices. By addressing this issue, this study contributes to the existing literature on educational technologies and the integration of AI into teaching, by providing an in-depth and empirical analysis of the impact of ChatGPT on pedagogical dynamics.
In the current academic landscape, the integration of ChatGPT into education is attracting increased research interest. However, an in-depth exploration of the literature reveals a lack of empirical studies on the use and actual impact of ChatGPT. To date, few field studies have been identified. Most of the published work takes a theoretical approach or focuses on the analysis of existing literature, offering conceptual perspectives on the potential applications and theoretical implications of ChatGPT without drawing on concrete empirical data.
This gap raises pertinent questions about the actual effectiveness of ChatGPT in practical contexts, particularly in teaching. The aim of this work is to fill in some of the gaps in the literature by means of an exploratory study of teachers who use or do not use AI in their daily work. This approach will not only enable existing theoretical hypotheses to be validated but will also identify ways of optimising the use of ChatGPT in various practical applications.
To achieve this objective, a qualitative approach was favoured (Bardin 2013). Semi-directive interviews were conducted with a targeted sample of university teachers who had or had not integrated ChatGPT into their teaching methods. The aim of this first stage was to gather detailed data on their perceptions, their experiences and the concrete changes in their teaching practices following the use of ChatGPT. Following the analysis of the interviews, this research proposes a version of the TAM model revisited and adapted to the adoption of ChatGPT in education.
First, we present an updated literature review on the use of ChatGPT in education, as well as on models of acceptance of the technology. The methodological choices are then presented. Finally, the main results of this exploratory study are highlighted, allowing us to propose a specific and adapted model.

2. Theoretical Framework

Our research is based on two main concepts: ChatGPT and technology acceptance models. The first part will discuss the evolution of ChatGPT, its impact on education and its opportunities and challenges. The second part will look at the models of the acceptance of the technology, highlighting how they have evolved over the years and the results obtained.

2.1. Evolution of ChatGPT: Impacts on Higher Education Sector

Humanity’s ongoing quest to simulate human intelligence is not new. For many years, the ambition of scientists has been to create systems capable of thinking, understanding and interacting like a human being. Since 1950, the first Turing experiments have been exploring the possibilities of thinking machines by attempting to model complex cognitive processes. The first neural network computer was created, and Turing published the Turing Test used to evaluate AIs. These major advances went on to form the foundations of artificial intelligence. But it was not until 1956 that the word ‘Artificial Intelligence’ was officially used.
In 1989, the Frenchman Yann Lecun developed the first neural network capable of recognising handwritten numbers, an invention that was to lead to the development of deep learning. In 1997, IBM’s Deep Blue system won against the world chess champion, the first time a machine had surpassed human capabilities. Over the following decades, AI continued to evolve, each time pushing the limits of what machines could achieve.
The launch of ChatGPT by OpenAI, in November 2022, represents a significant step forward in this evolution, illustrating both a great capacity to accumulate and process information and to generate a language identical to that of human beings. ChatGPT, derived from the GPT (Generative Pre-trained Transformer) family of models, is the fruit of several years’ research into automatic natural language processing (ANLP). It is an ‘encyclopaedic model that integrates a large number of references to the real world’ (Langlais 2023). Its operating mechanism is based on a large corpus of data, on which it is trained and which enables it to respond to user queries in a superficially coherent and nuanced way (Langlais 2023). One of the special features of ChatGPT is its ability to perform fine-tuning, which allows the model to be specialised on specific tasks or domains, by re-training it on targeted datasets.
This ability to process large amounts of data and produce text in a consistent and contextual way has led to and facilitated its rapid adoption in several fields. The integration of AI in small- and medium-sized enterprises is marking a profound transformation of their environment and their working conditions (Berger-Douce et al. 2023). Indeed, this tool has been used in the healthcare sector to support research and clinical decision-making (Garg et al. 2023), in improving medical documentation (Baker et al. 2024), in business to improve employee productivity and customer service (Zhu et al. 2023), in entrepreneurship education (Dabbous and Boustani 2023) and in education in general to foster student engagement and improve learning (Lo 2023; Montenegro-Rueda et al. 2023). Thus, according to Kwan (2023), ChatGPT is an assistance tool for teachers, enabling more effective lesson preparation (Rahman and Watanobe 2023; Lo 2023) and offering personalised support to students (Montenegro-Rueda et al. 2023). Thanks to ChatGPT, we can indeed witness a transformation in teaching and learning methods, including personalised assistance, the creation of educational content, learner engagement, automated formative assessment and the diversification of pedagogical approaches. The table below summarises the main opportunities and threats of using ChatGPT in education (Table 1).
In the higher education context, although ChatGPT has several advantages and provides a good opportunity (Adeshola and Adepoju 2023) to improve the educational experience (Grassini 2023), the integration of ChatGPT is double-edged. Although this technology has real potential for both teachers and students, it is not without risk and its integration is not as straightforward as one might think. Admittedly, in theory, this AI can be used by teachers as an assistant, a course creation tool, an assessment support tool (Klyshbekova and Abbott 2024), a tool for personalising learning (Mhlanga 2023; Grassini 2023; Karthikeyan 2023) and a tool for managing administrative tasks (Memarian and Doleck 2023). All of this makes it possible to enhance the teacher’s experience and place greater emphasis on interaction and exchange with students (Lo 2023; Montenegro-Rueda et al. 2023). However, the problems of data security (Al-Mughairi and Bhaskar 2024), plagiarism (Lo 2023; Grassini 2023; Rahman and Watanobe 2023; Yu 2024), intellectual property (Mhlanga 2023), addiction (Al-Mughairi and Bhaskar 2024) and algorithmic bias (Memarian and Doleck 2023) hamper its rapid implementation. Furthermore, the empirical effectiveness and veracity of such promises have yet to be demonstrated, and their real impact on the depersonalisation of teaching and on students’ cognitive abilities has yet to be proven (Rahman and Watanobe 2023). Hence, it is important to put into place a framework that optimises the benefits of ChatGPT while reducing the ethical, psychological, legal and technical risks.

2.2. Theories of Technology Acceptance: Evolution and Specificities

Our conceptual model is based on the Technology Acceptance Model (Davis 1986), which conceptualises the acceptance of technology by users (Venkatesh and Davis 2000). Meta-analyses of the TAM (King and He 2006; Yousafzai et al. 2007a, 2007b) show the extent to which this model is used in a variety of contexts and situations around the world. In the field of education, studies have tested the adoption and acceptance (or otherwise) of innovative technologies from the point of view of both teachers (Scherer et al. 2019) and students (Mohammadi 2015; Ibrahim et al. 2017; Rafique et al. 2020).
Under the name TAM, there are in fact several models of technology acceptance which have been enriched as studies have been carried out and results obtained, making it possible to obtain ‘enriched’ or ‘extended’ versions of the basic model. The first TAM model (Davis 1986, 1989) focuses on perceived usefulness and perceived ease of use, which determine behavioural intention. It is a simplified, even minimalist, model which has the advantage of being easy to understand and usable in many contexts. TAM 2 (Venkatesh and Davis 2000) proposes an enriched version of the model by highlighting seven antecedents of perceived usefulness. TAM 3 (Venkatesh and Bala 2008) proposes the addition of six antecedents of perceived ease of use. In parallel with the TAM models, the UTAUT model, known as the unified model (Venkatesh et al. 2003), allows several approaches to be considered to explain intention to use. It focuses on four moderating factors and four determining factors that provide a better understanding of intention to use. The UTAUT2 model (Venkatesh et al. 2012) is more consumer-oriented, and includes habits, price and pleasure as additional antecedents. The table below summarises the different models (Table 2).
This study is focused on the latest UTAUT model proposed by Venkatesh (2022), which focuses on the context of artificial intelligence. Indeed, the specificity of AI leads to significant modifications of the model that need to be considered. The author proposes an adapted research programme to study the adoption of AI by employees. Ten points of vigilance, grouped into three categories, are highlighted as follows: the intrinsic characteristics of artificial intelligence tools, the characteristics of the employees who will use them and the characteristics of the organisation in which the AI will be used. The figure below illustrates these different elements (Figure 1).
Given these specific features, the proposed model incorporates four new moderating factors: individual characteristics linked to the personality of employees (acceptance or even search for risk, tolerance of uncertainty or desire to learn), characteristics linked to artificial intelligence technology (quality, errors and transparency), characteristics linked to the organisation in which the tools will be used (climate conducive to innovation, number of people involved and incomplete or missing information), and characteristics linked to training (regular training sessions, initiation and gamification). These four moderating factors can influence the classic determining factors of the model, namely performance expectancy, effort expectancy, social influence and facilitating conditions, which themselves help to explain the intention to use AI. Menon and Shilpa (2023) tested the model on ChatGPT users (students and employees) and found results in line with previous literature: performance expectancy, effort expectancy, social influence, facilitating conditions and privacy concerns were the predominant factors influencing the usage of ChatGPT. What about others who have not yet accepted this technique, such as teachers who have complete control over their teachings?
In the context of our study, we are going to question these different models of the acceptance of technology. The organisation considered will be represented by the higher education institutions considered, the employees will be the teachers interviewed and the artificial intelligence tool studied will mainly be ChatGPT. To do this, we used a qualitative methodology presented in the following section.

3. Methodology: A Qualitative Study

In order to gain a better understanding of the factors influencing the adoption of ChatGPT in the context of higher education and more specifically by teachers, a qualitative approach was adopted. The authors have conducted semi-directive face-to-face interviews. This method provides interesting and nuanced data on the participants’ experiences, perceptions and attitudes.
The sample consisted of teachers working at various levels of university education and academic fields in public and private institutions across continental France. Participants were chosen progressively using purposive sampling, with the goal of increasing profile variety while maintaining qualitative representativeness (Paillé and Mucchielli 2021). Respondents possessed unique characteristics such as gender, age, seniority, area of competence, and interest in technology.
The individual semi-structured interviews lasted an average of one hour (the length of the interviews varied between 30 min and 90 min) and were conducted face-to-face by two interviewers between March 2024 and April 2024. After an initial introductory phase, during which the interviewers introduced themselves and explained the purpose of the study, the interviewees were asked to introduce themselves and their relationship with the technology. Seven main themes are then addressed (the interview guide is presented in Appendix A):
  • The experience of using ChatGPT;
  • Its perceived usefulness;
  • Its perceived ease of use;
  • Its perceived interest;
  • The users’ attitude to generative AI;
  • Their intentions to use;
  • Subjective opinions and standards.
The interviewers used the classic techniques of reminder and reformulation. Interviewees were encouraged to express themselves freely about their experience with ChatGPT and generative AI.
All of the qualitative data (transcribed verbatim) was subjected to content analysis (Bardin 2013; Paillé and Mucchielli 2021) using an abductive approach, an analytical approach that is both deductive, based on the enriched TAM framework, and inductive, aimed at bringing out new relevant themes from the field.
To comply with an ethical approach to data collection and use (Kozinets 2019), we explained the research objectives to the interviewees, specifying that the results would be used in a scientific and/or educational context. Oral consent was systematically obtained. Anonymity and confidentiality were guaranteed both during data collection and data processing and analysis.
An in-depth examination of the verbatim transcripts by the researchers enabled us to identify the main trends and factors explaining the adoption of ChatGPT. Our empirical results confirm the relevance of the enriched TAM theoretical framework, while highlighting certain specificities linked to the singular nature of this conversational AI technology.

4. Results

The ten people interviewed had a wide range of characteristics. They ranged in age from 25 to 69. Six men and four women made up our sample. We were able to interview teachers specialising in various fields such as economics, management, finance, communication, marketing, mathematics and chemistry (Table 3).
This diversity of profiles ensures that the results are representative. Some of the people interviewed use ChatGPT regularly, which can be seen as ‘a time-saving assistant for everyday tasks or tasks that we’re not used to doing’ (I10). The uses may be professional, but also private. It quickly becomes apparent that this tool easily penetrates the different spheres of a person’s life: ‘So, on the other hand, the uses are extremely varied, ranging from preparing my future maths homework to finding the recipe for I don’t know the last time a fish I don’t know any more, in short finding recipes, preparing a road trip to Scotland or writing a paragraph to present a project, etc.’ (I7). On the other hand, other people had not yet used ChatGPT, or had used it very little, and some had only tried it: ‘Apart from the test I’ve just told you about, I’ve never tried it’ (I8); they had no plans to use it for the time being: ‘In any case, the attempt I made was a trial’ (I6).
The content analysis reveals that teachers’ perception of the usefulness of ChatGPT is very mixed. While the productivity gains (the creation of resources, time saving, optimisation of research, reformulation and editorial benefits, etc.) are recognised, which is in line with the findings on the perceived usefulness of other educational technologies (Mugo et al. 2017), the profound pedagogical impact of ChatGPT on learning is divisive. Some see it as a tool that allows ‘more flexibility for students depending on the subject, of course for students so that they can acquire more autonomy [...] so that students are more involved in their own learning’ (I6); others remain highly critical of these effects on learning: ‘What I think about students is that they no longer think. They don’t reason any more. They think they know everything because they’ve got ChatGPT. I’m afraid of the day when ChatGPT will take over from humans..., and I think that with young people now, it’s OK, it’s taken over’ (I1). Another aspect directly concerns the teacher, who wants to ‘remain in control of my lesson. I’m not interested in being the spokesperson or microphone for a computerised lesson’ (I6). Teachers can thus feel that their role or status has been taken away from them. This profound questioning of the teaching profession goes hand in hand with a questioning of its credibility: ‘they [the students] have the opportunity to check everything you say and the drift of course, we’re not going to believe the teacher’ (I8). Between the leverage of personalisation and autonomy and the risk of disengaging students and calling into question the status of the teacher, this crucial dimension of perceived usefulness remains very mixed between advocates and opponents.
Despite an intuitive conversational interface, our results show that the optimal pedagogical integration of ChatGPT is perceived as complex by a majority of teachers, who may feel overwhelmed by the amount of information available: ‘I have the impression that you can’t get to grips with the subject because you haven’t mastered it’ (I6); ‘so there’s too much information, which perhaps requires time to reprocess or to be able to stand back’ (I8). ChatGPT responses depend on the relevance of the prompts, which many teachers do not master: ‘In fact, when the answer doesn’t suit me, it’s because the question was badly put. When it doesn’t suit me, it’s because the prompt was badly written and so I have to revise my prompt’ (I7); ‘the tool is fairly intuitive. However, to specify the results obtained effectively, resources on prompting may be useful’ (I5).
The people interviewed are aware of the importance of training: ‘If I had training, it would be easier to use it to master it’ (I8). They even went so far as to call for an appropriate training program: ‘I’d like us to have support in getting to grips with it, whether for teaching or for research techniques, in a gradual way’ (I9). Without this training, teachers run the risk of abandoning ChatGPT and giving up using it: ‘sometimes I’d give him a command and he’d give me the wrong answer, so I thought, «I don’t have to give him the right information so that he can be reactive afterwards, so that made me angry»’ (I9). While ChatGPT may be perceived as magical by novices, experts will have more distance from its use and the reliability of the results obtained: ‘We know that what ChatGPT gives us is an approximation of what we’re looking for, and it always needs to be checked, so it’s absolutely unreliable’ (I10); “as the data isn’t necessarily reliable, since it’s everything that’s on the Web, we’re not necessarily sure of the result, so we need to be sufficiently trained and sufficiently wary about using this tool” (I7). Training is therefore essential for the proper use of generative AI, and it must be progressive and ongoing. Indeed, after the ‘Wow’ effect of discovering the tool, there is always a ‘Down’ phase when the person realises that anything can be made to say anything and obtain completely false results. Training must help them get past this second stage to reach the ‘Stabilisation’ phase, during which users become aware of the tool’s limitations: ‘For me, that’s the point of training. For me, that’s the point of training. It’s to ensure that when people come out of a training course, they understand the limits, the benefits and the uses they’re going to be able to make of it’ (I7).
This highlights the limitations of the perceived ease of use construct as initially conceptualised in the TAM. A more explicit consideration of the systemic dimension and the profound adaptation of practices seems necessary. The use of ChatGPT in certain areas, coding for example, seems easier and more relevant. To reassure teachers, effective mastery of ChatGPT and its prompts is seen as a major prerequisite for its successful adoption. ‘So in fact, I discovered that it was a tool that could be interesting for giving you the first bits of structuring, I find that in terms of structuring, you give a subject, very quickly, it is able to propose a good structure.... I took part in two videoconference training sessions with some guys, who talked about using ChatGPT and how to challenge it with prompts. I found it very interesting to discover that it’s difficult to get out of Google where you type queries, but I realised as I used it that the question of exchange, challenge and prompts is important’ (I4).
The issue of the controllability and transparency of ChatGPT emerges as a central concern in all of the teachers’ comments, and some even consider it to be ‘a plundering or copying software. Moreover, there’s a whole problem with copyright, which poses a real problem at that level’ (I6). The problem of data and its sources is ‘an element that, in any case, will necessarily be considered. Depending on everyone’s interests, particularly those of individual countries. I think that this notion of intellectual property rights will be settled when the two powers that are going to be at the forefront of AI have made their move, namely the Chinese and then the Americans, and it is on the basis of this tug of war that a code of property rights will emerge, and as I see it, we’ll have two of them!’ (I3). This new construct that emerges from the data analysis is particularly illuminating in understanding the reservations and reticence observed. The ‘black box’ (Kleinpeter 2020) represented by the opaque internal workings of AI, the lack of traceability of the results generated and the unpredictability of possible biases or ethical abuses are sources of many questions and concerns. As a result, major efforts on algorithmic transparency (explanations of reasoning), data training (what happens to the information entered), data sources and auditability and user control seem essential to reassure the teaching profession. ‘I don’t use ChatGPT, because apparently there are risks in terms of personal data... I don’t know who can do what with this data. You’ve always got someone behind it who manipulates as they please and does what they want with the data to sell it, to exploit it … so even for pedagogy, I don’t really need it.’ (I2).

5. Discussion

Since its initial conception, TAM has been the subject to numerous theoretical extensions aimed at strengthening its explanatory power, particularly in specific contexts such as teaching (Scherer et al. 2019). However, the adoption of such an innovative and disruptive technology as ChatGPT raises singular issues that require some adjustments to the initial model.
Indeed, beyond its simple practical utility, ChatGPT raises ethical concerns and questions among teachers, such as plagiarism, intellectual property, the risks of dehumanization, the oppression of creativity and the reinforcement of biases and discriminations. This ethical dimension in the AIA2M model was considered as a complementary variable influencing the general attitude towards the tool. Moreover, the very nature of ChatGPT, as a conversational AI, implies specific issues of the controllability and transparency of data perceived by users. Understanding the internal functioning of ChatGPT, its training and the possibility of parameterizing and auditing its results, as well as the ability to explain its reasoning in an interpretable way would seem to be key factors in its effective adoption by teachers.
What emerges from this study is a TAM model that has been enriched and adjusted to the specific features of ChatGPT. This revisited model should enable a finer, contextualized analysis of the psychological, technical and ethical factors shaping teachers’ perceptions, attitudes and intentions towards this disruptive innovation (Figure 2).
The model proposed in this study represents a significant evolution of the classic Technology Acceptance Model (TAM), specifically adapted to the challenges posed by the adoption of ChatGPT in the educational context. The major contribution of this new model lies in the integration of new variables that reflect the specificities of generative artificial intelligence technologies in general and ChatGPT in particular. The introduction of the “Perceived controllability and transparency” construct at the heart of the model captures teachers’ concerns about mastering and understanding the tool. In addition, the addition of “ethical concerns” as a moderating variable of general attitude demonstrates the importance of ethical, moral and deontological considerations in the adoption process. Furthermore, the distinction between “perceived educational gains” and “perceived pedagogical risks” within perceived usefulness provides a better understanding of teachers’motivations and disincentives.
It is therefore important to define a very clear ethical framework for the use of this tool and to put in place a well-defined institutional policy. This framework can be implemented by a protocol involving teachers, students and AI experts. This charter should define the fundamental principles (academic integrity, student interests, the primacy of human subjectivity, non-discrimination and data transparency) and the conditions for AI use (prohibited during assessments, limited by level and mandatory training for students). A monitoring committee must be set up to ensure that the charter is implemented and respected.
It is also important to encourage experimentation with ChatGPT through training courses and subscriptions offered to teaching teams to promote its integration and capitalize on the gains identified in the model. These training courses can be customized to suit individual needs and profiles. Similarly, awareness campaigns are needed to limit the risks associated with generative AI, including data security, the verification of the information generated and the risk of addiction. And let us not forget that AI cannot replace the sensitivity of a teacher, nor can it feel emotions when managing students.
The introduction of AI in general and ChatGPT in particular will not mean the disappearance of the teacher. Nevertheless, it will lead to a transformation of this profession. The role of the teacher must no longer be solely that of transmission and assessment. They must evolve towards the skills of co-construction with the student, using the tools made available. Given the multiplicity of data sources, the teachers’ expectations must evolve from teaching the student to search for information to sorting out existing information, detecting the true from the false and developing a sense of discernment.
Our findings have important implications for the enterprise sector in higher education, particularly with regard to performance promotion. Indeed, generative AI tools, such as ChatGPT, can facilitate distance learning solutions while stimulating student engagement and personalized learning experiences (Mallah Boustani and Merhej Sayegh 2021). ChatGPT is identified as “an assistant” by some of our interviewees, who anticipate an AI tutoring role for students. AI could, for example, help them identify which concepts need to be explored further, which options need to be chosen, which exercises they still need to work on. AI could become a veritable personalized guidance counsellor. This ability to manage large quantities of information and automate routine processes can streamline administrative tasks (such as resource management), reducing the need for human and financial resources.
Furthermore, these solutions also support the ongoing professional development of instructors/teachers and stakeholders, as AI can provide real-time information on market trends and emerging corporate strategies, enabling educators to cultivate a better-prepared workforce for the future.
From an educational policy perspective, business schools using generative AI can lead the way in supporting digital transformation and sustainable practices in higher education. By integrating AI-based tools into their courses, they help students acquire skills that are essential in today’s working world, such as digital literacy, critical thinking and ethical decision-making. This is consistent with the broader aims of sustainable education, as it prepares students to tackle important global issues such as resource management, ethical business practices and environmental conservation.
It is important to emphasize here the impact that political leaders can have in the field: if school directors show a strong willingness and involvement in the use and adoption of these tools, then training sessions will be offered, and teachers will be accompanied and encouraged to use these tools, which corresponds to the traditional model of primary adoption (the organization decides to deploy a technology); then comes the decision of actual use by employees, which is secondary adoption (Gallivan 2001). However, Bidan et al. (2020) have proposed an inverted model, focusing on latent or dormant technologies. As the technology is available, the employee can seize it, accept it and use it without the organization even being aware of it. These are informal practices, born of the personal experience of the individual and/or those around him. Only then is the technology partly proposed by the organization, which is presented with a sort of fait accompli. The acceptance and adoption of generative AI seems to borrow from both models: indeed, the interviews enabled us to observe both the traditional and the inverted model, depending on the nature of the establishments concerned (private or public business schools), or the link with technology and the pedagogical innovation of the people interviewed.
In the long run, implementing AI in educational and business contexts can encourage a more responsible use of technology, instil a sense of accountability, and promote transparent decision-making processes, ultimately fostering a sustainability culture that extends beyond academic institutions and into the corporate world.

6. Conclusions

The main contribution of our study lies in questioning the traditional Technology Acceptance Model (TAM) and proposing a revisited TAM adapted to AI. This model takes into account ethics, data controllability, the control of the tool and its settings, knowledge of how personal data is used and intellectual property—elements often neglected in existing models. This enriched approach aims not only to address teachers’ concerns about data security and privacy (Grassini 2023; Rahman and Watanobe 2023), but also to strengthen transparency and user trust, key aspects for a sustainable and responsible adoption of AI in education (Memarian and Doleck 2023). By introducing this new version of the TAM, we aim, based on field results, to establish a theoretical framework to better understand the factors driving teachers’ acceptance of ChatGPT. This includes the identification of potential benefits such as time savings and the automation of repetitive tasks, but also the risks associated with plagiarism and the lowering of students’ cognitive abilities. This model could thus serve as a basis for future research exploring in greater detail the implications of AI in education, particularly in diverse and multicultural contexts.
Furthermore, the contribution of this article lies in its empirical approach, which responds to a gap identified in the literature. Most studies on ChatGPT in education are based on theoretical analysis or commentary, with little empirical data (Pradana et al. 2023; Memarian and Doleck 2023). By collecting qualitative data from teachers, this research enriches the understanding of the real dynamics of ChatGPT adoption, considering the reality of practices and expectations specific to the educational context (Abdaljaleel et al. 2024).
This work has enabled us to study teachers’ perceptions. Some see it as an opportunity for pedagogical enrichment, others as a risk to the quality of learning and the cognitive capacity of students. While ChatGPT is appreciated for its speed and ability to synthesize, the absence of critical reflection and the risk of cheating and algorithmic bias are causes for concern. This study thus clarified the role of ethics and the mastery of the tool in the acceptance of ChatGPT, highlighting the importance of appropriate, ongoing training and supervision. This ethical and pedagogical aspect is frequently highlighted in the literature as a condition for the beneficial integration of AI into educational practices, but remains underexplored. For a successful integration of ChatGPT, it becomes essential to formulate and formalize educational policies that take into consideration these dimensions of control and safety, thus offering a balance between technological innovation and the protection of fundamental pedagogical interests.
Despite the interest in this research on the integration of ChatGPT into teachers’ pedagogical practices, it nevertheless presents certain limitations that deserve to be highlighted in order to contextualize the results and guide future research.
First of all, this model adds other external variables than the basic model. It is important to test it empirically to validate the relationships between the different variables and the most influential factors. A large-scale questionnaire is needed to validate the model. Secondly, the exploratory work was carried out in a limited geographical area and reflects only local perceptions and experiences. It would be more interesting to extend the research to other geographical contexts (Boustani and Chammaa 2023). International comparative studies would enable us to assess other variables, such as cultural specificities, which may influence the adoption of ChatGPT. From a methodological point of view, this research was conducted on the basis of semi-structured interviews. Other methodologies, such as observation and longitudinal studies, could provide a better understanding of the conditions under which ChatGPT is used. This study does not take into account the diversity and specificity of the different domains. It would also be interesting to conduct targeted research on specific disciplines and measure the impact of ChatGPT adoption between the hard sciences and the humanities and social sciences, for example. Exploring these additional dimensions will enable us to better grasp the potential and challenges of using ChatGPT in particular, and generative AI in general, in the field of education.
In terms of perspectives, this article paves the way for future research to explore how students perceive the integration of ChatGPT into their educational pathways, examining in particular the influence of this technology on their autonomous learning (memorization, creativity, assimilation and comprehension) and their capacity for critical analysis. Indeed, the majority of respondents believe that ChatGPT could weaken critical thinking, the ability to reflect and solve problems autonomously, particularly when students become too dependent on AI-generated suggestions. These aspects call for more detailed explorations, particularly on the impact of the tool on their academic performance and results. Studying ChatGPT’s impact on students’ motivation and engagement, as well as their ability to learn self-discipline and digital ethics, could also enrich the understanding of the use of this technology in an educational context (Pradana et al. 2023).
In conclusion, this exploratory study marks the starting point for a structured research project articulated in several phases, each aimed at deepening our understanding of this pedagogical revolution. The first phase will consist of a large-scale qualitative study of a wide range of teachers (permanent or part-time teachers, with or without research activity). This survey will not only enable us to refine the conceptual model, but also to question the uses of ChatGPT either in professional or personal dimensions. It will also provide an opportunity to draw up typical teacher-user profiles, the outlines of which are already emerging from this exploratory study. The second phase will focus on students’ perceptions and practices through a quantitative study. The aim of this phase will be to shed light, from a dual perspective, on the dynamics of appropriating generative AI by comparing the views of teachers and students.
We are convinced that this ambitious study program will pave the way for a more holistic and in-depth understanding of the pedagogical mutations underway

Author Contributions

Conceptualization, methodology, validation, formal analyses, investigation, writing, N.A.-C. and L.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research and the APC were funded by the Oise University Institute of Technology (order number: 4500308244).

Institutional Review Board Statement

All participants were informed that their anonymity is assured. We explained why the research is being conducted and how their data will be used. All participants agreed.

Informed Consent Statement

We obtain verbal informed consent from all participants in this study.

Data Availability Statement

Data are available on request.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Interview Guide

  • Presentation of the interviewees
  • Personal information (Sex, Age, Subjects taught, Research activity,…)
  • Link to technology and to innovative pedagogical practices
  • The experience of using ChatGPT
  • Have you already used ChatGPT?
  • IF YES:
  • In what context(s)? Describe your experience
  • How often do you use it?
  • Are you satisfied with the results?
  • IF NO: why not?
  • FOR ALL (those who use it and those who do not):
  • What are the main advantages of ChatGPT for you?
  • What are its main drawbacks?
  • Its perceived usefulness
  • What impact do you think ChatGPT has had or could have on enriching your teaching techniques? Has it or could it transform the way you teach?
  • Can you give us specific examples of how ChatGPT has helped or could help you achieve your educational goals more quickly and easily (creation of exercises, MCQs, quizzes, scenarios, articles, critical thinking, etc.)?
  • Its perceived ease of use
  • Do you find ChatGPT easy to use?
  • Why or why not?
  • What obstacles did you encounter? How did you overcome them?
  • To master ChatGPT, did you feel the need to use specific educational resources, such as training courses or guides (tutorials)?
  • Its perceived Usefulness
  • In your life as a teacher, how do you perceive the importance of ChatGPT as a useful aid/asset?
  • The users’ attitude to generative AI
  • Are you enthusiastic or sceptical about using ChatGPT? Why
  • Their intentions to use
  • Do you intend to continue using ChatGPT for your teaching needs? What are the reasons behind your decision?
  • Would you recommend the use of ChatGPT? If so, in what area (pedagogical, personal, research, etc.)? under what conditions?
  • Subjective opinions and standards
  • What are the obstacles to using ChatGPT?

References

  1. Abdaljaleel, Maram, Muna Barakat, Mariam Alsanafi, Nesreen A. Salim, Husam Abazid, Diana Malaeb, Ali Haider Mohammed, Bassam Abdul Rasool Hassan, Abdulrasool M. Wayyes, Sinan Subhi Farhan, and et al. 2024. A multinational study on the factors influencing university students’ attitudes and usage of ChatGPT. Scientific Reports 14: 1983. [Google Scholar]
  2. Adeshola, Ibrahim, and Adeola Praise Adepoju. 2023. The opportunities and challenges of ChatGPT in education. Interactive Learning Environments, 1–14. [Google Scholar] [CrossRef]
  3. Al-Mughairi, Habiba, and Preeti Bhaskar. 2024. Exploring the factors affecting the adoption AI techniques in higher education: Insights from teachers’ perspectives on ChatGPT. Journal of Research in Innovative Teaching & Learning. [Google Scholar] [CrossRef]
  4. Baker, Hayden P., Emma Dwyer, Senthooran Kalidoss, Kelly Hynes, Jennifer Wolf, and Jason A. Strelzow. 2024. ChatGPT’s ability to assist with clinical documentation: A randomized controlled trial. JAAOS-Journal of the American Academy of Orthopaedic Surgeons 32: 123–29. [Google Scholar]
  5. Bardin, Laurence. 2013. L’analyse de contenu, 2nd ed. Paris: PUF. [Google Scholar]
  6. Berger-Douce, Sandrine, Régis Coeurderoy, Christina Constantinidis, Véronique Favre-Bonté, Olivier Germain, Gilles Guieu, Frank Janssen, Karim Messeghem, Christophe Schmitt, Josée St-Pierre, and et al. 2023. PME et entrepreneuriat: Ajuster la recherche aux réalités sociales, environnementales et économiques du monde des affaires en transformation. Revue internationale PME 36: 6–25. [Google Scholar]
  7. Bidan, Marc, Guillaume Biot-Paquerot, Mathieu-Claude Chaboud, and Frank-Mahé Lentz. 2020. Inversion du domaine de l’adoption: Les technologies latentes. Management & Datascience 4. [Google Scholar]
  8. Boustani, Nada Mallah, and Claude Chammaa. 2023. Youth Adoption of Innovative Digital Marketing and Cross-Cultural Disparities. Administrative Sciences 13: 151. [Google Scholar] [CrossRef]
  9. Dabbous, Amal, and Nada Mallah Boustani. 2023. Digital explosion and entrepreneurship education: Impact on promoting entrepreneurial intention for business students. Journal of Risk and Financial Management 16: 27. [Google Scholar] [CrossRef]
  10. Davis, Fred D. 1986. A Technology Acceptance Model for Empirically Testing New End-User Information Systems: Theory and Results. Cambridge: Massachusetts Institute of Technology. [Google Scholar]
  11. Davis, Fred D. 1989. Technology acceptance model: TAM. Al-Suqri, MN, Al-Aufi, AS: Information Seeking Behavior and Technology Adoption 205: 219. [Google Scholar]
  12. Gallivan, Michael J. 2001. Organizational adoption and assimilation of complex technological innovations: Development and application of a new framework. ACM SIGMIS Database: The DATABASE for Advances in Information Systems 32: 51–85. [Google Scholar] [CrossRef]
  13. Garg, Ravindra Kumar, Vijeth L. Urs, Akshay Anand Agarwal, Sarvesh Kumar Chaudhary, Vimal Paliwal, and Sujita Kumar Kar. 2023. Exploring the role of ChatGPT in patient care (diagnosis and treatment) and medical research: A systematic review. Health Promotion Perspectives 13: 183. [Google Scholar] [CrossRef]
  14. Grassini, Simone. 2023. Shaping the future of education: Exploring the potential and consequences of AI and ChatGPT in educational settings. Education Sciences 13: 692. [Google Scholar] [CrossRef]
  15. Ibrahim, Roslina, N. S. Leng, R. C. M. Yusoff, G. N. Samy, Suraya Masrom, and Z. I. Rizman. 2017. E-learning acceptance based on technology acceptance model (TAM). Journal of Fundamental and Applied Sciences 9: 871–89. [Google Scholar] [CrossRef]
  16. Karthikeyan, C. 2023. Literature review on pros and cons of ChatGPT implications in education. International Journal of Science and Research (IJSR) 12: 283–91. [Google Scholar] [CrossRef]
  17. King, William R., and Jun He. 2006. A meta-analysis of the technology acceptance model. Information & Management 43: 740–55. [Google Scholar]
  18. Kleinpeter, Édouard. 2020. Intelligence artificielle: L’hubris de contrôle des trans-humanistes en échec face à la boîte noire. Corps & Psychisme 1: 69–76. [Google Scholar]
  19. Klyshbekova, Maira, and Pamela Abbott. 2024. ChatGPT and Assessment in Higher Education: A Magic Wand or a Disruptor? Electronic Journal of e-Learning 22: 30–45. [Google Scholar]
  20. Kozinets, Robert. 2019. Netnography: The Essential Guide to Qualitative Social Media Research, 3rd ed. London: SAGE. [Google Scholar]
  21. Kwan, Christopher Chung Lim. 2023. Exploring ChatGPT-Generated assessment scripts of probability and engineering statistics from bloom’s taxonomy. In International Conference on Technology in Education. Singapore: Springer, pp. 275–86. [Google Scholar]
  22. Langlais, Pierre-Carl. 2023. ChatGPT: Comment ça Marche? Sciences Communes. Available online: https://scoms.hypotheses.org/1059 (accessed on 15 April 2024).
  23. Lo, Chung Kwan. 2023. What is the impact of ChatGPT on education? A rapid review of the literature. Education Sciences 13: 410. [Google Scholar] [CrossRef]
  24. Mallah Boustani, Nada, and May Merhej Sayegh. 2021. E-learning: Factors affecting students online learning during COVID-19 quarantine in a developing country. In International Conference on Information and Knowledge Systems. Cham: Springer International Publishing, pp. 17–28. [Google Scholar]
  25. Memarian, Bahar, and Tenzin Doleck. 2023. ChatGPT in education: Methods, potentials and limitations. Computers in Human Behavior: Artificial Humans 1: 100022. [Google Scholar] [CrossRef]
  26. Menon, Devadas, and Kalyan Shilpa. 2023. “Chatting with ChatGPT”: Analyzing the factors influencing users’ intention to Use the Open AI’s ChatGPT using the UTAUT model. Heliyon 9: e20962. [Google Scholar] [CrossRef]
  27. Mhlanga, David. 2023. Open AI in education, the responsible and ethical use of ChatGPT towards lifelong learning. In FinTech and Artificial Intelligence for Sustainable Development: The Role of Smart Technologies in Achieving Development Goals. Cham: Springer Nature, pp. 387–409. [Google Scholar]
  28. Mogavi, Reza Hadi, Chao Deng, Justin Juho Kim, Pengyuan Zhou, Young D. Kwon, Ahmed Hosny Saleh Metwally, Ahmed Tlili, Simone Bassanelli, Antonio Bucchiarone, Sujit Gujar, and et al. 2023. Exploring user perspectives on chatgpt: Applications, perceptions, and implications for ai-integrated education. arXiv arXiv:2305.13114. [Google Scholar]
  29. Mohammadi, Hossein. 2015. Investigating users’ perspectives on e-learning: An integration of TAM and IS success model. Computers in Human Behavior 45: 359–74. [Google Scholar] [CrossRef]
  30. Montenegro-Rueda, Marta, José Fernández-Cerero, José María Fernández-Batanero, and Eloy López-Meneses. 2023. Impact of the implementation of ChatGPT in education: A systematic review. Computers 12: 153. [Google Scholar] [CrossRef]
  31. Moussavou, Jean. 2023. ChatGPT dans la rédaction scientifique: Des perspectives prometteuses, des préoccupations légitimes. Management & Datascience 7: 23913. [Google Scholar]
  32. Mugo, David Gitumu, Kageni Njagi, Bernard Chemwei, and Jared Ochwagi Motanya. 2017. The technology acceptance model (TAM) and its application to the utilization of mobile learning technologies. British Journal of Mathematics & Computer Science 20: 1–8. [Google Scholar]
  33. Paillé, Pierre, and Alex Mucchielli. 2021. L’analyse qualitative en sciences humaines et sociales, 5th ed. Paris: Armand Colin. [Google Scholar]
  34. Pradana, Mahir, Hanifah Putri Elisa, and Syarifuddin Syarifuddin. 2023. Discussing ChatGPT in education: A literature review and bibliometric analysis. Cogent Education 10: 2243134. [Google Scholar] [CrossRef]
  35. Rafique, Hamaad, Alaa Omran Almagrabi, Azra Shamim, Fozia Anwar, and Ali Kashif Bashir. 2020. Investigating the acceptance of mobile library applications with an extended technology acceptance model (TAM). Computers & Education 145: 103732. [Google Scholar]
  36. Rahman, Md Mostafizer, and Yutaka Watanobe. 2023. ChatGPT for education and research: Opportunities, threats, and strategies. Applied Sciences 13: 5783. [Google Scholar] [CrossRef]
  37. Scherer, Ronny, Fazilat Siddiq, and Jo Tondeur. 2019. The technology acceptance model (TAM): A meta-analytic structural equation modeling approach to explaining teachers’ adoption of digital technology in education. Computers & education 128: 13–35. [Google Scholar]
  38. Venkatesh, Viswanath. 2022. Adoption and use of AI tools: A research agenda grounded in UTAUT. Annals of Operations Research 308: 641–52. [Google Scholar] [CrossRef]
  39. Venkatesh, Viswanath, and Fred D. Davis. 2000. A theoretical extension of the technology acceptance model: Four longitudinal field studies. Management Science 46: 186–204. [Google Scholar] [CrossRef]
  40. Venkatesh, Viswanath, and Hillol Bala. 2008. Technology acceptance model 3 and a research agenda on interventions. Decision Sciences 39: 273–315. [Google Scholar] [CrossRef]
  41. Venkatesh, Viswanath, James Y. L. Thong, and Xin Xu. 2012. Consumer Acceptance and Use of Information Technology: Extending the Unified Theory of Acceptance and Use of Technology. MIS Quarterly: Management Information Systems 36: 157–78. [Google Scholar] [CrossRef]
  42. Venkatesh, Viswanath, Michael G. Morris, Gordon B. Davis, and Fred D. Davis. 2003. User acceptance of information technology: Toward a unified view. MIS Quarterly 27: 425–78. [Google Scholar] [CrossRef]
  43. Yousafzai, Shumaila Y., Gordon R. Foxall, and John G. Pallister. 2007a. Technology acceptance: A meta-analysis of the TAM: Part 1. Journal of Modelling in Management 2: 251–80. [Google Scholar] [CrossRef]
  44. Yousafzai, Shumaila Y., Gordon R. Foxall, and John G. Pallister. 2007b. Technology acceptance: A meta-analysis of the TAM: Part 2. Journal of Modelling in Management 2: 281–304. [Google Scholar] [CrossRef]
  45. Yu, Hao. 2024. The application and challenges of ChatGPT in educational transformation: New demands for teachers’ roles. Heliyon 10: e24289. [Google Scholar] [CrossRef]
  46. Zhu, Chenjia, Meng Sun, Jiutong Luo, Tianyi Li, and Minhong Wang. 2023. How to Harness the Potential of ChatGPT in Education? Knowledge Management & E-Learning 15: 133–52. [Google Scholar]
Figure 1. Ten special issues of AI studies (adapted from Venkatesh 2022).
Figure 1. Ten special issues of AI studies (adapted from Venkatesh 2022).
Admsci 14 00325 g001
Figure 2. Generative AI Acceptance and Adoption Model (AIA2M).
Figure 2. Generative AI Acceptance and Adoption Model (AIA2M).
Admsci 14 00325 g002
Table 1. The opportunities and threats of using ChatGPT in education.
Table 1. The opportunities and threats of using ChatGPT in education.
Opportunities Threats
Assistant for teachersHelp with course preparation (Lo 2023; Rahman and Watanobe 2023)
Can generate plans and develop concepts (Memarian and Doleck 2023)
Help in creating educational content (Lo 2023)
Help with creating MCQs, QUIZZs (Yu 2024)
Help with evaluation and correction (Karthikeyan 2023)
Reduce teachers’ mental workload (Memarian and Doleck 2023)
Ethics and academic integrityReliability of information and data generated (Yu 2024; Rahman and Watanobe 2023)
The problem of intellectual property (Mhlanga 2023)
The problem of authenticity (Mhlanga 2023)
Integrity of examinations and assessments (Pradana et al. 2023)
Facilitates student plagiarism (Lo 2023; Grassini 2023; Rahman and Watanobe 2023; Yu 2024)
Threatens academic integrity (Lo 2023; Grassini 2023; Yu 2024and creativity (Karthikeyan 2023)
The learning experiencePromotes self-training: individualized, self-directed training (Yu 2024)
Can provide real-time feedback (Lo 2023; Grassini 2023; Rahman and Watanobe 2023; Memarian and Doleck 2023)
Task optimization and performance (Yu 2024)
Personalized learning (Mhlanga 2023; Grassini 2023; Karthikeyan 2023)
Improving access to information (Mhlanga 2023; Rahman and Watanobe 2023)
Enriches teachers’ pedagogical and educational experience (Karthikeyan 2023; Montenegro-Rueda et al. 2023)
Promoting educational digitization (Yu 2024)
Fast, easy access to information (Karthikeyan 2023; Rahman and Watanobe 2023)
Technical aspects: bias and discriminationCapabilities vary according to field (Lo 2023)
Lack of equity and non-discrimination (Mhlanga 2023)
Lack of transparency on algorithms (Memarian and Doleck 2023)
Presence of algorithmic bias (Mhlanga 2023)
Production of incorrect or fictitious information (Lo 2023; Grassini 2023; Yu 2024)
Student tutorCan question students and adapt to their evolving understanding (Montenegro-Rueda et al. 2023)
Personalized learning for students (Montenegro-Rueda et al. 2023)
Pedagogical support (Mhlanga 2023)
Developing students’ writing skills (Karthikeyan 2023; Montenegro-Rueda et al. 2023)
Legal aspect: privacy and securityProtection of privacy (Memarian and Doleck 2023)
Use of collected data for improvement purposes (Memarian and Doleck 2023)
The use of sensitive data (Al-Mughairi and Bhaskar 2024)
Teacher–student interactionFacilitates collaboration, discussions and debates (Lo 2023; Montenegro-Rueda et al. 2023)
Improves asynchronous communication (B. Memarian and Doleck 2023)
Psychological aspect: socio-psychological effectDependency (Mhlanga 2023; Karthikeyan 2023; Al-Mughairi and Bhaskar 2024)
Lack of human interaction (Al-Mughairi and Bhaskar 2024)
Decrease in students’ cognitive development and critical thinking (Rahman and Watanobe 2023)
Obsolescence of traditional teaching and learning skills (Karthikeyan 2023)
Table 2. Comparison of the different versions of the TAM and UTAUT models.
Table 2. Comparison of the different versions of the TAM and UTAUT models.
ModelModerator VariablesExplanatory VariablesExplained Variables
TAM 1
Davis (1986, 1989)
External Variables1. Perceived Usefulness (PU)
2. Perceived
Ease-of-Use (PEOU)
Attitude
Intention to Use
Actual Use/Usage Behaviour
TAM 2
Venkatesh and Davis (2000)
External Variables +
Seven antecedents of PU:
  • Subjective Norm
  • Image
  • Job Relevance
  • Output Quality
  • Result Demonstrability
  • Experience
  • Voluntariness
1. PU
2. PEOU
Behavioural
Intention
Use Behaviour
TAM 3
Venkatesh and Bala (2008)
External Variables
Seven antecedents of PU +
Six antecedents of PEOU:
  • Computer self-efficacy
  • Perceptions of external control
  • Computer anxiety
  • Computer playfulness
  • Perceived enjoyment
  • Objective usability
1. PU
2. PEOU
Behavioural
Intention
Use Behaviour
UTAUT 1
Venkatesh et al. (2003)
Four moderating variables:
  • Gender
  • Age
  • Experience
  • Voluntariness of use
1. Performance Expectancy
2. Effort Expectancy
3. Social Influence
4. Facilitating Conditions
Behavioural
Intention
Use Behaviour
UTAUT 2
Venkatesh et al. (2012)
Three moderating variables:
  • Gender
  • Age
  • Experience
1. Performance Expectancy
2. Effort Expectancy
3. Social Influence
4. Facilitating Conditions
+
5. Hedonic Motivation
6. Price Value
7. Habit
Behavioural
Intention
Use Behaviour
Table 3. Characteristics of the sample.
Table 3. Characteristics of the sample.
Interview NumberSexDateField of ExpertiseAge Bracket
I1M19 March 2024Economy50–60
I2M19 March 2024Negotiation, Management40–50
I3F21 March 2024Chemistry, Environment50–60
I4M18 March 2024Marketing60–70
I5F21 March 2024Expression communication50–60
I6M18 March 2024Management50–60
I7F23 April 2024Mathematics30–40
I8M2 April 2024Finance50–60
I9F23 April 2024Economy40–50
I10M16 April 2024Numerical mathematics20–30
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mourtajji, L.; Arts-Chiss, N. Unleashing ChatGPT: Redefining Technology Acceptance and Digital Transformation in Higher Education. Adm. Sci. 2024, 14, 325. https://doi.org/10.3390/admsci14120325

AMA Style

Mourtajji L, Arts-Chiss N. Unleashing ChatGPT: Redefining Technology Acceptance and Digital Transformation in Higher Education. Administrative Sciences. 2024; 14(12):325. https://doi.org/10.3390/admsci14120325

Chicago/Turabian Style

Mourtajji, Loubna, and Nathalie Arts-Chiss. 2024. "Unleashing ChatGPT: Redefining Technology Acceptance and Digital Transformation in Higher Education" Administrative Sciences 14, no. 12: 325. https://doi.org/10.3390/admsci14120325

APA Style

Mourtajji, L., & Arts-Chiss, N. (2024). Unleashing ChatGPT: Redefining Technology Acceptance and Digital Transformation in Higher Education. Administrative Sciences, 14(12), 325. https://doi.org/10.3390/admsci14120325

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop