A Decision-Making Model for Adopting Al-Generated News Articles: Preliminary Results

: This study was conducted to present a consumer acceptance model for artiﬁcial intelligence (AI)-generated news articles in review of the most signiﬁcant factors in the context of accepting AI-generated news articles. To this end, a survey was conducted regarding ﬁve key factors among individuals handling news production at media organizations who were familiar with AI-related news: beneﬁts of technology utilization, recognized values, media reliability, content quality, and information perception. These factors were classiﬁed into 19 speciﬁc evaluation items in order to determine relative importance. According to the results of pairwise comparison analysis among elements through the analytic hierarchy process, hereunder referred to as AHP (analytic hierarchy process), it turned out that the importance of media reliability and content levels was relatively high, while the importance of recognized values and information perception was relatively low. It also turned out that among 19 speciﬁc items, the importance of institutional reliability, which was a subordinate item under media reliability, was of utmost value. Even if news articles were based on AI technology, their readers were more likely to be a ﬀ ected by reliability, which is a major attribute of journalism, than technical factors.


Introduction
AI (artificial intelligence), which is a key element to the 4th Industrial Revolution, has started to affect the general media and journalism significantly under the name of "robot journalism" [1]. Robot journalism has been known to the general public since 2010 as a concept of generating news articles automatically by means of AI-based algorithms with no involvement of human journalist [2]. As AI extracts natural language phrases (NLP) from various data sets automatically, the traditional news industry has transformed [3]. When it comes to promptness and accuracy, which are regarded as critical in news articles, robot journalist now surpass human journalist.
Forbes, a U.S. Economist magazine, is the news agency that introduced AI-generated articles in 2012 for the first time. In 2013, the Associated Press introduced Wordsmith and promptly reported a large amount of data regarding the business performance of enterprises. The LA Times reported the earthquake situation real-time by utilizing "Quakebot". In 2016, the Washington Post started to adopt "Heliograf" in the Rio Olympic Games to report sports game results in a prompt and simple manner. In Korea, AI-generated news started to be used in 2015. In January 2016, the Financial News introduced IamFNBOT, an AI robot journalist. The Herald Business introduced HeRo, an AI robot journalist for English articles. eToday and Electronic Times utilized e2BOT and @NEWS respectively in their news article production. While these robot journalist handled existing news reports on corporate information and stocks, including current stock condition and fluctuation, corporate performance, etc., Yonhap News collected and analyzed data from every soccer game of the England Premier League and provided news reports by using SoccerBot.
At present, robot journalism based on AI-generated news articles is expected to be established over the general media market [4]. As such, media agencies now introduce an AI-based news generation system as an efficient media management system [5]. As demands for AI-generated news among news consumers are increasing, the introduction and commercialization of AI-generated articles is recognized, no longer as a matter of choice, but a matter of eventuality.
In its initial stage, research on AI-generated articles or robot journalism in the media industry focused on the area of technology development. Specifically, the focus was on technology development and the issue of whether AI-generated articles could replace, and be as significant as, human-written articles. In this process, one of the main subjects of research was the dynamic conflict between robot journalist and human journalist [6]. Since AI-generated articles started to be commercialized, the scope of research has expanded to cover feedbacks from news consumers [2], ethics of algorithms, and legal issues such as article copyright [7,8].
However, there has been little research on the market's perception of, or response to, AI, as well as on receptive attitudes in terms of news consumption or consumption values from the market's perspective while AI-generated articles are currently being commercialized. Also, in the news media environment with high technology and innovative businesses, the presses need the decision and creation of Al-generated news articles to communicate with people under uncertainty and information asymmetry. Accordingly, this study defines determinant factors of the intention to accept AI-generated news articles and presents a model for the consumers' decision to accept AI-generated news by analyzing determinant factors affecting such acceptance attitudes most significantly. To this end, the analytic hierarchy process (AHP) methodology related to decision-making models was selected in a survey conducted on an expert group in the media industry in order to present a model consisting of objective factors of acceptance and strategic significances for the popularization of AI-generated news.

Artificial Intelligence (AI) Technology and Journalism Change
AI is based on the continuous development of computers, big data, and algorithms. While it has long been in its dark age, it is now advancing in various aspects, such as external cognition, logics, inference, prediction, etc., in the utilization of NLP deep-learning. As its business potentials are shown, investment in AI areas is increasing, particularly among global ICT enterprises [9]. In the face of significant changes over the global society, such as the 4th Industrial Revolution and COVID-19, AI applications including robots and untact technology are spreading rapidly. Against this background, robot journalism is likewise strengthening as AI technology is adopted in the media industry.
Traditionally, humans have been the main subject in journalism from coverage to reporting. In this area, which has been monopolized completely by humans, AI is now intervening and expanding its boundary [10]. Robot journalism has begun initially on the level of computer-assisted reporting (CAR), but at present, the trend is changing: robot journalism determines whether to produce certain news based on mathematical algorithms, with robots serving as inhuman agents practicing acts of article generation, which have been exercised by human journalist [11].
As shown in Table 1, the period of computer-assisted reporting has been followed by computational journalism on the basis of big data and network information. Coleman [12] defined computational journalism as "complementing tasks of journalism by combining knowledge from social science with algorithms and data". Computational journalism is used as a supplementary mechanism of the media in the process of searching for new information through the analysis of patterns hidden in data [4]. Since AI-based robot journalism has spread recently, however, articles have been prepared by AI in such ways as machine learning and deep learning. Furthermore, editing, distribution, and service are automated [13].  [11] and Lee (2019) [14].
Accordingly, robot journalism is widely understood as a concept of generalizing a series of procedures and results for automated article preparation according to rules defined by AI algorithms with human intervention minimized [14]. In general, the process in which an AI algorithm generates a news article automatically is known to take the five following steps: (1) data crawling, (2) event extraction, (3) key event detection, (4) mood detection, and (5) news article generation [3]. In the case of AI-based robot journalist, it is possible to manage the distribution and advertisement of even customized articles based on customer information. Just like other industry sectors, the roles of traditional media organizations and journalists are affected by this changing paradigm in the field of the press [15].
Just as there have been various controversies on the evaluation of paintings by AI and human artists, there needs to be a discussion on a societal level now regarding the acceptance and values of articles generated by AI robots on the basis of big data and learning algorithms in comparison with articles written by humans on the basis of journalism and ideas [16]. Previous studies viewed robot-generated articles as more descriptive, informative, accurate, trustworthy, objective, and uninteresting than those written by humans. Meanwhile, articles written by human journalists were discerned to be coherent, well-written, clear, and pleasant to read [17,18].
Some previous studies also pointed out that the accuracy and reliability of AI-generated news articles were doubted [19]. The main point of critics was that there is limitation in generating accurate and analytic articles since the intellectual level of robot journalist is inferior to that of humans [20]. However, AI algorithms are advancing gradually, and some with a positive view prospect that robot journalists will be able to enhance the accuracy of articles, judge values, and understand contexts based on a thorough analysis of big data, thus producing more creative news articles [21]. After all, technology continues to advance and the spread of robot journalism will require a new standard and evaluation on the part of the general public when it comes to the acceptance of news articles in the global society [22,23].

Subsection
Previous studies on the acceptance and selection of news articles applied evaluation standards of journalism based on political tendencies, social ideas, critical views, and values of contents [24,25] or emphasized the reliability and inclination of entities producing news articles, such as journalists and media outlets [26,27]. In the case of AI-generated news articles, however, the receptive attitude towards technology and information, as well as the criteria of perception or appreciation of news content may affect the selection or acceptance of news articles. Thus, factors of news article selection need to be considered in a more complex perspective.

Technology Using Benefits
Basically, the acceptance of digital transformation, AI, and robotics in the 4th Industrial Revolution assumes the technology acceptance model (TAM) proven by Davis [28]. The TAM is viewed as a model with the most significant influence on studies of user behavior characteristics in the area of information technology (IT), which examines the intention of new technology adoption [29,30]. The TAM is a modified model [28] of the theory of reasoned action (TRA), appropriate for the context of new technology adoption. Proposed by Fishbein and Ajzen [31], the TRA is based on the assumption that an individual's intent of behavior is directly connected to actual behavior and is decided by one's attitude towards such behavior and relevant subjective norms. In contrast, the TAM emphasizes perceived usefulness and perceived ease of use based on the objectivity and attitude of behavioral intention. Among consumer acceptance studies on new technology-based media contents as well, the TAM was widely utilized either separately or in combination with another theory [32].

Perceived Value
The concept of "perceived values" became a major issue in the area of corporate strategies and marketing strategies aiming to improve customer values in the 1990s. Since then, marketing researchers have conducted studies for the explanation and analysis on customer behaviors [33][34][35] Perceived values are related not only to perceived quality or psychological states, but also to exchange of monetary sacrifices. In the context of service, quality is understood as a benefit and a consumed sacrifice for it on the part of customers. It affects a consumer's certain judgment or behavior intent [36]. The concept of perceived values is explained mainly in two perspectives: economic perspective and cognitive-emotional perspective. The economic perspective corresponds to effectiveness evaluation in relation to the comparison of cost and benefit. The cognitive-emotional perspective indicates evaluation in cognitive psychology regarding consumers' purchasing behaviors, as well as cognition and behavior characteristics in a certain emotional state. Whatever reason there may be to use media, recognizing the values is of importance. Value recognition can affect choices and use continuity [37]. From the users' perspective, in which useful information or joyful time is provided, value recognition affects the characteristics and use intentions of news articles [38,39].

Media Credibility
Reliability is defined as the probability of belief on a certain existence's intent at a certain time point. Reliability in certain information develops when the person perceives the appropriateness of the source of expertise, technology, and experience related to that information, as well as the objectivity and trustworthiness of the contents [40]. Previous studies from an information reliability perspective focused on two different aspects: professionalism and credibility [41]. With regard to news as well, usefulness or effectiveness assumes reliability from the consumers' perspective. Users would select certain news when they are certain about the reliability [42]. When it comes to the act of selecting news, the critical factor is the trust in the producer of articles, such as the media outlet and journalist [43]. The fact that reliability in news-generating entities counts in the process of article consumption is related to special attributes of the article production structure. Since news materials are processed as recognized by journalists, this process is viewed as an autonomous behavior of the news-producing subject [44,45]. News consumers expect that the reliability in a producing subject corresponds to that of the produced articles. When it comes to news articles, therefore, not merely the reliability of articles themselves but also the valuation of brands and the reliability of every individual involved in article production and distribution can affect article consumption directly [46].

Content Quality
"Content quality" means the quality of contents to be consumed. Content quality is acquired when an individual uses a service through a certain device or platform. Particularly, studies related to news contents have examined the relation between the quality of news contents and consumers' satisfaction with the news, the relation between motivation and use of news, and so forth [47][48][49]. According to previous studies, the quality of news contents affects the behavioral intention of users, and in the case of news services based on a new medium, such factors as variety, updatedness, professionalism, etc. affect users' intention, satisfaction, and use continuity. In addition, various factors of content quality evaluation have been introduced, including the timeliness, clarity, completeness, outward appearance, and understandability of news contents [50]. As news contents are digitalized, they are presented in various formats such as photos, texts, graphics, videos, etc. Accordingly, not only the quality of news contents but also various other factors related to visual expressions of contents, updatedness and genuineness of knowledge, social influence of contents, and reliability affect communication elements such as accuracy, understandability, and news acceptance [51,52].

Information Awareness System
Information is a variable since its meaning and value change depending on the living subject and external circumstances. It also can be interpreted differently in the context of users. Particularly in this digital era, participation through sharing and cooperation is more emphasized than accurate delivery when it comes to information [53]. Previous studies on information recognition and design were conducted for various purposes in different areas such as language, psychology, medical science, and design [54,55]. As ICT currently advances, computers, media, video devices, etc. collect, produce, process, preserve, deliver, and utilize information by means of software methods necessary to operate and manage information mechanisms. In this regard, there have been studies on user evaluation and attitudes towards information recognition and access systems [56]. In the case of interactive intellectual services, all information goes through the process of collection, interpretation, and inference. Throughout this process, differences in mutual communication with users, as well as information recognition and processing speed, can affect acceptance intention and satisfaction with services [57]. Since news is also digitalized in the media industry, users' information recognition and processing speed are likely to affect their news selection when the value of news increases as intangible information [58]. While users access information, grasp its flow, appreciate it, and absorb its context and meaning clearly, providing a systematic and highly approachable system becomes essential. The higher the ease of, and the keener the interest in, information access and processing through such an information recognition system are, the higher the likelihood of it affecting users' acceptance positively [59,60].

Analytic Hierarchy Process (AHP)
The analytic hierarchy process (AHP) was originally developed by Professor Thomas Saaty at the University of Pennsylvania in the 1970s [61,62]. This is a decision-making methodology to detect the evaluator's knowledge, experience, and intuition by way of pairwise comparison between elements that form the decision-making hierarchy. The AHP is widely utilized in deriving key factors, setting policy alternatives, and establishing strategies [63,64]. The AHP features the theory's simplicity, easy application, and universal use. The AHP methodology derives the evaluation results based on the total ranks of relative importance among elements determined by each evaluator [65,66]. Accordingly, this study utilized the AHP in evaluating the relative importance of factors for the purpose of clarifying what affects the acceptance of AI-generated news articles significantly.
There are two basic approaches to AHP analysis [67]: The "geometric mean of pairwise comparison" determines the relative importance of each element based on the geometric average, whereas the "arithmetic mean of importance weight" determines importance by calculating the arithmetic average. This study adopted the geometric mean of pairwise comparison, which is widely used to analyze the relative importance of factors comprehensively. In consideration of the fact that the AHP, proposed by Yoo and Kim [68], is a probabilistic methodology to determine the difference between an input variable and a certain model output variable, this study determined the reliability of results, as well as the calculation basis of input variables, by setting their specific weights.

Research Framework and Variables
In reference to previous studies, particularly regarding the five major factors affecting news article users' behavioral intentions, this study selected the following key variables for decision-making on AI-generated news article acceptance: benefits of technology utilization, recognized values, media reliability, content quality, and information perception. As to definitions of each variable used in this study: (1) "Benefits of technology utilization" indicates attributes of usefulness and ease of news consumption while using innovative and new technology that can affect one's decision on AI-generated article acceptance. (2) "Perceived values" indicates the value attributes that stimulate individuals' will and motivation, affecting users' selection of AI-generated articles. (3) "Media reliability" indicates the extent of reliability of various elements related to article generation that affect users' selection of AI-generated articles. (4) "Content quality" indicates the level of article content quality that affects users' acceptance of AI-generated articles. (5) Finally, "information perception" indicates the ease or significance of an access process or system that affects users' acceptance of AI-generated articles (see Figure 1). variable used in this study: (1) "Benefits of technology utilization" indicates attributes of usefulness and ease of news consumption while using innovative and new technology that can affect one's decision on AI-generated article acceptance. (2) "Perceived values" indicates the value attributes that stimulate individuals' will and motivation, affecting users' selection of AI-generated articles. (3) "Media reliability" indicates the extent of reliability of various elements related to article generation that affect users' selection of AI-generated articles. (4) "Content quality" indicates the level of article content quality that affects users' acceptance of AI-generated articles. (5) Finally, "information perception" indicates the ease or significance of an access process or system that affects users' acceptance of AI-generated articles (see Figure 1). Under these five key variables, there were 23 subordinate factors (see Table 2). First of all, subordinate factors of benefits of technology utilization were, based on the previous studies of Davis et al. [29] and Adriani [33], the four following items: effectiveness, usefulness, ease of use, and comprehensive reciprocity. Those of perceived values were, based on the studies of Koritos et al. [37], Mazurek and Malagocka [40], the following: financial value, social value, emotional value, and empirical value. Those of media reliability were, based on the studies of Rimmer and Weaver [42], Usher [47], the following: trust in institution, trust in journalist, trust in technology, and trust in platform. Those of content quality were, based on the studies of DeLone and McLean [48], Allen et al. [53], the following: expression, intelligence, influence, and communication. Finally, those of information perception were, based on the studies of Duchastel [56] and Cui et al. [60], the three following items: access system, concern system, and processing system.  Under these five key variables, there were 23 subordinate factors (see Table 2). First of all, subordinate factors of benefits of technology utilization were, based on the previous studies of Davis et al. [27] and Adriani [31], the four following items: effectiveness, usefulness, ease of use, and comprehensive reciprocity. Those of perceived values were, based on the studies of Koritos et al. [35], Mazurek and Malagocka [38], the following: financial value, social value, emotional value, and empirical value. Those of media reliability were, based on the studies of Rimmer and Weaver [40], Usher [46], the following: trust in institution, trust in journalist, trust in technology, and trust in platform. Those of content quality were, based on the studies of DeLone and McLean [46], Allen et al. [52], the following: expression, intelligence, influence, and communication. Finally, those of information perception were, based on the studies of Duchastel [54] and Cui et al. [58], the three following items: access system, concern system, and processing system.

Research Process and Data Collection
This study presents an objective acceptance model based on factors affecting the acceptance of AI-generated news articles. To this end, a pairwise comparison questionnaire was utilized in the application of the AHP methodology and on the basis of the research framework shown in Figure 1. The questionnaire consisted of 38 questions in total, including 1 subjective question and 37 pairwise comparison questions. A pilot test was conducted among five administrators at a media outlet.
In order to secure study subjects' professionalism and high understanding of AI-generated news, the subject group included 15 journalists working at one of the six major media outlets of Korea introducing and utilizing AI-generated news articles and 15 engineers handling media technology. In consideration of the necessary knowledge and experience regarding AI-generated news, survey participants were selected from those with a field career of at least 5 years. The survey was conducted for 2 months from May to June 2020 by way of one-on-one personal interviews lasting about 30 minutes. Table 3, 90% of the final data samples were male and 10% were female. A total of 26.6% were in their 30s, 36.7% in their 40s, and 36.7% in their 50s. As for experience, 16.7% of the participants had had a career in the media industry for 5-10 years, 30% 10-20 years, and 53.3% more than 20 years. MS Excel program was used for data analysis. In order to secure the reliability of survey results, the consistency ratio (CR) of the comparing agents was checked and found to be under 1, between 0.0015 and 0.0216.

Comparison of Evaluation Variables
As shown in Table 4, media reliability (0.373) was found to be the most significant factor among the five key factors. As for media articles, including AI-generated ones, it turned out that the basic reliability and content level of media articles affected the acceptance of AI-generated articles, rather than technical factors, in the following order: content level (0.197), benefits of technology utilization (0.184), perceived values (0.136), and information perception (0.110).
According to the analysis results of each area, it turned out that among the benefits of technology utilization factors, usefulness (0.364) and effectiveness (0.327) affected the decision of acceptance more significantly, and that among the perceived value factors, financial value (0.551) was of great importance, followed by empirical value (0.195) and then social value (0.157). The factor of the least effect turned out to be emotional value (0.098). Among media reliability factors, it turned out that trust in institution (0.385) was the most important factor affecting the decision of acceptance, and then trust in technology (0.286), trust in platform (0.205), and trust in journalist (0.123). As for the content level, the communication (0.333) was viewed as the most important factor of influence, followed by the intelligence (0.245), influence (0.233), and expression (0.190). Finally, among factors of information perception, system (0.424), processing system (0.356), and concern system (0.219) were the important factors.
In the total list of 19 specific factors, those affecting acceptance of AI-generated news articles most significantly were trust in institution (0.144), trust in technology (0.107), and trust in platform (0.077), all of which turned out to be related to reliability. Another important factor next to them was financial value (0.075). The following factors also turned out to be important: usefulness (0.067), communication (0.065), effectiveness (0.060), intelligence (0.048), and access system (0.046).

Comparison of Evaluation Areas between Journalist and Media Engineering Groups
According to the comparative analysis of evaluation areas among journalist and media engineers, both groups viewed media reliability as the most important factor. While the group of journalists selected the level of content (0.250) and benefits of technology utilization (0.130), the group of media engineers selected the benefits of technology utilization (0.247) and perceived values (0.205) as important factors. This result shows that the difference between the two groups of the critical factors affecting Al-generated news articles (see Table 5).

Comparison of Evaluation Factors between Journalist and Media Engineering Groups
As shown in Table 6, the result of the comparative analysis on each evaluation factor between the two groups indicates that both viewed trust in institution as the most important factor. In the comparison of the two groups based on the top five factors, it turned out that the journalist group recognized as most important in its acceptance of AI-generated news articles the following factors of reliability and article contents in order: trust in technology (0.116), communication (0.092), trust in platform (0.087), and intelligence (0.063). The media engineering group viewed the following factors in order as important: financial value (0.102), effectiveness (0.093), trust in technology (0.091), and usefulness (0.081). This result indicates the difference between the two groups, except in regards to the trust in technology factor. Particularly in the case of the media engineering group, it turned out that the two groups were keenly aware of the importance of the two following factors: benefits of technology utilization and financial value.

Discussion and Conclusions
This study defined acceptance models based on important factors affecting acceptance of AI-generated news and comparatively examined the importance of these factors. The following are the analysis results: First, it turned out that in the evaluation of five key factors, media reliability was the factor affecting the acceptance of AI-generated news most significantly. This corresponds to previous studies [69,70] where reliability was viewed as the most important element both in journalism and AI-generated news. As for AI-generated news articles, the accuracy of data and information is secured. However, there might be some differences even in the case of AI algorithms depending on the organization's learning method and scope, which indicates the limitation regarding complete reliance on the objectivity of intention in the process of generating news articles. Accordingly, although weather forecasts or articles of one-dimensional information may be more accurate and reliable, the reliability and objectivity of news articles are always an important factor in deciding whether they should be produced by humans or learned AI.
Second, among specific factors of media reliability, trust in institution turned out to be the most important factor. This suggests that the reliability of an institution, which manages AI and gives inputs to it can be a more sensitive factor rather than the reliability of robot journalists, AI technology, or relevant platforms. Previous studies likewise pointed out the significant relevance between news articles and the tendency or philosophy of media outlets [71][72][73]. After all, whether the journalist is a human or robot, the organization that supervises and decides news contents can affect robot journalism significantly.
Third, previous studies, in general, have suggested that the acceptance of technology-related products or services relies heavily on technological factors, such as ease and effectiveness [74,75]. In the case of news articles, however, the level of content is more significant than the benefits of technology acceptance according to the present study. Particularly, it turned out that rather than the expression and influence of articles, the level of expertise reflected in articles and their extent of communication were viewed as more important. In the case of AI-generated news articles, it is likely that in contrast with existing news articles, communication problems are due to the low level of knowledge or lack of communication skills on the part of robot journalist. This factor may induce rejection to AI news.
In view of practical significances based on the result above, media outlets need to emphasize organizational reliability in order to promote robot journalism and popularize AI-generated news articles. They need to put forth efforts to more actively share relevant information and promote the excellence and reliability of AI robot journalist generating news articles. In addition to that, efforts should continue to be put forth into sharing information transparently and promoting communication for the brand power and reliability of media outlets that the general public is aware of in consideration of the importance of reliability, not only on the part of robot journalist but also media outlets.
In terms of information perception, the importance of access systems has been emphasized more than that of interests in AI-generated news articles and relevant processing systems. Now it is time to emphasize the process of accessing AI-generated news articles. While AI-generated news is limited to mere objective facts, with the inclination of delivering information on weather, stock prices, sports game results, etc., it is likely that robot journalism will be popularized faster once attention is paid to the ease and variety of approaches to new areas of AI-generated news articles in the future, including society, politics, and economics.
Finally, research results show that in the context of acceptance of AI-generated news articles, reliability and level of contents are viewed as more important than benefits of technology and information perception. For this reason, media outlets need to consider establishing a complex system for both existing articles written by human journalists and AI-generated articles produced by robot journalists, rather than insisting on a company-wide press organization system on the basis of robot journalism in recognition of financial efficiency. While robot journalist involves geographical limits of coverage and interviews, human journalists need to focus on writing articles reflecting their insight on the actual context through comprehensive thinking by going around many different places and extending their networking of news sources. Humans and robots can complement each other efficiently while the former provide insight and the latter provide customized news services [31].
This study is of academic significance in that it defines factors related to the acceptance of AI-generated news and presents an acceptance model. At the same time, however, it involves limitations in model generalization since its subject group included only 30 related experts in Korea. At present, not only the current status of the media industry but also the level of robot journalism development and commercialization are different among countries. Thus, developmental research needs to be conducted among expert groups over this global society in order to develop an acceptance model that can be generalized properly. In addition, this study suggests a model by deriving factors of acceptance based on previous studies on general acceptance of technology and news articles. However, AI-generated news presents content with receptive and consumptive characteristics, unlike common products and services. There must be unique factors of acceptance or consumption value in AI-generated news articles. Accordingly, future studies need to clarify the differentiated factors of selection and stimulation regarding the acceptance of AI-generated news articles. Finally, this study suggests an acceptance model based on the opinions of an expert group, while an acceptance model needs to be developed in reflection of acceptance intention factors of the general public, the end users of articles. A future study needs to collect the opinions of general consumers on AI-generated news articles in such ways as managing big data from social network systems or collecting questionnaires from customers who have experienced Al-generated news articles to analyze them empirically.