Next Article in Journal
Thermodynamic Analysis of the Solubility of Propylparaben in Acetonitrile–Water Cosolvent Mixtures
Previous Article in Journal
Can Environmental Knowledge and Risk Perception Make a Difference? The Role of Environmental Concern and Pro-Environmental Behavior in Fostering Sustainable Consumption Behavior
Previous Article in Special Issue
Effects of Application of Information on the Expectations of Benefits from GaaP: Moderating Effects from Perceptions of IIT
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Artificial Intelligence and Public Values: Value Impacts and Governance in the Public Sector

1
School of Public Administration, University of Nebraska at Omaha, Omaha, NE 68182, USA
2
John W. McCormack Graduate School, University of Massachusetts Boston, Boston, MA 02125, USA
*
Author to whom correspondence should be addressed.
Sustainability 2023, 15(6), 4796; https://doi.org/10.3390/su15064796
Submission received: 13 January 2023 / Revised: 27 February 2023 / Accepted: 3 March 2023 / Published: 8 March 2023
(This article belongs to the Special Issue The AI-Augmented Smart Transformation for Sustainable Governance)

Abstract

:
While there has been growth in the literature exploring the governance of artificial intelligence (AI) and recognition of the critical importance of guiding public values, the literature lacks a systematic study focusing on public values as well as the governance challenges and solutions to advance these values. This article conducts a systematic literature review of the relationships between the public sector AI and public values to identify the impacts on public values and the governance challenges and solutions. It further explores the perspectives of U.S. government employees on AI governance and public values via a national survey. The results suggest the need for a broad inclusion of diverse public values, the salience of transparency regarding several governance challenges, and the importance of stakeholder participation and collaboration as governance solutions. This article also explores and reports the nuances in these results and their practical implications.

1. Introduction

Artificial intelligence (AI) has the potential to offer significant benefits and, at the same time, to pose challenges to our society [1]. The benefits of AI applications in the public sector include gains in the efficiency and effectiveness of AI-enabled rapid processing, abstracting, and reasoning based on a large amount of data [2,3,4]. Additionally, the government’s use of AI can assist in many areas of public service, such as citizen service, by using chatbots and security through the detection of and defense against cyberattacks [5,6]. AI-based decision support systems that are integrated into public service decision making can compensate for human vulnerability in decision making. Humans are susceptible to problematic decision making due to poor emotional states, personal bias, or limited cognitive capacity [7].
At the same time, there are concerns about AI’s potential to erode public values. Automation based on algorithms developed using existing data tends to perpetuate historical biases that are embedded in government data [8,9] and to magnify system tendencies for inequality [9,10,11]. AI can significantly reduce the transparency of data, data analysis, and decision making [12]. Additionally, the use of AI-enabled algorithms for administrative decision making about the production and delivery of public services has the potential to erode the discretion of street-level public service providers [13,14].
Understanding and addressing interconnections between public values and AI is crucial for advancing public values while minimizing negative consequences. The existing literature has identified the impacts of AI on public values that are focused on equity, human autonomy, efficiency, effectiveness, etc., e.g., [3,15,16]. However, the existing literature lacks a more systematic and comprehensive review of the relationships between AI and a broad set of public values, particularly at the level of AI system development.
Governance is central for addressing public value impacts of AI [12,17]. Governance regards government as the primary institution that makes and implements policy and rules [18] with the participation of stakeholders and citizens [19]. Governance is about producing and delivering publicly supported goods and services [18,20] for the purpose of advancing public values [21,22]. Therefore, public sector AI is an important focus because of the public mission of government in society and the vital role of government in regulating AI use [23]. Progress has been made in understanding potential governance challenges and solutions [12,24] and the use of AI in public governance [25]. Moreover, scholars have argued for the importance of considering public values for algorithmic and AI governance [15,26,27]. However, a significant gap remains in our knowledge about the impacts of AI on public values and the governance challenges and solutions associated with the interaction between AI and public values. Less is known about how government employees perceive the impacts of AI use in government on public values as well as the related governance challenges and solutions.
This article intends to make several contributions to academia. First, the systematic literature review suggests the need for the incorporation of a broader set of public values in AI systems in the public sector as well as the salience of these expanded values. Second, the systematic literature review provides the identification of the key governance challenges and solutions associated with public value creation. Such identification helps inform research and practice. Lastly, the investigation into U.S. government employees’ perspectives contributes to insights into the relevance of a broad outlook on public values, the salience of transparency regarding governance challenges, and the solutions involved. Differences in the stakeholders and productive stages for participation introduce opportunities for more targeted governance solutions.
This article first develops a conceptualization of AI and public values and the interactions between them in the public sector. It follows with a systematic literature review to understand the public values that are affected as well as the related governance challenges and solutions. Building on the results of the literature review, this article then explores the perspectives of government employees on governance to address public value impacts. Next, it discusses the features and implications of the literature review and a related survey of government employees. This article concludes with a summary of the main points and suggests avenues for future research.

2. Artificial Intelligence Systems and Public Values

2.1. Artificial Intelligence as Intelligent Systems

A broad notion of AI is “any technology that is able to function appropriately with foresight of its environment” [28]. Such technology is “inspired by … the ways people use their nervous systems and body to sense, learn, reason, and take actions” [29]. Consequently, AI can be characterized by the ability to perform tasks that, historically, have been performed by humans [30,31,32]. The capabilities of AI include perceiving, learning, abstracting, and reasoning, as articulated by the Defense Advanced Research Project Agency [5]. A more recent extension of the list adds communication to these capabilities [32].
Machine learning is one of the core capabilities of AI [32]; it allows for learning from data to achieve better performance in areas such as pattern recognition, classification, predictive analytics, and language processing. As a result, machine learning is a way to turn data into knowledge [33]. For administrative government agencies, machine learning is regarded as a core capability that aids in auditing and enforcing regulatory policies on taxation, environmental protection, and safety [34].
Until now, the performance of intelligent work has required human conditions and decision-making capabilities, but AI mostly requires supporting computational technologies. Today’s computer systems can perform automation and decision-making functions [35]. Robotic technologies are important for process automation [36] and are the embodiment of AI-enabled computer systems to perform motor functions [32].
To encompass the technologies needed to perform humanlike intelligent tasks, AI is better conceptualized as AI systems. Such a definition allows for the inclusion of data, algorithms, computer systems, and robotics to support the intelligent capabilities of perceiving, learning, reasoning, communicating, and taking actions. The uses of AI’s perception capability for the government could involve computer vision to monitor volcanic activity or surveillance to track potential threats to public safety. Regarding learning, the capability for image classification could help in analyzing cracks on bridges to evaluate transportation safety. For communication, an important AI capability might be interfacing with humans to understand their directions and to communicate actions back to them in an understandable way. For action, AI may be capable of making and executing decisions to implement regulatory policies or perform a rescue mission in a disaster area.

2.2. Public Values and Artificial Intelligence Systems in the Public Sector

This article focuses on the use of AI in the public sector and, consequently, on how this use aligns with specific public values—such as efficiency, effectiveness, accountability, transparency, and equity—that are central to public administration and governance [37,38]. These values guide the purposes of the government and the public institutions and organizations that are involved in public governance and service. Scholars aligned the government’s discretionary actions that are augmented by AI with public values [14,39,40].
This article adopts the taxonomy of values for assessing the impact of AI that was developed by Bannister and Connolly for information and communication technology in general [41]. This taxonomy classifies public values by their orientation: duty-oriented, service-oriented, and socially oriented. The duty-oriented values include responsibilities to citizens and elected officials, facilitation of democratic will, proper and efficient use of public funds, the economy, compliance with law, rectitude, integrity, and honesty. The service-oriented values include service to citizens in their various roles, efficiency, effectiveness, responsiveness, transparency, and respect for the individual. The socially oriented values are inclusiveness, justice, fairness, impartiality, equality of treatment and access, due process, protection of citizen security, protection of citizen privacy, protection of citizens from exploitation, accountability, and consultation.
This taxonomy provides a broader perspective that extends beyond the service orientation. Based on an extensive literature review, the efficiency and effectiveness of service have been demonstrated to be the most important public values regarding the use of technology in government [42]. However, such a narrow focus ignores the duty that government has in a democracy and societal concerns about justice, equity, and rights [41,43]. For the use of AI systems in government, efficiency and effectiveness can be achieved by enhancing and automating analytical capabilities [15,44]. As a parallel to information and communication technologies for government in general, a broader consideration of privacy, security, employment, and accountability is important for the use of AI in government [3,12]. The broad societal impacts of AI systems have created the need for ethical considerations beyond service-focused public values. The development of machine ethics and a framework for impact assessment is an effort to assist with the governance of AI systems [45,46].
This article adopts a sociotechnical perspective to understand the interactions between AI systems and public values for good AI governance. Governance involves government, citizens, and stakeholders in the development and implementation of policies and rules for the production and delivery of publicly supported goods and services [18,19,20]. Good AI governance advances public values via AI use in government. The development and use of AI systems could be understood to be interactions of social and technical components and processes [47], drawing from Latour [48]. From this perspective, the technical components of an AI system include data, algorithms, computer systems, and robotics [8,32,49,50], while the social components include people, organizations, governance rules, and interactions. These elements include both material and social entities such as data, humans, systems, and algorithms. The interactions include the impact of social structures and processes on the adoption and use of technology and vice versa.
The development and use of technologies in the public sector are shaped by organizational and institutional factors [51]. For AI, as an emerging technology, there is a need to ensure public good and establish public values [52,53]. The public values embedded in a government, as embodied in a national AI strategy, are likely to guide the development and implementation of AI. For instance, the U.S. National Artificial Intelligence Research and Development Strategic Plan (2019 Update) articulated the need for transparency and equity alongside efficiency [54]. This type of national policy framework shapes the development of AI in various government agencies.
At the same time, the adopted technologies can affect organizations and their governance [55]. The use of AI-enabled algorithms for policy implementation is likely to affect equality when socially constructed biases are embedded in the algorithms [9,10,56]. Additionally, human autonomy regarding administrative discretion in government is likely to be affected by the automation offered by AI [39,40]. Likewise, democratic governance is sure to be affected when the extensive use of algorithms allocates disproportionate power to groups with high levels of technical resources and expertise [56,57].
Currently, AI use in the public sector tends to focus on a few functions in the production or delivery of public service, though humans continue to play an important role as the ultimate decision makers or enforcers of rules. These AI functions could be audio or natural language processing that helps automate the understanding of inputs from humans [6]. Predictive analytics is another popular use of AI that helps prioritize resource allocation in the public sector for services such as public safety and the prevention of fraud [2,6]. Humans are still responsible for enforcing public safety rules and determining the subjects of fraud investigations. In the cybersecurity area, AI can fulfill the functions of security analytics and threat intelligence [5,6], while humans exercise judgment on the parameters of threat analysis and responses. Even regarding the use of chatbots for city services, when such chatbots are responsible for identifying relevant information, the actual production and delivery of services still fall on other government systems or personnel.

3. Artificial-Intelligence-Affected Public Values and Artificial Intelligence Governance Challenges and Solutions

3.1. Systematic Literature Review

This systematic literature review aims to identify and analyze the impacts of AI on public values as well as the governance challenges of addressing these impacts. This literature review provides a survey of how AI is related to public values. More importantly, it presents an empirical basis for understanding the relationships between AI and public values as well as the governance challenges and solutions. The literature begins with a broad search of AI and governance, to maintain the focus on governance, but remains open to the identification of possible interactions between AI and public values as well as the governance challenges.
The sources for the academic references are the Web of Science and Google Scholar, two popular academic literature databases, and recommendations from subject-area experts. The Web of Science provides coverage of publications in the fields of science, technology, social sciences, arts and humanities. Google Scholar complements the Web of Science with a broader coverage and an interdisciplinary research. As shown in Figure 1, we used the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) method to document the process of the systematic literature review. At the first stage, we used three clusters of keywords—“AI and Governance”, “AI and Dark Sides”, and “AI and Equity and Governance”—and set the time range as “between 2000 and 2022” (to December 2022) to search for articles in the two databases. That process identified 190 articles from the Web of Science and 215 from Google Scholar. Additionally, subject area experts recommended 23 articles regarding AI applications in the public sector. Most of these references were journal articles, but some were book chapters and research reports. Consequently, the acquired references, including 23 from expert recommendations and excluding duplicates and references that are not journal articles, totaled 289 pieces.
Then, the authors developed three criteria to determine the relevance scores for each article: “Does this article mention public values?”; “Does this article define governance challenges?”; “Does this article suggest governance solutions?” Next, we assigned relevance scores to each article on a scale of 1–10 and included articles with a score of 5 or higher. Additionally, only journal articles were included, which reduced the number of relevant references to 63. Then, we excluded five articles: the full texts for three journal articles could not be accessed, and the other two primarily discussed law and philosophy. Finally, that left 58 references to be analyzed.
For public values and AI systems, we used the taxonomy adopted from Bannister and Connolly [41], as discussed in Section 2.1. This taxonomy has three clusters of values: duty orientation, service orientation, and social orientation. We coded a reference into a particular value when it specifically referred to the values that are affected by the use of AI systems. For governance challenges, we conducted theme coding of the 58 references. Three themes emerged: public value challenges; data quality, processing, and outcome challenges; societal governance challenges. The themes emerged from the reviewed references based on a conventional content analysis, which aims to develop themes and codes from the documents and materials, especially when the existing theory and concepts are limited [58]. First, we listed all the mentioned challenges and solutions in the literature. Next, we discussed the findings from the literature review and then agreed on the three primary categories for the challenges and solutions listed above. Finally, we used the three themes to describe the challenges of AI governance as well as the solutions.

3.2. Public Values and Artificial Intelligence Systems

The results of the analysis of the references based on the taxonomy of public values are summarized in Table 1. For duty orientation, there is an emphasis on the duty of AI to comply with laws [1,35] and be responsible to citizens [1,23,59,60]. The application of AI to automate the decision-making process must meet the legal requirement of preventing malicious use of the technology [35]. When government agencies use algorithms and machine learning to deliver public services, they must ensure that AI is responsible for its actions and decisions [12,61]. Further, Larsson [60] indicates that responsibility is a crucial instrument for developing trustworthy AI in public governance in the European context.
For service-oriented values, the existing research primarily focuses on transparency, effectiveness, and efficiency. First, most of the references covering service-oriented values emphasize the importance of transparency, as shown in Table 1. Transparency is the basis for citizens to understand and monitor how AI delivers public service and what is provided [68,69]. Second, effectiveness is regarded as one of the criteria for evaluating AI performance in public service [39]. Additionally, effectiveness in governing AI can be achieved by a multilayered governance structure [23] or government regulation [75]. Third, the aim of applying AI in government is to enhance efficiency, such as automating routine tasks to reduce the time spent and costs involved [15,39].
The references analyzed in this literature review address several dimensions for socially oriented values, including accountability, equality, privacy, fairness, justice, due process, and security. This list is ordered by the number of references citing these values, from the greatest to the fewest. Accountability to the public is one of the core values when applying AI in various domains [77]. By enhancing accountability, AI security and quality can be assured [50]. How to ensure the equal treatment of citizens, civil servants, and applicants through AI-enabled services is another focus. Ensuring reliable and affordable access to public services is also discussed in the research [39,49,79]. Fairness in the administrative system, procedural due process, and justice in the judicial branch are crucial to improving the ethical performance of AI [60,79]. Finally, ways to protect personal data, especially sensitive information, are discussed in several of the articles [50,60,64,79,81]. For instance, Wirtz et al. [3] point out that AI systems carrying personal information may be compromised by cyberattacks.

3.3. Governance Challenges

The second focus of the literature review concerns governance challenges, which can be categorized into three themes: public values challenges; data quality, processing, and outcome challenges; societal governance challenges. The extensive list of references for each theme is presented in Table 2. Public values challenges mainly emphasize that the lack of appropriate governance mechanisms or actions could result in the erosion of public values. The most-cited challenges for AI systems are the lack of appropriate governance mechanisms to foster transparency and accountability (e.g., [50,65,70,73,83]). Governance challenges associated with ensuring equity are also highlighted [3,9,63,73,77,79]. Additionally, privacy and safety concerns pose governance challenges to minimize the negative impact of AI (e.g., [36,79,84]). Linked to the concern about the public sector is a concern about the lack of safeguards against administrative evil and the malicious use of AI systems [1,9,39,77].
AI introduces various governance challenges in the areas of data quality, information processing, and outcomes. Data quality and management problems underscore the governance challenges associated with the racial biases and discrimination in the data set [17,49,83,85]. When AI processes biased information, the government obtains biased results that could threaten individual rights. This problem is exacerbated when humans have limited or no control over AI [12,64]. The difficulty in measuring AI performance and outcome quality amplifies the risks [85].
The challenges of AI related to societal governance are substantial, such as collaborative governance problems, displacement of jobs, and threats to human autonomy. The use of AI may affect governmental structure, intergovernmental relationships, and network governance. Coordinating with other agencies, companies, and nonprofit organizations is a significant challenge for public institutions [70]. Conversely, although AI can enhance the efficiency in delivering public services through high-speed operations and rapid automation, humans can lose their jobs, and civil servants may have to sacrifice their discretion [24,39,64].

3.4. Governance Solutions

The systematic literature review suggests three themes for AI governance solutions and their supporting references, as shown in Table 3. First, the references introduce public values solutions as responses to governance challenges. The most-cited tactic is explainable AI for addressing the lack of transparency, emphasizing that the process and outcomes are understandable to and realizable by citizens [59,65,68,86]. Additionally, the research stresses the importance of including more public values in alleviating the negative impacts of AI on public governance [50,60,66,71,72]. Finally, as the system can be embedded with more moral values, ethical AI is regarded as a governance solution [46,77].
With respect to data quality and processing, the most-often-cited governance solutions are the use of regulations and the detection of biases. Governments can regulate behaviors related to data quality and processing through legal approaches [12,17,35,76,78,82,88]. For instance, many European countries have adopted the General Data Protection Regulations to alleviate privacy and human rights threats from AI [76]. Furthermore, including data auditing to eliminate AI biases is another governance solution [35,74,81,94].
Finally, scholars have proposed various societal governance solutions to the potential negative impacts of AI. The most-cited solutions are collaborative governance and the multilevel approach to AI governance. Collaborative governance encourages governments to establish horizontally collaborative relationships. For example, the same levels of government can work together to enact personal data protection regulations or to build mechanisms that address AI challenges [70,80,82,95]. The multilevel approach suggests that governments use a variety of strategies to alleviate the negative impacts at different levels [49,85].

4. Perspectives of Government Employees on Artificial Intelligence Value Impacts and Governance Challenges and Solutions

4.1. Artificial Intelligence Governance and Public Values from the Government Employees’ Perspective

This empirical investigation builds on the key themes of public values, governance challenges, and governance solutions, which emerged from the systematic literature review. There are many stakeholders of AI use in government who need to be considered, and their perspectives differ [97]. Among these various perspectives, in this study, we choose to focus on the perspective of public employees because their opinions are crucial for the implementation of AI in government. Additionally, such a perspective is consistent with the original formulation of the taxonomy that is grounded in the perspective of civil servants [41].
To understand the perspective of government employees, we conducted an exploratory national survey of government employees in the U.S. We used Amazon’s Mechanical Turk (MTurk) because of its cost-effectiveness and comparability in terms of the data collected via traditional methods [98,99]. The survey was conducted in February 2021 and yielded 566 total responses. The respondents were current employees at the local, state, and federal levels of government in the U.S. The survey included questions that serve as the basis for quality assurance. After removing duplications, nongovernment respondents, and inconsistent answers to the quality assurance questions, we arrived at a final data set of 323 responses. This study draws mostly from the section of the survey that addresses the governance of AI use in government.

4.2. Public Value Impacts of Artificial Intelligence Use in Government

Government employees have opinions about the impacts of AI use in government on public values. The existing literature discusses the concerns of government employees about the implementation of AI [97], but there is a gap in the empirical exploration of how these concerns are translated into the perceived impact of AI use on public values. Information and communication technologies hold the potential for digital transformation [100,101] and the creation of public values [102]. AI is also cited as an emerging technologies that has transformative potential due to the introduction of machine learning capabilities as well as its application to a wide range of public services. Given the current development of AI, a useful approach to understanding the impacts on public values is to explore AI’s perceived transformative potential.
The survey questions explored government employees’ perception of the extent to which AI technologies would affect public values as aspects of government. The respondents were asked to indicate the extent to which they believe AI technologies will improve decision making, effectiveness, efficiency, and accountability. For each public value, the survey respondents were asked to indicate the extent as “no improvement” (1), “modest” (2), “substantial” (3), or “transformative” (4). The decision-making question aligns with the government’s duty-oriented values, as the government is responsible for making satisfactory decisions to serve the public. This type of decision making by the government is manifested by the elected officials and legislators who represent the citizens. Efficiency and effectiveness are frequently cited public service values, as demonstrated in the earlier systematic literature review. Accountability is a society-oriented value that captures the government’s role of acting in the public’s best interest.
The responses regarding each public value, expressed in percentages, are presented in Table 4. Regarding decision making as a duty-oriented public value, approximately 61% of the respondents indicated that AI has substantial or transformative potential for improving decision making. The majority of the respondents believed in the potential for public value creation via AI use in government decision making. Regarding the service-oriented public values, 65.3% of respondents noted that the use of AI has substantial or transformative potential to improve government effectiveness, and 77.7% had the same opinion with regard to efficiency. As for the society-oriented public value of accountability, 53.8% of respondents believed there is substantial or transformative improvement in accountability with the use of AI in government.
Collectively, these results suggest a positive link between AI use in government and its potential for improving public values. These values include the duty-oriented one: the responsibility to citizens via better government decision making; service-oriented ones: efficiency and effectiveness; and the society-oriented one: accountability to the public. Moreover, the results show a favorable assessment of the potential for public value creation, with the majority of respondents indicating that each value has substantial or transformative potential. These results provide empirical evidence on the linkage and potential of AI government use for a positive impact on public values.
The difference in the percentage of respondents believing that AI government use can have substantial or transformative improvement potential for public values has important implications. These government employees see the relatively strong potential for AI’s role in improving government efficiency and effectiveness in comparison with decision making and accountability. Such relatively strong potential suggests the relative efficacy of the government’s use of AI in affecting service-oriented values than other duty-oriented or society-oriented values. Accountability to the public as a society-oriented value has the lowest percentage of respondents, indicating substantial or transformative improvement potential among all values examined. This suggests that AI use in government has less impact on broader societal values.

4.3. Governance Challenges of Artificial Intelligence Use in Government

Government employees are likely to have some recognition of the potential negative impact of AI on transparency. Government policies and documents regarding the use of AI in government, such as those in the U.S. [54] and in European countries [6], have articulated transparency as a challenge for AI. These documents have likely created some awareness among government employees about the transparency challenge. Moreover, government employees, in their own exposure to AI either through personal use or in conversation with colleagues, have probably found it difficult to understand how and why certain algorithms are developed from machine learning. Consequently, there is likely some awareness among government employees about the negative impact of AI use on transparency.
One important data processing and outcome challenge regarding the use of AI in government is the extent to which AI makes public service decisions [12,103] and how much it could constitute the loss of human control [64]. One of the significant direct impacts of AI use on government employees is the potential loss of their discretionary authority in the face of the automation that accompanies AI. Scholars have cautioned that the removal of discretionary authority from street-level bureaucrats may consolidate that authority in the hands of information-system-level bureaucrats [40]. Moreover, concerns have been expressed about the potential negative impacts created by unchecked AI automation [103]. Government employees are likely to be concerned about the loss of the discretionary authority on which they have traditionally relied to make personalized and responsive public service decisions. Such concern is also likely to be more prevalent for street-level bureaucrats who interact directly with people and make decisions that can have a significant impact on people’s lives.
The public value challenge associated with transparency Is captured by the following survey question: “The development of AI in government is likely to negatively impact transparency due to the technical nature of the algorithm”. Respondents could indicate the extent to which they agree with the statement on a scale of 1–7, with 1 corresponding to “strongly disagree” and 7 corresponding to “strongly agree”. The survey question addressing whether discretionary power is connected to the governance challenge of data processing and outcomes is “The use of AI in my government agency will take away your discretionary authority”. The respondents could select their level of agreement on the same 1–7 scale. The survey question for the societal governance challenge addresses a broader set of societal values: “The goal setting of AI use in government cannot fully consider societal values beyond technical efficiency”. Again, the respondents could select their level of agreement on the same 1–7 scale.
Regarding the likely negative impact of AI in government, as shown in Table 5, 68% of the respondents said they are either neutral toward or in agreement with the statement, with scores ranging from 4 to 7. The mean score for the transparency question is 4.32. The single sample t-test result indicates that such a mean score is statistically greater than the neutral position (a score of 4) at the 0.01 level of significance. From this, we can infer that government employees do perceive the likelihood that AI in government could have a negative impact on transparency. Such a result lends empirical support for the concern about the potential lack of transparency, as indicated earlier in the systematic literature review.
Regarding the question about the erosion of their discretionary authority, 68% of the respondents said they were either neutral toward or in agreement with the statement (scores of 4–7). The t-test result indicates that such a mean score is statistically greater than the neutral score of 4.0. The results indicate that government employees do perceive that the use of AI in their respective government agencies represents a threat to their discretionary authority. Such a result provides direct support to the concern about the erosion of discretionary power by bureaucrats as a result of AI use in government [40].
Regarding the inability of goal setting for AI use in government to fully consider societal values, 68% of the respondents were either neutral or in agreement with the statement (scores of 4–7). The mean score for the concern about losing discretionary authority is 4.372. The t-test result indicates that such a mean score is statistically greater than the neutral score of 4.0. We can infer that government employees do perceive that the use of AI in government cannot fully consider a range of societal values.
Such a result provides some of the first evidence regarding the limitations of AI in considering a broad range of societal values, especially from the perspective of government employees. The specific societal concerns and their governance challenges could include job loss [1,3,90], threats to human autonomy [39,64], and the disadvantages of AI making value judgments [12]. Additionally, the extent to which such a recognition regarding AI’s limitations is shared by government employees can be a facilitating condition for a dialogue about governance solutions.

4.4. Transparency and Participation as Governance Solutions to Artificial Intelligence Use in Government

The provision of transparency serves as a governance solution to public values. Explainable AI offers a form of transparency that helps humans understand why and how AI makes decisions. Such transparency can be further distinguished between transparency to whom and transparency at what stages of AI development and implementation. For government employees, the targets of transparency can be divided between internal-facing (government managers) and external-facing (the general public). To overcome the lack of transparency of government AI systems, government employees are expected to support making information about the development of the algorithms used in AI systems available to government managers. Government employees are also likely to be in favor of making such information available to the general public to achieve the goal of accountability.
For data quality, processing, and outcome solutions, transparency could focus on data, which are the materials from which the machine learns. Consequently, biases in data tend to be reinforced by the algorithms developed from them [49] and could have negative consequences for the welfare of individuals [104]. Data transparency is also foundational for audit efforts [105]. Government employees are likely to support such data transparency to inform themselves of any inherent bias in the data. With the government’s emphasis on being responsible for and accountable to the general public, it is also expected that government employees would support making data transparent to the public.
For societal governance solutions, collaborative governance is suggested by the systematic literature review as one of the main mechanisms [96]. Such collaboration can be among organizations and individuals in the public, private, and nonprofit sectors or among academia, government, and businesses. One of the less-explored dimensions of collaboration is between government and citizens regarding the development and implementation of AI, in which citizen participation is a critical step. Another dimension that lacks comprehensive research is the stages of AI development and implementation. In practice, government organizations have recognized the importance of process orientation for addressing stage-specific concerns [105]. However, scant scholarly research has explored the role of stages in addressing AI-related impacts. This study provides one of the first explorations of the important stage involving the participation of public officials and citizens. Such participation provides an important basis for collaboration.
The public value solution pertaining to transparency for government employees is captured by the survey question “Do you agree that the information on the development of algorithms used in the AI system should be made available to the government managers?” The respondents could indicate the extent to which they agree with the statement on a scale of 1–7, with 1 corresponding to “strongly disagree” and 7 corresponding to “strongly agree”. The survey question addressing the transparency of the AI system for the general public is “Do you agree that the information on the development of algorithms used in the AI system should be made available to the general public?” Again, the respondents could choose their level of agreement with the statement on the same 1–7 scale.
The data quality and processing governance solution pertaining to data transparency for government employees is captured by the survey question “Do you agree that the data the AI system uses should be made available to the government managers?” The question pertinent to data transparency to the general public is “Do you agree that the data the AI system uses should be made available to the general public?” For both questions, the respondents could select a level of agreement on the same 1–7 scale.
The questions about the societal governance solution focus on participation at a specific stage in the AI system life cycle. The life cycle of AI systems can include stages of goal setting, development, use, and impact assessment. Two survey questions explore government employees’ views regarding two groups of participants: public officials and the general public. More specifically, government employee respondents could answer the following survey question by indicating which stage is most important for public officials or whether all stages are important: “At which stage of AI development is it most important for public officials to participate? (a) goal setting for AI use in government, (b) development of AI systems, (c) use of AI systems in government decisions, (d) impact assessment of the AI-enabled decisions, (e) all of the above”. The second question replaces “public officials” with “the general public” and offers the same answer options to explore the respondents’ potentially different sets of expectations for participation by public officials and the general public.
As Table 6 indicates, the results reveal that the majority of respondents (74%) support making information on the development of algorithms available to government managers, but a lesser number (64%) also support sharing the same information with the general public. Based on single-sample t-tests, the mean values of the agreement scores are statistically larger than the neutral position (a score of 4) for both government managers and the general public. Therefore, we can infer that there is support among government employees for making the development of algorithms available as a transparency governance solution to advance transparency as a public value. Moreover, the mean level of agreement (5.285) for sharing information on the algorithms with government managers is higher than the mean level of agreement (4.932) for doing the same with the general public. Such a difference is statistically significant, and it suggests a position favoring more algorithm transparency for government managers than for the public.
Providing data transparency is a governance solution that belongs to the category of data quality, processing, and outcomes. The result reveals that a majority of the government employee respondents (80%) support providing information about the data used by the AI system to the government managers who are responsible for the service. Similarly, the result indicates that the majority of respondents (63%) also support sharing the same information with the general public. Based on single-sample t-tests, the mean values of the agreement scores are statistically larger than the neutral position (a score of 4) for both government managers and the general public. We can infer that there is support among government employees for making information about the data used by the AI system available as a solution to the governance challenge of the lack of transparency.
The mean level of agreement (5.285) for sharing information about the development of algorithms with government managers is higher than the mean level of agreement (4.932) for doing the same with the general public. The difference in these levels of agreement is statistically significant and suggests a position favoring more transparency for the data used by the AI systems for government managers than for the general public.
Participation by stage is the focus of exploration for the societal governance solution to a lack of transparency. The survey results reveal support for participation by both public officials and the general public. As shown in Table 7, approximately 36% of respondents consider it important for public officials to be involved in all the processes of the AI system. Among all the stages, 27% of respondents suggest that AI system development is the most important stage for public officials to be involved in. Other stages are also considered to be the most important stage for participation: goal setting (14.6%), decision making (15.5%), and impact assessment (7.1%). In addition, 25.7% of survey respondents regard the use of AI in government decision making as the most important stage for public participation. All of the above stages (23.8%) and the stage of AI system development (21.1%) follow closely as being considered as the most important for public participation.
The results reflect the opinions of government employees regarding what stages are the most important for participation by public officials and members of the general public. The survey respondents expect public officials, more than the public, to participate in all stages of the process. Such a view could stem from the expectation of public officials responsible for all stages of the life cycle to be accountable for AI use in government. Survey respondents ranked the use of AI in government decision making as the most important stage for public participation. This opinion could result from the perceived effectiveness of having the public participate directly in AI-enabled government decision making. Therefore, the decision is informed by the members of the public who have information relevant to the decision and provide the perspective of service recipients to ensure better decisions.

5. Discussions

5.1. Implications of the Results from the Systematic Literature Review

The results of the systematic literature review support a broad-based approach, as suggested by Bannister and Connolly, to considering the public values that are affected by AI. All three categories of the public values—duty-oriented, service-oriented, and socially oriented—have received scholarly attention. Such attention is congruent with the broad public value implications of AI systems in the public sector that extend beyond a narrow perspective of efficiency and effectiveness or the profit focus of the private sector [15,24]. The duty of government, government service, and societal considerations should serve as the broad but necessary set of public values that are prioritized in the development of public-sector AI systems. For the duty-oriented values, the results suggest the need for the thorough consideration of legal requirements and responsibility to citizens as important public values. For the service-oriented values, the results imply the salience of transparency followed by effectiveness and efficiency. For the society-oriented public values, further identification and articulation would be helpful for the inclusiveness of diverse public values. The recognition of the importance of these public values being connected to the governance of AI systems in the public sector is helpful for both their identification and use to guide AI system development and implementation. Doing so moves beyond a traditionally narrower focus on effectiveness and efficiency.
In practice, the comprehensive consideration of duty-oriented, service-oriented, and society-oriented public values is necessary for effective governance. To address the tendency to focus narrowly on efficiency and effectiveness, the consideration of societal values such as accountability, equity, fairness, justice, privacy, and security, as suggested in the literature, is recommended. Additionally, making these public value goals explicit as a guide helps identify the affected public values and advance them accordingly. Policy documents with AI goals and principles are useful sources of public values.
Moreover, the systematic literature review has implications for the key governance challenges and solutions associated with three categories: (a) public values; (b) data quality, processing, and outcomes; (c) societal governance. For public values, the prominence of the transparency challenge supports the need for further research into its governance as one of the main concerns about AI systems in the public sector. Moreover, the literature review underscores the need for research into a variety of public values and their trade-offs. For the public value governance solution, the literature review points to explainable AI to ensure further transparency for AI development and implementation. The practical implications of public value challenges and solutions include the rigorous pursuit of explainable AI as a means to advance transparency as well as the full consideration of various public values and their trade-offs.
For data quality, processing, and outcomes, the systematic literature review offers important research and practical implications. Research should focus more on understanding the sources of biases and inequality as well as governance solutions to address them. For AI-enabled algorithms, biases in the data can result in discrimination, wrongful judgments, and increases in inequality [49,89]. Regulations regarding data are considered to be a generic solution to potential biases and other public value impacts arising from the data used for AI development and implementation [35,82]. In practice, the development of rules and regulations that specifically address data biases and their implications would be productive. The specific regulations could focus on data disclosure, audits, or assessments, and data audits governed by an oversight committee are proposed as a governance solution [81].
For societal governance challenges and solutions, the development of broader frameworks to consider the societal challenges of AI is important. These challenges include job displacement [1,12] and the erosion of human autonomy in discretionary decisions [39,64,81]. A salient focal research area includes the dynamics and effectiveness of cross-sector and multilevel governance solutions to societal challenges posed by AI. In practice, the discipline to consider societal impacts is important for AI governance as well as the use of AI governance frameworks and solutions that consider multiple sectors [80,96] and levels [24].

5.2. Implications of the Results on the Perspective of Government Employees

The empirical survey results regarding the potential public value impacts of AI from the perspective of public employees have implications for both research and practice. Public employees are concerned that AI use in government targeted primarily at efficiency and effectiveness will affect important public values, which is in line with the tendency to emphasize efficiency and effectiveness in e-government [42]. In fact, the survey results of government employees indicate the relevance of a broader set of values, such as improvement in decision making, to be responsible to citizens as well as have accountability to broad societal concerns. Research into a wide range of public values while recognizing the emphasis on efficiency and effectiveness is needed for advancing scholarship. The practical implication of the survey results is on the value of understanding the sources of the government employees’ focus on the efficiency and effectiveness of AI use in government. Understanding the role that policies, guidelines, and programs play is also critical in realizing the substantial or transformative potential of AI on public values.
The empirical survey results on governance challenges underscore transparency as an important AI governance challenge, and they are among the first to provide evidence of government employees’ recognition of the existence of such a challenge. Moreover, the results indicate that the erosion of discretionary authority is an important AI governance challenge for government employees. The existing literature raises the concern about government employees gradually losing their discretionary authority when decisions are increasingly made by information systems and then by AI-enabled algorithms [39,40]. This study’s finding lends empirical support to that concern, especially as expressed by government employees, about AI use in government. Additionally, the results reveal the governance challenges associated with considering a broader set of social values regarding AI use in government. The identification of such a challenge furthers the need for research into how to move beyond a narrow set of public values such as efficiency and effectiveness. In practice, the survey findings on governance challenges reveal the usefulness of understanding the definition and perception of transparency by government employees. Furthermore, exploration of the ways in which AI use in government can erode or remove discretionary authority is likely to inform an appropriate division of authority in the automation of government decisions.
The survey results on transparency underscore the need for AI governance to consider multiple perspectives. Existing research into the governance of digital technologies indicates the diverse views of various stakeholders [97,101]. However, these survey results provide one of the important perspectives: that of government employees. Moreover, the results advance AI governance scholarship by providing empirical support for specific areas of transparency, including the development of algorithms and the data used by AI systems. Additionally, the findings further point out the prioritization of stakeholders in receiving information, with government employees placing their need for transparency above that of the general public. Such evidence supports the need to explore which stakeholder groups should be prioritized to receive information on AI algorithm development and data use. These are nuances that are important to gain a better understanding of the diversity of stakeholder views on AI governance solutions as well as the prioritization of various stakeholder groups in the delivery of solutions.
In practice, the survey results on transparency governance solutions offer specific areas of activities in which AI governance can engage. These include transparency of the information that is pertinent to the development of algorithms and the data used by AI systems. Such transparency is an important step toward addressing the negative impacts of AI use in government.
The results regarding the relative importance that government employees assign to stakeholder groups and stages for participation have implications for both research and practice. For research, the differentiation in the level of participation at all stages between government officials and the general public suggests the importance of exploring such nuances. The results provide initial empirical evidence of the need for making such distinctions between stakeholder groups. Moreover, the results suggest an important need for further distinguishing between the stages of AI development and implementation in government. In practice, the governance of government AI systems must consider both the stakeholder groups involved and the stages of development and operation. More specifically, the results suggest a stronger role for public officials than the public in all stages as well as a need for public officials to focus on the development stage of AI systems, while the public focuses on the use of AI systems. These suggestions could guide the development of recommendations for the design and implementation of participation initiatives.

6. Conclusions

The overarching goal of this article is to advance the knowledge of interconnections between public values and AI systems as well as the governance challenges of and solutions to public value creation. This study uses a systematic literature review to investigate these connections as well as governance challenges and solutions. The most-identified public values that are being affected are effectiveness, effectiveness, and transparency for service-oriented values and, compliance with laws, and responsiveness to citizens for duty-oriented values. For socially oriented values, accountability, equality, fairness, privacy, justice, due process, and security are mentioned. The inclusiveness of these values reflects the literature’s recognition of the types and range of public values.
The literature review outlines governance challenges and solutions with three focuses: public values, data processing and outcomes, and societal governance. The primary public value governance challenge is to advance transparency, and the main solution is the development and implementation of governance mechanisms to improve transparency. For data processing and outcomes, the primary governance challenges lie in the identification of any biases and rectification of these biases both in the development of AI algorithms and decision making. Governance solutions to improve practice include regulations and audits to assess and address biases as sources of negative impacts on public values. For societal governance, the governance challenges involve addressing the impact of AI use on human employment and autonomy, and the proposed solutions include cross-sector collaborative governance and a multilevel approach. The solutions in each of the three focuses serve as guides for a more effective practice of AI governance.
Moreover, this study explores the perspective of government employees via a national survey of U.S. government employees on the salient aspects identified by the literature review. The finding regarding public values supports the emphasis on efficiency and effectiveness as the main public values among those considered. Regarding governance challenges, government employees recognize the potential negative impact that AI has on transparency, and they underscore the potential loss of discretionary authority due to automation. The lack of full consideration of societal values beyond efficiency is also noted by government employees. The exploration of governance solutions reveals the relevance of transparency in the data used in AI and information on algorithm development as well as the priorities given to two stakeholder groups: government managers and the general public. Moreover, the survey results indicate the emphasis on public officials’ involvement in all stages of government AI development and implementation over that of the general public.
The survey findings complement the literature review by offering empirical data for a key stakeholder group, namely, government employees, regarding the use of AI in government. For research, the survey results support the need for the inclusive consideration of diverse public values, the pursuit of those values, and the potential trade-offs. Additionally, a productive research focus would be on the nature and types of transparency governance challenges as well as their respective solutions. Nuances in stakeholder-specific and stage-specific transparency and participation are also potential areas for research. In practice, the survey findings offer support for the comprehensive consideration of public values that are duty-oriented, service-oriented, and society-oriented. Efforts to provide governance solutions can be productive by focusing on the specific stakeholders and their transparency needs and solutions. Additionally, it would be useful to draw distinctions between the stages and stakeholders for targeted governance solutions.
The scope and focus of this research have limitations that also serve as research opportunities. The systematic literature review is based on relevancy scoring by the Web of Science and Google Scholar and then by the scoring of the authors. Future research can further refine the systematic literature review with a more focused and timely update, especially given the rapid advancement of research on the implementation and governance of AI in the public sector. Moreover, the focus on government employees limits the findings to that specific perspective. Future studies can consider the opinions of the general public to explore the salience of various public values affected by AI and the preferred AI governance challenges and solutions.
Collectively, the findings of the systematic literature review and the government employee survey suggest several productive research avenues. First, there is the examination of the extent to which scholarly attention on a wide range of public values has been translated into practice. Next, there is the exploration of governance challenges and solutions with various stakeholder groups beyond government employees. Such an advancement in understanding would assist in the design of transparency measures and institutional mechanisms that address the potentially different priorities of stakeholder groups. Last, there is the need to move from a generic governance solution to more specific ones to advance society’s overall knowledge of AI use. Focusing on a key capability of AI (e.g., predictive analytics) applied in a public service area (e.g., policing or healthcare) is likely to provide insights into how public values interact with AI systems and what configuration of governance design could be most effective.

Author Contributions

Conceptualization, Y.-C.C. and M.J.A.; Methodology, Y.-C.C. and M.J.A.; Investigation, Y.-F.W.; Data curation, Y.-F.W.; Writing—original draft, Y.-C.C.; Writing—review & editing, M.J.A. and Y.-F.W.; Project administration, Y.-C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This study (IRB Study Number: 2866) was determined to be exempt from review by the Institutional Review Board at University of Massachusetts Boston on 19 January 2021 because it fell under the category of “45 CFR 46.104(d)(2) Educational tests/survey/interview procedures, or observation of public behavior”.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Our survey was conducted through Amazon MTurk, and participation was voluntary and dependent on individual consent. Prior to beginning the survey, all potential participants were presented with a consent form outlining the study’s purpose, potential risks or discomforts, potential benefits, confidentiality protections, the voluntary nature of participation, and contact information for the principal investigator. Participants only proceeded to the survey if they indicated their informed consent by clicking “OK” at the bottom of the consent form. Importantly, all participants had the option to withdraw from the survey at any point before or after starting, without any negative consequences.

Data Availability Statement

The data that support the findings of this study are not publicly available.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Butcher, J.; Beridze, I. What is the State of Artificial Intelligence Governance Globally? RUSI J. 2019, 164, 88–96. [Google Scholar] [CrossRef]
  2. Engstrom, D.F.; Ho, D.E.; Sharkey, C.M.; Cuéllar, M.-F. Government by Algorithm: Artificial Intelligence in Federal Administrative Agencies; Administrative Conference of the United States: Washington, DC, USA, 2020.
  3. Wirtz, B.W.; Weyerer, J.C.; Geyer, C. Artificial Intelligence and the Public Sector—Applications and Challenges. Int. J. Public Adm. 2019, 42, 596–615. [Google Scholar] [CrossRef]
  4. Desouza, K.C. Delivering Artificial Intelligence in Government: Challenges and Opportunities; IBM Center for the Business of Government: Washington, DC, USA, 2018.
  5. Government Accountability Office. Artificial Intelligence: Emerging Opportunities, Challenges, and Implications; Government Accountability Office: Washington, DC, USA, 2018; p. 95.
  6. Misuraca, G.; van Noordt, C. Overview of the Use and Impact of AI in Public Services in the EU; JRC120399; Publications Office of the European Union: Luxembourg, 2020. [Google Scholar]
  7. Frederickson, G.H.; Smith, K.; Larimer, C.W.; Licari, M.J. Public Administration Theory Primer, 3rd ed.; Westview Press: Boulder, CO, USA, 2016. [Google Scholar]
  8. Janssen, M.; Kuk, G. The Challenges and Limits of Big Data Algorithms in Technocratic Governance. Gov. Inf. Q. 2016, 33, 371–377. [Google Scholar] [CrossRef]
  9. Fountain, J.E. The moon, the ghetto and artificial intelligence: Reducing systemic racism in computational algorithms. Gov. Inf. Q. 2022, 39, 101645. [Google Scholar] [CrossRef]
  10. Eubanks, V. Automating Inequality: How High-Tech Tools Profile, Police, and Punish the Poor; St. Martin’s Press: New York, NY, USA, 2017. [Google Scholar]
  11. European Political Strategy Centre. The Age of Artificial Intelligence. In Towards a European Strategy for Human-Centric Machines; European Commission: Brussels, Belgium, 2018. [Google Scholar]
  12. Wirtz, B.W.; Weyerer, J.C.; Sturm, B.J. The Dark Sides of Artificial Intelligence: An Integrated AI Governance Framework for Public Administration. Int. J. Public Adm. 2020, 43, 818–829. [Google Scholar] [CrossRef]
  13. Bovens, M.; Zourisdis, S. From Street-Level to System-Level Bureaucracies: How Information and Communication Technology is Transforming Administrative Discretion and Constitutional Control. Public Adm. Rev. 2002, 62, 174–184. [Google Scholar] [CrossRef]
  14. Busch, P.A.; Henriksen, H.Z. Digital Discretion: A Systematic Literature Review of ICT and Street-level Discretion. Inf. Polity 2018, 23, 3–28. [Google Scholar] [CrossRef]
  15. Young, M.M.; Bullock, J.; Lecy, J.D. Artificial Discretion as a Tool of Governance: A Framework for Understanding the Impact of Artificial Intelligence on Public Administration. Perspect. Public Manag. Gov. 2019, 2, 301–314. [Google Scholar] [CrossRef]
  16. Vogl, T.M.; Siedelin, C.; Ganesh, B.; Bright, J. Algorithmic Bureaucracy. In Proceedings of the dg.o 2019: 20th Annual International Conference on Digital Government Research (dg.o 2019), Dubai, United Arab Emirates, 18 June 2019; pp. 148–153. [Google Scholar]
  17. Madhavan, R.; Kerr, J.A.; Corcos, A.R.; Isaacoff, B.P. Toward Trustworthy and Responsible Artificial Intelligence Policy Development. IEEE Intell. Syst. 2020, 35, 103–108. [Google Scholar] [CrossRef]
  18. Fukuyama, F. What is Governance? Governance 2013, 26, 347–368. [Google Scholar] [CrossRef]
  19. Smith, S.R.; Ingram, H. Policy Tools and Democracy. In The Tools of Government: A Guide to the New Governance; Salamon, L.M., Ed.; Oxford University Press: New York, NY, USA, 2002; pp. 565–584. [Google Scholar]
  20. Lynn, L.E., Jr.; Heinrich, C.J.; Hill, C.J. Studying Governance and Public Management: Challenges and Prospects. J. Public Adm. Res. Theory 2000, 10, 233–262. [Google Scholar] [CrossRef]
  21. Chen, Y.-C. Managing Digital Governance: Issues, Challenges, and Solutions; Routledge: New York, NY, USA; Taylor & Francis Group: London, UK, 2017. [Google Scholar]
  22. Salamon, L.M. (Ed.) The Tools of Government: A Guide to the New Governance; Oxford University Press: Oxford, UK; New York, NY, USA, 2002. [Google Scholar]
  23. Shneiderman, B. Bridging the Gap Between Ethics and Practice. ACM Trans. Interact. Intell. Syst. 2020, 10, 1–31. [Google Scholar] [CrossRef]
  24. Wirtz, B.W.; Müller, W.M. An integrated artificial intelligence framework for public management. Public Manag. Rev. 2019, 21, 1076–1100. [Google Scholar] [CrossRef]
  25. Zuiderwijk, A.; Chen, Y.-C.; Salem, F. Implications of the use of artificial intelligence in public governance: A systematic literature review and a research agenda. Gov. Inf. Q. 2021, 38, 101577. [Google Scholar] [CrossRef]
  26. Shin, D.; Fotiadis, A.; Yu, H. Prospectus and Limitations of Algorithmic Governance: An Ecological Evaluation of Algorithmic Trends. Digit. Policy Regul. Gov. 2019, 21, 369–383. [Google Scholar] [CrossRef]
  27. Bannister, F.; Connolly, R. Administration by algorithm: A risk management framework. Inf. Polity 2020, 25, 471–490. [Google Scholar] [CrossRef]
  28. Nilsson, N. The Quest for Artificial Intelligence; Cambridge University Press: Cambridge, UK, 2010. [Google Scholar]
  29. Stone, P.; Brooks, R.; Brynjolfsson, E.; Calo, R.; Etzioni, O.; Hager, G.; Hirschberg, J.; Kalyanakrishnan, S.; Kamar, E.; Kraus, S.; et al. Artificial Intelligence and Life in 2030; Stanford University: Stanford, CA, USA, 2016. [Google Scholar]
  30. Dwivedi, Y.K.; Hughes, L.; Ismagilova, E.; Aarts, G.; Coombs, C.; Crick, T.; Duan, Y.; Dwivedi, R.; Edwards, J.; Eirug, A.; et al. Artificial Intelligence (AI): Multidisciplinary perspectives on emerging challenges, opportunities, and agenda for research, practice and policy. Int. J. Inf. Manag. 2021, 57, 101994. [Google Scholar] [CrossRef]
  31. Russell, S.J.; Norvig, P. Artificial Intelligence: A Modern Approach, 3rd ed.; Pearson India: Bengaluru, India, 2015. [Google Scholar]
  32. U.S. National Science Foundation. National Artificial Intelligence (AI) Research Institutes: Accelerating Research, Transforming Society, and Growing the American Workforce (Program Solicitation); NSF, Ed.; U.S. National Science Foundation: Alexandria, VA, USA, 2019.
  33. Bechmann, A.; Bowker, G.C. Unsupervised by any other name: Hidden layers of knowledge production in artificial intelligence on social media. Big Data Soc. 2019, 6, 205395171881956. [Google Scholar] [CrossRef] [Green Version]
  34. Coglianese, C.; Lehr, D. Regulating by Robot: Administrative Decision Making in the Machine-Learning Era. Georget. Law J. 2017, 105, 1147. [Google Scholar]
  35. Henman, P. Improving public services using artificial intelligence: Possibilities, pitfalls, governance. Asia Pac. J. Public Adm. 2020, 42, 209–221. [Google Scholar] [CrossRef]
  36. Davenport, T.; Kalakota, R. The potential for artificial intelligence in healthcare. Future Healthc. J. 2019, 6, 94–98. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Bryson, J.; Sancino, A.; Benington, J.; Sørensen, E. Towards a multi-actor theory of public value cocreation. Public Manag. Rev. 2017, 19, 640–654. [Google Scholar] [CrossRef]
  38. Nabatchi, T. Public Values Frames in Administration and Governance. Perspect. Public Manag. Gov. 2018, 1, 59–72. [Google Scholar] [CrossRef]
  39. Bullock, J. Artificial Intelligence, Discretion, and Bureaucracy. Am. Rev. Public Adm. 2019, 49, 751–761. [Google Scholar] [CrossRef]
  40. Barth, T.J.; Arnold, E. Artificial Intelligence and Administrative Discretion. Am. Rev. Public Adm. 1999, 29, 332–351. [Google Scholar] [CrossRef]
  41. Bannister, F.; Connolly, R. ICT, Public Values and Transformative Government: A Framework and Programme for Research. Gov. Inf. Q. 2014, 31, 119–128. [Google Scholar] [CrossRef]
  42. Pang, M.-S.; Lee, G.; Delone, W.H. IT Resources, Organizational Capabilities, and Value Creation in Public-Sector Organizations: A Public-value Management Perspective. J. Inf. Technol. 2014, 29, 187–205. [Google Scholar] [CrossRef]
  43. Cordella, A.; Bonina, C.M. A Public Value Perspective for ICT enabled Public Sector Reforms: A Theoretical Reflection. Gov. Inf. Q. 2012, 29, 512–520. [Google Scholar] [CrossRef]
  44. Chen, T.; Guo, W.; Gao, X.; Liang, Z. AI-based self-service technology in public service delivery: User experience and influencing factors. Gov. Inf. Q. 2020, 38, 101520. [Google Scholar] [CrossRef]
  45. Raab, C.D. Information privacy, impact assessment, and the place of ethics. Comput. Law Secur. Rev. 2020, 37, 105404. [Google Scholar] [CrossRef]
  46. Winfield, A.F.; Michael, K.; Pitt, J.; Evers, V. Machine Ethics: The Design and Governance of Ethical AI and Autonomous Systems [Scanning the Issue]. Proc. IEEE 2019, 107, 509–517. [Google Scholar] [CrossRef] [Green Version]
  47. Holton, R.; Boyd, R. ‘Where are the people? What are they doing? Why are they doing it?’(Mindell) Situating artificial intelligence within a socio-technical framework. J. Sociol. 2019, 57, 179–195. [Google Scholar] [CrossRef]
  48. Latour, B. Reassembling the Social: An Introduction to Actor-Network-Theory; Oxford University Press: Oxford, UK, 2005. [Google Scholar]
  49. Janssen, M.; Brous, P.; Estevez, E.; Barbosa, L.S.; Janowski, T. Data governance: Organizing data for trustworthy Artificial Intelligence. Gov. Inf. Q. 2020, 37, 101493. [Google Scholar] [CrossRef]
  50. Reddy, S.; Allan, S.; Coghlan, S.; Cooper, P. A governance model for the application of AI in health care. Journal of the American Med. Inform. Assoc. 2020, 27, 491–497. [Google Scholar] [CrossRef]
  51. Fountain, J. Building the Virtual State: Information Technology and Institutional Change; Brookings Institution Press: Washington, DC, USA, 2001. [Google Scholar]
  52. Desouza, K.C.; Dawson, G.S.; Chenok, D. Designing, developing, and deploying artificial intelligence systems: Lessons from and for the public sector. Bus. Horiz. 2020, 63, 205–213. [Google Scholar] [CrossRef]
  53. Fatima, S.; Desouza, K.C.; Dawson, G.S. National strategic artificial intelligence plans: A multi-dimensional analysis. Econ. Anal. Policy 2020, 67, 178–194. [Google Scholar] [CrossRef]
  54. Select Committee on Artificial Intelligence of the National Science & Technology Council. The National Artificial Intelligence Research and Development Strategic Plan: 2019 Update; Executive Office of the President: Washington, DC, USA, 2019.
  55. Bannister, F.; Connolly, R. Definiing E-Governance. E Serv. J. 2012, 8, 3–25. [Google Scholar] [CrossRef]
  56. Williamson, B. Governing software: Networks, databases and algorithmic power in the digital governance of public education. Learn. Media Technol. 2015, 40, 83–105. [Google Scholar] [CrossRef] [Green Version]
  57. Williamson, B. Knowing public services: Cross-sector intermediaries and algorithmic governance in public sector reform. Public Policy Adm. 2014, 4, 292–312. [Google Scholar] [CrossRef]
  58. Hsieh, H.-F.; Shannon, S.E. Three Approaches to Qualitative Content Analysis. Qual. Health Res. 2005, 15, 1277–1288. [Google Scholar] [CrossRef] [PubMed]
  59. König, P.D.; Wenzelburger, G. Opportunity for renewal or disruptive force? How artificial intelligence alters democratic politics. Gov. Inf. Q. 2020, 37, 101489. [Google Scholar] [CrossRef]
  60. Larsson, S. On the Governance of Artificial Intelligence through Ethics Guidelines. Asian J. Law Soc. 2020, 7, 437–451. [Google Scholar] [CrossRef]
  61. Wirtz, B.W.; Weyerer, J.C.; Kehl, I. Governance of artificial intelligence: A risk and guideline-based integrative framework. Gov. Inf. Q. 2022, 39, 101685. [Google Scholar] [CrossRef]
  62. Djeffal, C.; Siewert, M.B.; Wurster, S. Role of the state and responsibility in governing artificial intelligence: A comparative analysis of AI strategies. J. Eur. Public Policy 2022, 29, 1799–1821. [Google Scholar] [CrossRef]
  63. Erman, E.; Furendal, M. The global governance of artificial intelligence: Some normative concerns. Moral Philos. Politics 2022, 9, 267–291. [Google Scholar] [CrossRef]
  64. Floridi, L.; Cowls, J.; Beltrametti, M.; Chatila, R.; Chazerand, P.; Dignum, V.; Luetge, C.; Madelin, R.; Pagallo, U.; Rossi, F.; et al. AI4People—An Ethical Framework for a Good AI Society: Opportunities, Risks, Principles, and Recommendations. Minds Mach. 2018, 28, 689–707. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  65. Adadi, A.; Berrada, M. Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI). IEEE Access 2018, 6, 52138–52160. [Google Scholar] [CrossRef]
  66. Fisher, S.; Rosella, L.C. Priorities for successful use of artificial intelligence by public health organizations: A literature review. BMC Public Health 2022, 22, 2146. [Google Scholar] [CrossRef]
  67. Gulson, K.N.; Webb, P.T. Mapping an emergent field of ‘computational education policy’: Policy rationalities, prediction and data in the age of Artificial Intelligence. Res. Educ. 2017, 98, 14–26. [Google Scholar] [CrossRef]
  68. Coglianese, C.; Lehr, D. Transparency and Algorithmic Governance. Adm. Law Rev. 2019, 71, 1–56. [Google Scholar]
  69. Jimenez-Gomez, C.E.; Cano-Carrillo, J.; Falcone Lanas, F. Artificial Intelligence in Government. Computer 2020, 53, 23–27. [Google Scholar] [CrossRef]
  70. Lauterbach, A. Artificial intelligence and policy: Quo vadis? Digit. Policy Regul. Gov. 2019, 21, 238–263. [Google Scholar] [CrossRef]
  71. Lepri, B.; Staiano, J.; Sangokoya, D.; Letouzé, E.; Oliver, N. The Tyranny of Data? The Bright and Dark Sides of Data-Driven Decision-Making for Social Good; Springer International Publishing: Berlin/Heidelberg, Germany, 2017; pp. 3–24. [Google Scholar]
  72. Lodders, A.; Paterson, J.M. Scrutinising COVIDSafe: Frameworks for evaluating digital contact tracing technologies. Altern. Law J. 2020, 45, 153–161. [Google Scholar] [CrossRef]
  73. Montes, G.A.; Goertzel, B. Distributed, decentralized, and democratized artificial intelligence. Technol. Forecast. Soc. Chang. 2019, 141, 354–358. [Google Scholar] [CrossRef]
  74. Palladino, N. A ‘biased’emerging governance regime for artificial intelligence? How AI ethics get skewed moving from principles to practices. Telecommun. Policy 2022, 102479. [Google Scholar] [CrossRef]
  75. Wallach, W.; Marchant, G. Toward the Agile and Comprehensive International Governance of AI and Robotics [point of view]. Proc. IEEE 2019, 107, 505–508. [Google Scholar] [CrossRef]
  76. Butterworth, M. The ICO and artificial intelligence: The role of fairness in the GDPR framework. Comput. Law Secur. Rev. 2018, 34, 257–268. [Google Scholar] [CrossRef]
  77. Kerr, A.; Barry, M.; Kelleher, J.D. Expectations of artificial intelligence and the performativity of ethics: Implications for communication governance. Big Data Soc. 2020, 7, 205395172091593. [Google Scholar] [CrossRef]
  78. Pesapane, F.; Volonté, C.; Codari, M.; Sardanelli, F. Artificial intelligence as a medical device in radiology: Ethical and regulatory issues in Europe and the United States. Insights Into Imaging 2018, 9, 745–753. [Google Scholar] [CrossRef] [PubMed]
  79. Calo, R. Artificial Intelligence Policy: A Primer and Roadmap. Univ. Bologna Law Rev. 2018, 3, 180–218. [Google Scholar] [CrossRef]
  80. Campion, A.; Gasco-Hernandez, M.; Jankin Mikhaylov, S.; Esteve, M. Overcoming the Challenges of Collaboratively Adopting Artificial Intelligence in the Public Sector. Soc. Sci. Comput. Rev. 2020, 40, 462–477. [Google Scholar] [CrossRef]
  81. Winter, J.S.; Davidson, E. Governance of artificial intelligence and personal health information. Digit. Policy Regul. Gov. 2019, 21, 280–290. [Google Scholar] [CrossRef]
  82. Todolí-Signes, A. Algorithms, artificial intelligence and automated decisions concerning workers and the risks of discrimination: The necessary collective governance of data protection. Transfer 2019, 25, 465–481. [Google Scholar] [CrossRef]
  83. Ho, C.W.L.; Ali, J.; Caals, K. Ensuring trustworthy use of artificial intelligence and big data analytics in health insurance. Bull. World Health Organ. 2020, 98, 263–269. [Google Scholar] [CrossRef] [PubMed]
  84. Vanderelst, D.; Winfield, A. An architecture for ethical robots inspired by the simulation theory of cognition. Cogn. Syst. Res. 2018, 48, 56–66. [Google Scholar] [CrossRef]
  85. Nishant, R.; Kennedy, M.; Corbett, J. Artificial intelligence for sustainability: Challenges, opportunities, and a research agenda. Int. J. Inf. Manag. 2020, 53, 102104. [Google Scholar] [CrossRef]
  86. Meske, C.; Bunde, E.; Schneider, J.; Gersch, M. Explainable Artificial Intelligence: Objectives, Stakeholders, and Future Research Opportunities. Inf. Syst. Manag. 2020, 39, 53–63. [Google Scholar] [CrossRef]
  87. Kim, E.-S. Deep learning and principal–agent problems of algorithmic governance: The new materialism perspective. Technol. Soc. 2020, 63, 101378. [Google Scholar] [CrossRef]
  88. Linkov, I.; Trump, B.D.; Anklam, E.; Berube, D.; Boisseasu, P.; Cummings, C.; Ferson, S.; Marie-Valentine, F.; Goldstein, B.; Hristozov, D.; et al. Comparative, collaborative, and integrative risk governance for emerging technologies. Environ. Syst. Decis. 2018, 38, 170–176. [Google Scholar] [CrossRef]
  89. Engin, Z.; Treleaven, P. Algorithmic Government: Automating Public Services and Supporting Civil Servants in using Data Science Technologies. Comput. J. 2019, 62, 448–460. [Google Scholar] [CrossRef] [Green Version]
  90. Sachs, J.D. Some Brief Reflections on Digital Technologies and Economic Development. Ethics Int. Aff. 2019, 33, 159–167. [Google Scholar] [CrossRef]
  91. Popenici, S.A.D.; Kerr, S. Exploring the impact of artificial intelligence on teaching and learning in higher education. Res. Pract. Technol. Enhanc. Learn. 2017, 12, 22. [Google Scholar] [CrossRef] [PubMed]
  92. Harrison, T.M.; Luna-Reyes, L.F. Cultivating Trustworthy Artificial Intelligence in Digital Government. Soc. Sci. Comput. Rev. 2022, 40, 494–511. [Google Scholar] [CrossRef]
  93. Siau, K.; Wang, W. Building trust in artificial intelligence, machine learning, and robotics. Cut. Bus. Technol. J. 2018, 31, 47–53. [Google Scholar]
  94. Nazerdeylami, A.; Majidi, B.; Movaghar, A. Autonomous litter surveying and human activity monitoring for governance intelligence in coastal eco-cyber-physical systems. Ocean. Coast. Manag. 2021, 200, 105478. [Google Scholar] [CrossRef]
  95. deSousa, W.G.; deMelo, E.R.P.; Bermejo, P.H.D.S.; Farias, R.A.S.; Gomes, A.O. How and where is artificial intelligence in the public sector going? A literature review and research agenda. Gov. Inf. Q. 2019, 36, 101392. [Google Scholar]
  96. Mikhaylov, S.J.; Esteve, M.; Campion, A. Artificial Intelligence for the Public Sector: Opportunities and Challenges of Cross-sector Collaboration. Philos. Trans. R. Soc. A 2018, 376, 20170357. [Google Scholar] [CrossRef] [Green Version]
  97. Sun, T.Q.; Medaglia, R. Mapping the challenges of Artificial Intelligence in the public sector: Evidence from public healthcare. Gov. Inf. Q. 2019, 36, 362–383. [Google Scholar] [CrossRef]
  98. Stritch, J.M.; Pedersen, M.J.; Taggart, G. The Opportunities and Limitations of Using Mechanical Turk (MTURK) in Public Administration and Management Scholarship. Int. Public Manag. J. 2017, 20, 489–511. [Google Scholar] [CrossRef] [Green Version]
  99. Marvel, J.D.; Girth, A.M. Citizen Attributions of Blame in Third-Party Governance. Public Adm. Rev. 2016, 76, 96–108. [Google Scholar] [CrossRef]
  100. Pencheva, I.; Esteve, M.; Mikhaylov, S.J. Big Data and AI–A transformational shift for government: So, what next for research? Public Policy Adm. 2020, 35, 24–44. [Google Scholar] [CrossRef]
  101. Mergel, I.; Edelmann, N.; Haug, N. Defining digital transformation: Results from expert interviews. Gov. Inf. Q. 2019, 36, 101385. [Google Scholar] [CrossRef]
  102. Panagiotopoulos, P.; Klievink, B.; Cordella, A. Public value creation in digital government. Gov. Inf. Q. 2019, 36, 101421. [Google Scholar] [CrossRef]
  103. Young, M.M.; Himmelreich, J.; Bullock, J.B.; Kim, K.-C. Artificial Intelligence and Administrative Evil. Perspect. Public Manag. Gov. 2021, 4, 244–258. [Google Scholar] [CrossRef]
  104. Marsh, S. Councils Scrapping Use of Algorithms in Benefit and Welfare Decisions. The Guardian. 24 August 2020. Available online: https://www.theguardian.com/society/2020/aug/24/councils-scrapping-algorithms-benefit-welfare-decisions-concerns-bias (accessed on 12 January 2023).
  105. Government Accountability Office. Artificial Intelligence: An Accountability Framework for Federal Agencies and Other Entities; Government Accountability Office: Washington, DC, USA, 2021; p. 112.
Figure 1. Systematic literature review of artificial intelligence and governance: PRISMA.
Figure 1. Systematic literature review of artificial intelligence and governance: PRISMA.
Sustainability 15 04796 g001
Table 1. Public values related to artificial intelligence and its governance.
Table 1. Public values related to artificial intelligence and its governance.
Public Values
Duty OrientationResponsibility to citizens [1,12,23,35,59,60,61,62,63,64]
Compliance with laws [1,35]
Service OrientationTransparency [1,17,35,50,60,63,65,66,67,68,69,70,71,72,73,74]
Effectiveness [15,23,39,75]
Efficiency [15,39]
Social OrientationAccountability to public [1,12,15,35,49,50,61,63,66,69,71,72,74,76,77,78]
Privacy [3,12,49,50,60,64,74,79,80,81]
Equality of treatment and access [3,12,15,39,49,72,79,82]
Fairness [9,17,50,60,61,63,66,74,76]
Justice [35,60,64,79]
Due process [35]
Inclusiveness [23,74]
Security [64]
Table 2. Artificial intelligence governance challenges.
Table 2. Artificial intelligence governance challenges.
Governance Challenges
Public Value ChallengesLack of transparency [3,15,36,50,65,68,70,73,83]
Lack of accountability [1,12,15,24,36,59,61]
Privacy concerns [3,12,36,50,72,79,80,81]
Inequity concerns [3,12,73,77,79]
Lack of responsibility [1,3,12,61,64]
Safety concerns [3,12,84]
Malicious use of AI [1,77]
Moral dilemmas [3,12]
Social acceptance and trust concerns [3,12]
Administrative evil [9,39]
Cybersecurity risks [85]
Discrimination in recruitment, promotion, and dismissals in organizations [82]
Fairness concerns in data processing [76,86]
Lack of AI expertise and knowledge [3]
Lack of responsiveness [59]
Principal–agent problem [87]
Sustainability challenges [88]
Violation of laws [60]
Data Quality, Processing, and Outcome ChallengesData quality and management [3,17,49,50,71,83,85,89]
AI rule-making concerns [3,12]
Lost control of AI [12,64]
Adverse impacts, difficulty of measuring the performance of AI, and uncertain human behavioral responses to AI-based interventions [85]
Financial feasibility [3]
Interaction problem with humans [12]
Societal Governance ChallengesReplacing human jobs [1,3,12,90,91]
Insufficient regulation and “soft laws” [75]
Replacing human discretion [39,64]
Value judgment concerns [3,12]
Authoritarian abuses [17]
Cross-sector collaboration hardship [70]
Power asymmetry [86]
Threatening autonomy [81]
Table 3. Artificial intelligence governance solutions.
Table 3. Artificial intelligence governance solutions.
Governance Solutions
Public Values SolutionsExplainable AI [59,61,65,68,74,86]
Inclusion of more public values [50,60,66,71,72,74]
Ethical AI [46,61,74,77]
Distributed, decentralized, and democratized market [73]
Ethical agent [46]
Impacts on bureaucratic discretion [39]
Privacy by design [35]
Trustworthy AI [92,93]
Data Quality and Processing sSolutionsRegulation [12,17,35,74,76,78,82,88]
Bias assessment [35,74]
Data audit [74,81]
Data-driven digital government [69]
Data-sharing agreement [81]
Human audition [61,74,94]
Independent quality assurance [35]
Oversight committee [74,81]
Recognition and removal of bias [83]
Understanding of AI [64]
Societal Governance SolutionsCollaborative governance [66,70,80,82,95]
Multilevel approach [49,85]
Artificial discretion analysis [15]
Facilitative leadership, alignment of goals and objectives, shared knowledge, socialization, expert insights, and strategies [74,96]
Governance coordinating committee [75]
Holistic industrywide solution with governmental involvement [23]
Integration of workflow and governance [50]
Levels of governance [23]
Pluralist approach [87]
Risk governance [88]
Stakeholder participation [74,83]
Systems dynamics approaches [85]
Task characteristics (complexity and uncertainty) [39]
Table 4. Potential impacts of artificial intelligence on public values.
Table 4. Potential impacts of artificial intelligence on public values.
No ImprovementModestSubstantialTransformativeSubstantial and Transformative Combined
Decision Making24
(7.43%)
102
(31.58%)
126
(39.01%)
71
(21.98%)
60.99%
Effectiveness12
(3.72%)
100
(30.96%)
154
(47.68%)
57
(17.65%)
65.3%
Efficiency9
(2.79%)
63
(19.75%)
147
(45.51%)
104
(32.20%)
77.7%
Accountability43
(13.31%)
106
(32.82%)
123
(38.08%)
51
(15.79%)
53.8%
Table 5. Governance challenges.
Table 5. Governance challenges.
Development of AI in Government Is Likely to Negatively Impact Transparency Due to the Technical Nature of the AlgorithmUse of AI in My Government Agency Will Take Away Your Discretionary AuthorityGoal Setting of AI Use in Government Cannot Fully Consider Societal Values beyond Technical Efficiency
1 (Strongly Disagree)22 (6.81%)27 (8.36%)13 (4.02%)
2 (Disagree)40 (12.38%)29 (8.98%)33 (10.22%)
3 (Slightly Disagree)37 (11.46%)46 (14.24%)54 (16.72%)
4 (Neutral)68 (21.05%)55 (17.03%)70 (21.67%)
5 (Slightly Agree)57 (17.65%)69 (21.36%)64 (19.81%)
6 (Agree)67 (20.74%)54 (16.72%)52 (16.10%)
7 (Strongly Agree)32 (9.91%)43 (13.31%)37 (11.46%)
Mean4.324.374.37
Median454
Mode454
Table 6. Opinions on transparency governance solutions.
Table 6. Opinions on transparency governance solutions.
Do You Agree That the Information on the Development of Algorithms Used in the AI System Should Be Made Available to Government Managers?Do You Agree That the Information on the Development of Algorithms Used in the AI System Should Be Made Available to the General Public?Do You Agree That Information about the Data That the AI System Uses Should Be Made Available to the Government Managers Responsible for the Service?Do You Agree That Information about the Data That the AI System Uses Should Be Made Available to the General Public?
1 (Strongly Disagree)2 (0.62%)12 (3.72%)4 (1.24%)5 (1.55%)
2 (Disagree)13 (4.02%)23 (7.12%)7 (2.17%)22 (6.81%)
3 (Slightly Disagree)18 (5.57%)12 (3.72%)10 (3.10%)27 (8.36%)
4 (Neutral)49 (15.17%)68 (21.05%)43 (13.31%)63 (19.50%)
5 (Slightly Agree)86 (26.63%)77 (23.84%)101 (31.27%)70 (21.67%)
6 (Agree)86 (26.63%)75 (23.22%)92 (28.48%)71 (21.98%)
7 (Strongly Agree)69 (21.36%)56 (17.34%)66 (20.43%)65 (20.12%)
Mean5.2854.9325.3844.994
Median5555
Table 7. Most important stage for participation.
Table 7. Most important stage for participation.
Participants
/Stages
Public OfficialsMembers of the General Public
StagesCountPercentageCountPercentage
Goal Setting for Government AI use4714.55%4714.55%
Development of the AI System8726.93%6821.05%
Use of AI Systems in Government Decisions5015.48%8325.70%
Impact Assessment of AI-Enabled Decisions237.12%4814.86%
All of the Above11635.91%7723.84%
Total323100%323100%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Y.-C.; Ahn, M.J.; Wang, Y.-F. Artificial Intelligence and Public Values: Value Impacts and Governance in the Public Sector. Sustainability 2023, 15, 4796. https://doi.org/10.3390/su15064796

AMA Style

Chen Y-C, Ahn MJ, Wang Y-F. Artificial Intelligence and Public Values: Value Impacts and Governance in the Public Sector. Sustainability. 2023; 15(6):4796. https://doi.org/10.3390/su15064796

Chicago/Turabian Style

Chen, Yu-Che, Michael J. Ahn, and Yi-Fan Wang. 2023. "Artificial Intelligence and Public Values: Value Impacts and Governance in the Public Sector" Sustainability 15, no. 6: 4796. https://doi.org/10.3390/su15064796

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop