Anthropocentrism and Environmental Wellbeing in AI Ethics Standards: A Scoping Review and Discussion

: As AI deployment has broadened, so too has an awareness for the ethical implications and problems that may ensue from this deployment. In response, groups across multiple domains have issued AI ethics standards that rely on vague, high-level principles to ﬁnd consensus. One such high-level principle that is common across the AI landscape is ‘human-centredness’, though oftentimes it is applied without due investigation into its merits and limitations and without a clear, common deﬁnition. This paper undertakes a scoping review of AI ethics standards to examine the commitment to ‘human-centredness’ and how this commitment interacts with other ethical concerns, namely, concerns for nonhumans animals and environmental wellbeing. We found that human-centred AI ethics standards tend to prioritise humans over nonhumans more so than nonhuman-centred standards. A critical analysis of our ﬁndings suggests that a commitment to human-centredness within AI ethics standards accords with the deﬁnition of anthropocentrism in moral philosophy: that humans have, at least, more intrinsic moral value than nonhumans. We consider some of the limitations of anthropocentric AI ethics, which include permitting harm to the environment and animals and undermining the stability of ecosystems.


Introduction
As artificial intelligence (AI) deployment has broadened, so too has an awareness for the ethical implications and problems that may ensue from this deployment.In response, groups across multiple domains have issued AI ethics standards, ranging from short lists of principles to extensive documents explicating AI ethics problems and instructions on how to mitigate these.New AI ethics standards are continually being published, so it is difficult to ascertain exactly how many currently exist.As a general guide, the AI Ethics Guidelines Global Inventory counted 167 AI ethics guidelines in May 2022.
There are now so many individual AI ethics guidelines that groups have begun to survey standards, summarising and consolidating common themes or assertions about the AI ethics landscape [1].Surveying the AI ethics landscape is now a well-established methodology for examining the trends, themes, and values within current AI research and development (some examples include [2][3][4][5][6][7]).Surveys combine a range of sources, such as corporate, government, international coalitions, and non-profits, and include a range of formats, ranging from white-papers to professional codes of ethics and reports from private industry.
There is a lack of methodological and theoretical consensus across such a wide variety of AI ethics standards.Crawford [1] argues that this has led to considerable limitations, including conflicting or vague ideals and definitions, and a lack of overarching or standard protocols.Across the AI ethics space, there is little agreement on what is important and how ethical problems that relate to AI ought to be mitigated [8].The ethical positions that are used in AI ethics tend to be based on principles as mid-level, applicable, and action-guiding concepts [8].Mittelstadt argues that these 'broadly acceptable' and 'vague' principles are applied in order to find some consensus in an otherwise nebulous space [9] (p.503).Indeed, despite the vast number of AI ethics standards, and despite the lack of consensus, the range of principles applied in AI ethics standards is limited [8,10].In perhaps the most notable survey of AI ethics standards, Jobin et al. [5] (p.395) found the most prominent principles across AI ethics standards included transparency, justice and fairness, non-maleficence, and responsibility.
Such vague principles can oftentimes be prescribed without due investigation into their merit and limitations and without clear, established definitions.One such accepted ethical prescription that has seen a notable increase without due investigation is a commitment to 'human-centredness'. AI ethics principles are often directly linked to established human rights [6], and human-centredness is increasing as a prescribed ethical grounding for AI development and deployment [11].However, the application of 'human-centredness' comes with a range of different definitions [12] and without investigation or instruction on how this ought to interact with other commitments, such as environmental sustainability.
In what is the most prominent and perhaps only survey of AI ethics frameworks from an environmental lens, Owe and Baum [13] re-examined the surveys of Jobin et al. [5] and Baum [14] in terms of concern for nonhumans and anthropocentrism, as it is defined in moral philosophy.Owe and Baum found that, in theory, moral consideration for nonhumans is compatible and consistent with certain anthropocentric perspectives and with many key principles across existing AI ethics frameworks [13] (pp.519, 525).These principles include transparency and explainability, diversity, non-discrimination, fairness, privacy, and data governance-some of the major ethical principles noted by Jobin et al. [5].Nevertheless, they argued, "specific treatments of the principles commonly neglect nonhumans" and that concern for nonhumans is vastly outweighed in AI ethics by concern for humans [13] (pp.519-520).
Though Owe and Baum [13] uncover some key trends across the AI ethics landscape with regards to environmental wellbeing through their qualitative review, some questions remain.In particular, it is unclear whether broader applications of the term 'human-centred AI' within AI ethics standards are consistent with anthropocentrism, as defined in moral philosophy, and how these commitments to human-centred AI interact with the moral considerations for nonhumans.

Objectives and Contributions of This Work
The overall objective of this work was to examine the commitments to humancentredness within AI ethics standards and how these commitments interact with other ethical considerations, namely, environmental wellbeing.To approach this, we undertook a scoping review of AI ethics standards, particularly by examining the commitments to human-centredness and considerations for humans and nonhumans.We also critically analysed our findings within the context of human-centred AI more broadly and within the context of anthropocentrism.Overall, we found that the application of human-centredness in AI ethics conforms with the definition of anthropocentrism as established and applied in moral philosophy, namely, that humans have more moral value than nonhumans.The contributions of this work are: • A discussion of human-centredness in AI, particularly the historical roots, groundings, definitions, and applications of this commitment; • A scoping review of human-centredness and the moral considerations of humans and nonhumans across 146 AI ethics standards; • An examination of how applications of anthropocentrism, as defined in moral philosophy, would play out in the development and deployment of AI systems.
Section 2 discusses the background of human-centredness in AI and the philosophical groundings of anthropocentrism.Here, we outline the supposed gap between humancentredness, which is considered as a vague, high-level concept to unify and find consensus among the AI landscape, and anthropocentrism, which is considered as the well-established claim for moral hierarchies between humans, nonhumans, and the environment.Section 3 will outline the methodology employed to survey the AI ethics landscape for both humancentredness and environmental concern.The results of this survey will be discussed in Section 4. Section 5 will critically analyse the results within the context of anthropocentrism and human-centred AI more broadly.Here, we argue that human-centredness in AI ethics reflects the moral hierarchies of anthropocentrism.We also discuss how these anthropocentric moral hierarchies play out in practice in the regulation of AI systems in Section 5.1.

Background 2.1. Asimov
Early AI ethics standards drew inspiration from Isaac Asimov's laws of robotics from his 1950s science fiction novel, I, Robot.These are seen as the first governing laws for autonomous technology, explicating an implicit consensus to use and develop technology for the good of humanity [15].Asimov's three original laws for robotics are as follows: a robot may not injure a human being or, through inaction, allow a human being to come to harm; a robot must obey orders given by human beings except where such orders would conflict with the first law; and a robot must protect its own existence as long as such protection does not conflict with the first or second laws [16].A fourth law was later added by Asimov, which can be seen across the AI ethics landscape even today: a robot may not harm humanity, or, by inaction, allow humanity to come to harm [15,17].Together, these laws have been used as a benchmark for "mainstream" AI ethics [18] (p.209) or as the "essential requirement" for ethical AI [19] (p.1409).
Murphy and Woods [20] suggest that Asimov's laws have been so successfully inculcated into the public consciousness through entertainment that they now appear to have shaped society's expectations about how technology and humans ought to interact.Multiple domains, including philosophy, AI, and medicine, have discussed the ethics of robots in society using Asimov's laws as a reference [20].In fact, a draft of the European Civil Law Rules of Robotics prescribed Asimov's three laws to designers, producers, and operators of robots, including robots assigned with built-in autonomy and self-learning (sec.T. [21]).On reflection, a later iteration of this civil law conceded that it was culturally and scientifically "wrong" to mention Asimov's laws as the work of fiction could not be taken as true legal principles [22] (p.8).While Asmivo's laws were described as "unfit to protect humanity", it was nevertheless suggested that a framework could be drawn to ensure that future AI benefits humanity [22] (p.13).This later work posited new rules that remain hauntingly familiar: "The first principle of roboethics is to protect humans against any harm caused by a robot" [22] (p.20).

HCAI
Today, a commitment to protect humanity in AI development and deployment seems ubiquitous across the AI ethics landscape.Reflecting the increasing importance and integration of AI in peoples' lives, there is a move towards human-centred artificial intelligence (HCAI), where the goal is to put the human, rather than technology, at the centre of AI development [12] (pp.[25][26].
HCAI is arguably entrenched in and inspired by Asimov's laws.However, Asmiov made these laws inherently and purposefully vague, which adds to the plot of I, Robot but detracts from their usefulness as real action-guiding principles [22].Mittelstadt [9] notes that there is often no clear definition of the vague, and oftentimes contested, concepts in AI ethics, and the author also notes that without due definition, these concepts are not specific enough to be action-guiding.Indeed, there is no single, broadly accepted definition of HCAI, but many different definitions that combine the criteria from a human-centred design with AI-specific factors, such as data usage, bias, and uncertainty of outcomes [12].Some of the applications of the term HCAI include: augmenting rather than displacing human abilities; explainability and accountability; preserving human control; fair and trustworthy use of data; and providing efficient solutions to users' problems [12].
Without due analysis of the merits and drawbacks of human-centredness as well as an investigation of alternative approaches to AI ethics, it is unclear as to why this commitment ought to be adopted over any other normative grounding.Moreover, without clear definitions of human-centredness, it is unclear exactly what HCAI ought to refer to.This is particularly the case when human-centredness is applied to AI ethics as anthropocentrism has a unique, established definition in moral philosophy.

Anthropocentrism
In moral philosophy, anthropocentrism, which literally translates to human-centredness, is the view that only humans have intrinsic moral value or, at the very least, humans have more intrinsic moral value than nonhumans [23].In philosophy, intrinsic value is commonly defined as value that an object has for its own sake, whereas extrinsic value is value that an object has for the sake of something else or in relation to some other object [24].This definition of anthropocentrism dates back to antiquity, heralded most notably by Aristotle, who summated, "[nature] has made all animals specifically for the sake of man" [25] (p.1137).This definition of anthropocentrism has been consistently applied within philosophy and has been continued by, among others: Aquinas, who argued, "But plants exist for the sake of animals; indeed, some animals exist for the sake of others, and all exist for the sake of man" [26] (Bk.3, Pt. 2, Ch.127); Kant, who stated, "all animals exist only as a means, and not for their own sakes, in that they have no self-consciousness, whereas man is the end" (27:495) [27]; and, more recently, Passmore, who declared, "I treat human interests as paramount.I do not apologise for that fact" [28] (p.187).
Anthropocentrism can be divided into strong or moderate.Strong anthropocentrism is the claim that humans are the only bearers of intrinsic moral value: "On a bold version of the view, only humans count, morally speaking; nonhumans don't count at all" [29] (p.2).On this strong anthropocentric account, nonhumans have the most extrinsic value to humans: for survival, pleasure, monetary gain, and so on.Meanwhile, moderate anthropocentrism claims that nonhumans still have intrinsic value, though less so than humans [29] (p.2).This means that nonhumans or the environment are not solely valuable for the sake of humans but matter for their own sake, albeit, less than humans.
Simply put, both strong and moderate anthropocentrists hold that there is something morally special about humans that gives them significant moral value over nonhumans.This means that, under anthropocentrism, in cases where environmental or nonhuman wellbeing conflicts with human wellbeing, human wellbeing wins out.
Crucially, this definition of anthropocentrism differs from some uses of the term 'human-centred' in AI or in HCAI, particularly in the uses found in design or safety, which aim to ensure the safety and wellbeing of human users of AI; however, this prioritisation of humans does appear in other instances of HCAI.Consider, for instance, "empower both individuals and society" or "augment rather than displace human abilities, as HCAI seeks to enhance human performance and human-AI collaboration by integrating artificial and human intelligence" [12] (p.2).Here, the wellbeing of specifically human individuals is seen as of the utmost importance.Consider also the high-level expert group that states that reasoning between conflicting principles should never violate "fundamental rights and correlated principles" such as "human dignity", as such principles are "absolute and cannot be subject to a balancing exercise" [30] (p.13).These examples of prioritising human beings may intend to place human wellbeing over and above technology, monetary gain, or machine efficiency as opposed to over and above nonhuman wellbeing or environmental sustainability.However, this is not made explicit.Herein lies the problem of relying on vague, high-level principles: though they are likely to find consensus, in doing so, we are left without clear definitions of terms and without guidance on how to use and balance these principles with other commitments and concerns.

Environmental Ethics and the Value of Nonhumans
There is a long-standing debate within environmental ethics over anthropocentrism and the value of nonhumans: are nonhumans valuable in and of themselves, or are they valuable for the sake of something else, such as humans; does harming nonhumans in turn harm humans?
There are many reasons why humans would want to extend care to the nonhuman world.We rely on the nonhuman world for shelter, food, and medicine and, in many cases, environmental destruction motivated by monetary gain or pleasure has threatened not only the wellbeing of nonhumans, but our own survival as well.
Anthropocentrism can be consistent with extending moral consideration to the nonhuman world.Moderate anthropocentrists maintain that nonhumans have some intrinsic value, and so the wanton destruction of nonhumans would be impermissible in many cases.Moreover, even if humans are the sole object of intrinsic value, as strong anthropocentrists claim, we can still have indirect duties to nonhumans for the sake of humans.
For instance, Kant argued that our duties to nonhuman objects "allude, indirectly, to our duties towards men" (27:460) [27].Causing needless destruction to the natural world indirectly affects other humans, as others may have use for it (27:460) [27].Moreover, the wanton destruction of flora and harm caused to nonhuman animals will impede on our own aspirations towards moral perfection, and as this moral perfection is the overarching goal of humans, we have indirect duties to not cause needless harm to the nonhuman world [27,31].More recently, Passmore states that we ought not to impede on the freedom of any human by "destroying the natural world which makes that freedom possible" [28] (p.195).
Meanwhile, non-anthropocentric environmental ethicists argue that nonhumans do have intrinsic value, with varying degrees of inclusivity.Sentiocentrists argue that all sentient animals, that is, those with the capacity to suffer, deserve equal moral consideration with humans [32].Biocentrists argue, more inclusively, that all living beings have intrinsic moral value in so far as they have morally significant interests, including survival, wellbeing, and flourishing, which matter beyond what they provide for humans [33,34].Even more inclusive is a relational ecocentric ethic, which considers the interdependent relations between ecological objects, including humans, nonhuman animals, and plants, such that these individuals are considered part of one interacting whole [35,36].Humans are considered "plain" members of this community, just like any other organism, which grounds the extension of moral consideration from humans alone to the wider environment [36] (p.194).Deep ecology, one branch of ecocentrism, extends the "self" to identify with the broader environment, including animals, plants, and ecosystems [37].Deep ecology describes all parts of this broader "self" as having intrinsic value [37].
The value of nonhumans remains the object of debate within environmental ethics, and different branches have different reasons to consider nonhumans as worthy of moral concern.While strong anthropocentrists may care for the environment indirectly, for the sake of humans, moderate anthropocentrists and nonanthropocentrists will see parts and even all of the nonhuman world as deserving of moral concern for its own sake.

Background Conclusion
Human-centredness remains an under-examined concept in AI.It is commonly discussed, accepted, and prescribed without due investigation into its merits and limitations, nor with a clear, shared understanding of its meaning and how it ought to be embedded into the wider framework and commitments of AI ethics.This paper examines the conformance of AI ethics standards to human-centredness by undertaking a scoping review of 146 AI ethics standards.As anthropocentrism in moral philosophy discusses the moral considerations of humans in relation to that of nonhumans, this paper will also examine ethical considerations for nonhumans and the environment throughout these standards.Upon a critical analysis of our findings, we find that human-centredness in AI reflects the definition of anthropocentrism in moral philosophy, having higher levels of moral concern for humans than for nonhumans.

Materials and Methods
This review of AI ethics standards follows the five-step methodology for a scoping review as set out by Arksey and O'Malley [38] (pp.[22][23].Paterson et al. [39] note that although there is no universally accepted definition or purpose for a scoping review [40][41][42][43], they are particularly useful in providing an overview of a broad topic [42,44] and can be flexible when the literature is vast and complex [38].Because the AI ethics landscape is indeed vast, diverse, and complex, a broader scoping review was chosen over a systemic review, which can be more narrow and limited in its focus and inclusion criteria [39].
Scoping reviews differ substantially from systematic reviews in the methodology, expectations, tools, and outputs [45].For instance, because scoping reviews are designed to provide a broad and inclusive overview of the existing evidence base, formal assessments of the bias or quality of the included studies are generally not performed [45,46].Although there is no exacting set of procedures for scoping reviews, the following five-step methodology for a scoping review as set out by [38] is commonly applied [39,42]: 1.
Identifying the research question; 2.
Collating, summarising, and reporting the results; Section 3.1 outlines the research question.Section 3.2 identifies the relevant standards.Section 3.3 discusses the sources, including the inclusion and exclusion criteria, for the standard selection.Section 4 charts the data and collates and reports the results.

Identifying the Research Question
The research question this scoping review aims to answer is: how prevalent is humancentredness in AI ethics standards, along with moral considerations of humans and nonhumans?

Identifying Relevant Standards
Regulation, codes of conduct, standardisation, certification, and accountability and governance frameworks are some examples of the non-technical tools for practitioners to implement and conform to AI ethics principles [30] (pp.[22][23].This work surveys 146 examples of these non-technical tools, henceforth referred to as 'standards'.

Standard Selection
The sources used to collect AI ethics standards were the study by Jobin et al., (2019) [5] and The AI Ethics Guidelines Global Inventory [47] (accessed May 2022).Jobin et al. [5] was chosen as a primary source for guidelines as this work is so far the largest and most comprehensive survey of AI ethics standards.The AI Ethics Guidelines Global Inventory was chosen as a supporting source, as this inventory includes standards not included in and/or published after Jobin et al. [5].The citations identified from the sources were compiled with the identifying review criteria: Issuer, Year, and Domain.The compiled citations were then manually screened in accordance with predetermined exclusion criteria: foreign languages, unofficial blog posts that do not constitute non-technical AI ethics compliance tools [30], duplicates, and standards that have since been deleted.In total, we identified and surveyed 146 ethical AI standards across 8 domains and published over 10 years.Figure 1 abstracts this process of identifying relevant standards.

Review Criteria
A data extraction sheet was developed through iterative pilot testing on a small subset of included articles, which refined both the design and review criteria to be objective with the aim of reducing author bias and subjectivity.Table 1 explicates the final review criteria by which each framework was surveyed.Appendix A.1 lists all of the standards that we surveyed.We also reviewed the domains under which each framework was published, which included government, non-profit, professional association, research group, political party, private, religious, and research council.The domain under which each framework was published was identified by Jobin et al. [5] and The AI Ethics Guidelines Global Inventory [47].

Inclusion of Humans and Nonhumans
We examined whether the standards included humans and/or nonhumans within moral concern.Inclusion of humans and nonhumans was defined using broad criteria, namely, did the framework include humans and/or nonhumans anywhere within the text or principles?This broad criteria were used to offer a generous perspective over the concern for humans and nonhumans across AI ethics standards.

Humans and Nonhumans in Core Principles
A binary review was made as to whether standards included a set of core principles or values.Recommendations were not included as core principles as many standards had both principles and recommendations and this work is primarily interested in the normative motivations behind the recommendations.
The content of these core principles was examined.We examined whether any of these core principles were related to human wellbeing, respect, dignity, or equivalent phrasing.For example, "AI should positively contribute to the wellbeing of humans".If so, we stated that this standard included humans in the core principles.We also examined whether any of these core principles were related to nonhumans; this might be in addition to humans.For example, "AI should not harm humans or animals".In this case, this standard was defined as including humans and nonhumans in the core principles.Alternatively, the principles might be solely related to nonhumans.For example, "AI development should be environmentally sustainable".In this case, this standard would be defined as including nonhumans in the core principles.We also examined whether nonhumans were included in the core principles for their own sake or for the sake of humans.For example, "AI development should not damage the natural environment on which humans depend for survival" would be defined as including nonhumans in the core principles for the sake of humans.Appendix A.2 includes the sources with the core principles categorised by their inclusion of humans, nonhumans, nonhumans and humans, and nonhumans for humans within their core principles.
We can see the relationship between these criteria in Figure 2 below.Within the set of standards that include humans and nonhumans anywhere in the text are the standards with core principles.Within the set of standards with core principles are the standards that have core principles that relate to humans and core principles that relate to nonhumans.We have clarified this further using examples from the AI standards included in this survey in Table 2 below.

Human-Centred
This work defines 'human-centred' standards as those that used the term 'humancentred', 'human-centric', or equivalent phrasing.This definition of 'human-centred' therefore reduces to an explicit commitment to 'human-centredness'.This removes subjectivity and author bias and reduces doubt as to whether a framework actually supports human-centredness.This definition of 'human-centred' also differs from anthropocentrism as defined in Section 2.3 and may encompass some of the examples of HCAI as outlined in Section 2.2.

Limitations
There are several limitations to our survey.A language bias may have skewed our corpus towards English or English-translated results.Our survey took place in 2022 and thus excludes more recent standards.
Our review faces the typical limitations of quantitative analyses of reviewed studies, including that the themes that are numerically identified may not necessarily indicate actual weight or significance.A higher count of "inclusions" in this manner does not necessarily indicate significance but does indicate a form of prioritisation.This is further exacerbated by the fact that philosophy, particularly ethics, is not an exact science but inherently difficult to quantify and quantifiably compare.These limitations have been mitigated by developing review criteria that was explicit, objective, and data-driven, with the aim of reducing subjective author biases.We also address these limitations with an analysis of our results using qualitative data, drawing out quotes from the reports to substantiate our findings, and by critically analysing our findings within the wider context of HCAI and moral philosophy.

Findings
The range of AI ethics standards included in this survey reflect the broad and varying AI ethics landscape.Given increasing concern for the moral problems of AI systems, it is encouraging to see such a wide variety of AI ethics guidelines being offered from a range of disciplines across such a large time frame.The largest domain that published AI ethics guidelines is government bodies, which issued 30% of the standards included in this survey.This is followed by private industry, which issued 27%, and research groups, which published 16%.Professional associations and non-profits each contributed to 14% of the standards surveyed.Other domains were religious groups and research councils, which each issued one framework, and political parties, which issued two.There were also standards that came from a collective of various domains.For example, Fairness, Accountability, and Transparency in Machine Learning (FATML) [55] constitutes both a non-profit and professional association, and the Rome Call for AI Ethics was published by a collective of religious, private, and governmental groups [52].Figure 3a   There is a clear diversity in the parties involved in AI who wish to publish ethical guidelines over its development and usage.The AI ethics standards also heavily diverged with respect to the issuers and their focus or specific area of expertise, including business ethics, future employment trends, and human rights.The standards also mentioned a variety of AI systems, including machine learning and social embodied systems.The publishing year of the surveyed standards was also broad, spanning 2011 to 2021, although most standards surveyed were published between 2017 and 2019.Figure 3b charts the year distribution.

Concern for Humans and Nonhumans
Figure 4 charts the different results for the inclusion of humans and nonhumans both within and outside of the core principles across standards.Despite such a diversity of standards, the AI ethics landscape we surveyed as a whole extends concern for humans more so than concern for nonhumans.All standards included concern for humans and 26% of those surveyed included concern for nonhumans.Concern for nonhumans ranges from moral consideration of the nonhuman world to practical consideration of how the nonhuman world may undermine the success of AI deployment.Cigref's A Guide for Professionals of the Digital Age extended consideration towards the environmental footprint of digitisation [56] (p.12), and the AI Now 2019 Report includes concern for the environmental and labour costs of AI systems [1] (p. 6).Other standards considered the application of AI in pursuing environmental wellbeing.For instance, Korea's Mid-to Long-Term Master Plan in Preparation for the Intelligent Information Society includes environmental protection and energy as possible future applications for AI systems [57] (p.42).Others considered the environment only in terms of how this may affect the deployment of AI systems.For example, Safety First for Automated Driving includes natural landscapes and environmental conditions as factors worth considering in AI development to ensure robustness across a range of environments [50].
The majority (71%) of standards included core principles.Within this, 69% included a single principle for the benefit or respect of humans, whereas only 20% included concern for nonhumans within those core principles.This difference is charted in Figure 4.
We further examined whether these core principles included nonhumans for their own sake or for the sake of humans.Examples of standards that included principles relating to nonhumans for their own sake include Hochschule der Medien's 10 Ethical Guidelines for the Digitalisation of Companies.This standard includes the principle that "digitization should serve to conserve natural resources" [53].The Future of Life Institute also includes the conservation of natural resources in its longer-term principles: "Advanced AI could represent a profound change in the history of life on Earth, and should be planned for and managed with commensurate care and resources" [58].The Chinese National Governance Committee for the New Generation Artificial Intelligence, [59], IA-Latam [60], and the UNI Global Union [61] among others developed principles relating solely to the nonhuman world.
Others developed a principle relating to humans and nonhumans together.AI4Peoples' Ethical Framework for a Good AI Society includes in the principle of beneficence, "promoting well being, preserving dignity, and sustaining the planet" [54] (p.696).Similarly, the Alan Turing Institute promotes the protection of human individuals, "future generations", and the "biosphere as a whole" [62] (p.11).Meanwhile, the Beijing AI Principles "promote the sustainable development of nature and society, to benefit all mankind and the environment, and to enhance the well-being of society and ecology" [63].The Machine Intelligence Garage Ethics Committee [64], Itechlaw [65], Tieto [66], and the European Parliament [67] were among others to state that AI should be developed for the benefit of both humans and the environment.
Some extended extrinsic concern to the environment only in so far as this would benefit humans.For instance, the IEEE states that their principle of wellbeing "encompasses the full spectrum of personal, social, and environmental factors that enhance human life and on which human life depend" [68] (p.70).Meanwhile, the European Commission states, "AI technology must be in line with the human responsibility to ensure the preconditions for life on our planet, continued prospering of mankind, and preservation of a good environment for future generations" [69] (p.19).
Finally, some standards extended concern to humans alone.For instance, Vodafone states, "We will ensure that we respect international human rights standards" [70].Similarly, the National Research Council of Canada states that it will "preserve human and legal rights" [71].Meanwhile, the UK Department of Health and Social Care have two core principles relating to human wellbeing and respect: "respect for persons" and "respect for human rights" [72].A full list of all standards with their respective core principle categories is available in Appendix A.2.
Overall, as can be seen in Figure 4, though nonhumans are considered in some AI ethics standards for a variety of reasons, humans are included in the concern far more often than nonhumans.Moreover, the standards that adopt core principles include a core principle relating to humans more often than a core principles relating to nonhumans.Appendix A.2 shows that, where nonhumans are considered in core principles, this is most often in relation to or as an extension to humans; few standards have core principles solely relating to nonhumans.

Human-Centredness
Figure 5 charts the data on human-centredness in AI ethics standards.We found that 27% of the ethics standards we surveyed explicitly supported human-centred approaches.'Human-centric' and 'human-centred' were the most common examples of such language.Other examples included 'benefiting first and foremost humans'; 'bringing humans' and 'human rights to the centre'; and having a 'primacy fiduciary duty to humanity'.This figure shows the percentage of human-centred and nonhumancentred standards that include humans anywhere, include nonhumans anywhere, include humans in core principles, and include nonhumans in core principles.This graph expands on the data visualised in Figure 4, breaking this data further down by whether the standards are human-centred.So, for example, just under 80% of human-centred standards with core principles include humans within those core principles.
Much of this human-centredness is motivated by an interest in ensuring that AI benefits humans.Telefonica includes "human-centric AI" as a core principle to ensure AI benefits humanity [51] (p.3).The Institute for Information and Communications Policy (IICP) include the pursuit of a "human-centred society" where all humans benefit from and live in harmony with AI [73] (p.4).Meanwhile, Data Ethics' principles look for "sustainable solutions benefitting first and foremost humans" [74] (p. 7).
Others look to human-centredness as a means of aligning AI with the values of humans.The IEEE states that AI should remain "human-centric" to serve "humanity's values and ethical principles" [68] (p.2).IBM also states that AI should be "human-centric" to align with humanity's values [75] (p.8).Meanwhile, the Chinese AI Alliance's Joint Pledge on Artificial Intelligence Industry Self-Discipline states that AI should be human-oriented to uphold humanity's values and prevent the replacement of humans [59].Some standards have taken a 'human rights lens' or argued for a 'human rights perspective' of AI ethics.Examples of this included White Paper: How to Prevent Discriminatory Outcomes in Machine Learning [76] and Privacy and Freedom of Expression In the Age of Artificial Intelligence [77].
Human-centredness is also seen as essential for building trustworthy AI (TAI).The European Commission High-Level Expert group's four ethical principles of TAI are: respect for human autonomy; prevention of harm; fairness; and explicability [30] (pp.11-12).'Human-centredness' is cited as the unifying feature of these ethical principles, necessary for AI systems to be developed for human benefit and having the goal of improving the welfare and freedom of humans [30] (pp.[9][10][11][12][13][14][15][16][17][18][19].This sentiment is echoed across a range of AI ethics standards.For instance, the Personal Data Protection Commission of Singapore states that "human-centricity, as clear baseline requirements can build consumer trust in AI deployments" [78] (p.3).The Organisation for Economic Co-operation (OECD) also states, "there is a need for a stable policy environment that promotes a human-centric approach to trustworthy AI" [79].Meanwhile, the Telia Guided Principles on Trusted AI includes "human centric" as a key principle [80] (p.3).Fraunhofer IAIS also places humans "at the centre" of their Trustworthy Use of Artificial Intelligence framework [81] (pp.4-5, 12-13).
Figure 5 draws an overview of the impact that this commitment to human-centredness has on concern for humans and nonhumans.Accounting for the fact that there are more human-centred than nonhuman-centred standards, the percentage of human-centred standards which include nonhumans, include nonhumans in core principles, and included humans in core principles was calculated.This was compared with the percentage of nonhuman-centred standards that made the same considerations.Figure 5 shows that a higher percentage of human-centred standards include humans and exclude nonhumans in their core principles compared with nonhuman-centred standards.Moreover, nonhuman-centred standards more often include nonhumans compared with human-centred ones.
To examine this further, Table 3 shows the human-centred standards that include core principles relating to humans and nonhumans.We broke this down using the criteria set out in Section 3.4.2.In this table, human-centred standards are checked as having a core principle relating to nonhumans, to nonhumans and humans together, to nonhumans for the sake of humans, or for humans alone.A single framework may have multiple individual core principles relating to humans and nonhumans.For example, the Alan Turing Institute published a framework with two individual core principles relating to humans and nonhumans.
Based on the results in Figures 3 and 5, we can see that human-centred standards tend to more often include a single core principle relating to humans only, excluding nonhumans.This would accord with the strong anthropocentric view (as defined in Section 2.3) that only humans matter morally.The IEEE's Ethically Aligned Design: A Vision for Prioritizing Human Wellbeing with Autonomous and Intelligent Systems (versions 1 and 2) includes core principles relating to humans alone and to nonhumans only for the sake of humans.This would comply with the strong anthropocentric view that nonhumans have at most extrinsic value in how they may benefit humans.
Other human-centric standards that include core principles relating to nonhumans and humans together are compatible with moderate anthropocentrism (as defined in Section 2.3).These include the Alan Turing Institute's Understanding AI Ethics and Safety and OECD's Principles for Responsible Stewardship of Trustworthy AI.Note that no human-centred standard includes core principles that solely relate to nonhumans.
Overall, 27% of the ethics standards that we explicitly surveyed supported humancentred AI ethics approaches.Human-centred standards tend to include humans and exclude nonhumans in their core principles more than nonhuman-centred standards.In contrast, nonhuman-centred standards include nonhumans more often than human-centred standards do.Most human-centred AI ethics standards comply with the definition of strong anthropocentrism, including nonhumans only insofar as they benefit humans.Moreover, no human-centred standard included core principles solely relating to nonhumans.

Tensions between Human-Centredness and Environmental Wellbeing
In order to further ascertain whether human-centredness in AI ethics actually accords with the exclusion of nonhumans, trends in the standards were further analysed over time and by domain.
Figure 3b shows the standards included in this survey that followed a left-skewed distribution in the time that they were published, peaking in 2018.In accounting for the fact that there was not an equal count of standards per year, the various features of the standards (human-centredness, inclusion of humans and nonhumans, both within and outside of core principles) were divided by the number of occurrences per year.This is captured in Figure 6.Because the mention of humans is included in all standards and is thereby static, this was excluded.Figure 6 shows that from 2016 to 2020, which is when most of the ethics standards are drawn from, all three features (inclusion of nonhumans and inclusion of both humans and nonhumans within core principles) increase and decrease with each other, implying that they accord with each other.Between 2011 and 2016, and from 2020 onwards, there are fewer standards, and so it is less reliable to perform a concrete analysis of trends.Overall, the first finding was that the mentioning of nonhumans and inclusion of nonhumans in core principles closely follow each other in trends over time.The inclusion of humans in core principles also followed this same trend over time, albeit less so.The second finding from this analysis was that human-centredness diverges away from concern for nonhumans.Figure 6 shows that, from the same time period of 2016 to 2020, human-centredness as shown with small circle line markings increases where the other lines decrease and vice versa.Over time, mention of nonhumans and the inclusion of both nonhumans and humans in core principles follow the same pattern and seem to accord with each other, whereas human-centredness diverges from this pattern.
To further determine the effects of human-centredness upon extensions of concern, we also analysed the standards by the domains within which these are published.Again, in order to account for the fact that the number of standards published across domains was not equal, the various features of the standards were divided by the number of occurrences per domain.These data are visualised in Figure 7.Because the mention of humans is included in all standards and is therefore both static and equivalent to the number of instances per domain, this feature was again excluded.To improve clarity of the visualisation, the domains having only one or two standards each, religious group, research council, and Political party, were excluded.From Figure 7, we can see that across domains, increases in human-centredness accords with decreases in the inclusion of nonhumans in core principles.Private companies, nonprofits, and research groups support human-centredness the most.These domains also mention nonhumans the least.Private companies, in particular, include nonhumans the least and include nonhumans in their core principles the least of all domains.Professional associations and governments support human-centredness the least.Both of these domains also mention nonhumans the most.Professional associations, in particular, support humancentredness the least, mention nonhumans the most, and include nonhumans in their core principles the most out of all domains.Overall, across domains, support for humancentredness tends to relate to lack of concern for nonhumans, whereas lack of support for human-centredness tends to relate to concern for nonhumans.

Findings Summary
The following key findings of our scoping review of the AI ethics landscape were noted as follows: • The entire AI ethics landscape includes humans in concern more often than nonhumans and includes more core principles related to humans than nonhumans; • Wherever nonhumans are included within core principles, this is most often as an extension or in relation to humans; few standards have a core principle relating solely to nonhumans; • A total of 27% of standards support human-centredness, most of which comply with strong anthropocentrism as defined in Section 2.3; • The standards that support human-centredness tend to include humans and exclude nonhumans more than nonhuman-centred standards.
Overall, our findings indicate that human-centredness in AI ethics standards tends to correlate with increased concern for human wellbeing and decreased concern for nonhuman wellbeing.

Discussion
In Section 2, we discussed the support for human-centredness across the AI landscape and how this differs in the definition and application from security measures, limits of bias, or inequalities to being portrayed as 'ethical' or 'beneficial' for humans.We discussed this HCAI as inspired by, or perhaps accepting of, Asimov's laws of robotics, particularly the fourth law: a robot may not harm humanity or, by inaction, allow humanity to come to harm.Overall, we argued that applications and concepts of 'human-centredness' across AI were nebulous, without a clear, consistent definition, and lacking investigation into their normative groundings, merits, and limitations.Through our review, we aimed to examine the commitment of AI ethics standards to human centredness and how this commitment interacts with other ethical considerations for nonhumans and the environment.We asked: how prevalent is a commitment to human-centredness in AI ethics standards, and how does this relate to commitments to nonhuman animals and the environment?
Our findings showed that 27% of AI ethics standards explicated such a commitment to human-centredness.In contrast, 69% of AI ethics standards employ core principles relating to human wellbeing, dignity, respect, or similar phrasing that closely resemble Asimov's laws, particularly the fourth law."We will ensure that we respect international human rights standards" [70], "preserve human and legal rights" [71], "respect for persons", and "respect for human rights" [72] are some examples.This indicates a strong adherence to Asmiovian human-centredness and shows a clear gap between implicitly prioritising human wellbeing and explicating a commitment to human-centredness.
In Section 2.3, we also discussed anthropocentrism, defined in moral philosophy as the claim that humans have, at the very least, more intrinsic moral value than nonhumans, or, more strongly, that nonhumans have no intrinsic moral value.In our scoping review, we compared the commitment to humans and nonhumans across AI ethics standards.We found that overall, AI ethics standards include humans in concern more than nonhumans.
Moreover, where standards included core principles, these were more often related to human wellbeing than to nonhuman wellbeing.We found that very few standards included core principles that solely related to nonhumans.
Human-centred frameworks were defined somewhat broadly; that is, they explicated a commitment to 'human-centredness'.This means that the frameworks we defined as human-centred would not necessarily commit themselves to the definition of anthropocentrism as used in moral philosophy.As our background discussion showed, the use of the term 'human-centred' could relate to safety, bias, inclusion, or a range of other commitments and not necessarily to the claim that humans are more morally valuable than nonhumans.However, our findings show that AI ethics standards that explicate a commitment to human-centredness in whatever regard also tend to exclude nonhumans more often and include humans within concern.
Thus, even when human-centredness may not be explicitly used in the way that anthropocentrism is defined in moral philosophy, the implications of a human-centred AI are that humans are prioritised over nonhumans as in philosophical anthropocentrism.The use of human-centredness on the surface may seem to include a range of possible meanings and applications, not necessarily indicating a moral preference or prioritisation of humans over nonhumans but rather referencing a design preference or broader social justice.However, when interrogated and examined, as we have performed in this review, there comes to light an undercurrent that promotes, or at the very least, accords more strongly with, the moral hierarchies of humans, nonhumans, and the environment, which has been established in antiquity and carried through centuries of moral philosophical debate.
Overall, a commitment to human-centredness in AI may seem a nebulous, vague, or high-level principle that can be used to unite or find consensus among different bodies, stakeholders, ethical norms, and interests.Anthropocentrism is a well established concept used throughout moral philosophy with a specific technical definition and implications.However, when investigated, the seeming gap between human-centredness and anthropocentrism narrows, and the application of this high-level unifying concept is actually exposed to come with real moral implications in terms of who or what stands as objects of moral concern in the development and deployment of AI.Upon analysing our findings within the context of HCAI, Asimov's laws, and anthropocentrism in moral philosophy, we suggest that human-centredness in AI ethics standards accords with anthropocentrism in the moral prioritisation of humans over nonhumans.

Anthropocentrism and Environmental Wellbeing in AI Development and Deployment
The reader might reject the notion that AI ethics ought to extend intrinsic moral concern to nonhumans.This is because, one might argue, AI is created and used by humans for human specific purposes; nonhumans are not significant stakeholders that are worthy of the moral consideration of AI makers and users.
However, there is a range of AI systems for which nonhumans are significant stakeholders.There are AI systems that interact with the environment by traversing landscapes, monitoring wildlife, collecting data, and interacting with ecosystems and individuals for commercial, research, or conservation purposes [82].Other systems interact more directly with ecosystems through pest control, managing predator populations (i.e., killing certain overpowering species members), and chemically stabilising abiotic environments [82].Research shows that deploying embodied AI systems into established ecosystems can cause distress to certain species [82][83][84].In cases of malfunctioning, AI systems can crash, sometimes in delicate ecosystems [82,85,86].When wildlife are disturbed in this way and either flee or, in extreme cases, are killed, changes in species populations alter the entire structure of the ecosystem, oftentimes undermining its stability for generations to come [85,86].
Some AI systems interact with the environment on a more permanent basis as they are designed and built into an existing ecosystem.Examples include artificially intelligent biosystems [87], intelligent greenhouses [88], or, more rudimentarily, autonomous power stations [89].Again, ethical questions arise, such as whether it is right to introduce inorganic engineered elements into a natural environment [82].If such systems are designed to maintain an ecosystem that would otherwise destabilise, there may also be a sense in which it would be wrong to remove the system from the environment [82].
There is also growing concern over AI's environmental impact as estimates of the substantial energy required to run, as well as the emissions caused by, machine learning models have been released.At the sharp end of machine learning, models can contain billions or trillions of parameters and take months to train, carrying a significant environmental cost.For example, Chat GPT-3, with 175 billion parameters, was estimated to train on 1024 GPUs for 34 days at a cost of USD 4.6 million and an expected energy consumption of 936 MWh [90,91].For comparison, the average UK household uses between 8.5 kWh and 10 kWh of energy per day [92].Chat GPT-4 (released March 2023) has even more parameters and is expected to have had an even greater environmental impact [90].As well as requiring significant amounts of energy for training, machine learning models can lead to the production of large amounts of emissions [93].For example, the BERTlarge model was estimated to produce as much greenhouse gas emissions as a commercial flight between San Francisco and New York [93,94].Others have highlighted the environmental impact of an entire AI system's lifecycle, from extracting the necessary materials for hardware to waste production and recycling [95].
One might argue that anthropocentrism does not necessarily stand in opposition to moral concern for the environment.Moderate anthropocentrists maintain that nonhumans have intrinsic value, albeit less than humans.Moreover, caring for the environment and avoiding environmental harm need not rely on nonhumans having intrinsic value.As discussed, some strong anthropocentrists extend extrinsic moral concern to nonhumans for the sake of human wellbeing, survival, or flourishing.
However, under strong anthropocentrism, the harm to nonhuman animals, plants, and ecosystems that come with certain AI systems are only morally significant insofar as humans may be in turn negatively affected.Where humans are not expected to be negatively affected, or where the benefit to humans is significant, these environmental harms are not of moral concern.Harm to the environment may also be permissible under moderate anthropocentrism in cases where a significant benefit to humans outweighs the harm caused to nonhumans.Humans stand to greatly benefit from the development and deployment of AI systems, such as those wherein the environment stands as a significant stakeholder: unmanned delivery vehicles can offer safe delivery of essential materials in areas with low to no infrastructure; data collection systems can collect data from areas that are dangerous or inaccessible for humans; machine learning models can provide decision support for medical professionals, AI-generated code and fixes for programmers, and content for creators; not to mention the profits to shareholders from increased deployment.
In the case of AI systems that interact with the environment, such as unmanned vehicles and robots, anthropocentrism may permit the disturbance of wildlife, altering the population of species within an environment and undermining the stability of ecosystems over generations.As for AI systems that are built into an environment, landscapes may be irreversibly altered, with the future of ecosystems thereafter being dependent on the input of inorganic optimisation systems.An anthropocentric AI ethic may also permit devastating energy consumption and emissions for the advancement of machine learning models.
Strong anthropocentrists who either consider the environment as only extrinsically valuable to humans ought to be concerned about this, as humans will in turn be affected by damage to the environment.Meanwhile, moderate anthropocentrists and nonanthropocentrists ought to also be concerned for the sake of the environment.Therefore, whether the reader supports or opposes anthropocentrism, an anthropocentric AI ethic poses serious risks that are worth consideration.

Alternative Approaches
Environmental ethics supplies a range of well-established schools of non-anthropocentric ethics that could ground future work on AI ethics standards.Within philosophy, "the idea of drawing on environmental ethics to address the moral problem of artificial intelligence is not new" [96] (p.330).AI ethics approaches that are grounded in non-anthropocentric environmental ethics would not permit such harm to nonhumans and would avoid the risks identified in this work.
Recall, for instance, sentiocentrism, defined in Section 2.4 as viewing anything that suffers to possess intrinsic value [32].A sentiocentric AI ethic may extend moral concern to a wider range of nonhuman animals or study our relations to AI systems in terms of our relations with humans or other sentient animals [97,98].In terms of the examples discussed above, sentiocentric AI ethics standards may ban the use of rovers or unmanned vehicles that harm sentient wildlife.
Recall also a biocentric or 'life-centred' environmental ethic, which extends moral concern to the living environment and which has also been discussed as an alternative environmental foundation for AI ethics [99,100].Under biocentrism, our moral duties in the development and deployment of AI systems would extend to all living individuals and systems [99].In policy, this would force designers of environmentally interactive embodied systems to not only consider the individual harm caused to humans or sentient animals but also to the wider living community of plants and non-sentient animals that may in turn be affected.One example discussed is when AI systems are built into existing ecosystems to optimise environmental conditions or 'parameters'.In such a case, biocentric AI ethics would demand that the system be designed in such a way as to become a stable, contributive ecosystem member.A biocentric AI ethic may also ban the designing of AI systems into ecosystems to avoid the future dependencies of the living environment on artificial optimisation systems.
Finally, recall ecocentrism, which considers all interconnected parts of the environment, including the ecosystems themselves, to possess intrinsic value.In the examples discussed, an ecocentric AI ethic may ban AI systems that undermine ecosystem stability, such as unmanned vehicles that scare away animals, or force designers to consider renewable and sustainable materials for embodied system hardware.
Above are just some examples of how non-anthropocentric environmental ethics can avoid the limitations of anthropocentric AI ethics in considering nonhumans and how such ethics can inform approaches to the moral questions posed by AI systems both in AI ethics regulation and future research.

Conclusions
This work set out to examine human-centredness in AI ethics standards.We first introduced the topic of human-centredness in AI more broadly, outlining the range of possible uses, the vagueness of this term, and the lack of due investigation into its meaning and implications.We undertook a scoping review of 146 AI ethics standards, reviewing the commitment to human and/or nonhuman wellbeing as well as human-centredness.We then critically analysed our results within the wider context of HCAI, Asimovian laws, and anthropocentrism in moral philosophy.Overall, we found that the application of what may seem to be a high-level unifying concept actually comes with real moral implications in terms of who or what stands as objects of moral concern in the development and deployment of AI.In particular, human-centredness in AI ethics promotes, or at the very least accords more strongly with, a prioritisation of humans over nonhumans and the environment.This accords with anthropocentrism, as defined in moral philosophy.In fact, we found that most human-centred AI ethics standards complied with strong anthropocentrism: the claim that nonhumans have at most extrinsic moral value to humans.
We have also briefly discussed some of the ways in which anthropocentric approaches to AI ethics could permit harm to nonhuman animals and the environment, including the disturbance of animals, undermining the stability of ecosystems, alteration of landscapes, and devastating energy consumption and emissions.We recommend that the future use and application of 'human-centredness' in AI development and deployment be technically explicated, which should include sufficient discussion of the normative groundings and moral implications of theory and practice.Without this, the development and application of HCAI will continue without proper investigation of the potential environmental impact; namely, failing to consider in sufficient depth the impacts of AI on nonhumans and their moral standing.Some alternative approaches to anthropocentric AI ethics grounded in environmental philosophy were also considered; namely, sentiocentrism, biocentrism, and ecocentrism.We have shown the ways in which alternative groundings to AI ethics may avoid some of the limitations of anthropocentrism by extending moral concern to the wider nonhuman world in cases of AI systems that interact, are built in, and impact the environment.The motivation to design future AI systems in accordance with non-anthropocentric environmental ethics provides opportunities for future research across a range of disciplines, including philosophy, political science, social robotics, and computer science.Future environmental AI ethics policy development would benefit from these disciplines working together to widen the scope of who and what counts as deserving of moral consideration and protection.

Figure 1 .
Figure 1.PRISMA flow diagram for source identification using sources Jobin et al. [5] and The AI Ethics Guidelines Global Inventory [47].

Figure 2 .
Figure 2. Venn diagram of various possible relationships between review criteria represented by letters A-G.
charts the domain results as a horizontal bar graph.

Figure 5 .
Figure5.Human-centredness.This figure shows the percentage of human-centred and nonhumancentred standards that include humans anywhere, include nonhumans anywhere, include humans in core principles, and include nonhumans in core principles.This graph expands on the data visualised in Figure4, breaking this data further down by whether the standards are human-centred.So, for example, just under 80% of human-centred standards with core principles include humans within those core principles.

Figure 6 .
Figure 6.Frequency of trends in standards over time.

Figure 7 .
Figure 7. Frequency of standards across domains according to trends.
Human-centricExplicitly states 'human-centredness' or equivalent phrasing Inclusion of humans Extends concern to humans anywhere within the standard Inclusion of nonhumans Extends concern to nonhumans anywhere within the standard Humans in core principles Includes a single principle/foundations/priority/aim of human respect/care/wellbeing/values/or equivalent phrasing Nonhumans in core principles Extends concern to humans within one or more core principles

Table 2 .
Example text of relationship between review criteria.
(D) Has core principles relating to humans "AI should be at the service of society and generate tangible benefits for people"

Table 3 .
Core principles of human-centric standards.* is per core principle.Therefore, ** in a column means two individual core principles relating to that column.
Toward a G20 Framework for Artificial Intelligence in the Workplace CIGI Gentre for International Governance Innovation 2018 Data for the Benefit of the People: Recommendations from the Danish Expert Group on Data Ethics * Declaration on Ethics and Data Protection in Artificial Intelligence * Discussion Paper on Artificial Intelligence (AI) and Personal Data-Fostering Responsible Development and European Ethical Charter on the use of Artificial Intelligence in Judicial Systems and their Environment * Everyday Ethics for Artificial Intelligence.A practical Guide for Designers & Developers * Five Guiding Principles for Responsible use of AI in Healthcare and Healthy Living * For a Meaningful Artificial Intelligence.Towards a French and European Strategy * Governance Principles for a New Generation of Artificial Intelligence: Develop Responsible Artificial Intelligence Report with Recommendations to the Commission on Civil Law Rules on Robotics * * Responsible AI #AIFORALL Approach Document for India Part 1-Principles for Responsible AI * Responsible AI #AIFORALL Approach Document For India Part 2-Operationalizing Principles For Respon-Work in the Age of Artificial Intelligence.Four Perspectives on the Economy, Employment, Skills and Ethics *