Next Article in Journal
Ultra-Fast Polarity Switching, Non-Radioactive Drift Tube for the Miniaturization of Drift-Time Ion Mobility Spectrometer
Next Article in Special Issue
Deep-Learning-Based Character Recognition from Handwriting Motion Data Captured Using IMU and Force Sensors
Previous Article in Journal
Control Aspects of Shape Memory Alloys in Robotics Applications: A Review over the Last Decade
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Quality Models for Artificial Intelligence Systems: Characteristic-Based Approach, Development and Application

by
Vyacheslav Kharchenko
,
Herman Fesenko
and
Oleg Illiashenko
*
Department of Computer Systems, Networks and Cybersecurity, National Aerospace University “KhAI”, 17, Chkalov Str., 61070 Kharkiv, Ukraine
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(13), 4865; https://doi.org/10.3390/s22134865
Submission received: 5 June 2022 / Revised: 21 June 2022 / Accepted: 24 June 2022 / Published: 27 June 2022
(This article belongs to the Special Issue Smart Sensors and Devices in Artificial Intelligence II)

Abstract

:
The factors complicating the specification of requirements for artificial intelligence systems (AIS) and their verification for the AIS creation and modernization are analyzed. The harmonization of definitions and building of a hierarchy of AIS characteristics for regulation of the development of techniques and tools for standardization, as well as evaluation and provision of requirements during the creation and implementation of AIS, is extremely important. The study aims to develop and demonstrate the use of quality models for artificial intelligence (AI), AI platform (AIP), and AIS based on the definition and ordering of characteristics. The principles of AI quality model development and its sequence are substantiated. Approaches to formulating definitions of AIS characteristics, methods of representation of dependencies, and hierarchies of characteristics are given. The definitions and harmonization options of hierarchical relations between 46 characteristics of AI and AIP are suggested. The quality models of AI, AIP, and AIS presented in analytical, tabular, and graph forms, are described. The so-called basic models with reduced sets of the most important characteristics are presented. Examples of AIS quality models for UAV video navigation systems and decision support systems for diagnosing diseases are described.

1. Introduction

1.1. Motivation

Household comfort, quality of life, and safety of people are factors which are becoming increasingly dependent on information technology. Among such technologies are the most complex and promising means of artificial intelligence (AI). Evidence of the growing dynamics of the implementation of AI systems (AIS) in various fields, as well as the intensity of development and research, is the rapid increase in the number of publications during 2018–2021 [1], accepted and developed standards and guides EU Commission [2,3], ISO/IEC [4,5,6,7,8,9,10], IEEE [11,12], NIST [13,14,15,16,17,18], OECD [19,20,21], and UNESCO [22].
In industrial systems, healthcare, transportation, weapons systems, etc., the impact of AI is becoming increasingly tangible and sustainable, and on the other hand, it is very controversial in terms of implementation. This is due to:
  • The versatility and complexity of decisions taken in the development and application of systems in which AI tools are built;
  • The variability of the physical and information environments which are not always defined by the parameters in which they operate. The number and extent of external influences, such as cyberattacks aimed at artificial intelligence and based on AI methods are growing and expanding;
  • The accumulation of expert information and expansion of knowledge bases that can be used to improve the efficiency of these systems. The principle of human-centeredness in its creation and application must be balanced to reduce the risks of wrong decisions due to subjective reasons;
  • They increase the importance of ethical and safety aspects during use. This factor is especially important and specific to AISs. According to [22] and other documents focusing on the humanitarian aspects, human dignity, personal and collective security, and well-being are valued in the development and implementation of AISs.
These conditions, in contrast to “traditional” systems, complicate the formulation of specifications and verification of compliance with the requirements for creation and modernization. In addition, the number and variety of AI and AISs traits that need to be considered are growing, especially, ethics, clarity, credibility, etc. [23,24,25]. In turn, the methods of evaluation are diversified, which should be based on a clear idea of the nature and interdependence of the characteristics of artificial intelligence.
It should be emphasized that the increase in the number of publications and standards is accompanied by a significant disturbance in the characteristics of AI, which, on the one hand, determines, and on the other hand is due to a certain contradiction of definitions. Therefore, it is extremely important to research to harmonize and hierarchy characteristics, which objectifies and simplifies the development of tools for standardization, evaluation, and provision of requirements for the creation and implementation of AI systems.

1.2. Aim, Objectives, and Structure

The study aims to develop a model of the quality of artificial intelligence and AI systems based on the definition and ordering of traits.
Objectives:
  • Formulate the principles and justify the sequence of analysis and development of quality models of AI and SSI as an ordered sets of characteristics;
  • Analyze and classify the characteristics, determine their relationship and principles of use of AI and AI systems, provided in known sources, including standards and guides developed by leading institutions;
  • Propose quality models of AI, AI platforms (AIPs) as parts of AISs, and systems using different forms of representation for further use, primarily as an evaluation of individual characteristics and quality in general;
  • Demonstrate the profiling of AI and AISs quality models for systems using artificial intelligence (drone video navigation system (UAV) and urological disease diagnosis system).
The article is structured as follows. Section 2 grounds the principles of developing quality models and their sequence. Section 3 provides an overview of the references that provide definitions and describe the hierarchical relationships between characteristics. The codification of definitions is proposed, and the characteristics are separated from the principles of development and application of artificial intelligence. Section 4 proposes approaches to formulate definitions of AI characteristics based on the analysis of existing ones and their harmonization considering different groups of references, the final table with definitions and classification of AI quality traits is formed. The different forms of representation of dependencies and hierarchies of characteristics are presented in Section 5, and then it is used to analyze the relevant hierarchies based on the analysis of key sources. Section 6 describes the quality models of artificial intelligence and AI platforms, presented in parentheses, tables, and graphs. The so-called basic models with reduced sets of characteristics are given because of their importance. Section 7 describes examples of quality models for AI systems. Lastly, Section 8 provides conclusions and describes areas for further research.
The main contribution of the research includes a set of streamlined AI, AIP, and AIS characteristics, and quality models making it possible to specify requirements to systems and assess them during development and application.

2. Approach to Development of AI Quality Models

2.1. Principles and Concept

2.1.1. Quality as a Generated AI Characteristic

The set of characteristics of artificial intelligence systems analyzed in the article is united by the concept of “quality” similar to how it is usually performed for software, where there are stable quality models that have developed and improved over 50 years of evolution [26,27,28]. The concept of “quality” of AI, in our opinion, is an acceptable generalizing feature, even though in some works it is used as a partial feature of AI, or study considers the quality of artificial intelligence purely in the context of software quality [29].
The software quality context is indeed very important, but it should be used as an approach to form a more general AI quality model. In [30], a position similar to the position of the authors of this study on the importance of AI quality is formed, although this work narrows the content of quality somewhat, as the set of AI traits analyzed is limited. Therefore, we further use the concept of AI and AIS quality as a system-forming, top-level entity in the hierarchy of all characteristics according to the general interpretation of quality according to ISO 9001: 2015 as the extent to which a set of objects-specific characteristics (in this case AI) to which a set of inherent characteristics of an object meets requirements). It is a guarantee for the development of an orderly set of characteristics of artificial intelligence.
Importantly, the quality of AI and AI system should be distinguished from the quality of the product that is the result of the AIS application. For example, the quality of images should be distinguished from the quality of AI with which the image has been processed. Although, there is an effect of AI (AIS) quality on product quality.

2.1.2. AI, AI Platforms, and AI Systems

The quality of the AI system consists of the quality of AI as a generalized but specific object and the quality of the software and hardware platform (called as AIP), through which AI is implemented. This study considers only those components of the quality (characteristics) of AIPs that should somehow consider the specificity of AI as opposed to standard quality characteristics of software and hardware. It should be added, as the analysis showed, that there are quality characteristics that are common to AI and AIP. In this case, we will refer them to the characteristics of the quality of AI, but also consider them at the system level. The characteristics (models) of AI and AIP quality are combined in a general quality model. It is needed to note that requirements to common AI and AIP characteristics are implemented at the two stages: firstly, at the development of the models/algorithms considering specific features of AI functionality and design, and, secondly, at the development of hardware and software by the use of methods considering these features.
In addition, the ISO25010 standard [31], which describes the software quality model, divides the characteristics into two sets, namely subsets of the “product quality” and “quality in use” characteristics. In [30] there is a description of a fragment of such a distribution. Therefore, it may be the second feature to classify AI quality traits for more detailed analysis.

2.1.3. Characteristics as a Key Conception

The key concept used in the study is “characteristic”—a component of quality, which describes the different characteristics of the AI, AI platform, and AI system. The characteristic is the basis for formulating requirements for the AI system and its components by:
  • Considering the relevant characteristics in the development of the system specification, i.e., inclusion in the list of requirements;
  • The definition and choice of metrics according to which the value of the property is evaluated, namely scales and methods of measurement;
  • Substantiation of the necessary “limits” of this property, i.e., requirements to the qualitative or quantitative level thereof defined by the corresponding metrics.
If the characteristic Ch1 depends on the characteristic Ch2, then we will call the sub-characteristic of the characteristic Ch1. Ch1 and Ch2 should be located at the upper and next lower levels of the hierarchy of the quality model, respectively. For each of the characteristics (sub-characteristics), metrics must be defined for their evaluation, as well as formulated requirements for the values of these characteristics, and develop a profile of requirements and their quality [32].
Quality characteristics form a matrix classification, the general view of which is presented in Figure 1. We will call it the quality characteristics classification card (QCCC). QCCC columns correspond to the set of characteristics of artificial intelligence (AI characteristics), platform (AIP characteristics), including their common part (AI&PC), and the rows correspond to the sets of product quality characteristics (QPC), quality in use (QUC) and many characteristics that include both types of quality (QPUC).
Thus, the concept of artificial intelligence assessment is thus to develop a quality model as a general and orderly set of characteristics of the actual artificial intelligence, the corresponding software and hardware platform, and the AI system as a whole. Consider the stages of its implementation in the construction of quality models.

2.2. The Order of Quality Models Construction

Figure 2 illustrates the sequence, intermediate, and final results of constructing quality models of AI, as well as the AI platform and AI system. The results are given in the right part of the figure.
The main phases are as follows:
  • In the first stage, the set of characteristics and principles of use of AI is formed based on the analysis of references, their coding is performed and the corresponding table is developed (Table 1, Section 3.1);
  • Then, the separation of principles and guidelines for the use of AI from its characteristics, the results of which are given in Table 2 (Section 3.2);
  • The analysis and harmonization of characteristics definitions according to the presented approach is performed. The result is Table 3 (Section 4). So, at this point, the formation of both the list of characteristics and their definitions is complete;
  • To organize the characteristics according to their dependency, existing hierarchies are analyzed using their descriptions using simple brackets. The results of the analysis are presented in Table 4 (Section 5.1 and Section 5.2);
  • A partially formalized procedure for building a hierarchy of AI quality traits is being developed. Accordingly, hierarchies between brackets, table, and graph forms are represented (Section 6.1);
  • Then the models of AI, AI systems, and AI platforms quality are presented in visual graph form. Table forms of AI and AIP quality are given in Table 5 and Table 6 (Section 6.2), respectively. In addition, basic models are built, which are part of the overall quality models (Section 6.3);
  • Finally, examples of profiling of quality models for two UAV video navigation systems and decision support for the diagnosis of urological diseases are provided (Section 7.1 and Section 7.2).

3. Analysis of References Related to AI Characteristics and Principles Definitions

3.1. Forming and Codification of AI Characteristics

The order for solving this problem is as follows.
  • The forming of a set of characteristics related to or may be related to AI, its platforms, and systems is as follows: the entire reference base, formed based on selected publications, was analyzed, and selected 75 characteristics named or defined in these references (Table 1).
  • Selected AI characteristics were coded and provided alphabetically. The codification is performed using three Latin letters, which ensures their uniqueness.
  • All references to the definitions in Table 1 are provided in the two right-hand columns:
    • Column with an indication of the reference where the characteristic is mentioned “Referenced in” provides links to publications where the relevant AI characteristic is only mentioned but not defined;
    • Column with an indication of the reference where the definition of the characteristic is given “Defined in” provides references to publications where the characteristics of AI or AI systems are defined in any verbal version.
During the analysis and harmonization of the definitions of these characteristics, all references are divided into three groups:
  • (GS) normative documents—standards, guidelines, technical reports of leading institutes and organizations, namely ISO, UNESCO, NIST, etc., as well as professional dictionaries and guides. These documents define the relevant characteristics to which we refer and take;
  • (GR) scientific publications—articles, monographs, posts, etc., which provide their definitions of characteristics, or are based on the definitions provided in the links of the first group;
  • (GV) academic dictionaries are used to define characteristics in the general sense when their definitions are missing in the documents of the first and second groups, or when they need to be clarified due to inconsistencies in the definitions given in the references of these groups.
Consequently, for each of the attributes in Table 1, the link in the column «Reference where the definition of the characteristic is given» is provided in a separate line (from the first to the third line according to the groups: GS, GR, and GV). Table 1, thus, collected 75 different terms related to characteristics and systematized references to relevant, most representative sources. Then their substantive analysis is performed.

3.2. Analysis of AI Principles

In the first phase of the analysis of all terms collected in Table 1, those that are not real characteristics but can be attributed to the principles related to the development of requirements, creation, and use of AI systems were identified.
Principles are generalized attitudes (values) and views that underlie the development, evaluation, and application of artificial intelligence, AI platforms, and AI systems.
Based on this, a set of characteristics is formed, which directly determines the various components of quality in a broad sense, and which can be measurable. Therefore, the terms “quality” and “quality model” are used as generalizations by analogy with the quality models of software and information systems.
In the process of analyzing key references related to the principles and characteristics of AI [13,14,15,16,17,18,22], as well as other sources provided in Table 1, the following are identified:
  • The boundary between principles and characteristics is difficult to find because principles can dissolve into properties, and characteristics can be generalized into principles. Some characteristics are even identified with the principles;
  • Even between different principles (groups of principles) there is a significant intersection in their components, which are in fact characteristics;
  • The key difference between the characteristics and the principles is the possibility of their measurement (evaluation). Therefore, considering, firstly, the priorities of engineering practice, we have referred to the characteristics of the relevant concepts when such a possibility exists, i.e., there are or can be proposed criteria (scales and assessment methods) for it.
These conclusions are illustrated in Table 2, which provides the results of analyzing two important documents [2,22] and suggestions for the delineation of principles and characteristics.
In addition, in many sources and some sectoral regulations, such as [113,114], the principles, characteristics, and relevant requirements are formulated for both the AI systems themselves and for the personnel who develop and are responsible for their use of AI systems. It should also be noted that there are significant characteristics and differences in the list and detail of principles and characteristics for different industries, namely defense systems, healthcare, law, and education [13,19,20,113]. There are strong developments on standards for AI and AI systems in such areas, especially made by the IEEE [11,12] and other institutions, which is the subject of a separate analysis.
Thus, according to the analysis of Table 1, which contains 75 terms, with the excluded terms in Table 2 (seven terms related to the principles), it is necessary to analyze and harmonize the definitions for the remaining (68) characteristics.

4. Harmonization of AI Characteristics Definitions

The analysis and harmonization of the definitions of the characteristics of AI and AI platforms were performed as follows.
  • The analysis of definitions considered that individual characteristics can be identical, i.e., those that have different names but the same essence. Of the subsets of such properties, only one remained for further use. For example, of the characteristics “governance” and “controllability”, which means “ability to be controlled”, the characteristic “controllability” was left as more general. Plus, under the characteristics “explicability” and “explainability”, which means “ability to be explained”, for further consideration the characteristic “explainability” is chosen.
  • Some characteristics with insignificant differences were combined, and these differences were considered in the relevant definitions. For example, the characteristic “human oversight and determination” was absorbed by the characteristic “human oversight”, considering the characteristics of AI supervision suggested in the absorbed characteristic in the final definition.
  • Several characteristics have been excluded because, in our opinion, they do not have specific features for AI and AI platforms, but are common to technical systems or their software and hardware. Such characteristics include, in particular, “confidence” and “compliance”.
  • For characteristics relevant to AI and AI platforms, the definition is provided by:
    • Repetition (citation) or insignificant adjustment of the definition of one of the documents, which is the most adequate and accurate, according to the authors (marked with the letter R—referred). The definition of the attribute “integrity” was given, for example, by [31];
    • Harmonization of definitions based on definitions provided in various publications (marked with the letter H—harmonized). The essence of harmonization was to identify key terms and combine the essential components of different definitions of the characteristic being analyzed. The definition of the trait “integrity” was obtained, for example, by combining the essential components of the definitions of this trait proposed in [6,31,33];
    • Definition provided by the authors in the absence or in their opinion unsatisfactory wording for the description in the available sources (marked with the letter A—author). Thus, for example, the definition of the characteristic “resiliency” was obtained.
The relationship with the characteristics of different groups (AI, AI platforms, or AI and AI platforms) is determined in the wording itself, and the relationship with the type of quality characteristic (product quality QP and quality in use QU) is determined by a separate column.
The results of the analysis and harmonization of AI characteristics definitions are given in Table 3. Thus, 46 characteristics were selected, of which:
  • Thirty-two are the characteristics of AI (22—exclusively the characteristics of AI), 24—the characteristics of AI platforms (14—exclusively the characteristics of AI platforms), and 10—the characteristics of AI and AI platforms at the same time. This division is made based on the analysis of each of the characteristics. The common characteristics include the following: accuracy, diversity, integrity;
  • Twenty-nine characteristics are attributed to the quality characteristics of AI as a product (of which 7 are exclusively product quality characteristics), 39—to the quality characteristics in use (of which 17 are exclusively such characteristics), 22 characteristics are both characteristics of both types;
  • The definition of 6 characteristics was selected from the relevant sources without changes; the definition of 36 characteristics is harmonized, and the definition of 4 characteristics is provided by the authors.
The results of the analysis are presented by the classification map of quality characteristics (Figure 3) according to the template developed in Section 2.1 (Figure 1). In this map, the corresponding characteristics in each cell are given, considering two parameters according to the object of quality assessment and the type of quality characteristics.
Quality characteristics in the classification map of quality characteristics cells (Figure 2) are written alphabetically without considering their importance or interdependence. Therefore, the next step should be to determine the dependencies of the characteristics and build a model of AI quality.

5. Description and Analysis of AI Characteristics Hierarchies

5.1. Description of AI Characteristics Hierarchies

Given a large number of characteristics of AI, AI platforms, and AI systems (Table 3), we present a description of the hierarchies of characteristics, determine the presence, and hierarchy of relationships to organize, and provide ease of assessment. Let us analyze how in the well-known sources, firstly, such leading organizations as ISO, UNESCO, NIST, and the EU special commissions, form the corresponding dependencies between the characteristics of AI.
The relationship of dependence and hierarchy between characteristics can be described in the form of simple parenthesis. This form is a multiple representation based on a hierarchy of characteristics, which is as follows:
X   = { A ,   B ,   C   { D ,   E } ,   F ,   H   { G ,   I ,   J } } ,
where A, B, C, F, H—sub-characteristics of characteristic X (or quality characteristics of AI in general);
  • D, E—sub-characteristics of C;
  • G, I, J—sub-characteristics of H.
According to this notation, Table 4 provides the results of the analysis of those references where the verbal description of the already coded in Table 1, Table 2 and Table 3 characteristics’ hierarchy is taken place.
For example, for [2], Expression (1) is as follows:
TST   = { HMA ,   HMO ,   RBS ,   SFT ,   PRV ,   DGV ,   TRP   { TRC ,   EXP } ,   DVS ,   NDS ,   FRN ,   SWB ,   ACN   { ADT ,   RDR } }
where HMA, HMO, RBS, SFT, PRV, DGV, TRP, DVS, NDS, FRN, SWB, ACN—sub-characteristics of TST;
  • TRC, EXP—sub-characteristics of TRP;
  • ADT, RDR—sub-characteristics of ACN.
This expression is obtained through verbal analysis of the content of the document and its structuring.

5.2. Analysis of AI Characteristics Hierarchies

The conclusions of the analysis of Table 4 are as follows:
  • The number of references, where in some ways the ratios of dependence on properties are given, is rather small. Of the more than one hundred references analyzed in Table 1 and considered in Table 3, the number of distinctive hierarchies that can be identified is 15;
  • The most commonly used characteristics with described their hierarchies are trustworthiness TST (6 sources), explainability EXP (5 sources), responsibility RSP (3 sources), and ethics ETH (2 sources);
  • The maximum number of hierarchy levels is 3, for example in [2] the TST property depends on the sub-characteristics of NMA, NMO, ACN, and the sub-characteristics of TRP and ACN depend on TRC, EXP and ADT, and RDR, respectively;
There are many inconsistencies in the proposed hierarchies:
  • Firstly, the composition of the sub-characteristics, which are attributed to the relevant characteristics. For example, the (functional) safety of SFTs is attributed in some sources to the trustworthiness TST [2], which in our opinion is appropriate, and in others [38]—to the ethics of ETH, which is incorrect enough given their purpose;
  • Secondly, by assigning them to a certain level of the hierarchy. Some characteristics (e.g., explainability EXP) in some sources [36] are characteristics of the first level of the hierarchy, and in others [17]—the second level or even the third one [2];
  • Thirdly, the interpretation of these characteristics in general, due to inconsistencies in their definitions.
The discrepancies identified are most likely due to the fact that:
  • Different documents have different directions due to different target audiences and domains (technical, ethical, legal, etc.), and are therefore insufficiently consistent;
  • The definition of many characteristics is not provided in sufficient detail and clearly;
  • Documents were prepared in parallel for a relatively short period (2–3 years) for such a complex and important problem;
Authors of non-institutional publications have focused on solving partial problems, so hierarchical dependencies were either not considered at all or were fairly specific [115]. Therefore, it is necessary to build a correct and consistent hierarchy of characteristics based on definitions and analysis of the relationship of dependency.

6. Model of AIS Quality

6.1. The Order of Building and Forming of AI Systems Quality Models

Construction of the AI systems quality model is performed in the following order:
Stage 1. Dividing the set of properties of the AI systems S C h A I S (Table 3) on the set of characteristics of pure AI, i.e., those who have specific characteristics of artificial intelligence, and those who are characteristics of software and hardware platforms which implement AI S C h A I P :
S C h A I S = S C h A I     S C h A I P ,
it also true that
S C h A I S C h A I P   ,
since there are common characteristics for AI and AI platform.
According to the results of the analysis given in Section 4, we have two subsets of S C h A I , S C h A I P and subset S C h A I C h A I P , which consists of joint characteristics of AI and AI platform.
Stage 2. Building of hierarchy for the quality model Q S C h A I based on the analysis of properties set S C h A I . When building a hierarchy in these and subsequent stages, the following procedure is used:
Step 1. Each of the characteristics of S C h A I are compared with all others and choose those that are dependent on others, and are those on which all others are not dependent (the ratio of dependence is determined by experts). Such characteristics must be attributed to the first level of the hierarchy S C h A I 1 (with power m 1 );
Step 2. Characteristics not included in S C h A I 1 , that is, formed a set
S C h A I 2 = S C h A I   \   S C h A I 1
are divided into m 1 subsets of S C h A I 2 i , which do not intersect and affect the corresponding characteristics of the set S C h A I 1 :
S C h A I 2 i =   S C h A I 2 i ; i = { 1 , 2 , , m 1 } ,   A   i j :   S C h A I 2 i     S C h A I 2 j = .
Step 3. Steps 1 and 2 are repeated for each of the subsets S C h A I 2 i with power m 2 i , which makes it possible to form the second and third levels of the hierarchy.
This procedure continues, in the case of more levels in the hierarchy.
Stage 3. Building of hierarchy for the quality model Q S C h A I P based on the analysis of properties set S C h A I P according to the procedure described for the Stage 2.
Stage 4. Combining hierarchies of sets of attributes S C h A I     S C h A I P into a general quality model Q S C h A I S .
Thus, quality models are presented in three forms:
  • Parenthesis (analytical) form, already used in the analysis of hierarchies of characteristics described in well-known publications (Table 3);
  • Tabular form, in which the columns specify the levels of dependence on the characteristics, and their number is equal to the number of levels of the hierarchy for depth, and the rows specify the characteristics and sub-characteristics according to their dependence on individual groups;
  • Graph form, the most obvious and convenient for further use to assess the quality of AI. In the graph, the vertices correspond to the characteristics and sub-characteristics, and the edges correspond to the relationship between them.

6.2. Building of AI, AI Platform, and AI System Quality Models

6.2.1. AI Quality Model Q S C h A I

According to the step-by-step procedure, we form a set of characteristics of the first level. It includes the following characteristics S C h A I 1 = { E T H ,   E X P ,   L F L ,   R S P ,   T S T } because they are the most commonly used and directly affect the quality of AI.
The set of characteristics of the second layer (sub characteristics) are:
  • for ETH: S C h A I 21 = { FRN ,   GRS ,   HMA ,   HMO ,   RDR } ;
  • for EXP: S C h A I 22 = { ACN ,   CSL ,   CMT ,   CMH ,   TRP ,   INP ,   INR ,   VFB } ;
  • for LFL, RSP: S C h A I 23 = S C h A I 24 = ;
  • for TST: S C h A I 25 = { DVS ,   RSL ,   RBS ,   SFT ,   SCR ,   ACP ,   ACR } .
Next are the characteristics of the third level:
S C h A I 1 = { BIS ,   NDS ,   TRC ,   SWB ,   PRV ,   ING ,   OBC } .
Thus, the parenthesis form of the model has the following representation:
Q S C h A I   = { ETH   { FRN   { BIS ,   NDS } ,   GRS ,   HMA ,   HMO ,   RDR } ,   EXP   { ACN ,   CSL ,   CMT ,   CMH ,   TRP   { TRC } , INP ,   INR ,   VFB } ,   LFL ,   RSP ,   TST   { DVS ,   RSL ,   RBS ,   SFT   { SWB } ,   SCR   { PRV ,   ING ,   OBC }   ACP ,   ACR } } .
The tabular form of the model is presented in Table 5. The asterisks (*) in the table indicate the characteristics common to AI and AI platform.
The graph form of the model is shown in Figure 4.

6.2.2. AI Platform Quality Model Q S C h A I P

The peculiarity of this model is that it combines many of the actual platform characteristics of S C h A I P 1 , as well as characteristics of S C h A I P 2 , which is also part of many AI quality characteristics. At the platform level, these characteristics should also be considered when assessing the quality of AI platform.
Generally, the AI platform quality model is formed from a set of properties of S C h A I P 1 by the same procedure as the previous AI model. The set of the first level consists of the following characteristics
S C h A I P 1 = { ACS ,   ADT ,   AVL ,   CNT ,   EFS ,   INF ,   RLB ,   MNT ,   SST ,   USB } ,
and the second level—according to the characteristics S C h A I P 2 = { DGV ,   MTR ,   TRF ,   GRN } .
Thus, the parenthesis view of the model is as follows:
Q S C h A I P = { ACS ,   ADT ,   AVL ,   CNT   { DGV } ,   EFS ,   INF ,   RLB   { MTR } ,   MNT   { TRF } ,   SST   { GRN } ,   USB }
It should be noted that the models Q S C h A I P the following characteristics joint with Q S C h A I are added:
S C h A I C h A I P = { VFB ,   DVS ,   RSL ,   RBS ,   SFT ,   SCR ,   ACR } .
The tabular form of the model is presented in Table 6. The asterisks (*) in the table indicate the characteristics common to AI and AI platform.
The graph form of the model is shown in Figure 5, where the additional vertex AIG combines the characteristics of the set S C h A I C h A I P .

6.2.3. AI System Quality Model Q S C h A I S

AI system quality model combines the previous two (Figure 6).
The proposed quality models of AI, AI platform, and AI system can be extended and detailed due to additional characteristics that consider the specific characteristics of different disciplines. On the other hand, these models are quite extensive and can be simplified under certain circumstances and according to a certain procedure.

6.3. Developing of Basic AI and AI System Quality Models

6.3.1. Building of a Basic AI Quality Model Q S C h A I b

The basic AI quality model is developed to make it more compact and engineer-friendly to evaluate real AI systems. The basic model can be obtained by optimizing it “vertically” and “horizontally”.
The basic AI quality model differs from the original in the following:
  • Vertical optimization is performed by presenting the model in two levels. Sub-characteristics of the third level are considered at the level of metrics of the corresponding characteristics of the second level;
  • The relevant components of the characteristics that are removed or combined can be considered at the level of criteria used for evaluation and weighed accordingly when evaluating the top-level characteristics;
  • The RSP characteristic is removed because it intersects with other characteristics of this level:
    (a)
    ETN and LFL—in terms of responsibility for compliance with ethical and legal norms;
    (b)
    TST—in terms of responsibility for complying with the requirements of the user as a whole. Moreover, the requirement to inform him in case of possible breach, which is part of the responsibility, can be considered compulsory and therefore considered in the assessment of reliability;
    (c)
    EXP—in terms of suitability for verification and provision of information in the event of a breach of the relevant rules and requirements that are part of the characteristics of TRP, VFB;
  • LFL characteristics are combined with ETF, as they are similarly worded and differ only in references to ethical and legal norms, the boundary between which is not always clear. After unification, the characteristics of ETN are formulated as follows: ethics—the ability of AI to comply with current moral standards, laws and regulations regarding the results of surgery;
  • The characteristics of HMA and HMO are combined because they are usually considered together and can complement each other at the metric level. The new definition of HMA is the ability of AI to enable the user to make autonomous informed decisions about the use of AI based on control and to interfere in some way in its functioning;
  • Accountability of ACN and causability CSL causality are combined with transparency TRP as they can be considered as additional transparency metrics. Transparency can then be defined as the ability of AIs to describe, test and repeat models, individual components and algorithms according to which decisions are made, to determine cause-and-effect relationships and to report on performance in a defined form do;
  • Acceptability of ACP is excluded as separate, since it is in fact a “soft” component of the actual credibility, the definition of which does not require adaptation.
Thus, the parenthesis view of the basic AI quality model can be represented as follows:
Q S C h A I b = { ETH   { FRN ,   GRS ,   HMA ,   RDR } ,   EXP   { CMT ,   CMH ,   TRP ,   INP ,   INR ,   VFB } ,   TST   { DVS ,   RSL ,   RBS ,   SFT ,   SCR ,   ACR } } .
This model is also described by a graph (Figure 7), which is a sub-graph of the general Q S C h A I model and includes 19 characteristics.

6.3.2. Building of a Basic AI Platform Quality Model Q S C h A I P b

The optimization of the AI platform quality model was performed according to the same principles and the basic variant thereof was obtained (Figure 8).
Optimization was carried out as follows:
  • One lower level of detail (sub-characteristics: DGV, MTR, TRF, and GRN) is removed;
  • Accessibility of ACS characteristic is absorbed by availability AVL because availability is often seen as synonymous or as a component of readiness. The definition of AVL after this merge does not need to be adapted, as it includes the term “availability”;
  • The IFS feature is absorbed by the maintainability MNT and usability USB characteristics. First, it intersects with maintainability, as the provision and use of information is necessary for the recovery, prevention, and modernization of AI system. Second, the information provided to the user for the convenience of using AI system is an important component of the quality of human-machine interfaces.

6.3.3. Building of a Basic AI System Quality Model Q S C h A I S b

This basic model is a simple combination of basic models Q S C h A I b and Q S C h A I P b (Figure 9).

7. Case Study

Consider examples of building a quality model for real AI systems based on the proposed models. Such models can be used to justify the requirements for the developed systems or to check their implementation and adapt design solutions. The process of building models for real AI systems can be called profile development or requirements profiling. Profiling is implemented by defining quality characteristics at each level of the model hierarchy that is important to the system being analyzed.
This task is solved for the two systems. The basic model of AI system quality was chosen for further consideration and use (Figure 9).
The purpose of the case study is to illustrate how the proposed model can be used to obtain a profile, i.e., models for a particular system with artificial intelligence, considering the characteristics of its (system) use, as well as requirements for its non-functional characteristics. Thus, the solution to this problem is to determine the set of characteristics (formally separating the subgraph from the general graph of the quality model), which must be considered for a particular system (class of systems).
The procedure for obtaining a profile of quality characteristics (requirements) that are important for the system was as follows:
  • Each of the characteristics and sub-characteristics of the model (Figure 9) was analyzed separately;
  • The experts (authors of the article and developers of analyzed systems) determined the need to include the characteristics in the AIS quality model;
  • the decision to include these characteristics in the quality model was made by consensus after considering the opinions of the experts. From our point of view, such a consensus-based approach is acceptable considering the complexity of the tasks and the lack of principal differences in opinions of the experts.

7.1. Quality Model of the UAV’s Video Navigation System as an AI System

The first example relates to the UAV’s video navigation system (VNS), which can be used as a separate device or as part of the UAV fleet [116] in monitoring, reconnaissance, etc. The system is based on a convolutional neural network (CNN) with a multilevel structure [117]. The architecture consists of a convolutional neural network for visual features, an extreme learning machine for estimating position bias, and an advanced classifier of extreme information values to predict UAV barriers.
The quality model of VNS as an AI system with built-in CNN is shown in Figure 10. The characteristics which should be considered are marked in gray color. The model features are as follows:
  • At the first level, all three main characteristics of AI quality include: ethics ETH, explainability EXP, and trustworthiness TST;
  • Under the sub-characteristics of ethics the fairness FRN is left, because for VNS the component of FRN definition is important—to reduce the risks of anomalies due to erroneous assumptions and errors in the process of minimizing model setting; for explainability EXP all sub-characteristics are included except for interactivity INR, taking into account the autonomous operation of the UAV; for trustworthiness TST all sub-characteristics are also included, except for diversity DVS, as the application of the multi-versatility principle in on-board systems is limited by the need to minimize overall mass and energy performance;
  • For the AI platform level, all characteristics are included except for the audibility ADT, controllability CNT, and sustainability SST with respect to the purpose and functions of the VNS.

7.2. Quality Model of the Decision Support System for the Diagnosis of Urological Diseases as an AI System

The second case illustrates the construction of a quality profile for a decision support system for the diagnosis of urological diseases, based on a distributed three-tier system of neural network modules with additional training [118]. At the first level—the diagnostic urological office—the patient is examined directly and the relevant parameters of urological fluorograms are measured and processed to support the doctor’s decision about the patient’s diagnosis using neural network modules; at the second group level—the regional network of hospitals—the accumulation and exchange of cases of urological diseases is carried out to ensure intensive training of neural network modules of all hospitals; at the third level, the interregional level, information is exchanged between groups for further study, taking into account the details of each region.
The model of the quality of the decision support system for the diagnosis of urological diseases (SDUD) as an AI system with a built-in convolutional neural network is shown in Figure 11. Its features are as follows:
  • At the first level, the model includes all three characteristics (ethics ETH, explainability EXP, and trustworthiness TST);
  • For ethics, all sub-characteristics of fairness FRN are included for obvious reasons due to its definition; for explainability EXP all sub-characteristics are included: fairness FRN, graspability GRS, and human agency HMA except redress RDR, considering the fact that it is a decision support system for diagnosis;
  • Explainability of EXP is represented by all six sub-characteristics as it is important for the medical system;
  • Trustworthiness is represented by four of the six characteristics, except for diversity DVS and safety SFT, given the lack of functions for forming control effects on patients;
  • For the level of the AI platform, only controllability CNT is not included due to the lack of functions for forming control effects and the peculiarities of the purpose of VNS.
Therefore, if we compare the quality models for these systems, we can deduce a more saturated model for the decision support system for the diagnosis of urological diseases as an AI system relative to the basic—its profile contains 23 characteristics and sub-characteristics out of 27, and for VNS—19 out of 27.
Graph of quality model for VNS in fact is a subgraph of graph for SDUD. The graph of SDUD has the following additional characteristics and sub-characteristics which should be considered for assessment of AIS quality: characteristics ADT and SST, GRS, HMA, INR, and SFT.

8. Conclusions

8.1. Discussion

Despite a large number of publications and the availability of a number of high-level documents issued by reliable national and international institutions, it must be concluded that there is no structured and complete set of AI characteristics, which can be called a quality model similar to an existing and generally accepted software quality models.
The main result of the study is the set of quality models for artificial intelligence, AI platforms, and AI systems. These models are based on the analysis and harmonization of definitions and dependencies of quality traits specific to the AI itself and AI systems. According to our conclusions, some of the characteristics were common to these two entities in the sense that they could be provided at the level of development of the AI itself and its platform.
We tried to select characteristics and build quality models in such a way as to eliminate duplication, ensure completeness of presentation, determine the specific peculiarities of each of the characteristics, and distinguish between the characteristics of the AI and the platform on which AI has been deployed. Clearly, it is extremely difficult to create a model that fully meets such requirements, so the proposed options need to be supplemented and improved considering the rapid development of technologies and applications of AI.
Quality models are presented in this study in various forms—analytical (brackets), tabular, and graph forms, which are convenient for substantive and formal analysis. These models develop the results of [115] and offer the possibility to obtain partial quality profiles for specific developments, considering the details of the respective systems, as demonstrated in the two cases. They can be used as a basis for metric-based AI quality assessment.
The proposed quality models are open and can be supplemented and detailed according to the specific purpose and scope of the AIS. In our opinion, based on the proposed models, it is possible to develop an intersectoral quality standard and requirements for AI (AI systems and AI platforms).

8.2. Future Research

Further research needs to be conducted in the following areas:
  • Profiling (addition and detailing) of models for specific industries (healthcare, law, industrial systems, mobility, etc.) considering evolution issues [119]. Such profiling should be accompanied by an overview of the characteristics and sub-characteristics added based on experience in the development and use of AI, for further generalization in the quality models of AI, AI systems, and AI platforms;
  • Development of metrics and algorithms for evaluating AI and AI platforms for each of the proposed characteristics and quality in general. It is advisable to collect and analyze information on various criteria for their inclusion in the general database;
  • Development of tools and case-oriented methods for assessing the quality of AI, AI systems, and AI platforms [120]. They can be based on general assurance case approaches [121,122] as well as functional and cybersecurity assessment approaches [123];
  • Application of internal validation as an additional procedure which can be embedded into AIS assessment [124];
  • Development of content quality models including different aspects of image quality assessment and so on.

Author Contributions

Conceptualization, V.K., H.F. and O.I.; methodology, V.K., H.F. and O.I.; validation, V.K., H.F. and O.I.; formal analysis, V.K., H.F. and O.I.; investigation, V.K., H.F. and O.I.; resources, V.K., H.F. and O.I.; data curation, V.K., H.F. and O.I.; writing—original draft preparation, V.K., H.F. and O.I.; writing—review and editing, V.K., H.F. and O.I.; visualization, V.K., H.F. and O.I.; supervision, V.K.; project administration, V.K.; funding acquisition, O.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Acknowledgments

This work was supported by the ECHO project, which has received funding from the European Union’s Horizon 2020 research and innovation program under the grant agreement no 830943. The authors appreciate the scientific society of the consortium and in particular the staff of the Department of Computer Systems, Networks and Cybersecurity of the National Aerospace University “KhAI” for invaluable inspiration, hard work, and creative analysis during the preparation of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Islam, M.R.; Ahmed, M.U.; Barua, S.; Begum, S. A Systematic Review of Explainable Artificial Intelligence in Terms of Different Application Domains and Tasks. Appl. Sci. 2022, 12, 1353. [Google Scholar] [CrossRef]
  2. EU Commission. High-Level Expert Group on Artificial Intelligence. Ethics Guidelines for Trustworthy AI. Available online: https://www.aepd.es/sites/default/files/2019-12/ai-ethics-guidelines.pdf (accessed on 4 June 2022).
  3. EU Commission. High-Level Expert Group on Artificial Intelligence. The Assessment List for Trustworthy Artificial Intelligence (ALTAI). Available online: https://airegio.ems-carsa.com/nfs/programme_5/call_3/call_preparation/ALTAI_final.pdf (accessed on 4 June 2022).
  4. ISO/IEC TR 24372:2021. Information Technology—Artificial Intelligence (AI)—Overview of Computational Approaches for AI Systems. Available online: https://www.iso.org/standard/78508.html (accessed on 4 June 2022).
  5. ISO/IEC TR 24030:2021 Information Technology—Artificial Intelligence (AI)—Use Cases. Available online: https://www.iso.org/standard/77610.html (accessed on 4 June 2022).
  6. ISO/IEC TR 24028:2020. Information Technology—Artificial Intelligence—Overview of Trustworthiness in Artificial Intelligence. Available online: https://www.iso.org/standard/77608.html (accessed on 4 June 2022).
  7. ISO/IEC 38507:2022. Information Technology—Governance of IT—Governance Implications of the Use of Artificial Intelligence by Organizations. Available online: https://www.iso.org/standard/77608.html (accessed on 4 June 2022).
  8. ISO/IEC TR 24029-1:2021. Artificial Intelligence (AI). Assessment of the Robustness of Neural Networks. Overview. Available online: https://www.iso.org/standard/77609.html (accessed on 4 June 2022).
  9. IEC White Paper AI:2018. Artificial Intelligence Across Industries. Available online: https://www.en-standard.eu/iec-white-paper-ai-2018-artificial-intelligence-across-industries/ (accessed on 4 June 2022).
  10. ISO/IEC TR 24027:2021. Information Technology—Artificial Intelligence (AI)—Bias in AI Systems and AI Aided Decision Making. Available online: https://www.iso.org/standard/77607.html (accessed on 4 June 2022).
  11. IEEE 1232.3-2014—IEEE Guide for the Use of Artificial Intelligence Exchange and Service Tie to All Test Environments (AI-ESTATE). Available online: https://ieeexplore.ieee.org/document/6922153 (accessed on 4 June 2022).
  12. IEEE 2941-2021—IEEE Standard for Artificial Intelligence (AI) Model Representation, Compression, Distribution, and Management. Available online: https://ieeexplore.ieee.org/document/6922153 (accessed on 4 June 2022).
  13. NISTIR 8312 Four Principles of Explainable Artificial Intelligence (September 2021). Available online: https://doi.org/10.6028/NIST.IR.8312 (accessed on 4 June 2022).
  14. NISTIR 8367. Psychological Foundations of Explainability and Interpretability in Artificial Intelligence (April 2021). Available online: https://doi.org/10.6028/NIST.IR.8367 (accessed on 4 June 2022).
  15. NIST Special Publication 1270. Towards a Standard for Identifying and Managing Bias in Artificial Intelligence (March 2022). Available online: https://doi.org/10.6028/NIST.SP.1270 (accessed on 4 June 2022).
  16. Draft NISTIR 8269. A Taxonomy and Terminology of Adversarial Machine Learning (October 2019). Available online: https://doi.org/10.6028/NIST.IR.8269-draft (accessed on 4 June 2022).
  17. Draft NISTIR 8332. Trust and Artificial Intelligence (March 2021). Available online: https://doi.org/10.6028/NIST.IR.8332-draft (accessed on 4 June 2022).
  18. NIST. AI Risk Management Framework: Initial Draft (March 2022). Available online: https://www.nist.gov/system/files/documents/2022/03/17/AI-RMF-1stdraft.pdf (accessed on 4 June 2022).
  19. OECD. Trustworthy AI in Education: Promises and Challenges. Available online: https://www.oecd.org/education/trustworthy-artificial-intelligence-in-education.pdf (accessed on 4 June 2022).
  20. OECD. Trustworthy AI in Health: Promises and Challenges. Available online: https://www.oecd.org/health/trustworthy-artificial-intelligence-in-health.pdf (accessed on 4 June 2022).
  21. OECD. Tools for Trustworthy AI: A Framework to Compare Implementation Tools. Available online: https://www.oecd.org/science/tools-for-trustworthy-ai-008232ec-en.htm (accessed on 4 June 2022).
  22. UNESCO. Recommendation on the Ethics of Artificial Intelligence. Available online: https://unesdoc.unesco.org/ark:/48223/pf0000381137 (accessed on 4 June 2022).
  23. Christoforaki, M.; Beyan, O. AI Ethics—A Bird’s Eye View. Appl. Sci. 2022, 12, 4130. [Google Scholar] [CrossRef]
  24. Xu, F.; Uszkoreit, H.; Du, Y.; Fan, W.; Zhao, D.; Zhu, J. Explainable AI: A brief survey on history, research areas, approaches and challenges. In Natural Language Processing and Chinese Computing; Tang, J., Kan, M.Y., Zhao, D., Li, S., Zan, H., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2019; Volume 11839, pp. 563–574. [Google Scholar] [CrossRef]
  25. Chatila, R.; Dignum, V.; Fisher, M.; Giannotti, F.; Morik, K.; Russell, S.; Yeung, K. Trustworthy AI. In Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Braunschweig, B., Ghallab, M., Eds.; Springer International Publishing: Cham, Switzerland, 2021; Volume 12600, pp. 13–39. [Google Scholar] [CrossRef]
  26. Gordieiev, O.; Kharchenko, V. IT-oriented software quality models and evolution of the prevailing characteristics. In Proceedings of the 2018 IEEE 9th International Conference on Dependable Systems, Services and Technologies (DESSERT), Kyiv, Ukraine, 24–27 May 2018; pp. 375–380. [Google Scholar] [CrossRef]
  27. Gordieiev, O.; Kharchenko, V.; Fusani, M. Software quality standards and models evolution: Greenness and reliability issues. In Information and Communication Technologies in Education, Research, and Industrial Applications; Yakovyna, V., Mayr, H.C., Nikitchenko, M., Zholtkevych, G., Spivakovsky, A., Batsakis, S., Eds.; Springer: Berlin/Heidelberg, Germany, 2016; pp. 38–55. [Google Scholar] [CrossRef]
  28. Gerstlacher, J.; Groher, I.; Plösch, R. Green und Sustainable Software im Kontext von Software Qualitätsmodellen. HMD Prax. Wirtsch. 2021. [Google Scholar] [CrossRef]
  29. Lenarduzzi, V.; Lomio, F.; Moreschini, S.; Taibi, D.; Tamburri, D.A. Software quality for AI: Where we are now? In Lecture Notes in Business Information Processing; Winkler, D., Biffl, S., Mendez, D., Wimmer, M., Bergsmann, J., Eds.; Springer International Publishing: Cham, Switzerland, 2021; Volume 404, pp. 43–53. [Google Scholar] [CrossRef]
  30. Smith, A.L.; Clifford, R. Quality characteristics of artificially intelligent systems. CEUR Workshop Proc. 2020, 2800, 1–6. [Google Scholar]
  31. ISO/IEC 25010:2011. Systems and Software Engineering—Systems and Software Quality Requirements and Evaluation (SQuaRE)—System and Software Quality Models. Available online: https://www.iso.org/standard/35733.html (accessed on 4 June 2022).
  32. Gordieiev, O. Software individual requirement quality model. Radioelectron. Comput. Syst. 2020, 94, 48–58. [Google Scholar] [CrossRef]
  33. The Industrial Internet of Things. Trustworthiness Framework Foundations. An Industrial Internet Consortium Foundational Document. Version V1.00—2021-07-15. Available online: https://www.iiconsortium.org/pdf/Trustworthiness_Framework_Foundations.pdf (accessed on 4 June 2022).
  34. Morley, J.; Morton, C.; Karpathakis, K.; Taddeo, M.; Floridi, L. Towards a framework for evaluating the safety, acceptability and efficacy of AI systems for health: An initial synthesis. arXiv 2021, arXiv:2104.06910. [Google Scholar] [CrossRef]
  35. Antoniadi, A.M.; Du, Y.; Guendouz, Y.; Wei, L.; Mazo, C.; Becker, B.A.; Mooney, C. Current Challenges and Future Opportunities for XAI in Machine Learning-Based Clinical Decision Support Systems: A Systematic Review. Appl. Sci. 2021, 11, 5088. [Google Scholar] [CrossRef]
  36. Toreini, E.; Aitken, M.; Coopamootoo, K.; Elliott, K.; Zelaya, C.G.; van Moorsel, A. The relationship between trust in AI and trustworthy machine learning technologies. In Proceedings of the Conference on Fairness, Accountability, and Transparency (FAT), Barcelona, Spain, 27–30 January 2020; pp. 272–283. [Google Scholar]
  37. Cambridge Dictionary. Acceptability. Cambridge University Press. Available online: https://dictionary.cambridge.org/dictionary/english/acceptability (accessed on 4 June 2022).
  38. Barredo Arrieta, A.; Díaz-Rodríguez, N.; Del Ser, J.; Bennetot, A.; Tabik, S.; Barbado, A.; Garcia, S.; Gil-Lopez, S.; Molina, D.; Benjamins, R.; et al. Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Inf. Fusion 2020, 58, 82–115. [Google Scholar] [CrossRef] [Green Version]
  39. Zoldi, S. What is Responsible AI? Available online: https://www.fico.com/blogs/what-responsible-ai (accessed on 4 June 2022).
  40. Adadi, A.; Berrada, M. Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI). IEEE Access 2018, 6, 52138–52160. [Google Scholar] [CrossRef]
  41. Burciaga, A. Six Essential Elements of a Responsible AI Model. Available online: https://www.forbes.com/sites/forbestechcouncil/2021/09/01/six-essential-elements-of-a-responsible-ai-model/?sh=39e32be56cf4 (accessed on 4 June 2022).
  42. Cambridge Dictionary. Awareness. Cambridge University Press. Available online: https://dictionary.cambridge.org/dictionary/english/awareness (accessed on 4 June 2022).
  43. Smith, G.; Rustagi, I.; Haas, B. Mitigating Bias in Artificial Intelligence: An Equity Fluent Leadership Playbook. Available online: https://haas.berkeley.edu/wp-content/uploads/UCB_Playbook_R10_V2_spreads2.pdf (accessed on 4 June 2022).
  44. Dilmegani, C. Bias in AI: What it is, Types, Examples & 6 Ways to Fix it in 2022. Available online: https://research.aimultiple.com/ai-bias/ (accessed on 4 June 2022).
  45. Brotcke, L. Time to Assess Bias in Machine Learning Models for Credit Decisions. J. Risk Financ. Manag. 2022, 15, 165. [Google Scholar] [CrossRef]
  46. Alaa, M. Artificial intelligence: Explainability, ethical issues and bias. Ann. Robot. Autom. 2021, 5, 34–37. [Google Scholar] [CrossRef]
  47. Roselli, D.; Matthews, J.; Talagala, N. Managing bias in AI. In Proceedings of the 2019 World Wide Web Conference (WWW), San Francisco, CA, USA, 13–17 May 2019; pp. 539–544. [Google Scholar] [CrossRef]
  48. Weber, C. Engineering Bias in AI. IEEE Pulse 2019, 10, 15–17. [Google Scholar] [CrossRef] [PubMed]
  49. Sgaier, B.S.K.; Huang, V.; Charles, G. The Case for Causal AI. Soc. Innov. Rev. 2020, 18, 50–56. Available online: https://ssir.org/articles/entry/the_case_for_causal_ai# (accessed on 4 June 2022).
  50. Holzinger, A.; Langs, G.; Denk, H.; Zatloukal, K.; Müller, H. Causability and Explainability of Artificial Intelligence in Medicine. WIREs Data Min. Knowl. Discov. 2019, 9, e1312. [Google Scholar] [CrossRef] [Green Version]
  51. Shin, D. The effects of explainability and causability on perception, trust, and acceptance: Implications for explainable AI. Int. J. Hum.-Comput. Stud. 2020, 146, 102551. [Google Scholar] [CrossRef]
  52. Zablocki, É.; Ben-Younes, H.; Pérez, P.; Cord, M. Explainability of vision-based autonomous driving systems: Review and challenges. arXiv 2021, arXiv:2101.05307. [Google Scholar] [CrossRef]
  53. Linardatos, P.; Papastefanopoulos, V.; Kotsiantis, S. Explainable AI: A Review of Machine Learning Interpretability Methods. Entropy 2021, 23, 18. [Google Scholar] [CrossRef]
  54. Cambridge Dictionary. Comprehensibility. Cambridge University Press. Available online: https://dictionary.cambridge.org/dictionary/english/comprehensibility (accessed on 4 June 2022).
  55. Gohel, P.; Singh, P.; Mohanty, M. Explainable AI: Current status and future directions. arXiv 2021, arXiv:2107.07045. [Google Scholar] [CrossRef]
  56. Cambridge Dictionary. Confidence. Cambridge University Press. Available online: https://dictionary.cambridge.org/dictionary/english/confidence (accessed on 4 June 2022).
  57. Yampolskiy, R. On Controllability of AI. arXiv 2020, arXiv:2008.04071. [Google Scholar] [CrossRef]
  58. Markus, A.F.; Kors, J.A.; Rijnbeek, P.R. The role of explainability in creating trustworthy artificial intelligence for health care: A comprehensive survey of the terminology, design choices, and evaluation strategies. J. Biomed. Inform. 2021, 113, 103655. [Google Scholar] [CrossRef]
  59. Ghajargar, M.; Bardzell, J.; Renner, A.S.; Krogh, P.G.; Höök, K.; Cuartielles, D.; Boer, L.; Wiberg, M. From “Explainable AI” to “Graspable AI”. In Proceedings of the 15th International Conference on Tangible, Embedded, and Embodied Interaction (TEI), New York, NY, USA, 14–17 February 2021; pp. 1–4. [Google Scholar] [CrossRef]
  60. Gardner, A.; Smith, A.L.; Steventon, A.; Coughlan, E.; Oldfield, M. Ethical funding for trustworthy AI: Proposals to address the responsibilities of funders to ensure that projects adhere to trustworthy AI practice. AI Ethics 2022, 2, 277–291. [Google Scholar] [CrossRef] [PubMed]
  61. Baker-Brunnbauer, J. Management perspective of ethics in artificial intelligence. AI Ethics 2021, 1, 173–181. [Google Scholar] [CrossRef]
  62. Ryan, M. In AI We Trust: Ethics, Artificial Intelligence, and Reliability. Sci. Eng. Ethics 2020, 26, 2749–2767. [Google Scholar] [CrossRef] [PubMed]
  63. Bogina, V.; Hartman, A.; Kuflik, T.; Shulner-Tal, A. Educating Software and AI Stakeholders About Algorithmic Fairness, Accountability, Transparency and Ethics. Int. J. Artif. Intell. Educ. 2021. [Google Scholar] [CrossRef]
  64. Zhu, L.; Xu, X.; Lu, Q.; Governatori, G.; Whittle, J. AI and ethics—Operationalizing responsible AI. In Humanity Driven AI; Chen, F., Zhou, J., Eds.; Springer International Publishing: Cham, Switzerland, 2022; pp. 15–33. [Google Scholar] [CrossRef]
  65. Holzinger, A.; Carrington, A.; Müller, H. Measuring the Quality of Explanations: The System Causability Scale (SCS). KI—Künstliche Intell. 2020, 34, 193–198. [Google Scholar] [CrossRef] [Green Version]
  66. Sovrano, F.; Vitali, F. An Objective Metric for Explainable AI: How and Why to Estimate the Degree of Explainability. arXiv 2021, arXiv:2109.05327. [Google Scholar] [CrossRef]
  67. Cambridge Dictionary. Exactness. Cambridge University Press. Available online: https://dictionary.cambridge.org/dictionary/english/exactness (accessed on 4 June 2022).
  68. Vilone, G.; Longo, L. Classification of Explainable Artificial Intelligence Methods through Their Output Formats. Mach. Learn. Knowl. Extr. 2021, 3, 615–661. [Google Scholar] [CrossRef]
  69. Dosilovic, F.K.; Brcic, M.; Hlupic, N. Explainable artificial intelligence: A survey. In Proceedings of the 41st International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO), Opatija, Croatia, 21–25 May 2018; pp. 210–215. [Google Scholar] [CrossRef]
  70. Rai, A. Explainable AI: From Black Box to Glass Box. J. Acad. Mark. Sci. 2020, 48, 137–141. [Google Scholar] [CrossRef] [Green Version]
  71. Miró-Nicolau, M.; Moyà-Alcover, G.; Jaume-i-Capó, A. Evaluating Explainable Artificial Intelligence for X-ray Image Analysis. Appl. Sci. 2022, 12, 4459. [Google Scholar] [CrossRef]
  72. Vilone, G.; Longo, L. Explainable Artificial Intelligence: A Systematic Review. arXiv 2020, arXiv:2006.00093. [Google Scholar] [CrossRef]
  73. Umbrello, S.; Yampolskiy, R.V. Designing AI for Explainability and Verifiability: A Value Sensitive Design Approach to Avoid Artificial Stupidity in Autonomous Vehicles. Int. J. Soc. Robot. 2022, 14, 313–322. [Google Scholar] [CrossRef]
  74. Meske, C.; Bunde, E.; Schneider, J.; Gersch, M. Explainable Artificial Intelligence: Objectives, Stakeholders, and Future Research Opportunities. Inf. Syst. Manag. 2022, 39, 53–63. [Google Scholar] [CrossRef]
  75. Toward, X.A.I.; Xai, M.; Tjoa, E. A Survey on Explainable Artificial Intelligence. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 4793–4813. [Google Scholar] [CrossRef]
  76. Samek, W.; Müller, K.R. Towards explainable artificial intelligence. In Explainable AI: Interpreting, Explaining and Visualizing Deep Learning; Samek, W., Montavon, G., Vedaldi, A., Hansen, L.K., Müller, K.R., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2019; Volume 1, pp. 5–22. [Google Scholar] [CrossRef] [Green Version]
  77. Degas, A.; Islam, M.R.; Hurter, C.; Barua, S.; Rahman, H.; Poudel, M.; Ruscio, D.; Ahmed, M.U.; Begum, S.; Rahman, M.A.; et al. A Survey on Artificial Intelligence (AI) and eXplainable AI in Air Traffic Management: Current Trends and Development with Future Research Trajectory. Appl. Sci. 2022, 12, 1295. [Google Scholar] [CrossRef]
  78. Vassiliades, A.; Bassiliades, N.; Patkos, T. Argumentation and explainable artificial intelligence: A survey. Knowl. Eng. Rev. 2021, 36, e5. [Google Scholar] [CrossRef]
  79. Górski, Ł.; Ramakrishna, S. Explainable artificial intelligence, lawyer’s perspective. In Proceedings of the 18th International Conference on Artificial Intelligence and Law (ICAIL), New York, NY, USA, 21–25 June 2021; pp. 60–68. [Google Scholar] [CrossRef]
  80. Hanif, A.; Zhang, X.; Wood, S. A survey on explainable artificial intelligence techniques and challenges. In Proceedings of the 2021 IEEE 25th International Enterprise Distributed Object Computing Workshop (EDOCW), Gold Coast, Australia, 25–29 October 2021; pp. 81–89. [Google Scholar] [CrossRef]
  81. Omeiza, D.; Webb, H.; Jirotka, M.; Kunze, L. Explanations in Autonomous Driving: A Survey. IEEE Trans. Intell. Transp. Syst. 2021, 1–21. [Google Scholar] [CrossRef]
  82. Longo, L.; Goebel, R.; Lecue, F.; Kieseberg, P.; Holzinger, A. Explainable artificial intelligence: Concepts, applications, research challenges and visions. In Machine Learning and Knowledge Extraction; Holzinger, A., Kieseberg, P., Tjoa, A.M., Weippl, E., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; Volume 12279, pp. 1–16. [Google Scholar] [CrossRef]
  83. Gade, K.; Geyik, S.C.; Kenthapadi, K.; Mithal, V.; Taly, A. Explainable AI in industry. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 3203–3204. [Google Scholar] [CrossRef] [Green Version]
  84. Chaczko, Z.; Kulbacki, M.; Gudzbeler, G.; Alsawwaf, M.; Thai-Chyzhykau, I.; Wajs-Chaczko, P. Exploration of explainable AI in context of human–machine interface for the assistive driving system. In Intelligent Information and Database Systems; Nguyen, N.T., Jearanaitanakij, K., Selamat, A., Trawiński, B., Chittayasothorn, S., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; Volume 12034, pp. 507–516. [Google Scholar] [CrossRef]
  85. Kong, X.; Tang, X.; Wang, Z. A survey of explainable artificial intelligence decision. Syst. Eng. Theory Pract. 2021, 41, 524–536. [Google Scholar] [CrossRef]
  86. Sovrano, F.; Sapienza, S.; Palmirani, M.; Vitali, F. Metrics, Explainability and the European AI Act Proposal. J 2022, 5, 126–138. [Google Scholar] [CrossRef]
  87. Clinciu, M.A.; Hastie, H.F. A survey of explainable AI terminology. In Proceedings of the 1st Workshop on Interactive Natural Language Technology for Explainable Artificial Intelligence (NL4XAI), Tokyo, Japan, 29 October 2019; pp. 8–13. [Google Scholar]
  88. Confalonieri, R.; Coba, L.; Wagner, B.; Besold, T.R. A historical perspective of explainable Artificial Intelligence. WIREs Data Min. Knowl. Discov. 2020, 11, e1391. [Google Scholar] [CrossRef]
  89. Baum, K.; Mantel, S.; Schmidt, E.; Speith, T. From Responsibility to Reason-Giving Explainable Artificial Intelligence. Philos. Technol. 2022, 35, 12. [Google Scholar] [CrossRef]
  90. Zhang, Y.; Bellamy, R.K.E.; Singh, M.; Liao, Q.V. Introduction to AI fairness. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 25–30 April 2020; pp. 1–4. [Google Scholar] [CrossRef]
  91. Dignum, V. The myth of complete AI-fairness. In Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics; Tucker, A., Abreu, P.H., Cardoso, J., Rodrigues, P.P., Riaño, D., Eds.; Springer International Publishing: Cham, Switzerland, 2021; Volume 12721, pp. 3–8. [Google Scholar] [CrossRef]
  92. Bartneck, C.; Lütge, C.; Wagner, A.; Welsh, S. Chapter 4—Trust and fairness in AI systems. In Springer Briefs in Ethics. An Introduction to Ethics in Robotics and AI; Bartneck, C., Lütge, C., Wagner, A., Welsh, S., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 27–38. [Google Scholar]
  93. Hazirbas, C.; Bitton, J.; Dolhansky, B.; Pan, J.; Gordo, A.; Ferrer, C.C. Towards Measuring Fairness in AI: The Casual Conversations Dataset. arXiv 2021, arXiv:2104.02821. [Google Scholar] [CrossRef]
  94. Xivuri, K.; Twinomurinzi, H. A systematic review of fairness in artificial intelligence algorithms. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics; Dennehy, D., Griva, A., Pouloudi, N., Dwivedi, Y.K., Pappas, I., Mäntymäki, M., Eds.; Springer International Publishing: Cham, Switzerland, 2021; Volume 12896, pp. 271–284. [Google Scholar] [CrossRef]
  95. NIST. Glossary. Fit to Purpose. Available online: https://csrc.nist.gov/glossary/term/fit_for_purpose (accessed on 4 June 2022).
  96. Cambridge Dictionary. Fruitfulness. Cambridge University Press. Available online: https://dictionary.cambridge.org/dictionary/english/fruitfulness (accessed on 4 June 2022).
  97. Cambridge Dictionary. Governance. Cambridge University Press. Available online: https://dictionary.cambridge.org/dictionary/english/governance (accessed on 4 June 2022).
  98. Mathieson, S.A. How to Make AI Greener and More Efficient. Available online: https://www.computerweekly.com/feature/How-to-make-AI-greener-and-more-efficient (accessed on 4 June 2022).
  99. Cambridge Dictionary. Greenness. Cambridge University Press. Available online: https://dictionary.cambridge.org/dictionary/english/greenness (accessed on 4 June 2022).
  100. Cambridge Dictionary. Informativeness. Cambridge University Press. Available online: https://dictionary.cambridge.org/dictionary/english/informativeness (accessed on 4 June 2022).
  101. Cambridge Dictionary. Impartiality. Cambridge University Press. Available online: https://dictionary.cambridge.org/dictionary/english/impartiality (accessed on 4 June 2022).
  102. Weld, D.S.; Bansal, G. The challenge of crafting intelligible intelligence. arXiv 2018, arXiv:1803.04263. [Google Scholar] [CrossRef] [Green Version]
  103. Gilpin, L.H.; Bau, D.; Yuan, B.Z.; Bajwa, A.; Specter, M.; Kagal, L. Explaining explanations: An overview of interpretability of machine learning. In Proceedings of the 2018 IEEE 5th International Conference on Data Science and Advanced Analytics (DSAA) 2018, Turin, Italy, 1–4 October 2018; pp. 80–89. [Google Scholar] [CrossRef] [Green Version]
  104. Cambridge Dictionary. Literacy. Cambridge University Press. Available online: https://dictionary.cambridge.org/dictionary/english/literacy (accessed on 4 June 2022).
  105. Wright, D. Understanding “Trustworthy” AI: NIST Proposes Model to Measure and Enhance User Trust in AI Systems. Available online: https://www.jdsupra.com/legalnews/understanding-trustworthy-ai-nist-6387341 (accessed on 4 June 2022).
  106. Cambridge Dictionary. Similarity. Cambridge University Press. Available online: https://dictionary.cambridge.org/dictionary/english/similarity (accessed on 4 June 2022).
  107. Cambridge Dictionary. Suitability. Cambridge University Press. Available online: https://dictionary.cambridge.org/dictionary/english/suitability (accessed on 4 June 2022).
  108. Kharchenko, V.; Illiashenko, O. Concepts of green IT engineering: Taxonomy, principles and implementation. In Green IT Engineering: Concepts, Models, Complex Systems Architectures; Kharchenko, V., Kondratenko, Y., Kacprzyk, J., Eds.; Springer International Publishing: Cham, Switzerland, 2017; Volume 74, pp. 3–19. [Google Scholar] [CrossRef]
  109. Mora-Cantallops, M.; Sánchez-Alonso, S.; García-Barriocanal, E.; Sicilia, M.-A. Traceability for Trustworthy AI: A Review of Models and Tools. Big Data Cogn. Comput. 2021, 5, 20. [Google Scholar] [CrossRef]
  110. Zhang, C.; Wang, J.; Yen, G.G.; Zhao, C.; Sun, Q.; Tang, Y.; Qian, F.; Kurths, J. When Autonomous Systems Meet Accuracy and Transferability through AI: A Survey. Patterns 2020, 1, 100050. [Google Scholar] [CrossRef]
  111. Twin, A. Value Proposition. Available online: https://www.investopedia.com/terms/v/valueproposition.asp (accessed on 4 June 2022).
  112. Cambridge Dictionary. Verifiability. Cambridge University Press. Available online: https://dictionary.cambridge.org/dictionary/english/verifiability (accessed on 4 June 2022).
  113. Defense Innovation Board: AI Principles: Recommendations on the Ethical Use of Artificial Intelligence by the Department of Defense. Available online: https://media.defense.gov/2019/Oct/31/2002204458/-1/-1/0/DIB_AI_PRINCIPLES_PRIMARY_DOCUMENT.PDF (accessed on 4 June 2022).
  114. European Commission for the Efficiency of Justice (CEPEJ). European Ethical Charter on the Use of Artificial Intelligence in Judicial Systems and Their Environment. Available online: https://rm.coe.int/ethical-charter-en-for-publication-4-december-2018/16808f699c (accessed on 4 June 2022).
  115. Kharchenko, V.; Fesenko, H.; Illiashenko, O. Basic model of non-functional characteristics for assessment of artificial intelligence quality. Radioelectron. Comput. Syst. 2022, 2, 1–14, in print. [Google Scholar]
  116. Fesenko, H.; Kharchenko, V.; Sachenko, A.; Hiromoto, R.; Kochan, V. Chapter 9—An internet of drone-based multi-version post-severe accident monitoring system: Structures and reliability. In Dependable IoT for Human and Industry: Modeling, Architecting, Implementation; Kharchenko, V., Kor, A.L., Rucinski, A., Eds.; River Publishers: Gistrup, Denmark, 2018; pp. 197–217. [Google Scholar]
  117. Moskalenko, V.; Moskalenko, A.; Korobov, A.; Semashko, V. The Model and Training Algorithm of Compact Drone Autonomous Visual Navigation System. Data 2019, 4, 4. [Google Scholar] [CrossRef] [Green Version]
  118. Fedorenko, M.; Kharchenko, V.; Lutay, L.; Yehorova, Y. The processing of the diagnostic data in a medical information-analytical system using a network of neuro modules with relearning. In Proceedings of the 2016 IEEE East-West Design & Test Symposium (EWDTS), Yerevan, Armenia, 14–17 October 2016; pp. 381–383. [Google Scholar] [CrossRef]
  119. Gordieiev, O.; Kharchenko, V.; Fominykh, N.; Sklyar, V. Evolution of software quality models in context of the standard ISO 25010. In Advances in Intelligent Systems and Computing; Zamojski, W., Mazurkiewicz, J., Sugier, J., Walkowiak, T., Kacprzyk, J., Eds.; Springer International Publishing: Cham, Switzerland, 2014; Volume 286, pp. 223–232. [Google Scholar] [CrossRef]
  120. Felderer, M.; Ramler, R. Quality assurance for AI-based systems: Overview and challenges (introduction to interactive session). In Lecture Notes in Business Information Processing; Winkler, D., Biffl, S., Mendez, D., Wimmer, M., Bergsmann, J., Eds.; Springer International Publishing: Cham, Switzerland, 2021; Volume 404, pp. 33–42. [Google Scholar] [CrossRef]
  121. Potii, O.; Illiashenko, O.; Komin, D. Advanced security assurance case based on ISO/IEC 15408. In Advances in Intelligent Systems and Computing; Zamojski, W., Mazurkiewicz, J., Sugier, J., Walkowiak, T., Kacprzyk, J., Eds.; Springer International Publishing: Cham, Switzerland, 2015; Volume 365, pp. 391–401. [Google Scholar] [CrossRef]
  122. Bloomfield, R.; Netkachova, K.; Stroud, R. Security-informed safety: If it’s not secure, it’s not safe. In Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Tonetta, S., Schoitsch, E., Bitsch, F., Eds.; Springer International Publishing: Cham, Switzerland, 2013; Volume 8166, pp. 17–32. [Google Scholar] [CrossRef] [Green Version]
  123. Illiashenko, O.O.; Kolisnyk, M.A.; Strielkina, A.E.; Kotsiuba, I.V.; Kharchenko, V.S. Conception and application of dependable Internet of Things based systems. Radio Electron. Comput. Sci. Control 2020, 4, 139–150. [Google Scholar] [CrossRef]
  124. Siebert, J.; Joeckel, L.; Heidrich, J.; Trendowicz, A.; Nakamichi, K.; Ohashi, K.; Namba, I.; Yamamoto, R.; Aoyama, M. Construction of a quality model for machine learning systems. Software Qual. J. 2021. [Google Scholar] [CrossRef]
Figure 1. Quality characteristics classification card general view.
Figure 1. Quality characteristics classification card general view.
Sensors 22 04865 g001
Figure 2. The sequence of AI quality models construction.
Figure 2. The sequence of AI quality models construction.
Sensors 22 04865 g002
Figure 3. Classification map of AI quality characteristics according to Table 3.
Figure 3. Classification map of AI quality characteristics according to Table 3.
Sensors 22 04865 g003
Figure 4. The graph representation of the AI quality model Q S C h A I .
Figure 4. The graph representation of the AI quality model Q S C h A I .
Sensors 22 04865 g004
Figure 5. The graph representation of the AI platform quality model Q S C h A I P .
Figure 5. The graph representation of the AI platform quality model Q S C h A I P .
Sensors 22 04865 g005
Figure 6. The graph representation of the AI system quality model Q S C h A I S .
Figure 6. The graph representation of the AI system quality model Q S C h A I S .
Sensors 22 04865 g006
Figure 7. The graph representation of the basic AI quality model Q S C h A I b .
Figure 7. The graph representation of the basic AI quality model Q S C h A I b .
Sensors 22 04865 g007
Figure 8. The graph representation of the basic AI platform quality model   Q S C h A I P b .
Figure 8. The graph representation of the basic AI platform quality model   Q S C h A I P b .
Sensors 22 04865 g008
Figure 9. The graph representation of the basic AI system quality model Q S C h A I S b .
Figure 9. The graph representation of the basic AI system quality model Q S C h A I S b .
Sensors 22 04865 g009
Figure 10. The graph representation of the quality model of VNS as an AI system with built-in CNN.
Figure 10. The graph representation of the quality model of VNS as an AI system with built-in CNN.
Sensors 22 04865 g010
Figure 11. The graph representation of the quality model of the SDUD as an AI system with a built-in convolutional neural network.
Figure 11. The graph representation of the quality model of the SDUD as an AI system with a built-in convolutional neural network.
Sensors 22 04865 g011
Table 1. AI characteristics references.
Table 1. AI characteristics references.
NoCharacteristicCodeReferences
Referenced InDefined In
(First Row—GS, Second Row—GR, Third Row—GV)
1acceptabilityACP[33,34,35,36,37]
[37]
2accessibilityACS[3,31][3,31]
3accountabilityACN[2,3,6,33,38,39,40][2,3,6,31,33]
[40]
4accuracyACR[2,3,17,33][2,3,33]
5assuranceASR[33][33]
6auditabilityADT[2,3,33,41][2,3]
7authenticityATH[6,31][31]
8availabilityAVL[6,31][31]
9awarenessAWN[36,38,42]
[42]
10biasBIS[2,3,10,15,43,44,45,46,47,48][2,3,10,15]
[44]
11causabilityCSL[49,50,51]
[50]
12completenessCMT[31,52][31]
13complianceCML[33]
14comprehensibilityCMH[36,53,54]
[54]
15communicationCMN[2,3][2,3]
16confidenceCNF[36,55,56]
[56]
17controllabilityCNT[57][57]
18data governanceDGV[2,3][2,3]
19diversityDVS[2,3,33][2,3,33]
20effectivenessEFC[31,58,59][31]
21ethicsETH[2,3,22,23,60,61,62,63,64][2,3]
22environmental well-beingEWB[2,3][2,3]
23exactnessEXC[65,66,67]
[67]
24explainabilityEXP[1,2,3,13,14,17,24,38,40,46,50,51,52,53,55,58,59,66,68,69,70,71,72,73,74,75,76,77,78,79,80,81,82,83,84,85,86,87,88,89][2,3,13]
25explicabilityEXL[2,3][2,3]
26fairnessFRN[2,3,36,40,41,63,90,91,92,93,94][2,3]
27fairness and non-discriminationFND[22,38][22]
28fidelityFDL[53,58]
[58]
29fit to purposeFTP[33,95]
[95]
30fruitfulnessFRT[66,96]
[96]
31governanceGVN[39,97]
[97]
32graspabilityGRS[59]
[59]
33greennessGRN[98,99]
[98,99]
34human agencyHMA[2,3][2,3]
35human oversightHMO[2,3][2,3]
36human oversight and determinationHOD[22][22]
37informativenessINF[36,55,100]
[100]
38impartialityIMP[39,101]
[101]
39integrityING[6,31,33][31]
40intelligibilityINL[40,53,102]
[102]
41interactivityINR[36,38]
[38]
42interpretabilityINP[36,53,73,103]
[103]
43lawfulnessLFL[2,3][2,3]
44literacyLTR[38,104]
[104]
45maintainabilityMNT[31,33][31]
46maturityMTR[31,33][31]
47minimization and reporting of negative impactsMNI[2,3][2,3]
48multi-stakeholder and adaptive governance and collaborationMGC[22,38][22]
49non-discriminationNDS[2,3,22][2,3,22]
50objectivityOBC[17,105][17]
[105]
51prevention of harmPRH[2,3][2,3]
52privacyPRV[2,3,17,22,31,33,36,38,105][2,3,17,22,31]
[105]
53proportionality and do no harmPNH[22][22]
54qualityQLT[6,30,31,33][31]
55redressRDR[2,3][2,3]
56reliabilityRLB[6,17,31,33][31]
57resiliencyRSL[6,17,31,33,39,105]
[105]
58respect for human autonomyRHA[2,3][2,3]
59responsibilityRSP[22,31,38,39,40,58][22]
[40]
60robustnessRBS[2,3,25,58][2,3]
61safetySFT[2,3,17,22,25,38,41][2,3,22]
62similaritySML[65,66,106]
[106]
63securitySCR[6,17,22,25,31,33,38,39][22,31]
64societal well-beingSWB[2,3][2,3]
65suitabilitySTB[31,107]
[107]
66sustainabilitySST[22,33,38][22]
[108]
67traceabilityTRC[2,3,38,109][2,3]
68transferabilityTRF [36,110]
[110]
69transparencyTRP[2,3,6,25,33,36,38,39,40,55][2,3]
[40]
70trade-offsTRO[2,3][2,3]
71trustworthinessTST[2,3,6,17,25,31,33,36,41,55][6,33]
72usabilityUSB[6,31,33][6,31]
73understandabilityUND[36,57]
[36]
74value propositionVPR[33,111]
[111]
75verifiabilityVFB[25,73]
[112]
Table 2. Analysis of principles and characteristics of AI and AI systems.
Table 2. Analysis of principles and characteristics of AI and AI systems.
Group of Principles and ValuesReference PrinciplesComment
Ethical principles[2]Respect for human autonomyCan be defined as HMA and HMO
Prevention of harm
FairnessCan be defined as FRN
ExplicabilityCan be defined as EXP
Values[22]Respect, protection, and promotion of human rights and fundamental freedom, and human dignity
Environment and ecosystem flourishing
Ensuring diversity and inclusiveness
Living in peaceful, just, and interconnected societies
Principles of activity[22]Proportionality and non-infliction of harm
Safety and securityCan be defined as SFT and SCR
Fairness and non-discriminationCan be defined as FRN and NDS
SustainabilityCan be defined as SST
Right to privacy and data protectionCan be defined as PRV
Human oversight and determinationCan be defined as HMO, HMA
Transparency and explainabilityCan be defined as TRP, EXP
Responsibility and accountabilityCan be defined as RSP, ECM
Awareness and literacy
Multi-stakeholder and adaptive governance and collaboration
Table 3. Analysis of harmonization of AI characteristics definitions.
Table 3. Analysis of harmonization of AI characteristics definitions.
NoCharacteristicQP/QUDefinitionWayReferences
1Acceptability
ACP
QUThe ability of the AI to ensure at least partial compliance with customer requirements or consumer expectations H[37]
2Accessibility
ACS
QP, QUThe ability of the AI platform to enable a user or process that has the appropriate authority to use AI by established rules without waiting longer than a specified or acceptable timeH[3,31]
3Accountability
ACN
QPThe ability of AIs to report in a defined form on the results of operations in a transparent mannerR[31]
4Accuracy
ACR
QUThe ability of the AI and the AI platform to ensure that the results of the requirements and/or functions presented by certain data are close to their true valuesH[2,3,33]
5Auditability
ADT
QPThe ability of the AI platform, which is characterized by the degree of suitability for the audit of its ethical and technical components using certain methods and toolsH[2,3]
6Availability
AVL
QUThe ability of the AI platform, which is determined by the degree of its operability, availability, and recoverabilityH[31]
7Bias
BIS
QUIt is an AI characteristic that determines the risks of results that are biased due to erroneous assumptions and errors in the process of tuning models (e.g., machine learning)H[2,3,10,15]
8Causability
CSL
QUThe ability of the AI to determine the cause and effect relationships between events that occur during its useH[50]
9Completeness
CMT
QUThe ability of the AI to be holistic in terms of compliance with all customer requirementsH[31]
10Comprehensibility CMH QUThe ability of the AI to provide the user (or facilitate the user) with an understanding of the explanations sufficient to enable the use of the AI or the information obtained through it to perform other tasksA
11Controllability
CNT
QPThe ability of the AI platform to provide opportunities to control and manage the processes of functioning as intendedH[57]
12Data governance
DGV
QPThe ability of the AI platform to provide data control and management capabilitiesH[2,3]
13Diversity
DVS
QPThe ability of the AI and AI platforms to minimize the risk of failure to perform specified (defined as necessary) functions or tasks due to failures due to physical and informational factors, using a variety of models, algorithms, and other meansA
14Effectiveness
EFS
QP,
QU
The ability of the AI platform, which is determined by the degree of achievement of the planned results, considering the number of resources spentH[31]
15Ethics
ETH
QUThe ability of the AI to meet current standards of morality on the results of functioning H[2,3]
16Explainability
EXP
QP,
QU
The ability of the AI to be understood and predictable in terms of purpose and behaviorH[2,3,13]
17Fairness
FRN
QUThe ability of the AI to minimize the risk of biased anomalies in ethical decisions (including lack of favoritism, discrimination on religious, racial, or other grounds, etc.), as well as misconceptions and errors in the modeling processH
[2,3]
18Graspability
GRS
QUThe ability of the AI to provide the user with opportunities for the critical perception of AI in an open and democratic environmentH[59]
19Greenness
GRN
QP,
QU
The ability of the AI platform to consume a minimum of energy or other resources, be it energy and/or resource-efficient, and not have an unacceptable impact on the environmentA
20Human agency
HMA
QUThe ability of the AI to enable the user to make autonomous informed decisions about the use of AIH[2,3]
21Human oversight
HMO
QUThe ability of the AI to enable the user to control and, if necessary, interfere in a certain way with the functioning of AIH[2,3]
22Informativeness
INF
QP,
QU
The ability of the AI platform to provide the user with useful information to gain new knowledge and generate ideasH[100]
23Integrity
ING
QP,
QU
The ability of the AI and AI platform, which is characterized by the degree of prevention of unauthorized access to modify algorithms or data used by the system R[31]
24Interactivity
INR
QP,
QU
The ability of the AI to provide effective and proactive interaction with the user H[38]
25Interpretability
INP
QP,
QU
The ability of the AI to provide and interpret information in a user-friendly wayH[103]
26Lawfulness
LFL
QUThe ability of the AI to comply with laws and regulations R[2,3]
27Maintainability
MNT
QP,
QU
The ability of the AI platform, which is determined by the degree of efficiency and effectiveness with which it repairs, is maintained and can be modified according to customer requirements (for example through pre- or retraining models)H[31]
28Maturity
MTR
QP,
QU
The ability of the AI platform, which is determined by the degree of compliance with the reliability requirements set by the clientR[31]
29Non-discrimination
NDS
QUThe ability of the AI to enforce ethical standards on non-discrimination on any groundsH[2,3,22]
30Objectivity
OBC
QP,
QU
The ability of the AI and AI platform to prevent the use of compromised or falsified data R[105]
31Privacy
PRV
QP,
QU
The ability of the AI and AI platform to ensure the right to have personal information by user requirementsH[2,3,105]
32Redress
RDR
QUThe ability of the AI to provide available mechanisms to ensure adequate compensation for the effects of adverse effects on humansH[2,3]
33Reliability
RLB
QP,
QU
The ability of the AI platform to perform certain functions for a specified period under certain conditions of use, maintenance, and repair H[31]
34Resiliency
RSL
QP,
QU
The ability of the AI and AI platform to continue to function amid changing requirements, parameters of the physical and information environment, as well as the emergence of unspecified violations and failuresA
35Responsibility
RSP
QUThe ability of the AI to function considering the expectations of the client (user) by ethical norms, legal regulations, as well as to inform him in case of a possible violation H[22,40]
36Robustness
RBS
QP,
QU
The ability of the AI and AI platform to operate correctly in a wide range of input data and operating conditions and to enter a state of system shutdown if these data and conditions exceed the specified limitsH[2,3]
37Safety
SFT
QP,
QU
The ability of the AI and AI platform to avoid the risk of unacceptable damage and loss due to failures due to internal and external causes, and to reduce its consequences with the use of tools built into the AI H[2,3,22]
38Security
SCR
QP,
QU
(Including information and cybersecurity)—the ability of the AI and AI platforms to protect information and physical assets in such a way that other unidentified (unauthorized) persons or systems, including AI and AI platforms, do not have access to them or have such access. such as specified type and level of authorizationH[22,31]
39Societal
well-being
SWB
QUThe ability of the AI to take social processes into account and not to harm the physical and mental well-being of people and the well-being of society as a whole H[2,3]
40Sustainability
SST
QP,
QU
The ability of the AI platform to positively influence the sustainable development of economic, social, and environmental environments, as well as to develop without the use of additional resources that are unacceptable in terms of volume or security factors H[22,108]
41Traceability
TRC
QP,
QU
The ability of the AI to track user compliance in a user-friendly way, to search for and document errors and inconsistencies at every stage of the life cycle H[2,3]
42Transferability (or interoperability as portability for SW quality stds.)
TRF
QP,
QU
The ability of the AI platform or software and/or hardware to implement AI, which is determined by the degree of adaptability for the transfer of information, knowledge, and tools to other systems or platforms H[111]
43Transparency
TRP
QP,
QU
The ability of the AI to describe, test, and reproduce models, individual components, and decision-making algorithmsH[2,3,40]
44Trustworthiness
TST
QP,
QU
The ability of the AI, which is characterized by the degree of confidence of the user or other stakeholder (developer, auditor, etc.) that the AI meets the requirements and performs its functions in a predictable mannerH[6,33]
45Usability
USB
QP,
QU
The ability of the AI platform, which is characterized by the extent to which it can be used by users to achieve specific goals with efficiency, effectiveness, and satisfaction in a given usage context R[31]
46Verifiability
VFB
QPThe ability of the AI and AI platform, characterized by the degree of suitability for verification by different methodsH[113]
Table 4. Results of the coded references analysis.
Table 4. Results of the coded references analysis.
ReferenceTop-CharacteristicDefinition
[2]TrustworthinessTST = {HMA, HMO, RBS, SFT, PRV, DGV, TRP {TRC, EXP}, DVS, NDS, FRN, SWB, ACN {ADT, RDR}}
[6]TrustworthinessTST = {RLB, AVL, RSL, SCR, PRV, SFT, ACN, TRP, ING, USB}
[17]TrustworthinessTST = {ACR, RLB, RSL, OBC, SCR, EXP, SFT, ACN, PRV}
[33]TrustworthinessTST = {ACN {RSP, TRP, ADT, MTR}, MNT, ING, SST, ACR, DVS, CML, ACP, USB}
[36]ExplainabilityEXP = {CMH, INP, TST, CSL, TRF, INF, FRN, ACS, INR, PRV, TRP}
[38]EthicsETH = {SFT, SCR, SST, PRV, TRP, EXP, RSP, ACN}
[39]ResponsibilityRSP = {ACN, RSL, TRP, SCR}
[40]ResponsibilityRSP = {ACN, TRP, FRN, ETH}
[41]TrustworthinessTST = {FRN, EXP, ADT, SFT}
[52]ExplainabilityEXP = {CMT, INP}
[53]ExplainabilityEXP = {INP{CMH}}
[55]ExplainabilityEXP = {TST, INF, TRP}
[57]ExplainabilityEXP = {INP, CMH}
[58]ResponsibilityRSP = {RBS, EXP, ETH, EFC}
[106]TrustworthinessTST = {TRP, VFB, EXP, SCR, RBS, SFT}
Table 5. The tabular representation of the AI quality model Q S C h A I .
Table 5. The tabular representation of the AI quality model Q S C h A I .
1st Level2nd Level3rd Level
Ethics (ETH)Fairness (FRN)Bias (BIS)
Non-discrimination (NDS)
Graspability (GRS)
Human agency (HMA)
Human oversight (HMO)
Redress (RDR)
Explainability (EXP)Accountability (ACN)
Causability (CSL)
Completeness (CMT)
Comprehensibility (CMH)
Interactivity (INR)
Interpretability (INP)
Transparency (TRP)Traceability (TRS)
Verifiability * (VFB)
Lawfulness (LFL)
Responsibility (RSP)
Trustworthiness (TST)Acceptability (ACP)
Accuracy * (ACR)
Diversity * (DVS)
Resiliency * (RSL)
Robustness * (RBS)
Security * (SCR)Integrity (ING)
Objectivity (OBC)
Privacy (PRV)
Safety * (SFT)Societal well-being (SWB)
Table 6. The tabular representation of the AI platform quality model Q S C h A I P .
Table 6. The tabular representation of the AI platform quality model Q S C h A I P .
1st Level2nd Level
Accessibility (ACS)
Accuracy * (ACR)
Auditability (ADT)
Availability (AVL)
Controllability (CNT)Data governance (DGV)
Diversity * (DVS)
Effectiveness (EFC)
Informativeness (INF)
Maintainability (MNT)Transferability (TRF)
Reliability (RLB)Maturity (MTR)
Resiliency * (RSL)
Robustness * (RBS)
Security * (SCR)
Safety * (SFT)
Sustainability (SST)Greenness (GRN)
Verifiability * (VFB)
Usability (USB)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kharchenko, V.; Fesenko, H.; Illiashenko, O. Quality Models for Artificial Intelligence Systems: Characteristic-Based Approach, Development and Application. Sensors 2022, 22, 4865. https://doi.org/10.3390/s22134865

AMA Style

Kharchenko V, Fesenko H, Illiashenko O. Quality Models for Artificial Intelligence Systems: Characteristic-Based Approach, Development and Application. Sensors. 2022; 22(13):4865. https://doi.org/10.3390/s22134865

Chicago/Turabian Style

Kharchenko, Vyacheslav, Herman Fesenko, and Oleg Illiashenko. 2022. "Quality Models for Artificial Intelligence Systems: Characteristic-Based Approach, Development and Application" Sensors 22, no. 13: 4865. https://doi.org/10.3390/s22134865

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop