AI Trustworthiness in Manufacturing: Challenges, Toolkits, and the Path to Industry 5.0
Abstract
1. Introduction
- Critically examine the role of AI in the transition from Industry 4.0 to Industry 5.0, with a focus on the technical, ethical, and organisational challenges specific to manufacturing.
- Assess the effectiveness and limitations of existing toolkits for ensuring AI trustworthiness—specifically transparency, fairness, robustness, and accountability in manufacturing contexts.
- Formulate targeted research questions and methodological approaches to address the most pressing challenges of AI adoption in manufacturing, drawing on industry case studies
2. Background
2.1. Industry 5.0 Capabilities and Challenges
2.2. Explaining Industry 5.0
2.3. AI Use Cases in Industry 5.0
- Digital Twin: AI is utilised to create virtual representations of processes, production systems, factories, and supply chains, referred to as digital twins. These virtual models are employed to simulate, evaluate, and predict performance in real-time. By replicating the physical environment, digital twins allow manufacturers to monitor and improve operations without needing direct engagement with the physical assets. They depend on data from IoT sensors, programmable logic controllers (PLCs), deep learning techniques, and AI algorithms to continuously update the digital model with real-time information, ensuring an up-to-date and accurate virtual replica.
- Predictive maintenance: AI processes sensor data from machinery to predict potential failures before they happen. By utilising a digital twin to examine patterns in equipment behaviour and performance, these systems can notify operators of potential issues in advance, enabling them to prevent breakdowns before they worsen. For instance, automotive manufacturers use predictive maintenance on assembly-line robots, greatly decreasing unplanned downtime and leading to significant cost savings. This method also allows manufacturers to schedule maintenance during off-peak hours, minimising disruptions to production timelines [43].
- Custom Manufacturing: AI empowers manufacturers to provide mass customisation, enabling products to be tailored to individual customer preferences without disrupting production speed. By incorporating AI into the design process, companies can swiftly adjust designs in response to real-time consumer feedback. For example, clothing manufacturers utilise AI algorithms to personalise products, allowing customers to select designs that align with their unique tastes. This adaptability not only improves customer satisfaction but also boosts engagement by offering a more personalised shopping experience.
- Generative Design: This technology allows manufacturers to explore numerous design possibilities by considering factors like materials and manufacturing limitations. This approach accelerates the design process by enabling the rapid evaluation of multiple iterations. Generative AI design tools are already being utilised in industries like the aerospace and automotive industries, where companies use them to develop optimised parts. Although the technology is already in use, its complete potential is still not being explored within the dynamic landscape of modern manufacturing.
- Quality Control: AI improves quality control by using computer vision and machine learning, often supported by a digital twin, to detect defects in real-time. These systems examine product images during the manufacturing process, identifying inconsistencies or faults with greater precision than human inspectors. For example, electronic manufacturers utilise AI-driven quality control to ensure components meet stringent specifications. This results in higher product quality, reduced waste, and greater customer satisfaction.
- Supply Chain Management: AI streamlines supply chain operations by analysing large volumes of data to forecast demand, manage stock levels, and improve logistics. When coupled with a digital twin, AI can build a virtual model of the entire supply chain, enabling manufacturers to predict and simulate disruptions or shortages in real-time. Machine learning assists with demand predictions and automates procurement, ensuring that manufacturers receive materials precisely when needed. AI-driven order management systems also optimise order fulfilment, ensuring deliveries are made on time. For example, food manufacturers use AI to anticipate seasonal shifts in demand, allowing them to better manage resources and reduce waste. This ultimately boosts operational efficiency and enhances responsiveness to market fluctuations.
- Inventory management: AI enhances inventory management by analysing data to predict stock requirements and streamline replenishment. By forecasting demand and tracking inventory in real-time, manufacturers can ensure optimal stock levels, lowering storage costs and improving cash flow. For instance, food and beverage manufacturers use AI systems to monitor ingredient consumption as it happens. This enables them to predict future needs based on production timelines, seasonal factors, and historical usage, helping to avoid production disruptions and minimise waste from excess stock.
- Energy Management: AI systems track energy consumption in real-time to pinpoint inefficiencies. These systems can suggest changes that help cut energy costs and reduce environmental impact. For example, electronic manufacturers use AI-driven energy management solutions to improve their operations, leading to substantial cost reductions and a smaller carbon footprint.
3. Methodology
- Literature Review and Data Sources: This research begins with an extensive literature review, targeting peer-reviewed journal articles, conference proceedings, and authoritative industry reports. Sources are drawn from high-impact databases including IEEE Xplore, Scopus, Web of Science, and the ACM Digital Library. To ensure relevance and currency, this review is limited to works published within the last decade, with a particular emphasis on studies addressing AI trustworthiness in manufacturing. In addition, regulatory documents and guidelines—such as ISO/IEC standards, the European Union (EU) AI Act, and the Assessment List for Trustworthy Artificial Intelligence (ALTAI)—are included to capture the evolving landscape of ethical and legal requirements.
- Systematic Search and Selection: A systematic search strategy is employed, using targeted keywords such as “AI trustworthiness”, “Industry 5.0”, “ethical AI”, “transparency in AI”, “Toolkits in AI”, and “manufacturing.” The selection process prioritises studies that address the core dimensions of trustworthy AI—transparency, fairness, robustness, and accountability—within manufacturing contexts. The inclusion of case studies, both of AI successes and failures, provides practical grounding and validation for the findings. To ensure a comprehensive and transparent literature review, the search strategy involved querying databases using specific search strings like “AI trustworthiness” AND “manufacturing”, “AI ethics” AND “smart manufacturing”, and “responsible AI” AND “industry 5.0”. Inclusion criteria were applied to select peer-reviewed articles, conference papers, and relevant reports published between 2015 and 2024, focusing on AI trustworthiness in manufacturing applications. Exclusion criteria were used to filter out studies that were not directly relevant to the manufacturing sector or did not address AI trustworthiness. The initial search yielded 500 of articles, which were then screened based on their titles and abstracts. Full-text reviews were conducted on 300 articles, resulting in a final selection of 200 articles that met the inclusion criteria.
- Toolkit Analysis: The analysis is structured around four key dimensions of AI trustworthiness: transparency, fairness, robustness, and accountability. For each dimension, this study critically examines a range of prominent toolkits and frameworks, including but not limited to AI Explainability 360 (AIX360), SHapley Additive exPlanations (SHAP), Local Interpretable Model-agnostic Explanations (LIME), AI Fairness 360 (AIF360), FairLearn, IBM Adversarial Robustness Toolbox (IBM ART), and CleverHans. The discussion evaluates the strengths, limitations, and practical applications of these tools, offering a comprehensive perspective on their contributions to trustworthy AI in manufacturing. The evaluation criteria included: (1) transparency mechanisms (e.g., explainable AI (XAI) techniques), (2) fairness metrics and mitigation strategies, (3) robustness testing and validation methods, (4) accountability frameworks, and (5) ethical guidelines and compliance support. The toolkits were assessed based on their functionalities, ease of use, and applicability to manufacturing contexts. The evaluation involved a qualitative, comparative analysis, drawing upon expert judgment to assess the toolkits’ strengths and weaknesses in addressing AI trustworthiness concerns. A formal numerical scoring system was not used due to the diversity of toolkit functionalities and the context-dependent nature of manufacturing applications. Instead, the evaluation focused on providing a nuanced understanding of each toolkit’s capabilities and limitations in promoting AI trustworthiness.
- Pipeline and Practical Considerations: The methodology explicitly addresses challenges across the entire AI pipeline—from data collection and preprocessing to model training, deployment, and post-deployment monitoring. Special attention is given to issues such as data quality, interoperability, bias, concept drift, and the integration of domain expertise.
- Interdisciplinary and Human-Centric Approach: Recognising the complexity of manufacturing environments, the methodology emphasises interdisciplinary collaboration among AI developers, domain experts, and end-users. This ensures that technical solutions are both practically relevant and ethically aligned. The approach is further informed by the human-centric and sustainable ethos of Industry 5.0, integrating ethical considerations and stakeholder perspectives at every stage. No new stakeholder interviews or primary qualitative data were collected; instead, the human-centric perspective is embedded through the integration of ethical considerations and stakeholder insights from published qualitative studies and documented experiences. To illustrate the potential challenges and implications of AI trustworthiness in manufacturing, this study employs a series of hypothetical case studies. These scenarios are not based on specific real-world implementations but are carefully constructed to represent common AI applications across diverse manufacturing sectors such as automotive, aerospace, and electronics. The purpose of these illustrative cases is to explore potential issues related to transparency, fairness, robustness, and accountability that could arise when deploying AI solutions in these contexts. By analysing these hypothetical scenarios, this study aims to provide insights into the proactive measures and strategies that manufacturing organisations can adopt to ensure AI trustworthiness.
- Limitations and Bias Mitigation Strategies: As with any research, this study is subject to certain limitations. To address potential biases, several mitigation strategies were implemented throughout the research process. The possibility of selection bias in the illustrative case studies was reduced by ensuring a diverse representation of manufacturing sectors and AI application areas. To mitigate publication bias in the literature review, both peer-reviewed articles and grey literature sources (e.g., industry reports; white papers) were considered. The analytical subjectivity inherent in the toolkit evaluation was addressed through the use of a structured evaluation framework, clear evaluation criteria, and the involvement of multiple researchers in the analysis process to promote inter-rater reliability. While these strategies do not eliminate bias entirely, they significantly reduce its impact on this study’s findings.
- Synthesis and Research Questions: Findings from the literature, toolkit evaluations, and case studies are synthesised to identify persistent gaps and emerging best practices. This study formulates open research questions to guide future inquiry, particularly regarding the operationalisation of trustworthy AI in dynamic, real-world manufacturing settings.
4. Challenges in AI Adoption for Industry 5.0
- Technical Challenges: A primary technical challenge is the “black box” nature of many AI models, which lack transparency and make it difficult for operators to trust or verify their decisions, raising concerns about accountability. The European Commission’s Ethics Guidelines stress the need for explainable and transparent AI to foster user trust [45,46]. Another issue is the shortage of high-quality, relevant data for training AI models. Poor or biased data can lead to inaccurate results and reinforce existing biases, limiting AI’s effectiveness in manufacturing [47]. Reliability is also a concern, as AI models that perform well in controlled settings may not replicate their success under real-world conditions due to variations in data distributions and unforeseen operational challenges, leading to inconsistent performance and affecting production quality and efficiency [48].
- Security and Cybersecurity: Security concerns are paramount, as AI systems are susceptible to cyber threats that can compromise sensitive industrial data and disrupt operations. For example, adversarial attacks on machine learning models can manipulate outputs or cause system failures. The NIST Cybersecurity Framework and ISO/IEC 27001 provide standards for securing industrial AI systems, but their implementation in dynamic manufacturing environments remains challenging [49,50].
- Ethical and Regulatory Challenges: Ethical considerations further complicate AI adoption. The potential for AI systems to perpetuate biases or make decisions that lack fairness necessitates the development of robust ethical frameworks. The European Commission’s guidelines advocate for AI that is lawful, ethical, and robust, ensuring adherence to principles such as fairness, accountability, and respect for privacy [45]. Regulatory uncertainty is a significant barrier, particularly where existing regulations conflict with AI optimisation. For instance, the General Data Protection Regulation (GDPR) mandates the right to explanation for automated decisions, which can conflict with the opacity of some machine learning models. This tension between data privacy and model transparency creates compliance challenges for manufacturers seeking to deploy advanced AI solutions [51]. The absence of standardised regulations and governing bodies for AI in manufacturing further exacerbates uncertainty, making it difficult for companies to ensure compliance and align with best practices. The European Commission’s ALTAI aims to provide actionable guidance to address these issues [52].
- Organisational and Workforce Barriers: The human-centric approach of Industry 5.0 emphasises the importance of collaboration between AI systems and human workers. Bridging the skills gap through targeted education and training programs is vital to equip the workforce with the necessary competencies to effectively interact with AI technologies [53]. Organisational resistance to change, lack of digital maturity, and insufficient leadership support can also hinder successful AI adoption.
- Barriers for SMEs versus Large Manufacturers: Small and medium-sized enterprises (SMEs) face unique barriers compared to large manufacturers. SMEs often lack the financial resources, technical expertise, and access to high-quality data required for effective AI implementation. The cost of acquiring, integrating, and maintaining AI systems can be prohibitive, and SMEs may struggle to attract or retain skilled personnel. In contrast, large manufacturers typically have greater capacity to invest in digital infrastructure, data management, and workforce development, enabling them to overcome many of these barriers more readily. As a result, the digital divide between SMEs and large enterprises may widen, limiting the broader impact of AI in the manufacturing sector [54].
5. Importance of AI Trustworthiness
6. Factors Defining AI Trustworthiness
- Technical Robustness and Safety: AI systems must be reliable and function as intended. They should be capable of recovering from failures without harm and handle errors throughout the AI lifecycle. The system must also resist external threats and produce reproducible results [29].
- Privacy and Data Governance: AI systems must safeguard user data throughout its lifecycle, ensuring compliance with data protection regulations like the General Data Protection Regulation (GDPR). Sensitive data must be protected from misuse [74].
- Diversity, Non-discrimination, and Fairness: AI systems must ensure fairness, treating all societal groups equally and avoiding any form of discrimination, whether direct or indirect [77].
- Societal and Environmental Well-being: AI systems should not harm society or the environment during their development, operation, or use [61].
- Accountability: AI systems must be capable of justifying their decisions. There should be mechanisms for assigning responsibility for both correct and incorrect outcomes, along with regular audits to prevent harm [78].
6.1. Explainability
6.2. Accountability
6.3. Fairness
6.4. Robustness
- Data Level Robustness: A model trained on limited datasets that do not reflect real-world variations may suffer significant performance degradation. One major challenge is a distributional shift, where the data seen during deployment differs from the training data, affecting model reliability [147]. This issue is particularly concerning in safety-critical domains. For example, in autonomous driving, AI models must function under a range of environmental conditions. While a system trained in sunny weather may perform well, its effectiveness in night time or rainy conditions could be severely reduced. To address this, researchers and industry professionals employ extensive testing and development strategies to improve AI perception under varying weather conditions, ensuring consistent performance [148,149].
- Algorithm-Level Robustness: AI models can be vulnerable to adversarial attacks, where maliciously modified inputs deceive the system. These attacks have raised concerns in both academia and industry, leading to extensive research on threat classification and defence mechanisms [150,151,152,153,154]. Adversarial attacks can be categorised based on their timing:
- Decision-Time Attacks: These involve modifying input samples in real-time to manipulate the model’s predictions. Attackers may use such methods to bypass security mechanisms or impersonate legitimate users [155].
- Training-Time Attacks (Poisoning Attacks): In this approach, adversaries introduce deceptive samples into the training data, influencing the model’s learning process and altering its behaviour in specific situations [155].
Another important classification is based on the space in which attacks are conducted:- Feature-Space Attacks: Traditional adversarial methods directly alter input features to deceive the model.
- Problem-Space Attacks (Entity-Based Attacks): Instead of modifying digital data, attackers alter physical objects to manipulate AI recognition. For example, a person wearing specially designed adversarial glasses could bypass a facial recognition system [156,157]. Apart from adversarial attacks, model stealing (exploratory attacks) is another significant threat. These attacks do not directly alter model behaviour but extract knowledge about the AI system, which can later be exploited to craft more effective adversarial samples [158].
- System-Level Robustness: AI systems must be designed to handle a wide range of unexpected or illegal inputs in real-world applications. Practical cases include the following:
- Unanticipated Inputs: For instance, an image with an extremely high resolution might cause an AI-based image recognition system to crash.
- Sensor Interference: In autonomous vehicles, a lidar system might misinterpret signals from other vehicles, leading to corrupted input data.
- Spoofing Attacks: Attackers may use fake inputs—such as printed photos or masks—to deceive biometric authentication systems, raising security concerns [159]. To mitigate these risks, defensive mechanisms are categorised as either proactive or reactive [160]. Proactive defences aim to strengthen AI models against diverse inputs, making them inherently robust. Reactive defences focus on detecting adversarial samples or identifying anomalies in data distribution.
- Evaluating Robustness: Assessing robustness is crucial for detecting vulnerabilities and managing risks. Two primary evaluation methods are robustness testing and mathematical verification.
- Robustness Testing Testing plays a key role in validating AI robustness, just as it does in traditional software development. Techniques such as monkey testing—which uses randomised inputs to check system stability—can be applied to AI models [161]. Additionally, software testing methodologies have been adapted to assess AI resilience against adversarial attacks [162,163].Another common method is performance testing (benchmarking), which evaluates model robustness using test datasets with varying distributions. One widely used metric is the minimal adversarial perturbation, which measures the smallest modification needed to mislead an AI model. Another key evaluation metric is the attack success rate, which reflects how easily an adversary can compromise the system [164,165].
- Mathematical Verification Borrowed from formal verification methods, mathematical validation techniques are increasingly used to assess AI robustness. For instance, researchers derive certified lower bounds on the minimum distortion required for an adversarial attack—a measure of how resistant a model is to adversarial manipulations [166,167].
7. Challenges in the AI Pipeline: From Data Collection to Model Deployment
7.1. Data Collection
- RQ1: How can data producers and owners implement interoperable data schemas to ensure data integrity?
- RQ2: What mechanisms best facilitate the extraction of unbiased, informative datasets from complex environments?
- RQ3: What types of biases are present in manufacturing datasets and data collection processes, and what strategies can be used to detect and address them while maintaining optimal performance?
- RQ4: What methods can be employed to gather unbiased and informative datasets from shop floor environments where human involvement is significant?
- RQ5: How can workers with limited AI expertise effectively evaluate algorithms for bias and fairness?
7.2. Data Preprocessing
- Labelling Issues: The company needs labelled data to train its AI model, but quality labels are subjective and depend on the expertise of the quality inspectors. For instance, one inspector might classify a product as “acceptable”, while another might label it as “defective” under similar conditions. This inconsistency leads to uncertain labels, which can affect the model’s accuracy.
- Data Imbalance: Most of the products meet quality standards, resulting in an imbalanced dataset where defective products are under-represented. This imbalance can cause the AI model to overlook rare but critical defects.
- Data Cleaning: The IoT sensors occasionally produce noisy or incomplete data due to hardware malfunctions or network issues. For example, temperature readings from a sensor might show sudden spikes that do not reflect actual conditions, leading to incorrect predictions.
- Feature Engineering: The production team hypothesises that the humidity level in the factory might influence product quality. However, this data is not directly available and needs to be derived from existing environmental data, requiring domain expertise to create meaningful features.
- RQ6: What methodologies can be developed to ensure consistent and accurate labelling of manufacturing data, particularly in scenarios where labels are subjective or context-dependent?
- RQ7: What are the most effective practices for addressing imbalanced datasets in manufacturing, and how can synthetic data generation techniques be optimised for such applications?
- RQ8: How can automated approaches be designed to detect and rectify noisy or incomplete data originating from IoT sensors and other manufacturing data sources?
- RQ9: What strategies can be employed to integrate domain knowledge into feature engineering processes while minimising the risk of introducing additional biases?
- RQ10: What frameworks or methodologies can be developed to identify and mitigate biases in manufacturing datasets, ensuring fair and unbiased AI-driven predictions?
7.3. Developing AI Models
- Model selection is a crucial phase where the type of machine learning model is chosen, as it directly influences the model’s interpretability and performance. Simpler models, such as decision trees, are often preferred in manufacturing due to their ease of understanding and transparency. However, this preference can sometimes lead to reduced accuracy, creating a trade-off between interpretability and performance. Additionally, computational limitations, particularly in resource-constrained environments, can restrict the use of more advanced models, further complicating the selection process [181].
- Model training is another critical phase, where hyperparameters such as the depth of decision trees or the number of layers in a neural network are fine-tuned to enable the model to learn patterns effectively from the data. This phase is resource-intensive, with high computational and environmental costs, especially for large-scale models. The process also requires skilled professionals to manage training effectively and avoid suboptimal results. While automated machine learning (AutoML) tools can simplify the training process, they often introduce challenges such as reduced transparency, potential bias, and overfitting, which can compromise the trustworthiness of the model [185].
- Deployment phase focuses on integrating the model into real-world operations and ensuring its continued relevance. A significant challenge in this phase is addressing concept drift, where changes in the data distribution over time can render the model’s predictions inaccurate. To mitigate this, organisations must implement mechanisms to detect when updates are needed and determine whether incremental updates or full retraining is more appropriate. This phase requires careful monitoring and adaptation to ensure the model remains effective in dynamic manufacturing environments [186]. See Box 3.
- RQ11: How can manufacturing practitioners balance the trade-off between model interpretability and performance?
- RQ12: What are the most effective methods for detecting and managing concept drift in industrial applications?
- RQ13: How can organisations account for the environmental and financial costs of training AI models?
8. Conclusions and Future Directions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
AI | Artificial Intelligence |
GDPR | General Data Protection Regulation |
AIX360 | AI Explainability 360 |
SHAP | SHapley Additive exPlanations |
LIME | Local Interpretable Model-agnostic Explanations |
AIF360 | AI Fairness 360 |
IBM ART | IBM Adversarial Robustness Toolbox |
NIST | National Institute of Standards and Technology |
OECD | Organisation for Economic Co-operation and Development |
ISO | International Organization for Standardization |
IEC | International Electrotechnical Commission |
IEEE | Institute of Electrical and Electronics Engineers |
ALTAI | Assessment List for Trustworthy Artificial Intelligence |
EU | European Union |
IoT | Internet of Things |
ML | Machine Learning |
XAITK | Explainable AI Toolkit |
HVMC | High-Value Manufacturing Catapult |
COMPAS | Correctional Offender Management Profiling for Alternative Sanctions |
AutoML | Automated Machine Learning |
NLP | Natural Language Processing |
DNN | Deep Neural Network |
CNN | Convolutional Neural Network |
RMF | Risk Management Framework |
References
- Lu, Y. Industry 4.0: A survey on technologies, applications and open research issues. J. Ind. Inf. Integr. 2017, 6, 1–10. [Google Scholar] [CrossRef]
- Xu, L.D.; Duan, L. Big data for cyber physical systems in industry 4.0: A survey. Enterp. Inf. Syst. 2019, 13, 148–169. [Google Scholar] [CrossRef]
- Olsen, T.L.; Tomlin, B. Industry 4.0: Opportunities and challenges for operations management. Manuf. Serv. Oper. Manag. 2020, 22, 113–122. [Google Scholar] [CrossRef]
- Russell, S.J.; Norvig, P. Artificial Intelligence: A Modern Approach; Pearson: London, UK, 2016. [Google Scholar]
- Mitchell, T.M. Machine Learning; McGraw-Hill: New York, NY, USA, 1997. [Google Scholar]
- Lee, J.; Bagheri, B.; Kao, H.A. A cyber-physical systems architecture for industry 4.0-based manufacturing systems. Manuf. Lett. 2015, 3, 18–23. [Google Scholar] [CrossRef]
- Xu, X.; Lu, Y.; Vogel-Heuser, B.; Wang, L. Industry 4.0 and Industry 5.0—Inception, conception and perception. J. Manuf. Syst. 2021, 61, 530–535. [Google Scholar] [CrossRef]
- Ullah, A.S. Fundamental Issues of Concept Mapping Relevant to Discipline-Based Education: A Perspective of Manufacturing Engineering. Educ. Sci. 2019, 9, 228. [Google Scholar] [CrossRef]
- Doshi-Velez, F.; Kim, B. Towards a rigorous science of interpretable machine learning. arXiv 2017, arXiv:1702.08608. [Google Scholar]
- Hassija, V.; Chamola, V.; Mahapatra, A.; Singal, A.; Goel, D.; Huang, K.; Scardapane, S.; Spinelli, I.; Mahmud, M.; Hussain, A. Interpreting black-box models: A review on explainable artificial intelligence. Cogn. Comput. 2024, 16, 45–74. [Google Scholar] [CrossRef]
- Greenberg, M.R. Energy policy and research: The underappreciation of trust. Energy Res. Soc. Sci. 2014, 1, 152–160. [Google Scholar] [CrossRef]
- Floridi, L.; Cowls, J.; Beltrametti, M.; Chatila, R.; Chazerand, P.; Dignum, V.; Luetge, C.; Madelin, R.; Pagallo, U.; Rossi, F.; et al. AI4People—An ethical framework for a good AI society: Opportunities, risks, principles, and recommendations. Minds Mach. 2018, 28, 689–707. [Google Scholar] [CrossRef]
- Veale, M. A critical take on the policy recommendations of the EU high-level expert group on artificial intelligence. Eur. J. Risk Regul. 2020, 11, e1. [Google Scholar] [CrossRef]
- AI, N. Artificial Intelligence Risk Management Framework (AI RMF 1.0). 2023; pp. 1–42. Available online: https://nvlpubs.nist.gov/nistpubs/ai/nist.ai.100-1.pdf (accessed on 7 July 2025).
- Jobin, A.; Ienca, M.; Vayena, E. The global landscape of AI ethics guidelines. Nat. Mach. Intell. 2019, 1, 389–399. [Google Scholar] [CrossRef]
- Butt, J. A Strategic Roadmap for the Manufacturing Industry to Implement Industry 4.0. Designs 2020, 4, 11. [Google Scholar] [CrossRef]
- European Commission: DG Research and Innovation; Dixson-Declève, S.; Balland, P.-A.; Bria, F.; Charveriat, C.; Dunlop, K.; Giovannini, E.; Tataj, D.; Hidalgo, C.; Huang, A.; et al. Industry 5.0, A Transformative Vision for Europe—Governing Systemic Transformations Towards a Sustainable Industry; Publications Office of the European Union: Luxembourg, 2021. [Google Scholar]
- Misra, N.N.; Dixit, Y.; Al-Mallahi, A.; Bhullar, M.S.; Upadhyay, R.; Martynenko, A. IoT, Big Data, and Artificial Intelligence in Agriculture and Food Industry. IEEE Internet Things J. 2022, 9, 6305–6324. [Google Scholar] [CrossRef]
- Zheng, T.; Marco Ardolino, A.B.; Perona, M. The applications of Industry 4.0 technologies in manufacturing context: A systematic literature review. Int. J. Prod. Res. 2021, 59, 1922–1954. [Google Scholar] [CrossRef]
- Angelopoulos, A.; Michailidis, E.T.; Nomikos, N.; Trakadas, P.; Hatziefremidis, A.; Voliotis, S.; Zahariadis, T. Tackling Faults in the Industry 4.0 Era—A Survey of Machine-Learning Solutions and Key Aspects. Sensors 2020, 20, 109. [Google Scholar] [CrossRef]
- European Commission: DG Research and Innovation; Müller, J. Enabling Technologies for Industry 5.0—Results of a Workshop with Europe’s Technology Leaders; Publications Office: Luxembourg, 2020. [Google Scholar]
- European Commission. Skills Agenda for Europe: Promoting Skills and Talent in the European Union. 2020. Available online: https://research-and-innovation.ec.europa.eu/knowledge-publications-tools-and-data/publications/all-publications/enabling-technologies-industry-50_en (accessed on 7 July 2025).
- Claeys, G.; Tagliapietra, S.; Zachmann, G. How to Make the European Green Deal Work; JSTOR: New York, NY, USA, 2019; Volume 13. [Google Scholar]
- Puaschunder, J.M. The legal and international situation of AI, robotics and big data with attention to healthcare. In Report on Behalf of the European Parliament European Liberal Forum; European Liberal Forum: Brussels, Belgium, 2019. [Google Scholar]
- Marcus, J.S.; Martens, B.; Carugati, C.; Bucher, A.; Godlovitch, I.; The European Health Data Space. IPOL | Policy Department for Economic, Scientific and Quality of Life Policies, European Parliament Policy Department Studies. 2022. Available online: https://ssrn.com/abstract=4300393 (accessed on 7 July 2025).
- Nikolinakos, N.T. A European approach to excellence and trust: The 2020 white paper on artificial intelligence. In EU Policy and Legal Framework for Artificial Intelligence, Robotics and Related Technologies—The AI Act; Springer: Cham, Switzerland, 2023; pp. 211–280. [Google Scholar]
- Ai, H. High-level expert group on artificial intelligence. Ethics Guidel. Trust. AI 2019, 6. [Google Scholar]
- Mesarčík, M.; Solarova, S.; Podroužek, J.; Bieliková, M. Stance on the proposal for a regulation laying down harmonised rules on artificial intelligence—artificial intelligence act. In OSF Preprints; Center for Open Science: Charlottesville, VA, USA, 2022. [Google Scholar]
- Cannarsa, M. Ethics guidelines for trustworthy AI. In The Cambridge Handbook of Lawyering in the Digital Age; Cambridge University Press: Cambridge, UK, 2021; pp. 283–297. [Google Scholar]
- Ryan, M. The social and ethical impacts of artificial intelligence in agriculture: Mapping the agricultural AI literature. AI Soc. 2023, 38, 2473–2485. [Google Scholar] [CrossRef]
- Wang, J.; Chuqiao, X.; Zhang, J.; Zhong, R. Big data analytics for intelligent manufacturing systems: A review. J. Manuf. Syst. 2021, 62, 738–752. [Google Scholar] [CrossRef]
- Nahavandi, S. Industry 5.0—A human-centric solution. Sustainability 2019, 11, 4371. [Google Scholar] [CrossRef]
- Rikap, C. Same End by Different Means: Google, Amazon, Microsoft and Facebook’s Strategies to Dominate Artificial Intelligence. 2023. Available online: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4472222 (accessed on 26 February 2025).
- Habib ur Rehman, M.; Dirir, A.; Salah, K.; Damiani, E.; Svetinovic, D. TrustFed: A Framework for Fair and Trustworthy Cross-Device Federated Learning in IIoT. IEEE Trans. Ind. Inform. 2021, 8485–8494. [Google Scholar] [CrossRef]
- Vyhmeister, E.; Castane, G.; Östberg, P.O.; Thevenin, S. A responsible AI framework: Pipeline contextualisation. AI Ethics 2022, 3, 175–197. [Google Scholar] [CrossRef]
- Vyhmeister, E.; Gonzalez-Castane, G.; Östberg, P.O. Risk as a driver for AI framework development on manufacturing. AI Ethics 2022, 3, 155–174. [Google Scholar] [CrossRef]
- Wang, Q.; Liu, D.; Carmichael, M.; Aldini, S.; Lin, C.T. Computational Model of Robot Trust in Human Co-Worker for Physical Human-Robot Collaboration. IEEE Robot. Autom. Lett. 2022, 7, 3146–3153. [Google Scholar] [CrossRef]
- Wu, J.; Drew, S.; Dong, F.; Zhu, Z.; Zhou, J. Topology-aware Federated Learning in Edge Computing: A Comprehensive Survey. arXiv 2023. [Google Scholar] [CrossRef]
- Reddy, P.; Pham, V.; B, P.; Deepa, N.; Dev, K.; Gadekallu, T.; Ruby, R.; Liyanage, M. Industry 5.0: A Survey on Enabling Technologies and Potential Applications. J. Ind. Inf. Integr. 2021, 26, 100257. [Google Scholar] [CrossRef]
- Breque, M.; De Nul, L.; Petridis, A.; Directorate-General for Research and Innovation (European Commission). Industry 5.0: Towards a Sustainable, Human-Centric and Resilient European Industry; Publications Office of the European Union: Luxembourg, 2021; ISBN 978-92-76-25308-2. Available online: https://op.europa.eu/en/publication-detail/-/publication/468a892a-5097-11eb-b59f-01aa75ed71a1/ (accessed on 7 July 2025).
- Fahle, S.; Prinz, C.; Kuhlenkötter, B. Systematic review on machine learning (ML) methods for manufacturing processes—Identifying artificial intelligence (AI) methods for field application. Procedia CIRP 2020, 93, 413–418. [Google Scholar] [CrossRef]
- IBM. How Is AI Being Used in Manufacturing? AI in Manufacturing. Available online: https://www.ibm.com/think/topics/ai-in-manufacturing (accessed on 7 July 2025).
- Zonta, T.; da Costa, C.A.; da Rosa Righi, R.; de Lima, M.J.; da Trindade, E.S.; Li, G.P. Predictive maintenance in the Industry 4.0: A systematic literature review. Comput. Ind. Eng. 2020, 150, 106889. [Google Scholar] [CrossRef]
- Qarout, Y.; Begg, M.; Fearon, L.; Russell, D.; Pietrow, N.; Rahman, M.; McLeod, S.; Chakravorti, N.; Winter, T.; Fortune, J.; et al. Trustworthy AI Framework: A Comprehensive Review of AI Standards Policies and a Practical Guideline to Their Application in Manufacturing. 2024. Available online: https://www.the-mtc.org/api/sites/default/files/2024-07/Trustworthy%20AI%20Framework.pdf (accessed on 26 February 2025).
- European Commission; High-Level Expert Group on AI. Ethics Guidelines for Trustworthy AI. 2019. Available online: https://digital-strategy.ec.europa.eu/en/library/ethics-guidelines-trustworthy-ai (accessed on 29 January 2025).
- Reinhardt, K. Trust and trustworthiness in AI ethics. AI Ethics 2023, 3, 735–744. [Google Scholar] [CrossRef]
- Idamia, S.; Benseddik, H. Advancing Industry 5.0: An Extensive Review of AI Integration. In Industry 5.0 and Emerging Technologies: Transformation Through Technology and Innovations; Springer: Berlin/Heidelberg, Germany, 2024; pp. 3–21. [Google Scholar]
- Dimitrakopoulos, G.; Varga, P.; Gutt, T.; Schneider, G.; Ehm, H.; Hoess, A.; Tauber, M.; Karathanasopoulou, K.; Lackner, A.; Delsing, J. Industry 5.0: Research Areas and Challenges With Artificial Intelligence and Human Acceptance. IEEE Ind. Electron. Mag. 2024, 18, 43–54. [Google Scholar] [CrossRef]
- Keller, N. Cybersecurity Framework. 2013. Available online: https://www.nist.gov/cyberframework (accessed on 7 July 2025).
- International Organization for Standardization. ISO/IEC 27001:2022; Information Security, Cybersecurity and Privacy Protection—Information Security Management Systems—Requirements. ISO: Geneva, Switzerland, 2022. Available online: https://www.iso.org/standard/27001 (accessed on 16 June 2025).
- Wachter, S.; Mittelstadt, B.; Floridi, L. Why a right to explanation of automated decision-making does not exist in the general data protection regulation. Int. Data Priv. Law 2017, 7, 76–99. [Google Scholar] [CrossRef]
- Ala-Pietilä, P.; Bonnet, Y.; Bergmann, U.; Bielikova, M.; Bonefeld-Dahl, C.; Bauer, W.; Bouarfa, L.; Chatila, R.; Coeckelbergh, M.; Dignum, V.; et al. The Assessment List for Trustworthy Artificial Intelligence (ALTAI); European Commission: Brussels, Belgium, 2020. [Google Scholar]
- Martini, B.; Bellisario, D.; Coletti, P. Human-Centered and Sustainable Artificial Intelligence in Industry 5.0: Challenges and Perspectives. Sustainability 2024, 16, 5448. [Google Scholar] [CrossRef]
- Yusuf, S.O.; Durodola, R.L.; Ocran, G.; Abubakar, J.E.; Echere, A.Z.; Paul-Adeleye, A.H. Challenges and opportunities in AI and digital transformation for SMEs: A cross-continental perspective. World J. Adv. Res. Rev. 2024, 23, 668–678. [Google Scholar] [CrossRef]
- Wilson, H.J.; Daugherty, P.R. Collaborative intelligence: Humans and AI are joining forces. Harv. Bus. Rev. 2018, 96, 114–123. [Google Scholar]
- Thomas, M. Dangerous Risks of Artificial Intelligence. Volume 6. Available online: https://builtin.com/artificial-intelligence/risks-of-artificial-intelligence (accessed on 1 March 2021).
- Angwin, J.; Larson, J.; Mattu, S.; Kirchner, L. Machine bias. In Ethics of Data and Analytics; Auerbach Publications: Boca Raton, FL, USA, 2022; pp. 254–264. [Google Scholar]
- Zhang, M. Google Photos Tags Two African-Americans as Gorillas Through Facial Recognition Software. 2015. Available online: https://www.forbes.com/sites/mzhang/2015/07/01/google-photos-tags-two-african-americans-as-gorillas-through-facial-recognition-software/ (accessed on 7 July 2025).
- Dastin, J. Amazon scraps secret AI recruiting tool that showed bias against women. In Ethics of Data and Analytics; Auerbach Publications: Boca Raton, FL, USA, 2022; pp. 296–299. [Google Scholar]
- Kohli, P.; Chadha, A. Enabling pedestrian safety using computer vision techniques: A case study of the 2018 uber Inc. self-driving car crash. In Proceedings of the Future of Information and Communication Conference, San Francisco, CA, USA, 14–15 March 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 261–279. [Google Scholar]
- Fan, W.; Liu, J.; Zhu, S.; Pardalos, P.M. Investigating the impacting factors for the healthcare professionals to adopt artificial intelligence-based medical diagnosis support system (AIMDSS). Ann. Oper. Res. 2020, 294, 567–592. [Google Scholar] [CrossRef]
- Elliott, A. The Culture of AI: Everyday Life and the Digital Revolution; Routledge: London, UK, 2019. [Google Scholar]
- Marcus, G.; Davis, E. Rebooting AI: Building Artificial Intelligence We Can Trust; Vintage: London, UK, 2019. [Google Scholar]
- Floridi, L.; Cowls, J. A unified framework of five principles for AI in society. In Machine Learning and the City: Applications in Architecture and Urban Design; Wiley Online Library: Hoboken, NJ, USA, 2022; pp. 535–545. [Google Scholar]
- Floridi, L.; Cowls, J.; King, T.C.; Taddeo, M. How to design AI for social good: Seven essential factors. In Ethics, Governance, and Policies in Artificial Intelligence; Springer: Cham, Switzerland, 2021; pp. 125–151. [Google Scholar]
- Pichai, S. AI at Google: Our principles. Keyword 2018, 7, 1–3. [Google Scholar]
- UNI Global Union. Top 10 principles for ethical artificial intelligence. The FutureWorld ofWork, UNI Global Union: Nyon, Switzerland, 2017. Available online: https://uniglobalunion.org/report/10-principles-for-ethical-artificial-intelligence/ (accessed on 7 July 2025).
- Zeng, Y.; Lu, E.; Huangfu, C. Linking artificial intelligence principles. arXiv 2018, arXiv:1812.04814. [Google Scholar]
- Hagendorff, T. The ethics of AI ethics: An evaluation of guidelines. Minds Mach. 2020, 30, 99–120. [Google Scholar] [CrossRef]
- Kaur, D.; Uslu, S.; Durresi, A. Requirements for trustworthy artificial intelligence–a review. In Proceedings of the Advances in Networked-Based Information Systems: The 23rd International Conference on Network-Based Information Systems (NBiS-2020) 23, Victoria, BC, Canada, 31 August–2 September 2020; Springer: Berlin/Heidelberg, Germany, 2021; pp. 105–115. [Google Scholar]
- Kumar, A.; Braud, T.; Tarkoma, S.; Hui, P. Trustworthy AI in the age of pervasive computing and big data. In Proceedings of the 2020 IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops), Austin, TX, USA, 23–27 March 2020; pp. 1–6. [Google Scholar]
- Smuha, N.A. The EU approach to ethics guidelines for trustworthy artificial intelligence. Comput. Law Rev. Int. 2019, 20, 97–106. [Google Scholar] [CrossRef]
- Daugherty, P.R.; Wilson, H.J. Human+ Machine: Reimagining Work in the Age of AI; Harvard Business Press: Brighton, MA, USA, 2018. [Google Scholar]
- Goodman, B.; Flaxman, S. European Union regulations on algorithmic decision-making and a “right to explanation”. AI Mag. 2017, 38, 50–57. [Google Scholar] [CrossRef]
- Dignum, V. Responsible Artificial Intelligence: Designing AI for Human Values. ITU J. ICT Discov. 2018, 1, 1–8. Available online: https://www.itu.int/dms_pub/itu-s/opb/journal/S-JOURNAL-ICTF.VOL1-2018-1-P01-PDF-E.pdf (accessed on 5 March 2025).
- Shin, D.; Park, Y.J. Role of fairness, accountability, and transparency in algorithmic affordance. Comput. Hum. Behav. 2019, 98, 277–284. [Google Scholar] [CrossRef]
- Flores, A.W.; Bechtel, K.; Lowenkamp, C.T. False positives, false negatives, and false analyses: A rejoinder to machine bias: There’s software used across the country to predict future criminals. and it’s biased against blacks. Fed. Probat. 2016, 80, 38. [Google Scholar]
- Wieringa, M. What to account for when accounting for algorithms: A systematic literature review on algorithmic accountability. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, Barcelona, Spain, 27–30 January 2020; pp. 1–18. [Google Scholar]
- Islam, M.R.; Ahmed, M.U.; Barua, S.; Begum, S. A systematic review of explainable artificial intelligence in terms of different application domains and tasks. Appl. Sci. 2022, 12, 1353. [Google Scholar] [CrossRef]
- Chou, Y.L.; Moreira, C.; Bruza, P.; Ouyang, C.; Jorge, J. Counterfactuals and causability in explainable artificial intelligence: Theory, algorithms, and applications. Inf. Fusion 2022, 81, 59–83. [Google Scholar] [CrossRef]
- Ahmed, I.; Jeon, G.; Piccialli, F. From artificial intelligence to explainable artificial intelligence in industry 4.0: A survey on what, how, and where. IEEE Trans. Ind. Inform. 2022, 18, 5031–5042. [Google Scholar] [CrossRef]
- Khosravi, H.; Shum, S.B.; Chen, G.; Conati, C.; Tsai, Y.S.; Kay, J.; Knight, S.; Martinez-Maldonado, R.; Sadiq, S.; Gašević, D. Explainable artificial intelligence in education. Comput. Educ. Artif. Intell. 2022, 3, 100074. [Google Scholar] [CrossRef]
- Rawal, A.; McCoy, J.; Rawat, D.B.; Sadler, B.M.; Amant, R.S. Recent advances in trustworthy explainable artificial intelligence: Status, challenges, and perspectives. IEEE Trans. Artif. Intell. 2021, 3, 852–866. [Google Scholar] [CrossRef]
- Albahri, A.S.; Duhaim, A.M.; Fadhel, M.A.; Alnoor, A.; Baqer, N.S.; Alzubaidi, L.; Albahri, O.S.; Alamoodi, A.H.; Bai, J.; Salhi, A.; et al. A systematic review of trustworthy and explainable artificial intelligence in healthcare: Assessment of quality, bias risk, and data fusion. Inf. Fusion 2023, 96, 156–191. [Google Scholar] [CrossRef]
- Albahri, A.; Al-Qaysi, Z.; Alzubaidi, L.; Alnoor, A.; Albahri, O.; Alamoodi, A.; Bakar, A.A. A Systematic Review of Using Deep Learning Technology in the Steady-State Visually Evoked Potential-Based Brain-Computer Interface Applications: Current Trends and Future Trust Methodology. Int. J. Telemed. Appl. 2023, 2023, 7741735. [Google Scholar] [CrossRef]
- Velmurugan, M.; Ouyang, C.; Moreira, C.; Sindhgatta, R. Evaluating stability of post-hoc explanations for business process predictions. In Proceedings of the International Conference on Service-Oriented Computing, Dubai, United Arab Emirates, 22–25 November 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 49–64. [Google Scholar]
- Selbst, A.; Powles, J. “Meaningful information” and the right to explanation. In Proceedings of the Conference on Fairness, Accountability and Transparency, New York, NY, USA, 23–24 February 2018; p. 48. [Google Scholar]
- Velmurugan, M.; Ouyang, C.; Moreira, C.; Sindhgatta, R. Evaluating fidelity of explainable methods for predictive process analytics. In Proceedings of the International Conference on Advanced Information Systems Engineering, Melbourne, VIC, Australia, 28 June–2 July 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 64–72. [Google Scholar]
- Sreedharan, S.; Srivastava, S.; Kambhampati, S. Using state abstractions to compute personalized contrastive explanations for AI agent behavior. Artif. Intell. 2021, 301, 103570. [Google Scholar] [CrossRef]
- Shin, D. The effects of explainability and causability on perception, trust, and acceptance: Implications for explainable AI. Int. J. Hum.-Comput. Stud. 2021, 146, 102551. [Google Scholar] [CrossRef]
- Lim, W.X.; Chen, Z.; Ahmed, A. The adoption of deep learning interpretability techniques on diabetic retinopathy analysis: A review. Med Biol. Eng. Comput. 2022, 60, 633–642. [Google Scholar] [CrossRef] [PubMed]
- Huang, Y.; Chen, D.; Zhao, W.; Lv, Y.; Wang, S. Deep patch learning algorithms with high interpretability for regression problems. Int. J. Intell. Syst. 2022, 37, 8239–8276. [Google Scholar] [CrossRef]
- Yang, C.; Rangarajan, A.; Ranka, S. Global model interpretation via recursive partitioning. In Proceedings of the 2018 IEEE 20th International Conference on High Performance Computing and Communications; IEEE 16th International Conference on Smart City; IEEE 4th International Conference on Data Science and Systems (HPCC/SmartCity/DSS), Exeter, UK, 28–30 June 2018; pp. 1563–1570. [Google Scholar]
- Moreira, C.; Chou, Y.L.; Velmurugan, M.; Ouyang, C.; Sindhgatta, R.; Bruza, P. LINDA-BN: An interpretable probabilistic approach for demystifying black-box predictive models. Decis. Support Syst. 2021, 150, 113561. [Google Scholar] [CrossRef]
- Lyu, D.; Yang, F.; Kwon, H.; Dong, W.; Yilmaz, L.; Liu, B. Tdm: Trustworthy decision-making via interpretability enhancement. IEEE Trans. Emerg. Top. Comput. Intell. 2021, 6, 450–461. [Google Scholar] [CrossRef]
- Reed, C. How should we regulate artificial intelligence? Philos. Trans. R. Soc. A: Math. Phys. Eng. Sci. 2018, 376, 20170360. [Google Scholar] [CrossRef]
- Wickramanayake, B.; He, Z.; Ouyang, C.; Moreira, C.; Xu, Y.; Sindhgatta, R. Building interpretable models for business process prediction using shared and specialised attention mechanisms. Knowl.-Based Syst. 2022, 248, 108773. [Google Scholar] [CrossRef]
- Thiebes, S.; Lins, S.; Sunyaev, A. Trustworthy artificial intelligence. Electron. Mark. 2021, 31, 447–464. [Google Scholar] [CrossRef]
- Sindhgatta, R.; Ouyang, C.; Moreira, C. Exploring interpretability for predictive process analytics. In Proceedings of the International Conference on Service-Oriented Computing, Dubai, United Arab Emirates, 14–17 December 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 439–447. [Google Scholar]
- Bhatt, U.; Xiang, A.; Sharma, S.; Weller, A.; Taly, A.; Jia, Y.; Ghosh, J.; Puri, R.; Moura, J.M.; Eckersley, P. Explainable machine learning in deployment. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, Barcelona, Spain, 27–30 January 2020; pp. 648–657. [Google Scholar]
- IBM. AI Explainability 360. 2025. Available online: https://research.ibm.com/blog/ai-explainability-360 (accessed on 18 March 2025).
- Ribeiro, M.T.; Singh, S.; Guestrin, C. Why should I trust you? Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; ACM: New York, NY, USA, 2016; pp. 1135–1144. Available online: https://dl.acm.org/doi/10.1145/2939672.2939778 (accessed on 7 July 2025).
- Lundberg, S.M.; Lee, S.I. A unified approach to interpreting model predictions. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30, pp. 4765–4774. Available online: https://arxiv.org/abs/1705.07874 (accessed on 7 July 2025).
- Chen, J.; Kanan, C.; Fowlkes, C.C. XAITK Saliency: An Open Source Explainable AI Toolkit for Visual Saliency. Available online: https://ojs.aaai.org/index.php/AAAI/article/view/26871 (accessed on 18 March 2025).
- Yin, H.; Xie, K.; Zhou, Y.; Liu, J.; Zhang, Z.; Liu, P.; Wang, W.; Zhao, Y. Fidelity in Explanation Methods: A Comprehensive Study. arXiv 2022, arXiv:2202.06861. [Google Scholar]
- Ieracitano, C.; Mammone, N.; Hussain, A.; Morabito, F.C. A novel explainable machine learning approach for EEG-based brain-computer interface systems. Neural Comput. Appl. 2022, 34, 11347–11360. [Google Scholar] [CrossRef]
- Ras, G.; Xie, N.; Van Gerven, M.; Doran, D. Explainable deep learning: A field guide for the uninitiated. J. Artif. Intell. Res. 2022, 73, 329–396. [Google Scholar] [CrossRef]
- De Waal, A.; Joubert, J.W. Explainable Bayesian networks applied to transport vulnerability. Expert Syst. Appl. 2022, 209. [Google Scholar] [CrossRef]
- Mao, C.; Lin, R.; Towey, D.; Wang, W.; Chen, J.; He, Q. Trustworthiness prediction of cloud services based on selective neural network ensemble learning. Expert Syst. Appl. 2021, 168, 114390. [Google Scholar] [CrossRef]
- Srinivasan, R.; González, B.S.M. The role of empathy for artificial intelligence accountability. J. Responsible Technol. 2022, 9, 100021. [Google Scholar] [CrossRef]
- Choudhury, A.; Asan, O. Impact of accountability, training, and human factors on the use of artificial intelligence in healthcare: Exploring the perceptions of healthcare practitioners in the US. Hum. Factors Healthc. 2022, 2, 100021. [Google Scholar] [CrossRef]
- Fong, S.; Dey, N.; Joshi, A. ICT analysis and applications. In Proceedings of the ICT4SD, Goa, India, 23–24 July 2020; Volume 2. [Google Scholar]
- Cruz, B.S.; de Oliveira Dias, M. Crashed boeing 737-max: Fatalities or malpractice. GSJ 2020, 8, 2615–2624. [Google Scholar]
- Schwartz, R.; Schwartz, R.; Vassilev, A.; Greene, K.; Perine, L.; Burt, A.; Hall, P. Towards a Standard for Identifying and Managing Bias in Artificial Intelligence; US Department of Commerce; National Institute of Standards and Technology: Gaithersburg, MD, USA, 2022; Volume 3. [Google Scholar]
- Gevaert, C.M.; Carman, M.; Rosman, B.; Georgiadou, Y.; Soden, R. Fairness and accountability of AI in disaster risk management: Opportunities and challenges. Patterns 2021, 2, 100363. [Google Scholar] [CrossRef]
- Königstorfer, F.; Thalmann, S. AI Documentation: A path to accountability. J. Responsible Technol. 2022, 11, 100043. [Google Scholar] [CrossRef]
- Rahwan, I.; Cebrian, M.; Obradovich, N.; Bongard, J.; Bonnefon, J.F.; Breazeal, C.; Crandall, J.W.; Christakis, N.A.; Couzin, I.D.; Jackson, M.O.; et al. Machine behaviour. Nature 2019, 568, 477–486. [Google Scholar] [CrossRef]
- Morley, J.; Floridi, L.; Kinsey, L.; Elhalal, A. From what to how: An initial review of publicly available AI ethics tools, methods and research to translate principles into practices. Sci. Eng. Ethics 2020, 26, 2141–2168. [Google Scholar] [CrossRef] [PubMed]
- Guidotti, R.; Monreale, A.; Ruggieri, S.; Turini, F.; Giannotti, F.; Pedreschi, D. A survey of methods for explaining black box models. ACM Comput. Surv. (CSUR) 2018, 51, 1–42. [Google Scholar] [CrossRef]
- Omeiza, D.; Web, H.; Jirotka, M.; Kunze, L. Towards accountability: Providing intelligible explanations in autonomous driving. In Proceedings of the 2021 IEEE Intelligent Vehicles Symposium (IV), Nagoya, Japan, 11–17 July 2021; pp. 231–237. [Google Scholar]
- Kadambi, A. Achieving fairness in medical devices. Science 2021, 372, 30–31. [Google Scholar] [CrossRef] [PubMed]
- Mehrabi, N.; Morstatter, F.; Saxena, N.; Lerman, K.; Galstyan, A. A survey on bias and fairness in machine learning. ACM Comput. Surv. (CSUR) 2021, 54, 1–35. [Google Scholar] [CrossRef]
- Madaio, M.; Egede, L.; Subramonyam, H.; Wortman Vaughan, J.; Wallach, H. Assessing the fairness of ai systems: Ai practitioners’ processes, challenges, and needs for support. In Proceedings of the ACM on Human-Computer Interaction, New Orleans, LA, USA, 29 April–5 May 2022; Volume 6, pp. 1–26. [Google Scholar]
- Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 2019, 1, 206–215. [Google Scholar] [CrossRef]
- von Zahn, M.; Feuerriegel, S.; Kuehl, N. The cost of fairness in AI: Evidence from e-commerce. Bus. Inf. Syst. Eng. 2022, 64, 335–348. [Google Scholar] [CrossRef]
- Pastaltzidis, I.; Dimitriou, N.; Quezada-Tavarez, K.; Aidinlis, S.; Marquenie, T.; Gurzawska, A.; Tzovaras, D. Data augmentation for fairness-aware machine learning: Preventing algorithmic bias in law enforcement systems. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, Seoul, Republic of Korea, 21–24 June 2022; pp. 2302–2314. [Google Scholar]
- Chouldechova, A.; Benavides-Prado, D.; Fialko, O.; Vaithianathan, R. A case study of algorithm-assisted decision making in child maltreatment hotline screening decisions. In Proceedings of the Conference on Fairness, Accountability and Transparency, New York, NY, USA, 23–24 February 2018; pp. 134–148. [Google Scholar]
- Dwork, C.; Hardt, M.; Pitassi, T.; Reingold, O.; Zemel, R. Fairness through awareness. In Proceedings of the 3rd Innovations in Theoretical Computer Science Conference, Cambridge, MA, USA, 8–10 January 2012; pp. 214–226. [Google Scholar]
- Oneto, L.; Chiappa, S. Recent Trends in Learning from Data; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
- Kusner, M.J.; Loftus, J.; Russell, C.; Silva, R. Counterfactual fairness. Adv. Neural Inf. Process. Syst. 2017, 30, 1–11. [Google Scholar]
- Grari, V.; Lamprier, S.; Detyniecki, M. Adversarial learning for counterfactual fairness. Mach. Learn. 2023, 112, 741–763. [Google Scholar] [CrossRef]
- Santos, F.P.; Santos, F.C.; Paiva, A.; Pacheco, J.M. Evolutionary dynamics of group fairness. J. Theor. Biol. 2015, 378, 96–102. [Google Scholar] [CrossRef]
- Khalili, M.M.; Zhang, X.; Abroshan, M. Fair sequential selection using supervised learning models. Adv. Neural Inf. Process. Syst. 2021, 34, 28144–28155. [Google Scholar]
- Zheng, Y.; Wang, S.; Zhao, J. Equality of opportunity in travel behavior prediction with deep neural networks and discrete choice models. Transp. Res. Part C Emerg. Technol. 2021, 132, 103410. [Google Scholar] [CrossRef]
- Besse, P.; del Barrio, E.; Gordaliza, P.; Loubes, J.M.; Risser, L. A survey of bias in machine learning through the prism of statistical parity. Am. Stat. 2022, 76, 188–198. [Google Scholar] [CrossRef]
- Feuerriegel, S.; Dolata, M.; Schwabe, G. Fair AI: Challenges and opportunities. Bus. Inf. Syst. Eng. 2020, 62, 379–384. [Google Scholar] [CrossRef]
- Chen, Y.; Huerta, E.; Duarte, J.; Harris, P.; Katz, D.S.; Neubauer, M.S.; Diaz, D.; Mokhtar, F.; Kansal, R.; Park, S.E.; et al. A FAIR and AI-ready Higgs boson decay dataset. Sci. Data 2022, 9, 31. [Google Scholar] [CrossRef]
- AI Fairness 360. AI Fairness 360. 2025. Available online: https://research.ibm.com/blog/ai-fairness-360 (accessed on 7 July 2025).
- Fairlearn. Fairlearn. 2025. Available online: https://fairlearn.org/ (accessed on 13 March 2025).
- What-If Tool. What-If Tool. 2025. Available online: https://pair-code.github.io/what-if-tool/ (accessed on 7 July 2025).
- Aequitas. Aequitas. 2025. Available online: https://arxiv.org/abs/1811.05577 (accessed on 13 March 2025).
- Google Research. ML Fairness Gym: A Tool for Exploring Long-Term Impacts of Machine Learning Systems. 2019. Available online: https://research.google/blog/ml-fairness-gym-a-tool-for-exploring-long-term-impacts-of-machine-learning-systems/ (accessed on 7 July 2025).
- AI Now Institute. Algorithmic Impact Assessments: A Practical Framework for Public Agency Accountability. 2022. Available online: https://ainowinstitute.org/publications/algorithmic-impact-assessments-report-2 (accessed on 7 July 2025).
- UK Government. Data Ethics Framework 2020. 2020. Available online: https://assets.publishing.service.gov.uk/media/5f74a4958fa8f5188dad0e99/Data_Ethics_Framework_2020.pdf (accessed on 7 July 2025).
- Swiss Re. A Journey into Responsible AI. 2023. Available online: https://www.swissre.com/institute/ (accessed on 7 July 2025).
- Hauer, M.P.; Adler, R.; Zweig, K. Assuring fairness of algorithmic decision making. In Proceedings of the 2021 IEEE International Conference on Software Testing, Verification and Validation Workshops (ICSTW), Porto de Galinhas, Brazil, 12–16 April 2021; pp. 110–113. [Google Scholar]
- Amodei, D.; Olah, C.; Steinhardt, J.; Christiano, P.; Schulman, J.; Mané, D. Concrete problems in AI safety. arXiv 2016, arXiv:1606.06565. [Google Scholar]
- Tan, X.; Xu, K.; Cao, Y.; Zhang, Y.; Ma, L.; Lau, R.W. Night-time scene parsing with a large real dataset. IEEE Trans. Image Process. 2021, 30, 9085–9098. [Google Scholar] [CrossRef]
- Zhang, M.; Zhang, Y.; Zhang, L.; Liu, C.; Khurshid, S. Deeproad: Gan-based metamorphic testing and input validation framework for autonomous driving systems. In Proceedings of the 33rd ACM/IEEE International Conference on Automated Software Engineering, Montpellier, France, 3–7 September 2018; pp. 132–142. [Google Scholar]
- Akhtar, N.; Mian, A. Threat of adversarial attacks on deep learning in computer vision: A survey. IEEE Access 2018, 6, 14410–14430. [Google Scholar] [CrossRef]
- Chakraborty, A.; Alam, M.; Dey, V.; Chattopadhyay, A.; Mukhopadhyay, D. Adversarial attacks and defences: A survey. arXiv 2018, arXiv:1810.00069. [Google Scholar] [CrossRef]
- Li, Y.; Jiang, Y.; Li, Z.; Xia, S.T. Backdoor learning: A survey. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 5–22. [Google Scholar] [CrossRef]
- Silva, S.H.; Najafirad, P. Opportunities and challenges in deep learning adversarial robustness: A survey. arXiv 2020, arXiv:2007.00753. [Google Scholar]
- Yuan, X.; He, P.; Zhu, Q.; Li, X. Adversarial examples: Attacks and defenses for deep learning. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 2805–2824. [Google Scholar] [CrossRef] [PubMed]
- Vorobeychik, Y.; Kantarcioglu, M. Adversarial Machine Learning; Morgan & Claypool Publishers: San Rafael, CA, USA, 2018. [Google Scholar]
- Tong, L.; Li, B.; Hajaj, C.; Xiao, C.; Zhang, N.; Vorobeychik, Y. Improving robustness of {ML} classifiers against realizable evasion attacks using conserved features. In Proceedings of the 28th USENIX Security Symposium (USENIX Security 19), Santa Clara, CA, USA, 14–16 August 2019; pp. 285–302. [Google Scholar]
- Wu, T.; Tong, L.; Vorobeychik, Y. Defending against physically realizable attacks on image classification. arXiv 2019, arXiv:1909.09552. [Google Scholar]
- Tramèr, F.; Zhang, F.; Juels, A.; Reiter, M.K.; Ristenpart, T. Stealing machine learning models via prediction {APIs}. In Proceedings of the 25th USENIX security symposium (USENIX Security 16), Austin, TX, USA, 10–12 August 2016; pp. 601–618. [Google Scholar]
- Ramachandra, R.; Busch, C. Presentation attack detection methods for face recognition systems: A comprehensive survey. ACM Comput. Surv. (CSUR) 2017, 50, 1–37. [Google Scholar] [CrossRef]
- Machado, G.R.; Silva, E.; Goldschmidt, R.R. Adversarial machine learning in image classification: A survey toward the defender’s perspective. ACM Comput. Surv. (CSUR) 2021, 55, 1–38. [Google Scholar] [CrossRef]
- Exforsys. What Is Monkey Testing. 2011. Available online: http://www.exforsys.com/tutorials/testing-types/monkey-testing.html (accessed on 7 July 2025).
- Ma, L.; Juefei-Xu, F.; Zhang, F.; Sun, J.; Xue, M.; Li, B.; Chen, C.; Su, T.; Li, L.; Liu, Y.; et al. Deepgauge: Multi-granularity testing criteria for deep learning systems. In Proceedings of the 33rd ACM/IEEE International Conference on Automated Software Engineering, Montpellier, France, 3–7 September 2018; pp. 120–131. [Google Scholar]
- Pei, K.; Cao, Y.; Yang, J.; Jana, S. Deepxplore: Automated whitebox testing of deep learning systems. In Proceedings of the 26th Symposium on Operating Systems Principles, Shanghai, China, 28–31 October 2017; pp. 1–18. [Google Scholar]
- Carlini, N.; Wagner, D. Towards evaluating the robustness of neural networks. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–26 May 2017; pp. 39–57. [Google Scholar]
- Su, D.; Zhang, H.; Chen, H.; Yi, J.; Chen, P.Y.; Gao, Y. Is robustness the cost of accuracy?–a comprehensive study on the robustness of 18 deep image classification models. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 631–648. [Google Scholar]
- Boopathy, A.; Weng, T.W.; Chen, P.Y.; Liu, S.; Daniel, L. Cnn-cert: An efficient framework for certifying robustness of convolutional neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 3240–3247. [Google Scholar]
- Zhang, H.; Weng, T.W.; Chen, P.Y.; Hsieh, C.J.; Daniel, L. Efficient neural network robustness certification with general activation functions. Adv. Neural Inf. Process. Syst. 2018, 31. [Google Scholar]
- Nicolae, M.I.; Sinn, M.; Tran, M.N.; Buesser, B.; Rawat, A.; Wistuba, M.; Zantedeschi, V.; Baracaldo, N.; Chen, B.; Ludwig, H.; et al. Adversarial Robustness Toolbox v1.0.0. arXiv 2018, arXiv:1807.01069. [Google Scholar]
- Papernot, N.; Faghri, F.; Carlini, N.; Goodfellow, I.; Feinman, R.; Kurakin, A.; Xie, C.; Sharma, Y.; Brown, T.; Roy, A.; et al. Technical Report on the CleverHans v2.1.0 Adversarial Examples Library. arXiv 2018, arXiv:1610.00768. [Google Scholar]
- Rauber, J.; Brendel, W.; Bethge, M. Foolbox: A Python toolbox to benchmark the robustness of machine learning models. arXiv 2018, arXiv:1707.04131. [Google Scholar]
- Soklaski, R.; Goodwin, J.; Brown, O.; Yee, M.; Matterer, J. Tools and Practices for Responsible AI Engineering. arXiv 2022, arXiv:2201.05647. [Google Scholar]
- Li, Y.; Jin, W.; Xu, H.; Tang, J. DeepRobust: A PyTorch Library for Adversarial Attacks and Defenses. arXiv 2020, arXiv:2005.06149. [Google Scholar]
- Chen, J.; Su, Y.; Song, Z.; Wang, B.; Tang, R.; Zhao, S.; Yang, Y.; Zhang, J.; Kumar, M.; Xu, H. Model-agnostic meta-learning for fast adaptation of deep networks. arXiv 2019, arXiv:1902.07623. [Google Scholar]
- Radford, A.; Kim, J.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, K.; Clark, J.; Krueger, G.; Chen, M.; Sutskever, I. Learning Transferable Visual Models From Natural Language Supervision. arXiv 2021, arXiv:2101.04840. [Google Scholar]
- Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; Shinn, E.; Ibarz, J.; Wu, L.; et al. Language Models are Unsupervised Multitask Learners. arXiv 2019, arXiv:1901.08573. [Google Scholar]
- Liu, Z.; Lin, Y.; Cao, Y.; Zhang, F.; Zhang, Z.; Wang, X.; Zhang, T.; Chen, X.; Zeng, R.; Yu, L. Data-efficient image transformers for downstream tasks. arXiv 2022, arXiv:2203.05154. [Google Scholar]
- Yu, H.; Zheng, F.; Hu, H.; Gu, S.; Zhang, S.; Li, H. Learning to Learn with Conditional Class Dependencies. arXiv 2020, arXiv:2010.09670. [Google Scholar]
- He, M.; Li, Z.; Liu, C.; Shi, D.; Tan, Z. Deployment of Artificial Intelligence in Real-World Practice: Opportunity and Challenge. Asia-Pac. J. Ophthalmol. 2020, 9, 299–307. [Google Scholar] [CrossRef]
- Peres, R.S.; Jia, X.; Lee, J.; Sun, K.; Colombo, A.W.; Barata, J. Industrial artificial intelligence in industry 4.0-systematic review, challenges and outlook. IEEE Access 2020, 8, 220121–220139. [Google Scholar] [CrossRef]
- Ahsan, M.; Hon, S.T.; Albarbar, A. Development of Novel Big Data Analytics Framework for Smart Clothing. IEEE Access 2020, 8, 146376–146394. [Google Scholar] [CrossRef]
- Wang, P.; Wang, K.; Wang, D.; Liu, H. The Impact of Manufacturing Transformation in Digital Economy Under Artificial Intelligence. IEEE Access 2024, 12, 63417–63424. [Google Scholar] [CrossRef]
- Colombi, L.; Gilli, A.; Dahdal, S.; Boleac, I.; Tortonesi, M.; Stefanelli, C.; Vignoli, M. A Machine Learning Operations Platform for Streamlined Model Serving in Industry 5.0. In Proceedings of the NOMS 2024-2024 IEEE Network Operations and Management Symposium, Seoul, Republic of Korea, 6–10 May 2024; pp. 1–6. [Google Scholar] [CrossRef]
- Yu, W.; Liu, Y.; Dillon, T.; Rahayu, W.; Mostafa, F. An Integrated Framework for Health State Monitoring in a Smart Factory Employing IoT and Big Data Techniques. IEEE Internet Things J. 2022, 9, 2443–2454. [Google Scholar] [CrossRef]
- Hariharakrishnan, J.; Mohanavalli, S.; Srividya; Sundhara Kumar, K.B. Survey of pre-processing techniques for mining big data. In Proceedings of the 2017 International Conference on Computer, Communication and Signal Processing (ICCCSP), Chennai, India, 10–11 January 2017; pp. 1–5. [Google Scholar] [CrossRef]
- Shahriar, S.; Allana, S.; Hazratifard, S.M.; Dara, R. A Survey of Privacy Risks and Mitigation Strategies in the Artificial Intelligence Life Cycle. IEEE Access 2023, 11, 61829–61854. [Google Scholar] [CrossRef]
- Kemnitz, J.; Weissenfeld, A.; Schoeffl, L.; Stiftinger, A.; Rechberger, D.; Prangl, B.; Kaufmann, T.; Hiessl, T.; Holly, S.; Heistracher, C.; et al. An Edge Deployment Framework to Scale AI in Industrial Applications. In Proceedings of the 2023 IEEE 7th International Conference on Fog and Edge Computing (ICFEC), Bangalore, India, 1–4 May 2023; pp. 24–32. [Google Scholar] [CrossRef]
Toolkit | Pros | Cons | Use Cases |
---|---|---|---|
AI Explainability 360 (AIX360) [101] | Comprehensive set of algorithms covering various explanation dimensions. | Steep learning curve due to broad feature set. | Understanding and interpreting predictions from complex machine learning models. |
Supports multiple data types, enhancing versatility. | Some algorithms may require substantial computational resources. | Ensuring transparency and trustworthiness in AI-driven decision-making processes. | |
Developed by IBM, ensuring reliability and community support. | |||
LIME (Local Interpretable Model-agnostic Explanations) [102] | Provides local explanations by approximating complex models with simpler ones. | Local explanations may not fully capture global model behaviour. | Explaining individual predictions in domains like healthcare and finance. |
Model-agnostic; applicable to various machine learning models. | Performance can be affected by noisy data, leading to inconsistent interpretations. | Assisting in debugging and improving model performance by understanding specific decision paths. | |
Enhances user trust through understandable explanations. | |||
SHAP (Shapley Additive Explanations) [103] | Offers both global and local explanations, providing a comprehensive view of model behaviour. | Computationally intensive, especially with large datasets and complex models. | Assessing feature importance in predictive models. |
Based on solid game-theoretic foundations, ensuring consistent and fair feature importance values. | Requires careful handling of feature interactions to avoid misleading interpretations. | Enhancing model transparency in sectors like finance and healthcare by elucidating the impact of individual features on predictions. | |
XAITK (Explainable AI Toolkit) [104] | Provides a suite of tools for analysing and understanding complex machine learning models. | May require integration efforts with existing workflows. | Analysing and interpreting complex machine learning models across various domains. |
Includes analytics tools and methods for interpreting models, supporting various explanation techniques. | Documentation and community support might be less extensive compared to more established toolkits. | Supporting research and development in AI transparency and accountability. | |
Quantus [105] | Offers a collection of evaluation metrics for assessing the quality of explanations. | Primarily focused on evaluating explanations rather than generating them. | Evaluating the effectiveness of different explainability methods. |
Facilitates the comparison of different explanation methods. | May require additional tools or methods for generating explanations. | Assisting researchers in selecting appropriate explanation techniques for their models. | |
Aids in identifying the most effective explanation techniques for specific models. |
Toolkit Name | Advantages | Disadvantages | Use Cases |
---|---|---|---|
IBM AI Fairness 360 (AIF360) [138] | Open-source toolkit for bias detection and mitigation. | Focuses primarily on fairness, may not address other ethical AI dimensions. | Evaluating and mitigating bias in hiring, loan approvals, and other decision-making systems. |
Provides a comprehensive set of metrics and algorithms for bias mitigation. | Can be complex to implement in production systems. | Improving fairness in public services, education, and financial systems. | |
Highly customisable and suitable for large-scale applications. | May require significant computational resources. | Ensuring fairness in automated systems such as hiring and loan approvals. | |
Microsoft FairLearn [139] | Provides fairness assessment and bias mitigation tools. | Requires technical expertise for implementation. | Fairness assessment in machine learning models, especially in enterprise settings. |
Supports multiple fairness metrics and mitigation strategies. | Limited documentation for non-technical users. | Ensuring fairness in predictive models used in hiring, finance, and healthcare. | |
Enables easy integration with scikit-learn models. | Limited flexibility in non-enterprise applications. | Used for bias mitigation in sensitive decision-making systems. | |
Google What-If Tool [140] | Interactive tool for exploring model predictions and fairness. | Limited to TensorFlow models. | Analyzing fairness in predictive models, such as fraud detection and medical diagnoses. |
Allows visual exploration of bias and fairness across model predictions. | Not suitable for non-TensorFlow based models. | Used for exploring fairness in machine learning models for fraud detection and healthcare applications. | |
Easy-to-use interface for non-technical stakeholders. | Can be time-consuming for large datasets. | Assessing fairness in machine learning applications for public policy and social justice. | |
Aequitas [141] | Focuses on bias and fairness in decision-making systems. | Limited scope for other ethical AI principles. | Bias detection in public policy and social justice applications. |
Designed to be used with real-world decision-making data. | Does not provide technical tools for bias mitigation. | Used for fairness assessments in hiring, criminal justice, and education systems. | |
Strong documentation and user support. | Not highly customisable for complex AI systems. | Ensuring fairness in decision-making systems related to government and social issues. | |
ML Fairness Gym [142] | Simulates long-term impacts of fairness interventions. | Requires expertise in simulation modeling. | Evaluating fairness in dynamic systems like credit scoring and hiring. |
Provides an interactive environment for testing fairness interventions. | High computational cost for large simulations. | Used for understanding fairness in long-term decision-making systems. | |
Supports a range of fairness interventions for experimentation. | Simulation results may not generalise to all real-world scenarios. | Assessing fairness in evolving applications such as credit scoring and insurance. |
Toolkit Name | Advantages | Disadvantages | Use Cases |
---|---|---|---|
AINow Algorithmic Impact Assessment Toolkit. [143] | Engages stakeholders in assessing fairness and ethical implications. | Limited technical tools for bias mitigation. | Assessing fairness in community-focused AI applications. |
Provides a structured framework for ethical AI assessments. | Limited to ethical assessments rather than technical solutions. | Used for impact assessments in AI systems affecting marginalised communities. | |
Emphasises the importance of human oversight in AI decision-making. | Lacks deep technical fairness metrics and algorithms. | Assisting in responsible AI implementation in public services. | |
DotEveryone Consequence Scanning Toolkit | Open-source; minimal resources required; focuses on societal impacts. | Requires a strong facilitator, which may be a barrier for SMEs. | Conceptualising AI systems with societal and environmental considerations. |
Helps in identifying the societal consequences of AI deployments. | Primarily focused on societal impact rather than technical fairness. | Used for ethical evaluations of AI systems in public policy, education, and healthcare. | |
Allows for early detection of ethical and social risks in AI systems. | Lacks comprehensive tools for technical fairness evaluation. | Ensuring societal considerations are addressed in AI-based systems. | |
Data Ethics Impact Assessment [144] | Integrates data ethics into AI development processes. | Limited to ethical assessments, not technical bias mitigation. | Ethical assessments in AI systems for public and private sectors. |
Focuses on the ethical implications of data usage and AI models. | May not address technical fairness challenges directly. | Used in ensuring data usage complies with ethical standards in sectors like healthcare and government. | |
Provides an important framework for responsible AI development. | Does not provide tools for bias correction or mitigation. | Ensuring ethical and fair data usage in AI systems for social justice. | |
Veritas Fairness Assessment Methodology. [145] | Developed for financial systems; focuses on fairness and transparency. | Limited adoption outside the finance industry. | Fairness assessments in credit scoring and insurance systems. |
Strong focus on transparency and accountability in financial applications. | Not widely applicable to non-financial systems. | Used in fairness evaluations of automated financial decision-making systems. | |
Tailored for regulatory and compliance environments. | Limited support for non-financial use cases. | Ensuring fairness in automated decision-making systems in banking and finance. | |
Assurance Cases for Fairness. [146] | Provides structured arguments for fairness claims. | Requires domain expertise and collaboration for effective implementation. | Ensuring fairness in AI systems for healthcare and education. |
Useful in establishing transparency and accountability for fairness claims. | May require significant resources to build effective assurance cases. | Used in verifying fairness in AI-based healthcare and educational systems. | |
Focuses on providing evidence-based assurance for fairness. | Limited scalability to large AI systems. | Ensuring trust and accountability in AI systems in sectors with high public scrutiny. |
Toolkit | Pros | Cons | Use Cases |
---|---|---|---|
IBM ART [168] | Supports multiple ML frameworks | Can be complex to set up | Security testing for AI in finance & healthcare |
Offers both attacks & defences | Some attacks are slow | Adversarial training for robust AI models | |
Includes explainability tools | |||
CleverHans [169] | Well-documented | Limited defence techniques | Evaluating deep learning model security |
Strong focus on adversarial attacks | Primarily TensorFlow-focused | Research on adversarial robustness | |
Open-source and widely used | |||
Foolbox [170] | Simple API for adversarial attacks | Lacks built-in defence mechanisms | Red teaming for AI security |
Works with PyTorch, TensorFlow, JAX | Not as actively maintained as ART | Benchmarking model vulnerability | |
Optimised for speed | |||
MIT Robustness [171] | Designed for adversarial training | Focused on image classification tasks | Adversarial training in computer vision |
PyTorch support | Limited support for non-vision models | Research in robustness techniques | |
Provides pre-trained robust models | |||
DeepRobust [172] | Supports both graph and image-based AI models | Requires deep understanding of adversarial learning | AI security in social networks (graph AI) |
Covers attacks and defences | Not as widely adopted as ART or CleverHans | Robustness evaluation for medical AI models | |
Provides benchmark datasets | |||
AdverTorch [173] | PyTorch-based attack toolkit | Native PyTorch support | Developing adversarial defences in PyTorch |
Various attack implementations | Limited to PyTorch ecosystem | Generating adversarial examples | |
Focus on adversarial training | Robust training pipeline integration | ||
Robustness Gym [174] | Robustness benchmarking suite | Modular and extensible | Evaluating NLP model robustness |
Supports data transformations | Primarily NLP-focused | Stress-testing AI models in production | |
Model evaluation techniques | Enhancing general AI model reliability | ||
TRADES [175] | Trade-off between robustness and accuracy | Strong theoretical backing | Improving robustness in DNNs |
Adversarial training framework | Requires deep ML expertise | Research on robust AI training methods | |
Defence against adversarial perturbations | Defending against adversarial attacks | ||
AutoAttack [176] | Ensemble adversarial attack method | Strong attack performance | Benchmarking adversarial robustness |
Automates attack selection | Limited flexibility for defences | Validating adversarial defences | |
Works on image classifiers | Testing robustness in computer vision | ||
RobustBench [177] | Maintains a leaderboard of robust models | Limited set of attacks compared to ART | Benchmarking AI robustness in academia |
Easy benchmarking of defences | Mostly vision-focused | Comparing adversarial defences | |
Open-source |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Ahangar, M.N.; Farhat, Z.A.; Sivanathan, A. AI Trustworthiness in Manufacturing: Challenges, Toolkits, and the Path to Industry 5.0. Sensors 2025, 25, 4357. https://doi.org/10.3390/s25144357
Ahangar MN, Farhat ZA, Sivanathan A. AI Trustworthiness in Manufacturing: Challenges, Toolkits, and the Path to Industry 5.0. Sensors. 2025; 25(14):4357. https://doi.org/10.3390/s25144357
Chicago/Turabian StyleAhangar, M. Nadeem, Z. A. Farhat, and Aparajithan Sivanathan. 2025. "AI Trustworthiness in Manufacturing: Challenges, Toolkits, and the Path to Industry 5.0" Sensors 25, no. 14: 4357. https://doi.org/10.3390/s25144357
APA StyleAhangar, M. N., Farhat, Z. A., & Sivanathan, A. (2025). AI Trustworthiness in Manufacturing: Challenges, Toolkits, and the Path to Industry 5.0. Sensors, 25(14), 4357. https://doi.org/10.3390/s25144357