Towards a Conceptual Modeling of Trustworthiness in AI-Based Big Data Analysis
Abstract
:1. Introduction and Approach
- RQ1.
- How can the three fundamental concepts validity, capability, and reproducibility of trustworthiness be mathematically modeled?
- RQ2.
- How can the relationship between the TAI-BDM-based data exploration process and the dynamics of these three concepts be integrated into that model?
- RQ3.
- How can the current overall trustworthiness of the AI-based data exploration system be integrated into that model?
2. Selected State of the Art
2.1. The TAI-BDM Reference Model
2.2. Mathematical Foundations
3. Towards a Conceptual Modeling of Trustworthiness
- Commutativity. If in a data exploration situation for a manipulation step there exists an alternative algorithm leading to exactly the same output configuration as also the same produced insights and derived knowledge, their related trustworthiness state paths must end up at the same final point.
- Reversibility. If in an early TAI-BDM stage before any insight generation a data manipulation step can be fully reversed to exactly the initial data exploration configuration by an adapted redo operation, the corresponding compound trustworthiness mapping must act as the identity function.
4. Conclusions and Future Work
- Completing the model for learning trustworthiness: Developing a model that can learn and adapt trustworthiness based on acquired facts, potentially using, e.g., Bayesian methods, could enhance the dynamic nature of trust in AI systems. This model may also take dependencies between the concepts’ capability, validity, and reproducibility into consideration. The state space may be something different than the unit cube for advanced use cases, e.g., another convex subset of by giving the three concepts weights with .
- Implementing the trust bus architecture: In fact, coding and integrating the trust bus architecture into AI systems could significantly improve the user trust by providing a systematic framework for trust management. A new prototypical data analytics pipeline is currently being built according to TAI-BDM. It mimics the functionality of DARIA-S, but will be open source. A trust-bus component will be integrated. This prototype leverages existing workflow management systems (WMSs). One challenge is to replace the rather heavy Airflow in our prototypical AI2VIS4BigData analytics pipeline with a simpler modular system such as Luigi [25] or Prefect [26]. These should reduce maintenance overhead compared to Airflow.
- Evaluating the model of trustworthiness: Conducting thorough evaluations of the proposed trustworthiness model is crucial to ensure its effectiveness and applicability across different AI-based big data analysis scenarios. This involves assessing how well the model captures the dynamics of trust and its impact on user acceptance and system performance. This should also take the impact of dependencies between the concepts into consideration. A dual-phase validation approach stands to reason: Controlled scenario testing using synthetic datasets with engineered validity-capability-reproducibility profiles across the state space enabling the systematic analysis of trustworthiness trajectories through sequential manipulation steps . Secondly, empirical validation leveraging pseudonymized or masked longitudinal data from DARIA’s operational deployment. Structured expert interviews should be conducted in parallel as a complementary evaluation method.
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Glossary of Mathematical Notations
Trustworthiness state vector | |
: Capability component, | |
: Reproducibility component, | |
: Validity component | |
Trustworthiness state space defined as | |
State update function | |
: Feature vector of manipulation step, | |
: Current trustworthiness state, | |
: Environmental conditions factor | |
Trustworthiness norm function | |
Maps state vector to scalar trustworthiness measure |
References
- KPMG AG Wirtschaftsprüfungsgesellschaft. In Trust in Artificial Intelligence: A Five Country Study; The University of Queensland: Brisbane, Australia, 2021. [CrossRef]
- Cui, Q.; Erfani, A. Automatic Detection of Construction Risks. In Proceedings of the ECPPM 2021—eWork and eBusiness in Architecture, Engineering, and Construction, Moscow, Russia, 15–17 September 2021; CRC Press: Boca Raton, FL, USA, 2021; pp. 184–189. [Google Scholar]
- De Marco, A.; Thaheem, J. Risk Analysis in Construction Projects: A Practical Selection Methodology. Am. J. Appl. Sci. 2014, 11, 74–84. [Google Scholar] [CrossRef]
- FFG—The Austrian Research Promotion Agency. Data-driven Risk Analysis in the Construction Industry. In General Programmes 2022, FFG Project Number: FO999897631, Proposal Acronym: DARIA; FFG—The Austrian Research Promotion Agency: Vienna, Austria, 2022. [Google Scholar]
- Ivanova, M.; Grims, M.; Karas, D.; Höfinger, G.; Bornschlegl, M.X.; Hemmje, M.L. An AI-Based Approach to Identify Financial Risks in Transportation Infrastructure Construction Projects. In Artificial Intelligence Applications and Innovations; Maglogiannis, I., Iliadis, L., Macintyre, J., Avlonitis, M., Papaleonidas, A., Eds.; Springer: Cham, Switzerland, 2024; pp. 158–173. [Google Scholar] [CrossRef]
- Grims, M.; Karas, D.; Ivanova, M.; Bruchhaus, S.; Höfinger, G.; Bornschlegl, M.X.; Hemmje, M.L. Implementation and Rollout of a Trusted AI-based Approach to Identify Financial Risks in Transportation Infrastructure Construction Projects. In Proceedings of the Innovate 2025: World AI, Automation and Technology Forum, Davos, Switzerland, 16–18 April 2025. [Google Scholar]
- Wischnewski, M.; Krämer, N.; Müller, E. Measuring and Understanding Trust Calibrations for Automated Systems: A Survey of the State-Of-The-Art and Future Directions. In CHI ’23: Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems; ACM: New York, NY, USA, 2023; pp. 1–16. [Google Scholar] [CrossRef]
- Horvat, S.; Horvat, T. Artificial Intelligence, Neurodiversity, and Trust: An Exploratory Overview. Humanit. Soc. Sci. Commun. 2024, 11, 1568. [Google Scholar] [CrossRef]
- Kortum, H.; Rebstadt, J.; Böschen, T.; Meier, P.; Thomas, O. Towards the Operationalization of Trustworthy AI: Integrating the EU Assessment List into a Procedure Model for the Development and Operation of AI-Systems. In Proceedings of the INFORMATIK 2023: 53rd Annual Conference of the Gesellschaft für Informatik e.V. (GI), Berlin, Germany, 26–29 September 2023; Gesellschaft für Informatik e.V.: Bonn, Germany, 2023; pp. 289–300. [Google Scholar] [CrossRef]
- Dekate, C. Gartner Predicts the Future of AI Technologies. 2020. Available online: https://www.cdotrends.com/story/14686/gartner-predicts-future-ai-technologies (accessed on 1 May 2025).
- Bornschlegl, M.X. Towards Trustworthiness in AI-Based Big Data Analysis; Research Report; FernUniversität: Hagen, Germany, 2024. [Google Scholar] [CrossRef]
- Nunamaker, J.F.; Chen, M.; Purdin, T.D. Systems Development in Information Systems Research. J. Manag. Inf. Syst. 1990, 7, 89–106. [Google Scholar] [CrossRef]
- ISO 9001:2015; Quality Management Systems—Requirements. International Organization for Standardization: Geneva, Switzerland, 2015.
- Zhang, J.M.; Harman, M.; Ma, L.; Liu, Y. Machine Learning Testing: Survey, Landscapes and Horizons. IEEE Trans. Softw. Eng. 2022, 48, 1–36. [Google Scholar] [CrossRef]
- Myllyaho, L.; Raatikainen, M.; Männistö, T.; Mikkonen, T.; Nurminen, J.K. Systematic literature review of validation methods for AI systems. J. Syst. Softw. 2021, 181, 111050. [Google Scholar] [CrossRef]
- European Parliament. Proposal for a Regulation of the European Parliament and of the Council Laying Down Harmonised Rules on Artificial Intelligence (Artificial Intelligence Act) and Amending Certain Union Legislative Acts; European Parliament: Strasbourg, France, 2024. [Google Scholar]
- Vaccaro, M.; Almaatouq, A.; Malone, T. When combinations of humans and AI are useful: A systematic review and meta-analysis. Nat. Hum. Behav. 2024, 8, 2293–2303. [Google Scholar] [CrossRef] [PubMed]
- Desai, A.; Abdelhamid, M.; Padalkar, N.R. What is reproducibility in artificial intelligence and machine learning research? AI Mag. 2025, 46, e70004. [Google Scholar] [CrossRef]
- Easton, D. A Re-Assessment of the Concept of Political Support. Br. J. Political Sci. 1975, 5, 435–457. [Google Scholar] [CrossRef]
- Kaufmann, M. Towards a Reference Model for Big Data Management; FernUniversität: Hagen, Germany, 2016. [Google Scholar]
- Abbass, H.A.; Leu, G.; Merrick, K. A Review of Theoretical and Practical Challenges of Trusted Autonomy in Big Data. IEEE Access 2016, 4, 2808–2830. [Google Scholar] [CrossRef]
- Reis, T.; Bornschlegl, M.X.; Hemmje, M.L. AI2VIS4BigData: A Reference Model for AI-Based Big Data Analysis and Visualization. In Advanced Visual Interfaces. Supporting Artificial Intelligence and Big Data Applications; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2021; Volume 12585, pp. 1–18. [Google Scholar] [CrossRef]
- Royden, H.L.; Fitzpatrick, P. Real Analysis, 4th ed.; Prentice Hall: Hoboken, NJ, USA, 2010. [Google Scholar]
- Manziuk, E.; Barmak, O.; Radiuk, P.; Kuznetsov, V.; Krak, I. Representation of Trustworthiness Components in AI Systems: A Formalized Approach Considering Interconnections and Overlap. In Proceedings of the ProfIT AI 2024: 4th International Workshop of IT-professionals on Artificial Intelligence (ProfIT AI 2024), Cambridge, MA, USA, 25–27 September 2024. [Google Scholar]
- Spotify. Luigi Documentation, Version 3.6.0; Spotify: Stockholm, Sweden, 2024. [Google Scholar]
- Prefect Technologies, Inc. Prefect: Pythonic, Modern Workflow Orchestration For Resilient Data Pipelines. 2024. Available online: https://www.prefect.io (accessed on 1 May 2025).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Bruchhaus, S.; Kreibich, A.; Reis, T.; Bornschlegl, M.X.; Hemmje, M.L. Towards a Conceptual Modeling of Trustworthiness in AI-Based Big Data Analysis. Information 2025, 16, 455. https://doi.org/10.3390/info16060455
Bruchhaus S, Kreibich A, Reis T, Bornschlegl MX, Hemmje ML. Towards a Conceptual Modeling of Trustworthiness in AI-Based Big Data Analysis. Information. 2025; 16(6):455. https://doi.org/10.3390/info16060455
Chicago/Turabian StyleBruchhaus, Sebastian, Alexander Kreibich, Thoralf Reis, Marco X. Bornschlegl, and Matthias L. Hemmje. 2025. "Towards a Conceptual Modeling of Trustworthiness in AI-Based Big Data Analysis" Information 16, no. 6: 455. https://doi.org/10.3390/info16060455
APA StyleBruchhaus, S., Kreibich, A., Reis, T., Bornschlegl, M. X., & Hemmje, M. L. (2025). Towards a Conceptual Modeling of Trustworthiness in AI-Based Big Data Analysis. Information, 16(6), 455. https://doi.org/10.3390/info16060455