CAPTURE: A Stakeholder-Centered Iterative MLOps Lifecycle
Abstract
1. Introduction
Problem Statement and Contributions
2. Background and Related Work
Related Work
3. Methods—Deriving the CAPTURE Framework
3.1. Framework Overview
| CONSULT | Stakeholder identification and mapping: who should be involved and what do they need? |
| ARTICULATE | Requirements formalization: translating stakeholder needs into SMART specifications, data contracts, and KPIs. |
| PROTOCOL | Decision provenance: documenting how and why lifecycle decisions are made. |
| TERRAFORM | Infrastructure lifecycle: treating infrastructure as a governed entity with explicit versioning. |
| UTILIZE | Model development: building, training, and validating ML components. |
| REIFY | Deployment and evidence collection: observing system behavior in real stakeholder contexts. |
| EVOLVE | Evaluation and iteration: assessing impact and selecting appropriate evolution paths. |
3.2. Framework Design Rationale
3.2.1. Phase Alignment with Established Frameworks
3.2.2. Phase Selection Rationale
Stakeholder Engagement as Scarce Resource
Upstream Stakeholder Engagement
Necessity of the Seven-Phase Structure
3.3. Phase Transition Model
- Iterative Refinement
3.3.1. CONSULT: Engage and Define
Decision Gate 1: Transition to ARTICULATE
- measures stakeholder mapping completeness;
- represents the fraction of stakeholder groups with identified KPIs;
- assesses documentation completeness; and
- are project-specific weights satisfying .
3.3.2. ARTICULATE: Formalize Requirements
Decision Gate 2: Transition to PROTOCOL
- represents the fraction of stakeholders validating requirements;
- measures trade-off documentation completeness;
- assesses requirement matrix quality; and
- are project-specific weights.
3.3.3. PROTOCOL: Document Decisions
Decision Gate 3: Transition to TERRAFORM
- measures the completeness of stakeholder-requirement-decision links;
- assesses versioning schema quality;
- evaluates governance documentation completeness;
- quantifies data quality metric coverage across all six dimensions;
- evaluates completeness of executable data rules enforcing contracts;
- assesses provenance tracking specification completeness;
- are project-specific weights.
3.3.4. TERRAFORM: Build Infrastructure
Decision Gate 4: Transition to UTILIZE
- measures infrastructure uptime stability over the observation period;
- assesses versioning system functionality;
- evaluates performance benchmark completeness;
- are project-specific weights.
3.3.5. UTILIZE: Implement ML Modules
Decision Gate 5: Transition to REIFY
- is the achievement level for KPI i in validation (fraction of target met);
- measures data quality during model development;
- assesses requirement-to-model traceability completeness;
- indicates successful reproducibility verification;
- are project-specific weights.
3.3.6. REIFY: Apply and Obtain Feedback
Decision Gate 6: Transition to EVOLVE
- is the ratio of observation duration to required period;
- is the ratio of interactions to required count;
- is the compliance level for KPI i in production (fraction of target met);
- measures stakeholder feedback completeness; and
- are project-specific weights.
3.3.7. EVOLVE: Evaluate and Iterate
Decision Gate 7: Strategic Iteration Path Selection
- is the overall stakeholder satisfaction from REIFY feedback;
- is the ratio of KPIs meeting thresholds to total KPIs;
- measures fraction of requirements requiring modification;
- thresholds are project-specific, reflecting organizational risk tolerance and resource availability.
3.3.8. Decision Gate Threshold Determination
3.3.9. Lifecycle Transition Tracking
4. Results
4.1. Conceptual Cognitive Walkthrough
- (1)
- Smooth Forward ProgressionA computer vision model for defect detection passes all gates sequentially.Interpretation: This ideal case occurs when CONSULT feasibility analysis accurately predicts technical and organizational constraints.
Gates 1–3 Requirements stable, data contracts verified. Gate 4 Infrastructure stress test passed (72 h stability). Gate 5 Model F1-score exceeds 0.95 on held-out test set. Gate 6 Deployment successful, user feedback positive. Gate 7 System transitions to continuous monitoring. - (2)
- Infrastructure Inadequacy DiscoveredDuring UTILIZE, training repeatedly exceeds allocated memory, causing job failures.Interpretation: Gate 5 rejection addresses the root cause (compute capacity) through backward transition to TERRAFORM.
Decision Rejected due to mandatory criterion violation. Transition Return to TERRAFORM. Action Infrastructure team provisions GPUs with higher memory capacity. Re-entry Gate 4 re-evaluated for new infrastructure, then UTILIZE repeated. - (3)
- Requirements Gap Discovered in ProductionAfter Gate 6 approval, field usage reveals that lighting conditions are degrading performance.Interpretation: Gate 7 triggers minor iteration (7b) because the issue stems from an incomplete requirement, not model degradation. Multi-phase traversal demonstrates traceability.
Trigger Gate 7 evaluation shows a KPI dip due to unforeseen conditions. Decision Option 7b (minor iteration) returns to ARTICULATE. Rationale Low-light robustness was not specified; this is a requirement gap, not drift. Action New requirement “low-light robustness” added. Flow ARTICULATE → PROTOCOL → UTILIZE (fine-tune on dark images). - (4)
- System Retirement Due to Persistent DriftA prediction model degrades over 6 months despite two retraining cycles.Interpretation: Gate 7 identifies concept drift acceleration where retraining no longer maintains KPI thresholds.
Trigger Gate 7 fitness threshold failure (KPI below 0.7 for 3 consecutive months). Decision Retirement path selected. Action System decommissioned, data archived, stakeholders notified. - (5)
- Observation–Evaluation Conflation (REIFY + EVOLVE)A recommendation system is deployed and the team evaluates early feedback.Interpretation: Without distinct REIFY and EVOLVE phases, teams make iteration decisions before statistically valid evidence accumulates.
Issue Team decides to retrain after 48 h of initial user data. Consequence Early adopter behavior is non-representative; model oscillates through multiple retraining cycles. Root Cause No temporal separation between evidence collection and evaluation. - (6)
- Stakeholder–Specification Conflation (CONSULT + ARTICULATE)Stakeholder feedback is gathered and immediately formalized into requirements.Interpretation: Without distinct CONSULT and ARTICULATE phases, informal statements are prematurely hardcoded into requirements.
Issue Stakeholder says “we need fast predictions”; formalized as “latency below 100 ms”. Consequence Accuracy sacrificed for latency; stakeholder actually meant “fast enough for interactive use” (500 ms acceptable). Root Cause No trade-off analysis or stakeholder validation before formalization. - (7)
- Specification–Documentation Conflation (ARTICULATE + PROTOCOL)Requirements are captured as informal specifications without explicit traceability.Interpretation: Without distinct ARTICULATE and PROTOCOL phases, requirements lack traceable design rationale.
Issue Model performs poorly after 6 months; team cannot trace which requirement drove problematic key design decisions. Consequence Root cause analysis impossible; requires re-interviewing stakeholders. Root Cause No versioned decision tracking linking requirements to design rationale.
Synthesis: Validating Phase Granularity
4.2. Empirical Case Study: Longitudinal Framework Application
4.2.1. Cycles 1–2 (2021–2022): Foundation and Infrastructure
4.2.2. Cycles 3–5 (2023–2025): Iterative Refinement and CAPTURE-Enabled Success
Validation Synthesis
4.3. Expert Interview Validation
- Stakeholder Engagement (CONSULT)
- Participatory Modeling (ARTICULATE)
- Decision Tracking (PROTOCOL)
- Accessibility (UTILIZE)
- Iteration and Disagreement (EVOLVE)
- Summary
5. Discussion
5.1. Interpretation of Results
5.1.1. Stakeholder-Centered Foundation (RO1)
5.1.2. Sensor-Specific Engineering (RO2)
5.1.3. Continuous Verification and Validation (RO3)
5.1.4. Infrastructure Lifecycle and Evidence-Based Governance (RO4, RO5)
5.2. Implications for Intelligent Software Engineering
5.3. Limitations and Threats to Validity
- Construct Validity
- Internal Validity
- External Validity
- Calibration Burden
- Comparison with Existing Approaches
5.4. Applicability and Adoption Considerations
- Sensor-Specific Scope Criteria
- Threshold-Calibrated Applicability
- Reduction to Simpler Frameworks
- Trade-Off Acknowledgment
- Generalizability
- Organizational Readiness
6. Conclusions
6.1. Contributions
6.2. Future Work
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
| AI | Artificial Intelligence |
| CI/CD | Continuous Integration, Continuous Deployment |
| HCD | Human-Centered Design |
| HITL | Human-in-the-loop |
| HOTL | Human-on-the-loop |
| ISE | Intelligent Software Engineering |
| KPI | Key Performance Indicator |
| ML | Machine Learning |
| MLOps | Machine Learning Operations |
| RE4ML | Requirement Engineering for Machine Learning |
| SE4ML | Software Engineering for Machine Learning |
| V&V | Verification and Validation |
| XAI | Explainable AI |
Appendix A. Decision Gate Template
| Phase Transition: ___________ → ___________ |
| Date: ___________ Approvers: _______________________ |
| Mandatory Criteria |
| □ _______________________ |
| □ _______________________ |
| Quality Gates |
| □ _______________________ |
| □ _______________________ |
| Warnings/Deviations: ________________________________________ |
| Documentation Gates |
| □ _______________________ |
| □ _______________________ |
| Decision |
| □APPROVED - Proceed to next phase |
| □CONDITIONAL - Proceed with documented risks/deviations |
| □REJECTED - Return to phase [____] for [reason] |
| Rationale: __________________________________________________ |
Appendix B. CAPTURE Adoption Checklist
CONSULT—Engage and Define
|
ARTICULATE—Formalize Requirements
|
PROTOCOL—Document Decisions
|
TERRAFORM—Build Infrastructure
|
UTILIZE—Implement ML Modules
|
REIFY—Apply and Obtain Feedback
|
EVOLVE—Evaluate and Iterate
|
Cross-Cutting Guidance
|
References
- Yalley, P.P. Limitations of Maximizing Shareholder Value in the Era of Artificial Intelligence and Machine Learning. World J. Adv. Res. Rev. 2025, 28, 1816–1825. [Google Scholar] [CrossRef]
- Studer, S.; Bui, T.B.; Drescher, C.; Hanuschkin, A.; Winkler, L.; Peters, S.; Müller, K.R. Towards CRISP-ML(Q): A Machine Learning Process Model with Quality Assurance Methodology. Mach. Learn. Knowl. Extr. 2021, 3, 392–413. [Google Scholar] [CrossRef]
- ISO 9241-210:2019; Human-Centred Design for Interactive Systems. ISO Publishing: Geneva, Switzerland, 2019.
- ModelOP. Responsible AI-Benchmark Report 2024. 2024. Available online: https://www.modelop.com/resources-ebooks/responsible-ai-report-2024 (accessed on 18 December 2025).
- Sharma, A.; Sharma, V.; Jaiswal, M.; Wang, H.C.; Jayakody, D.N.K.; Basnayaka, C.M.W.; Muthanna, A. Recent Trends in AI-Based Intelligent Sensing. Electronics 2022, 11, 1661. [Google Scholar] [CrossRef]
- Rogan, J.; Bucci, S.; Firth, J. Health Care Professionals’ Views on the Use of Passive Sensing, AI, and Machine Learning in Mental Health Care: Systematic Review with Meta-Synthesis. JMIR Ment. Health 2024, 11, e49577. [Google Scholar] [CrossRef]
- Henzler, D.; Schmidt, S.; Koçar, A.; Herdegen, S.; Lindinger, G.L.; Maris, M.T.; Konopka, M.J. Healthcare professionals’ perspectives on artificial intelligence in patient care: A systematic review of hindering and facilitating factors on different levels. BMC Health Serv. Res. 2025, 25, 633. [Google Scholar] [CrossRef]
- Mat Sanusi, K.A.; Iren, D.; Fanchamps, N.; Geisen, M.; Klemke, R. Virtual virtuoso: A systematic literature review of immersive learning environments for psychomotor skill development. Educ. Technol. Res. Dev. 2025, 73, 909–949. [Google Scholar] [CrossRef]
- Karampatzakis, D.; Fominykh, M.; Fanchamps, N.; Firssova, O.; Amanatidis, P.; Van Lankveld, G.; Klemke, R. Educational Robotics at Schools Online with Augmented Reality. In Proceedings of the 2024 IEEE Global Engineering Education Conference (EDUCON), Kos Island, Greece, 8–11 May 2024; IEEE: New York, NY, USA, 2024; pp. 1–10. [Google Scholar] [CrossRef]
- Chander, B.; Pal, S.; De, D.; Buyya, R. Artificial Intelligence-based Internet of Things for Industry 5.0. In Artificial Intelligence-Based Internet of Things Systems; Pal, S., De, D., Buyya, R., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2022; pp. 3–45. [Google Scholar] [CrossRef]
- Nguyen, T.T.H.; Nguyen, P.T.L.; Wachowicz, M.; Cao, H. MACeIP: A Multimodal Ambient Context-Enriched Intelligence Platform in Smart Cities. In Proceedings of the 2024 IEEE International Conference on Consumer Electronics-Asia (ICCE-Asia), Danang, Vietnam, 3–6 November 2024; IEEE: New York, NY, USA, 2024; pp. 1–4. [Google Scholar] [CrossRef]
- Chen, G.; Wang, X. Performance Optimization of Machine Learning Inference under Latency and Server Power Constraints. In Proceedings of the 2022 IEEE 42nd International Conference on Distributed Computing Systems (ICDCS), Bologna, Italy, 10–13 July 2022; IEEE: New York, NY, USA, 2022; pp. 325–335. [Google Scholar] [CrossRef]
- Lin, K.; Li, Y.; Sun, J.; Zhou, D.; Zhang, Q. Multi-sensor fusion for body sensor network in medical human–robot interaction scenario. Inf. Fusion 2020, 57, 15–26. [Google Scholar] [CrossRef]
- Marotta, L.; Buurke, J.H.; van Beijnum, B.J.F.; Reenalda, J. Towards Machine Learning-Based Detection of Running-Induced Fatigue in Real-World Scenarios: Evaluation of IMU Sensor Configurations to Reduce Intrusiveness. Sensors 2021, 21, 3451. [Google Scholar] [CrossRef]
- Höschler, L.; Halmich, C.; Schranz, C.; Koelewijn, A.D.; Schwameder, H. Evaluating the Influence of Sensor Configuration and Hyperparameter Optimization on Wearable-Based Knee Moment Estimation During Running. Int. J. Comput. Sci. Sport 2025, 24, 80–106. [Google Scholar] [CrossRef]
- ISO/IEC 22989:2022; Artificial Intelligence Concepts and Terminology. ISO Publishing: Geneva, Switzerland, 2022.
- Amershi, S.; Begel, A.; Bird, C.; DeLine, R.; Gall, H.; Kamar, E.; Zimmermann, T. Software Engineering for Machine Learning: A Case Study. In Proceedings of the 2019 IEEE/ACM 41st International Conference on Software Engineering: Software Engineering in Practice (ICSE-SEIP), Montreal, QC, Canada, 25–31 May 2019; IEEE: New York, NY, USA, 2019; pp. 291–300. [Google Scholar] [CrossRef]
- Lima, A.; Monteiro, L.; Furtado, A. MLOps: Practices, Maturity Models, Roles, Tools, and Challenges—A Systematic Literature Review. In Proceedings of the 24th International Conference on Enterprise Information Systems, Virtual Event, 25–27 April 2022; SCITEPRESS-Science and Technology Publications: Setubal, Portugal, 2022; pp. 308–320. [Google Scholar] [CrossRef]
- Diaz-de Arcaya, J.; Torre-Bastida, A.I.; Zárate, G.; Miñón, R.; Almeida, A. A Joint Study of the Challenges, Opportunities, and Roadmap of MLOps and AIOps: A Systematic Survey. ACM Comput. Surv. 2024, 56, 1–30. [Google Scholar] [CrossRef]
- Hanchuk, D.; Semerikov, S. Automating Machine Learning: A Meta-Synthesis of MLOps Tools, Frameworks and Architectures. In Proceedings of the 7th Workshop for Young Scientists in Computer Science & Software Engineering (CS&SE@SW 2024). CEUR Workshop Proceedings, Kryvyi Rih, Ukraine, 27 December 2024; Volume 3917, pp. 362–414. [Google Scholar]
- Strong, B.; Boyda, E.; Kruse, C.; Ingold, T.; Maron, M. Digital applications unlock remote sensing AI foundation models for scalable environmental monitoring. Front. Clim. 2025, 7, 1520242. [Google Scholar] [CrossRef]
- Memarian, B.; Doleck, T. Fairness, Accountability, Transparency, and Ethics (FATE) in Artificial Intelligence (AI) and higher education: A systematic review. Comput. Educ. AI 2023, 5, 100152. [Google Scholar] [CrossRef]
- Ali, S.; Abuhmed, T.; El-Sappagh, S.; Muhammad, K.; Alonso-Moral, J.M.; Confalonieri, R.; Herrera, F. Explainable Artificial Intelligence (XAI): What we know and what is left to attain Trustworthy Artificial Intelligence. Inf. Fusion 2023, 99, 101805. [Google Scholar] [CrossRef]
- Eken, B.; Pallewatta, S.; Tran, N.; Tosun, A.; Babar, M.A. A Multivocal Review of MLOps Practices, Challenges and Open Issues. ACM Comput. Surv. 2025, 58, 1–35. [Google Scholar] [CrossRef]
- Truong, H.L.; Comerio, M.; de Paoli, F.; Gangadharan, G.R.; Dustdar, S. Data contracts for cloud-based data marketplaces. Int. J. Comput. Sci. Eng. 2012, 7, 280. [Google Scholar] [CrossRef]
- Wider, A.; Harrer, S.; Dietz, L.W. AI-Assisted Data Governance with Data Mesh Manager. In Proceedings of the 2025 IEEE International Conference on Web Services (ICWS), Helsinki, Finland, 7–12 July 2025; IEEE: New York, NY, USA, 2025; pp. 963–965. [Google Scholar] [CrossRef]
- Rangineni, S. An Analysis of Data Quality Requirements for Machine Learning Development Pipelines Frameworks. Int. J. Comput. Trends Technol. 2023, 71, 16–27. [Google Scholar] [CrossRef]
- ISO/IEC 5259:2024; Data Quality for Analytics and Machine Learning (ML). ISO Publishing: Geneva, Switzerland, 2024.
- Star, S.L. The Ethnography of Infrastructure. Am. Behav. Sci. 1999, 43, 377–391. [Google Scholar] [CrossRef]
- Pipek, V.; Wulf, V. Infrastructuring: Toward an Integrated Perspective on the Design and Use of Information Technology. J. Assoc. Inf. Syst. 2009, 10, 447–473. [Google Scholar] [CrossRef]
- Sculley, D.; Holt, G.; Golovin, D.; Davydov, E.; Phillips, T.; Ebner, D.; Dennison, D. Hidden Technical Debt in Machine Learning Systems. In Proceedings of the 29th International Conference on Neural Information Processing Systems-Volume 2, Montreal, QC, Canada, 7–12 December 2015; MIT Press: Cambridge, MA, USA, 2015; pp. 2503–2511. [Google Scholar]
- Volmar, A. From Systems to “Infrastructuring”: Infrastructure Theory and Its Impact on Writing the History of Media. In Rethinking Infrastructure Across the Humanities; Pinnix, A., Volmar, A., Esposito, F., Binder, N., Eds.; Transcript Verlag: Bielefeld, Germany, 2023; pp. 51–64. [Google Scholar] [CrossRef]
- Lwakatare, L.E.; Raj, A.; Bosch, J.; Olsson, H.H.; Crnkovic, I. A Taxonomy of Software Engineering Challenges for Machine Learning Systems: An Empirical Investigation. In Agile Processes in Software Engineering and Extreme Programming; Kruchten, P., Fraser, S., Coallier, F., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2019; Volume 355, pp. 227–243. [Google Scholar] [CrossRef]
- Martínez-Fernández, S.; Bogner, J.; Franch, X.; Oriol, M.; Siebert, J.; Trendowicz, A.; Wagner, S. Software Engineering for AI-Based Systems: A Survey. ACM Trans. Softw. Eng. Methodol. 2022, 31, 1–59. [Google Scholar] [CrossRef]
- Kreuzberger, D.; Kühl, N.; Hirschl, S. Machine Learning Operations (MLOps): Overview, Definition, and Architecture. IEEE Access 2023, 11, 31866–31879. [Google Scholar] [CrossRef]
- Zarour, M.; Alzabut, H.; Al-Sarayreh, K.T. MLOps best practices, challenges and maturity models: A systematic literature review. Inf. Softw. Technol. 2025, 183, 107733. [Google Scholar] [CrossRef]
- Serban, A.; van der Blom, K.; Hoos, H.; Visser, J. Software engineering practices for machine learning—Adoption, effects, and team assessment. J. Syst. Softw. 2024, 209, 111907. [Google Scholar] [CrossRef]
- Singh, J.; Cobbe, J.; Norval, C. Decision Provenance: Harnessing Data Flow for Accountable Systems. IEEE Access 2019, 7, 6562–6574. [Google Scholar] [CrossRef]
- Amershi, S.; Weld, D.; Vorvoreanu, M.; Fourney, A.; Nushi, B.; Collisson, P.; Horvitz, E. Guidelines for Human-AI Interaction. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, Glasgow, UK, 4–9 May 2019; ACM: New York, NY, USA, 2019; pp. 1–13. [Google Scholar] [CrossRef]
- van der Stappen, A.; Funk, M. Towards Guidelines for Designing Human-in-the-Loop Machine Training Interfaces. In Proceedings of the 26th International Conference on Intelligent User Interfaces, Paphos, Cyprus, 23–26 March 2026; ACM: New York, NY, USA, 2021; pp. 514–519. [Google Scholar] [CrossRef]
- Cheruvu, R. Human-in/on-the-Loop Design for Human Controllability. In Handbook of Human-Centered Artificial Intelligence; Xu, W., Ed.; Springer Nature: Singapore, 2025; pp. 1–47. [Google Scholar] [CrossRef]
- Shin, H.; Park, J.; Yu, J.; Kim, J.; Kim, H.Y.; Oh, C. Looping In: Exploring Feedback Strategies to Motivate Human Engagement in Interactive Machine Learning. Int. J. Hum.-Comput. Interact. 2025, 41, 8666–8683. [Google Scholar] [CrossRef]
- Ahmed, A.M.A. Exploring MLOps Dynamics: An Experimental Analysis in a Real-World Machine Learning Project. arXiv 2023, arXiv:2307.13473. [Google Scholar] [CrossRef]
- Boukerche, A.; Coutinho, R.W.L. Design Guidelines for Machine Learning-based Cybersecurity in Internet of Things. IEEE Netw. 2021, 35, 393–399. [Google Scholar] [CrossRef]
- Cigoj, M. Achieving EU AI Act Compliance by Integrating Governance as Code Within MLOps. 2025. Available online: https://htec.com/insights/whitepapers/achieving-eu-ai-act-compliance/ (accessed on 18 December 2025).
- ISO/IEC 42001:2023; Artificial Intelligence—Management System. International Organization for Standardization: Geneva, Switzerland, 2023.
- Singh, P. Systematic Review of Data-Centric Approaches in Artificial Intelligence and Machine Learning. Data Sci. Manag. 2023, 6, 144–157. [Google Scholar] [CrossRef]
- Habiba, U.e.; Haug, M.; Bogner, J.; Wagner, S. How mature is requirements engineering for AI-based systems? A systematic mapping study on practices, challenges, and future research directions. Requir. Eng. 2024, 29, 567–600. [Google Scholar] [CrossRef]
- Munappy, A.R.; Bosch, J.; Olsson, H.H. Data Pipeline Management in Practice: Challenges and Opportunities. In Product-Focused Software Process Improvement; Morisio, M., Torchiano, M., Jedlitschka, A., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; Volume 12562, pp. 168–184. [Google Scholar] [CrossRef]
- Ashmore, R.; Calinescu, R.; Paterson, C. Assuring the Machine Learning Lifecycle. ACM Comput. Surv. 2022, 54, 1–39. [Google Scholar] [CrossRef]
- Nelson, K.; Corbin, G.; Anania, M.; Kovacs, M.; Tobias, J.; Blowers, M. Evaluating model drift in machine learning algorithms. In Proceedings of the 2015 IEEE Symposium on Computational Intelligence for Security and Defense Applications (CISDA), Verona, NY, USA, 26–28 May 2015; IEEE: New York, NY, USA, 2015; pp. 1–8. [Google Scholar] [CrossRef]
- Lu, J.; Liu, A.; Dong, F.; Gu, F.; Gama, J.; Zhang, G. Learning under Concept Drift: A Review. IEEE Trans. Knowl. Data Eng. 2018, 31, 2346–2363. [Google Scholar] [CrossRef]
- Villamizar, H.; Escovedo, T.; Kalinowski, M. Requirements Engineering for Machine Learning: A Systematic Mapping Study. In Proceedings of the 2021 47th Euromicro Conference on Software Engineering and Advanced Applications (SEAA), Palermo, Italy, 1–3 September 2021; IEEE: New York, NY, USA, 2021; pp. 29–36. [Google Scholar] [CrossRef]
- Demir, E. The protection of human biodata: Is there any role for data ownership? Comput. Law Secur. Rev. 2023, 51, 105905. [Google Scholar] [CrossRef]
- Doran, G.T. There’s a SMART Way to Write Managements’s Goals and Objectives. Manag. Rev. 1981, 70, 35. [Google Scholar]
- Kruchten, P. An Ontology of Architectural Design Decisions in Software Intensive Systems. In Proceedings of the 2nd Groningen Workshop on Software Variability, Groningen, The Netherlands, 20–24 August 2012; pp. 54–61. [Google Scholar]
- Arif, S.; Amjad, M.U.; Faisal, M. AI-Driven Decision Support Systems for Software Architecture: A Framework for Intelligent Design Decision-Making (2025). J. Cloud Artif. Intell. 2025, 3, 1–32. [Google Scholar] [CrossRef]
- Zhou, Y.; Tu, F.; Sha, K.; Ding, J.; Chen, H. A Survey on Data Quality Dimensions and Tools for Machine Learning Invited Paper. In Proceedings of the 2024 IEEE International Conference on Artificial Intelligence Testing (AITest), Shanghai, China, 15–18 July 2024; IEEE: New York, NY, USA, 2024; pp. 120–131. [Google Scholar] [CrossRef]
- Wasser, J.; Kumara, I.; Monsieur, G.; Van Den Heuvel, W.J.; Tamburri, D.A. Data Contracts in Data Mesh: A Systematic Gray Literature Review. In Business Modeling and Software Design; Shishkov, B., Ed.; Springer Nature: Geneva, Switzerland, 2025; Volume 559, pp. 21–38. [Google Scholar] [CrossRef]
- Heltweg, P.; Riehle, D. A Systematic Analysis of Problems in Open Collaborative Data Engineering. J. Data Inf. Qual. 2023, 6, 1–30. [Google Scholar] [CrossRef]
- Leventon, J.; Fleskens, L.; Claringbould, H.; Schwilch, G.; Hessel, R. An applied methodology for stakeholder identification in transdisciplinary research. Sustain. Sci. 2016, 11, 763–775. [Google Scholar] [CrossRef]
- Miller, G.J. Stakeholder roles in artificial intelligence projects. Proj. Leadersh. Soc. 2022, 3, 100068. [Google Scholar] [CrossRef]
- Brugha, R.; Varvasovszky, Z. Stakeholder analysis: A review. Health Policy Plan. 2000, 15, 239–246. [Google Scholar] [CrossRef]
- Suresh, H.; Gomez, S.R.; Nam, K.K.; Satyanarayan, A. Beyond Expertise and Roles: A Framework to Characterize the Stakeholders of Interpretable Machine Learning and their Needs. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, Yokohama, Japan, 8–13 May 2021; ACM: New York, NY, USA, 2021; pp. 1–16. [Google Scholar] [CrossRef]
- Walker, D.H.T.; Bourne, L.M.; Shelley, A. Influence, stakeholder mapping and visualization. Constr. Manag. Econ. 2008, 26, 645–658. [Google Scholar] [CrossRef]
- Ninan, J.; Mahalingam, A.; Clegg, S.; Sankaran, S. ICT for external stakeholder management: Sociomateriality from a power perspective. Constr. Manag. Econ. 2020, 38, 840–855. [Google Scholar] [CrossRef]
- Poszler, F.; Portmann, E.; Lütge, C. Formalizing ethical principles within AI systems: Experts’ opinions on why (not) and how to do it. AI Ethics 2025, 5, 937–965. [Google Scholar] [CrossRef]
- Das, S.; Hossain, M.; Shiva, S.G. Requirements Elicitation and Stakeholder Communications for Explainable Machine Learning Systems: State of the Art. In Proceedings of the 2023 International Conference on Information Technology (ICIT), Amman, Jordan, 9–10 August 2023; IEEE: New York, NY, USA, 2023; pp. 561–566. [Google Scholar] [CrossRef]
- Imrie, F.; Davis, R.; van der Schaar, M. Multiple stakeholders drive diverse interpretability requirements for machine learning in healthcare. Nat. Mach. Intell. 2023, 5, 824–829. [Google Scholar] [CrossRef]
- Cantor, A.; Kiparsky, M.; Hubbard, S.S.; Kennedy, R.; Pecharroman, L.C.; Guivetchi, K.; Bales, R. Making a Water Data System Responsive to Information Needs of Decision Makers. Front. Clim. 2021, 3, 761444. [Google Scholar] [CrossRef]
- Asswad, J.; Marx Gómez, J. Data Ownership: A Survey. Information 2021, 12, 465. [Google Scholar] [CrossRef]
- Martens, B. The Importance of Data Access Regimes for Artificial Intelligence and Machine Learning. SSRN Electron. J. 2018, 1–23. [Google Scholar] [CrossRef]
- Maharana, K.; Mondal, S.; Nemade, B. A review: Data pre-processing and data augmentation techniques. Glob. Transitions Proc. 2022, 3, 91–99. [Google Scholar] [CrossRef]
- Borrohou, S.; Fissoune, R.; Badir, H. The role of data transformation in modern analytics: A comprehensive survey. J. Comput. Lang. 2025, 84, 101329. [Google Scholar] [CrossRef]
- Shamsinejad, E.; Banirostam, T.; Pedram, M.M.; Rahmani, A.M. A Review of Anonymization Algorithms and Methods in Big Data. Ann. Data Sci. 2025, 12, 253–279. [Google Scholar] [CrossRef]
- Hutchinson, B.; Smart, A.; Hanna, A.; Denton, R.; Greer, C.; Kjartansson, O.; Mitchell, M. Towards Accountability for Machine Learning Datasets: Practices from Software Engineering and Infrastructure. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, Toronto, ON, Canada, 3–10 March 2021; ACM: New York, NY, USA, 2021; pp. 560–575. [Google Scholar] [CrossRef]
- Werder, K.; Ramesh, B.; Zhang, R. Establishing Data Provenance for Responsible Artificial Intelligence Systems. ACM Trans. Web 2022, 13, 1–23. [Google Scholar] [CrossRef]
- Janssen, M.; Brous, P.; Estevez, E.; Barbosa, L.S.; Janowski, T. Data governance: Organizing data for trustworthy Artificial Intelligence. Gov. Inf. Q. 2020, 37, 101493. [Google Scholar] [CrossRef]
- Gjorgjevikj, A.; Mishev, K.; Antovski, L.; Trajanov, D. Requirements Engineering in Machine Learning Projects. IEEE Access 2023, 11, 72186–72208. [Google Scholar] [CrossRef]
- Dey, S.; Lee, S.W. A Multi-layered Collaborative Framework for Evidence-driven Data Requirements Engineering for Machine Learning-based Safety-critical Systems. In Proceedings of the 38th ACM/SIGAPP Symposium on Applied Computing, Tallinn, Estonia, 27–31 March 2023; ACM: New York, NY, USA, 2023; pp. 1404–1413. [Google Scholar] [CrossRef]
- Umm-E-Habiba; Habibullah, K.M. Explainable AI: A Diverse Stakeholder Perspective. In Proceedings of the 2024 IEEE 32nd International Requirements Engineering Conference (RE), Reykjavik, Iceland, 24–28 June 2024; IEEE: New York, NY, USA, 2024; pp. 494–495. [Google Scholar] [CrossRef]
- Tun, H.T.; Husen, J.H.; Yoshioka, N.; Washizaki, H.; Fukazawa, Y. Goal-Centralized Metamodel Based Requirements Integration for Machine Learning Systems. In Proceedings of the 2021 28th Asia-Pacific Software Engineering Conference Workshops (APSEC Workshops), Taipei, Taiwan, 6–9 December 2021; IEEE: New York, NY, USA, 2021; pp. 13–16. [Google Scholar] [CrossRef]
- Alsaggaf, F.; Eijkelenboom, A.; Lugten, M.; Turrin, M. A machine learning framework for early-stage dementia-friendly architectural design evaluation using visual access metrics. Comput. Intell. 2025, 4, 11. [Google Scholar] [CrossRef]
- Kücherer, C.; Gerasch, L.; Junger, D.; Burgert, O. Elicitation and Documentation of Explainability Requirements in a Medical Information Systems Context. In Proceedings of the 27th International Conference on Enterprise Information Systems, Porto, Portugal, 4–6 April 2025; SCITEPRESS-Science and Technology Publications: Setubal, Portugal, 2025; pp. 83–94. [Google Scholar] [CrossRef]
- Alebrahim, A.; Choppy, C.; Faßbender, S.; Heisel, M. Optimizing Functional and Quality Requirements According to Stakeholders’ Goals. In Relating System Quality and Software Architecture; Elsevier: Amsterdam, The Netherlands, 2014; pp. 75–120. [Google Scholar] [CrossRef]
- Schlegel, M.; Sattler, K.U. Management of Machine Learning Lifecycle Artifacts. ACM SIGMOD Rec. 2023, 51, 18–35. [Google Scholar] [CrossRef]
- Polyzotis, N.; Roy, S.; Whang, S.E.; Zinkevich, M. Data Lifecycle Challenges in Production Machine Learning: A Survey. ACM SIGMOD Rec. 2018, 47, 17–28. [Google Scholar] [CrossRef]
- Ryan, S.; Cai, W.; Bowman, R.; Doherty, G. Fairness Challenges in the Design of Machine Learning Applications for Healthcare. ACM Trans. Comput. Hum. Interact. 2025, 6, 1–26. [Google Scholar] [CrossRef]
- Puyt, R.W.; Lie, F.B.; Wilderom, C.P. The origins of SWOT analysis. Long Range Plan. 2023, 56, 102304. [Google Scholar] [CrossRef]
- Sanaei, R.; Otto, K.; Hölttä-Otto, K.; Luo, J. Trade-Off Analysis of System Architecture Modularity Using Design Structure Matrix. In Proceedings of the Volume 2B: 41st Design Automation Conference, Boston, MA, USA, 2 August 2015; American Society of Mechanical Engineers: New York, NY, USA, 2016; p. V02bt03a037. [Google Scholar] [CrossRef]
- Yoshinaga, K.; Kato, T.; Kai, Y. Clustering Method Of Design Structure Matrix For Trade-off Relationships. J. Sci. Des. 2017, 1, 1_77–1_86. [Google Scholar] [CrossRef]
- Luo, Y.; Tseng, H.H.; Cui, S.; Wei, L.; ten Haken, R.K.; El Naqa, I. Balancing accuracy and interpretability of machine learning approaches for radiation treatment outcomes modeling. BJR Open 2019, 1, 20190021. [Google Scholar] [CrossRef]
- Li, J.; Li, G. Triangular Trade-off between Robustness, Accuracy, and Fairness in Deep Neural Networks: A Survey. ACM Comput. Surv. 2025, 57, 1–40. [Google Scholar] [CrossRef]
- Pham, N.D.; Phan, K.T.; Chilamkurti, N. Enhancing Accuracy-Privacy Trade-Off in Differentially Private Split Learning. IEEE Trans. Emerg. Top. Comput. Intell. 2025, 9, 988–1000. [Google Scholar] [CrossRef]
- Brownlee, A.E.; Adair, J.; Haraldsson, S.O.; Jabbo, J. Exploring the Accuracy–Energy Trade-off in Machine Learning. In Proceedings of the 2021 IEEE/ACM International Workshop on Genetic Improvement (GI), Madrid, Spain, 30 May 2021; IEEE: New York, NY, USA, 2021; pp. 11–18. [Google Scholar] [CrossRef]
- Bhat, M.; Shumaiev, K.; Biesdorf, A.; Hohenstein, U.; Matthes, F. Automatic Extraction of Design Decisions from Issue Management Systems: A Machine Learning Based Approach. In Software Architecture; Lopes, A., de Lemos, R., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2017; Volume 10475, pp. 138–154. [Google Scholar] [CrossRef]
- Leest, J.; Gerostathopoulos, I.; Raibulet, C. Evolvability of Machine Learning-based Systems: An Architectural Design Decision Framework. In Proceedings of the 2023 IEEE 20th International Conference on Software Architecture Companion (ICSA-C), L’Aquila, Italy, 13–17 March 2023; IEEE: Berlin/Heidelberg, Germany, 2023; pp. 106–110. [Google Scholar] [CrossRef]
- Charoenwut, P.; Drobnjakovic, M.; Oh, H.; Nikolov, A.; Kulvatunyou, B. Machine Learning Lifecycle Explorer: Leverage the Lifecycle Metadata (onto) Logically. In Advances in Production Management Systems. Cyber-Physical-Human Production Systems: Human-AI Collaboration and Beyond; Mizuyama, H., Morinaga, E., Nonaka, T., Kaihara, T., von Cieminski, G., Romero, D., Eds.; Springer Nature: Geneva, Switzerland, 2025; Volume 769, pp. 370–385. [Google Scholar] [CrossRef]
- Mitchell, M.; Wu, S.; Zaldivar, A.; Barnes, P.; Vasserman, L.; Hutchinson, B.; Gebru, T. Model Cards for Model Reporting. In Proceedings of the Conference on Fairness, Accountability, and Transparency, Atlanta, GA, USA, 29–31 January 2019; ACM: New York, NY, USA, 2019; pp. 220–229. [Google Scholar] [CrossRef]
- Nicoletti, M.; Diaz-Pace, J.A.; Schiaffino, S.; Tommasel, A.; Godoy, D. Personalized architectural documentation based on stakeholders’ information needs. J. Enterp. Inf. Manag. 2014, 2, 9. [Google Scholar] [CrossRef]
- Parlak, İ.E. Blockchain-assisted explainable decision traces (BAXDT): An approach for transparency and accountability in artificial intelligence systems. Knowl.-Based Syst. 2025, 329, 114402. [Google Scholar] [CrossRef]
- Karkoskova, S.; Novotny, O. Design and Application on Business Data Lineage as a part of Metadata Management. In Proceedings of the 2021 International Conference on Computers and Automation (CompAuto), Paris, France, 7–9 September 2021; IEEE: New York, NY, USA, 2021; pp. 34–39. [Google Scholar] [CrossRef]
- Silva, C.; Abbasi, M.; Martins, P.; Silva, J.; Váz, P.; Mota, D. Enhancing Organizational Data Integrity and Efficiency through Effective Data Lineage. In Proceedings of the 2023 Second International Conference on Smart Technologies for Smart Nation (SmartTechCon), Singapore, 18–19 August 2023; IEEE: New York, NY, USA, 2023; pp. 312–317. [Google Scholar] [CrossRef]
- Wohlrab, R.; Knauss, E.; Steghöfer, J.P.; Maro, S.; Anjorin, A.; Pelliccione, P. Collaborative traceability management: A multiple case study from the perspectives of organization, process, and culture. Requir. Eng. 2020, 25, 21–45. [Google Scholar] [CrossRef]
- Stone, J.; Patel, R.; Ghiasi, F.; Mittal, S.; Rahimi, S. Navigating MLOps: Insights into Maturity, Lifecycle, Tools, and Careers. In Proceedings of the 2025 IEEE Conference on Artificial Intelligence (CAI), Santa Clara, CA, USA, 5–7 May 2025; IEEE: New York, NY, USA, 2025; pp. 643–650. [Google Scholar] [CrossRef]
- Mavromatis, I.; Katsaros, K.; Khan, A. Computing Within Limits: An Empirical Study of Energy Consumption in ML Training and Inference. In Proceedings of the 1st International Workshop on Artificial Intelligence for Sustainable Development, Pescara, Italy, 3 July 2024. [Google Scholar] [CrossRef]
- Tyagi, A.J. Scaling deep learning models: Challenges and solutions for large-scale deployments. World J. Adv. Eng. Technol. Sci. 2025, 16, 010–020. [Google Scholar] [CrossRef]
- Vijayan, N.E. Building Scalable MLOps: Optimizing Machine Learning Deployment and Operations. Int. J. Sci. Res. Eng. Manag. 2024, 8, 1–9. [Google Scholar] [CrossRef]
- Du, K.L.; Zhang, R.; Jiang, B.; Zeng, J.; Lu, J. Understanding Machine Learning Principles: Learning, Inference, Generalization, and Computational Learning Theory. Mathematics 2025, 13, 451. [Google Scholar] [CrossRef]
- Voinov, A.; Jenni, K.; Gray, S.; Kolagani, N.; Glynn, P.D.; Bommel, P.; Smajgl, A. Tools and methods in participatory modeling: Selecting the right tool for the job. Environ. Model. Softw. 2018, 109, 232–255. [Google Scholar] [CrossRef]
- Sakovich, N.; Aksenov, D.; Pleshakova, E.; Gataullin, S. A Neural Operator Using Dynamic Mode Decomposition Analysis to Approximate Partial Differential Equations. SSRN Electron. J. 2025, 10, 22432–22444. [Google Scholar] [CrossRef]
- Slupczynski, M.; Klamma, R. Vorschlag einer Infrastruktur zur verteilten Datenanalyse von multimodalen psychomotorischen Lerneraktivitäten; Technical Report; RWTH Aachen University: Aachen, Germany, 2021. [Google Scholar] [CrossRef]
- Slupczynski, M.; Klamma, R. MILKI-PSY Cloud: Facilitating Multimodal Learning Analytics by Explainable AI and Blockchain. In Proceedings of the Multimodal Immersive Learning Systems 2021 (MILeS 2021), Bozen-Bolzano, Italy, 20–24 September 2021; Klemke, R., Di Mitri, D., Ciordas-Hertel, G.P., Eds.; CEUR Workshop Proceedings. Volume 2979, pp. 22–28. [Google Scholar]
- Slupczynski, M.; Klamma, R. MILKI-PSY Cloud: MLOps-based Multimodal Sensor Stream Processing Pipeline for Learning Analytics in Psychomotor Education. In Proceedings of the Multimodal Immersive Learning Systems 2022 (MILeS 2022), Toulouse, France, 12–16 September 2022; Sanusi, K.A.M., Limbu, B., Schneider, J., Eds.; CEUR Workshop Proceedings. pp. 8–14. [Google Scholar]
- Mat Sanusi, K.A.; Slupczynski, M.; Geisen, M.; Iren, D.; Klamma, R.; Klatt, S.; Klemke, R. IMPECT-Sports: Using an Immersive Learning System to Facilitate the Psychomotor Skills Acquisition Process. In Proceedings of the Multimodal Immersive Learning Systems 2022, Touluse, France, 12–16 September 2022; Sanusi, K.A.M., Limbu, B., Schneider, J., Di Mitri, D., Klemke, R., Eds.; CEUR Workshop Proceedings. pp. 34–39. [Google Scholar]
- Gao, C. Sensor Based Human Motion Comparison. Master Thesis, RWTH Aachen University, Aachen, Germany, 2023. [Google Scholar]
- Slupczynski, M.P.; Decker, S.J. Semantic Motion Learning Analytics and Feedback System; Universitätsbibliothek der RWTH Aachen: Aachen, Germany, 2024. [Google Scholar] [CrossRef]
- Slupczynski, M.; Mat Sanusi, K.A.; Majonica, D.; Klemke, R.; Decker, S. Implementing Cloud-Based Feedback to Facilitate Scalable Psychomotor Skills Acquisition. In Proceedings of the Multimodal Immersive Learning Systems 2023, Aveiro, Portugal, 4–8 September 2023; Sanusi, K.A.M., Limbu, B., Schneider, J., Kravcik, M., Klemke, R., Eds.; CEUR Workshop Proceedings. Volume 3499, pp. 36–43. [Google Scholar]
- Li, S. From Still to Dynamic: A Comparative Analysis of Visualization Techniques for Human Movements. Bachelor’s Thesis, RWTH Aachen University, Aachen, Germany, 2024. [Google Scholar]
- Geisen, M.; Seifriz, F.; Fasold, F.; Slupczynski, M.; Klatt, S. A Novel Approach to Sensor-Based Motion Analysis for Sports: Piloting the Kabsch Algorithm in Volleyball and Handball. IEEE Sens. J. 2024, 24, 35654–35663. [Google Scholar] [CrossRef]
- Slupczynski, M.; Nekhviadovich, A.; Duong-Trung, N.; Decker, S. Analyzing Exercise Repetitions: YOLOv8-Enhanced Dynamic Time Warping Approach on InfiniteRep Dataset. In Sensor-Based Activity Recognition and Artificial Intelligence; Konak, O., Arnrich, B., Bieber, G., Kuijper, A., Fudickar, S., Eds.; Springer Nature: Potsdam, Germany, 2024; Volume 15357, pp. 94–110. [Google Scholar] [CrossRef]
- Slupczynski, M.; Mat Sanusi, K.A.; Klemke, R.; Decker, S. SpaceStriker: A Peer-Assisted Sensor-Based Exergame with Real-Time ML-Driven Feedback. In Sensor-Based Activity Recognition and Artificial Intelligence; Konak, O., Fudickar, S., Arnrich, B., Bieber, G., Kuijper, A., Eds.; Springer Nature: Geneva, Switzerland, 2025; Volume 16292. [Google Scholar] [CrossRef]
- Singh, D.; Slupczynski, M.; Pillai, A.G.; Pandian, V.P.S. Grounding Explainability Within the Context of Global South in XAI. arXiv 2022, arXiv:2205.06919. [Google Scholar] [CrossRef]
- Cooper, R.G. The 5-Th Generation Stage-Gate Idea-to-Launch Process. IEEE Eng. Manag. Rev. 2022, 50, 43–55. [Google Scholar] [CrossRef]
- Sadek, M.; Constantinides, M.; Quercia, D.; Mougenot, C. Guidelines for Integrating Value Sensitive Design in Responsible AI Toolkits. In Proceedings of the CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 11–16 May 2024; pp. 1–20. [Google Scholar] [CrossRef]




| Framework | Key Elements | Strengths | Gaps and Limitations |
|---|---|---|---|
| ISO/IEC 22989 [16]: Artificial intelligence Standardized system-level AI lifecycle framework |
|
|
|
| ISO 9241 [3]: Human-Centered Design (HCD) HCD process standard for interactive systems |
|
|
|
| CRISP-ML(Q) [2] Quality-assured ML workflow extending CRISP-DM |
|
|
|
| SE4ML [17] Empirically observed industrial ML practices |
|
|
|
| NIST AI RMF Risk-centric governance framework for trustworthy AI | Non-sequential risk functions:
|
|
|
| Phase | HCD Principle | Focus | Risk if Conflated |
|---|---|---|---|
| CONSULT | Understand context | Stakeholder mapping | Unidentified stakeholders excluded from governance |
| ARTICULATE | Specify requirements | SMART formalization | Unmeasurable requirements evade validation |
| PROTOCOL | Produce solutions | Decision provenance | Untraceable decisions undermine accountability |
| Aspect | REIFY | EVOLVE |
|---|---|---|
| Purpose | Deployment, evidence collection | Impact assessment, drift analysis |
| Stakeholder role | Observed/feedback providers | Decision authorities |
| ISO 22989 alignment | Operation/Monitor | Re-evaluate |
| Risk if conflated | Premature decisions before sufficient evidence; reactive rather than evidence-based iteration | |
| Phase | CONSULT–Engage and Define |
| Extends |
– ISO/IEC 22989: Inception; – CRISP-ML(Q): Business understanding (+sensor-derived constraints); – ISO 9241-210: Understand and specify context of use (Principle 1); – SE4ML: Problem formulation (+multi-stakeholder elicitation) |
| Data Context | Data as a collaborative artifact [60]; stakeholder context defining who stakeholders are, what constraints govern data collection/use, and what trade-offs are acceptable |
| Core Activities | Stakeholder analysis and mapping; constraint identification (ethics, legal, domain-specific, latency); model input/output requirements; preliminary KPI definition; collaborative requirements derivation |
| V&V Focus | Verification conditions framing; Explicit mapping to ISO/IEC 22989 inception phase |
| Decision Question | Who are the stakeholders and what are their requirements, information needs, constraints, and success criteria? |
| Key Outputs | Stakeholder maps; design principle collection; system and data requirements list |
| Transition | CONSULT → ARTICULATE |
| ISE Challenge | Traditional Requirements Engineering lacks systematic capture of ML-specific constraints (sensor calibration, data ownership, multi-stakeholder structures, AI ethics) |
| Gate Trigger Criteria |
– Mandatory: All stakeholder groups identified, representatives consulted, domain constraints catalogued; project initiation approved – Quality: Stakeholder power mapping completed, KPIs identified for majority of groups – Documentation: Stakeholder map, design principle collection, requirements list |
| Decision Question | Do we understand the stakeholder landscape sufficiently to formalize requirements? |
| Approvers | Project lead + senior stakeholder representative |
| Phase | ARTICULATE—Formalize Requirements |
| Extends |
– ISO/IEC 22989: Early design; – CRISP-ML(Q): Data understanding (+formalization layer); – ISO 9241-210: Specify user requirements (Principle 2); – SE4ML: Data quality assessment (+formal requirements and contracts) |
| Data Context | Data as a requirement; translation layer converting stakeholder-defined constraints into measurable, testable acceptance criteria (SMART [55]) |
| Core Activities | Formalization of design principles and requirements; translation of stakeholder goals to SMART requirements; trade-off optimization; data contract establishment; KPI formalization |
| V&V Focus | Requirement verifiability; KPI definition for data, code, and models; technical and non-technical metrics specification |
| Decision Question | Are the requirements SMART? What are the trade-offs? What are the KPIs? |
| Key Outputs | Requirement matrices (SWOT, DSM, decision matrices, QFD); trade-off models (accuracy vs. interpretability, robustness vs. fairness, accuracy vs. privacy/energy) |
| Transition | ARTICULATE → PROTOCOL |
| ISE Challenge | RE4AI gaps in formalizing ML-specific requirements (explainability, fairness, data quality); trade-offs require explicit stakeholder negotiation |
| Gate Trigger Criteria |
– Mandatory: Requirements are SMART, data contracts established, KPIs formalized with measurement methods, trade-offs documented – Quality: Requirements validated by decision authorities, data contracts include schema and quality thresholds – Documentation: Requirement matrix, trade-off models, data quality metric definitions |
| Decision Question | Can we commit to these requirements for this iteration, understanding they may evolve in EVOLVE phase? |
| Approvers | Requirements owners + data governance lead + stakeholder representatives |
| Phase | PROTOCOL—Document Decisions |
| Extends |
– ISO/IEC 22989: Design and documentation; – CRISP-ML(Q): New capability (traceability not in CRISP-ML(Q)); – ISO 9241-210: Produce design solutions (Principle 3); – SE4ML: New capability (explicit traceability and decision tracking) |
| Data Context | Data as a traceable artifact |
| Core Activities | Decision tracking using semantic data structures; traceability maintenance across artifacts; template-based communication (model cards, dataset cards); metadata and label synchronization; data quality metrics definition; data provenance and lineage tracking |
| V&V Focus | Requirements verification; traceability validation |
| Decision Question | Are design decisions documented and traceable to stakeholder requirements? |
| Key Outputs | Versioned design documentation; traceability links to stakeholders and requirements; defined data rules enforcing data contracts |
| Transition | PROTOCOL → TERRAFORM |
| ISE Challenge | Decision provenance tracking gap in ML frameworks; lack of explicit design rationale documentation |
| Gate Trigger Criteria |
– Mandatory: Decision tracking mechanism established, data quality metrics defined, data rules specified, provenance requirements documented – Quality: Traceability links created, sensor physical-world attributes documented, versioning schema designed – Documentation: Versioned design docs, traceability matrix (stakeholders → requirements → decisions), data governance policies |
| Decision Question | Do we have sufficient design documentation to build infrastructure that enforces our data contracts and governance policies? |
| Approvers | System architect + data governance lead |
| Phase | TERRAFORM—Build Infrastructure |
| Extends |
– ISO/IEC 22989: Development and integration; – CRISP-ML(Q): No analogue (infrastructure lifecycle not specified); – ISO 9241-210: Produce design solutions (Principle 3); – SE4ML: New capability (sensor pipelines and MLOps infrastructure) |
| Data Context | Data as an infrastructural flow; focus on data pipeline architecture and movement rather than model-specific processing |
| Core Activities | MLOps automation setup (CI/CD); data pipeline architecture design; ETL pipeline setup; hardware and software provisioning; monitoring infrastructure establishment; governance framework implementation |
| V&V Focus | Dataflow integrity verification; automated governance flows |
| Decision Question | Is the infrastructure ready to support ML development and deployment according to stakeholder requirements? |
| Key Outputs | Functional infrastructure prototypes; versioning schema; deployment pipelines; benchmark data (latency, scalability) |
| Transition | TERRAFORM → UTILIZE |
| ISE Challenge | ML frameworks treat infrastructure as implicit precondition; governance implemented reactively rather than proactively |
| Gate Trigger Criteria | – Mandatory: Data pipeline functional, data rules enforced, KPI monitoring operational, data quality metrics tracked, infrastructure passes dataflow integrity tests – Quality: Versioning operational, audit trails functional, benchmark data collected, sensor streaming/buffering/synchronization operational – Documentation: Infrastructure docs, deployment pipelines, operational runbooks |
| Decision Question | Can this infrastructure reliably support ML model training and/or inference at required scale/performance? |
| Approvers | Infrastructure lead + MLOps engineer + data governance lead |
| Phase | UTILIZE—Implement ML Components |
| Extends |
– ISO/IEC 22989: Development and V&V; – CRISP-ML(Q): Model engineering (+embedded V&V); – ISO 9241-210: Evaluate solutions against requirements (Principle 4, part 1: model validation); – SE4ML: Model evaluation (+continuous V&V across infrastructure and data) |
| Data Context | Data as training input or inference input; focus on model-specific data processing |
| Core Activities | ML component development; model selection, training, fine-tuning, or inference; data management for ML tasks; traceability maintenance |
| V&V Focus | Model behavior validation; KPI verification; data quality monitoring during training and inference; drift detection |
| Decision Question | Do the ML models and applications meet the specified requirements and KPIs? |
| Key Outputs | Integrated dataflow pipelines; ML models, components, and decision mechanisms |
| Transition | UTILIZE → REIFY |
| ISE Challenge | Traditional ML lifecycles assume training-centric workflows; fail to accommodate pre-trained model selection and domain-specific validation |
| Gate Trigger Criteria |
– Mandatory: ML models trained/fine-tuned/selected and validated, KPIs met (or deviations documented with stakeholder approval), data quality metrics acceptable, model behavior validated, reproducibility verified – Quality: Traceability links established (requirements → data → code → models), model cards created, explainability and fairness requirements satisfied – Documentation: Integrated dataflow pipelines, ML model documentation, test reports |
| Decision Question | Is this model ready for deployment in real stakeholder contexts, understanding that real-world feedback may require iteration? |
| Approvers | ML engineer + domain expert + stakeholder representative |
| Phase | REIFY—Apply and Obtain Feedback |
| Extends |
– ISO/IEC 22989: Deployment and monitoring; – CRISP-ML(Q): Deployment (+feedback loops); – ISO 9241-210: Evaluate solutions against requirements (Principle 4, part 2: real-world feedback); – SE4ML: Monitoring (+systematic real-world feedback collection) |
| Data Context | Data as feedback artifact [24,58]; real-world evidence collection |
| Core Activities | Stakeholder context deployment; automated insights capture; stakeholder/user feedback collection; real-world KPI measurement; production data quality monitoring |
| V&V Focus | Real-world outcome confirmation |
| Decision Question | Do we have sufficient real-world evidence to evaluate system performance and stakeholder satisfaction? |
| Key Outputs | Case studies and use cases; feedback datasets; updated stakeholder requirements |
| Transition | REIFY → EVOLVE |
| ISE Challenge | ML frameworks conflate deployment with evaluation; missing separation between feedback collection (observation) and impact assessment (decision-making) |
| Gate Trigger Criteria |
– Mandatory: System operational in at least one stakeholder context, real-world KPIs measured over statistically significant period, stakeholder feedback gathered – Quality: Automated insights captured, feedback in decision tracking system, at least one complete use case documented – Documentation: Feedback datasets, real-world performance reports, updated stakeholder requirements (if emerged) |
| Decision Question | Do we have enough real-world data to evaluate system success and plan next iteration? |
| Approvers | Product owner + stakeholder representatives |
| Phase | EVOLVE—Evaluate and Iterate |
| Extends |
– ISO/IEC 22989: Re-evaluate (+triggers retirement); – CRISP-ML(Q): Monitoring and maintenance (+systematic re-evaluation); – ISO 9241-210: Evaluate solutions against requirements (Principle 4, part 3: iteration decisions); – SE4ML: Retraining (+drift detection and structured iteration paths) |
| Data Context | Data as a learning opportunity; evidence for iteration decisions |
| Core Activities | Data-driven evaluation (technical and social impact); KPI achievement assessment; data quality trend analysis; requirements gathering for next iteration |
| V&V Focus | Model and concept drift assessment; decision impact evaluation |
| Decision Question | What action does the evidence warrant (major/minor iteration, model update, steady-state, or retirement)? |
| Key Outputs | Evaluation reports (quantitative and qualitative); lessons learned and best practices; updated models and data/system requirements |
| Transition | EVOLVE → Multiple strategic paths |
| ISE Challenge | Traditional software maintenance models do not capture multi-modal ML iteration strategies (continuous drift, automated updates, graceful retirement) |
| Gate Trigger Criteria |
– Mandatory: Impact evaluation completed (technical + social), KPI achievement assessed, data quality trends analyzed, drift analysis performed – Quality: Lessons learned documented, stakeholder satisfaction assessed – Documentation: Evaluation reports (quantitative and qualitative), updated models and requirements (if iterating), retirement plan (if retiring) |
| Strategic Options |
– 7(a): Major Iteration → CONSULT – 7(b): Minor Iteration → ARTICULATE – 7(c): Model Update → UTILIZE – 7(d): Continuous Monitoring – 7(e): Retirement |
| Decision Question | What iteration strategy does the evidence warrant? |
| Approvers (varies by option) | 7(a): Steering committee; 7(b): Product owner + stakeholders; 7(c): ML engineer + operations lead; 7(d): Operations lead; 7(e): Steering committee + compliance officer |
| Year | Publication | CAPTURE Phase(s) | Key Activities |
|---|---|---|---|
| 2021 | [112,113] | CONSULT, ARTICULATE, PROTOCOL, TERRAFORM | Initial stakeholder consultation, infrastructure conceptualization, security standards analysis, technology evaluation |
| [114] | TERRAFORM, REIFY | MLOps pipeline implementation, real-time communication infrastructure | |
| 2022 | [115] | ARTICULATE, REIFY | Feedback model design guidelines, ILS prototype deployment |
| [116] | UTILIZE, EVOLVE | Motion comparison algorithm development for whole-skeleton similarity, limitation identification | |
| [117] | ARTICULATE | Conceptual semantic motion notation proposal (not implemented) | |
| 2023 | [118] | TERRAFORM, REIFY | Session management infrastructure, ML-based feedback generation |
| [119] | ARTICULATE, EVOLVE | Visualization techniques research and comparative analysis | |
| [120] | UTILIZE, EVOLVE | Motion comparison algorithm validation in sports, key pose limitation identification | |
| 2024 | [121] | UTILIZE, REIFY, EVOLVE | Rule-based feedback engine implementation, keypoint detection integration |
| 2025 | [122] | REIFY, EVOLVE | Peer-assisted exergame prototype, gamification integration |
| Phase | Tier 1: Lightweight (Research Prototypes) | Tier 2: Standard (Industrial-Grade) | Tier 3: Full (Safety-Critical) |
|---|---|---|---|
| CONSULT | Informal mapping | Documented mapping | Formal audit trail |
| ARTICULATE | Lightweight reqs | Data contracts | Full V&V criteria |
| PROTOCOL | Minimal tracking | Decision tracking | Rigorous provenance |
| TERRAFORM | Ad hoc setup | CI/CD pipelines | Full MLOps governance |
| UTILIZE | Experimental | Validated models | Certified models |
| REIFY | Pilot deployment | Monitored ops | Audited operations |
| EVOLVE | Informal review | Structured gates | Governed transitions |
| Gate thresholds | = Low | = Medium | = High |
| Domain | Tier | (Stakeholder) | (KPI) | (Doc) | (Threshold) |
|---|---|---|---|---|---|
| Research prototype | Lightweight | 0.2 | 0.3 | 0.5 | 0.5 |
| Industrial analytics | Standard | 0.4 | 0.3 | 0.3 | 0.7 |
| Medical rehabilitation | Full | 0.5 | 0.3 | 0.2 | 0.85 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Slupczynski, M.; Reiners, R.; Decker, S. CAPTURE: A Stakeholder-Centered Iterative MLOps Lifecycle. Appl. Sci. 2026, 16, 1264. https://doi.org/10.3390/app16031264
Slupczynski M, Reiners R, Decker S. CAPTURE: A Stakeholder-Centered Iterative MLOps Lifecycle. Applied Sciences. 2026; 16(3):1264. https://doi.org/10.3390/app16031264
Chicago/Turabian StyleSlupczynski, Michal, René Reiners, and Stefan Decker. 2026. "CAPTURE: A Stakeholder-Centered Iterative MLOps Lifecycle" Applied Sciences 16, no. 3: 1264. https://doi.org/10.3390/app16031264
APA StyleSlupczynski, M., Reiners, R., & Decker, S. (2026). CAPTURE: A Stakeholder-Centered Iterative MLOps Lifecycle. Applied Sciences, 16(3), 1264. https://doi.org/10.3390/app16031264

