Next Article in Journal
A Lightweight Degradation-Aware Framework for Robust Object Detection in Adverse Weather
Previous Article in Journal
Correction: Abbas, Z.; Myeong, S. Enhancing Industrial Cyber Security, Focusing on Formulating a Practical Strategy for Making Predictions Through Machine Learning Tools in Cloud Computing Environment. Electronics 2023, 12, 2650
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CEDAR: An Ontology-Based Framework Using Event Abstractions to Contextualise Financial Data Processes

School of Computer Science and Engineering, University of New South Wales, Sydney, NSW 2052, Australia
*
Author to whom correspondence should be addressed.
Electronics 2026, 15(1), 145; https://doi.org/10.3390/electronics15010145
Submission received: 4 September 2025 / Revised: 24 December 2025 / Accepted: 24 December 2025 / Published: 29 December 2025
(This article belongs to the Special Issue Visual Analysis of Software Engineering Data)

Abstract

Financial institutions face data quality (DQ) challenges in regulatory reporting due to complex architectures where data flows through multiple systems. Data consumers struggle to assess quality because traditional DQ tools operate on data snapshots without capturing temporal event sequences and business contexts that determine whether anomalies represent genuine issues or valid behavior. Existing approaches address either semantic representation (ontologies for static knowledge) or temporal pattern detection (event processing without semantics), but not their integration. This paper presents CEDAR (Contextual Events and Domain-driven Associative Representation), integrating financial ontologies with event-driven processing for context-aware DQ assessment. Novel contributions include (1) ontology-driven rule derivation that automatically translates OWL business constraints into executable detection logic; (2) temporal ontological reasoning extending static quality assessment with event stream processing; (3) explainable assessment tracing anomalies through causal chains to violated constraints; and (4) standards-based design using W3C technologies with FIBO extensions. Following the Design Science Research Methodology, we document the first, early-stage iteration focused on design novelty and technical feasibility. We present conceptual models, a working prototype, controlled validation with synthetic equity derivative data, and comparative analysis against existing approaches. The prototype successfully detects context-dependent quality issues and enables ontological root cause exploration. Contributions: A novel integration of ontologies and event processing for financial DQ management with validated technical feasibility, demonstrating how semantic web technologies address operational challenges in event-driven architectures.

1. Introduction

The 2007–2008 Global Financial Crisis (GFC) was a turning point for the financial sector. Poor data and risk management led to the bankruptcy of the Lehman Brothers. The international Basel Committee on Banking Supervision raised concerns regarding banks’ abilities to accurately and timely define their risk exposures across processes.
Financial institutions (FIs) are regulated domestically to maintain financial market stability and submit standardized reports to regulators, as shown in Table 1.
Current challenges in regulatory reporting stem from the fact that the data are sourced from multiple legacy systems, requires complex transformation requirements, and reporting adheres to strict accuracy requirements and short submission deadlines, with heavy penalities for errors. Post-GFC, regulators have focused on operational risk as a key driver in the accuracy of regulatory reports required. These regulatory reports are derived using enterprise-level data processes, driving FIs towards a data-driven mindset. Regulators are evermore data-centric, adding explicit data risk requirements and overseeing operational reviews. A key aspect of data management is ensuring Data Quality (DQ), i.e., data fit for purpose, throughout data processes. Modern financial markets and FIs create, use and distribute Big Data (highly heterogeneous, quickly generating, large size), complicating the DQ management process.
Data consumers, defined as those receiving data to be used to create the regulatory reports, are at the end of the data process. By this point, several business and technical transformations have been applied to the source data, making it difficult for data consumers to understand the data received and therefore assess its quality. Globally, tech debt-driven issues around complex data architecture, low interoperability and a mixed IT infrastructure led to difficulties in implementing cost effective, efficient and effective DQ management practices.
This challenge is exacerbated by growing technical debt in financial institutions that has led to complex and fragmented data architectures, with limited system interoperability and a mixed technology infrastructure. These factors make it difficult to implement cost-effective, efficient and standardized data quality management practices.
As FIs have evolved into data-driven enterprises, there is an opportunity to uplift DQ processes. Current research has focused on using machine learning methods as part of automated, enterprise DQ management tools. In parallel, FIs, regulators and researchers are currently developing financial ontologies to standardise definitions and taxonomies in the sector. This need has risen as financial markets and instruments have become much more complicated and aggregated sector risk has started to become a key focus for regulators. However, there is a gap as to how these two approaches can be utilised together.
The main contribution of this paper is to present a novel framework underpinned by event-based abstractions to bridge the gap between the use of DQ methods and ontologies in financial data processes. The focus is on data consumers needing to understand what data represents (i.e., its context) before its use. Therefore, the use case of the framework for this paper will be to support data contextualisation as a key process in the DQ management activities of data consumers producing regulatory reports.
The specific problem this paper addresses is the challenge faced by data consumers who receive processed data at the end of complex financial data pipelines. These consumers, typically business users creating regulatory reports, must undertake the following: (1) Understand what the received data represents within their business context. (2) Assess if there are data quality issues before using it for regulatory purposes. (3) Do this assessment efficiently despite limited visibility into upstream transformations. (4) Handle increasing data volumes and complexity while maintaining accuracy.
The structure of this paper is as follows. Section 2 presents a review of the literature on FI architectures, discusses the opportunities and shortcomings of current DQ management methods, and identifies research gaps; Section 3 describes the methodology and design of the proposed framework; Section 4 details the architecture and operation of the framework; Section 5 covers the implementation for the iteration; Section 6 discusses the results of the implementation and evaluates it against discussed criteria; and Section 7 presents future work and closing remarks.

2. Literature Review

2.1. Academic Review

FIs are creating and using big data [1], which is defined as large volumes of heterogeneous data that grow rapidly and is characterised by the ‘V’s of Big Data’ (volume, variety, velocity and veracity) [2]. In this paper, we will focus primarily on veracity, as a indicator of data quality (DQ). DQ is data that are fit for use, and generally assessed from the perspective of the data user [3]. DQ is assessed in accordance with defined dimensions (e.g., accuracy, completeness, consistency) and associated metrics [4]. In academic and industrial literature, many DQ frameworks have been proposed, ranging from generic high-level to industry-specific ones that can be applied throughout different steps of the DQ management process and data lineage [5]. While these frameworks define a DQ management method, their application is limited because of the context of the application of specific dimensions and metrics is generally dataflow dependent and relies on structured data. These frameworks tend to focus on specific points of the data lifecycle as opposed to holistic DQ management [2]. DQ assessment methods also tend to use profiling and data sampling to reduce processing requirements [6]. Automated assessment processes have been designed for enterprise use but are more effective at detecting context independent DQ issues than context-dependent ones [7]. DQ detection techniques, seen in many enterprise tooling [8], have also used models such as machine learning models for pattern detection or rule-based models. However, these methods alone may not be efficient or accurate enough when applied to Big Data due to the curse of dimensionality, difficulty in detecting relevant data attributes, and the speed at which data are created [9].
FIs may experience DQ issues across the enterprise, ranging from untimely reporting to inaccurate or incomplete data capture. These issues are further exasperated through tech debt that has risen from ongoing technological adoption. A key limitation of DQ management for FIs is the data architecture resulting from complex and everchanging business processes creating data and the resulting data lineage [10]. Generally, FIs will have a bow-tie architecture, with many diverse inputs from data producers being standardised via a data processing teams, and then shared with data consumers producing diverse range of outputs (regulatory or otherwise) [11]. This leads to high data heterogeneity and a need to orchestrate and manage metadata as processes that involve different teams and data roles. FIs typically lean towards a centralised enterprise data management process to support Big Data DQ management [1]. As a result, a data consumer producing a regulatory report may not be able to assess the quality of the data received and may therefore be unable to determine what is unusual data, and what the data represents.
With the increase in big data, taxonomies and business knowledge stored via ontologies have become valuable additions to the DQ management toolkit [12]. In parallel, the finance sector is developing open-source ontologies to standardise definitions and processes across institutions and enable regulators to aggregate risk at a sector level [13].
Financial Domain Ontologies. Several ontologies have been developed to formalize financial concepts and support regulatory compliance. The Financial Industry Business Ontology (FIBO) [14] provides machine-readable definitions of financial instruments, organizations, and contractual relationships using OWL-DL, enabling semantic consistency across financial systems. For regulatory reporting, specialised ontologies have been developed for Basel III capital requirements [15], defining concepts such as risk-weighted assets and capital buffers with logical axioms that capture regulatory constraints.
Data Quality Ontologies. Complementing domain-specific ontologies, several frameworks address data quality assessment across domains. The Data Quality Ontology (DQO) [16] models quality dimensions (accuracy, completeness, timeliness) and their relationships, providing a vocabulary for expressing quality metadata. The Data Quality Assessment Methodology (DQAM) [17] builds on this foundation by incorporating semantic reasoning to automatically detect quality issues based on ontological constraints. Similarly, the Total Data Quality Management (TDQM) framework [18] has been formalized as an ontology to support systematic quality improvement processes.
Integration for Financial Data Quality. The relationship between these two categories is synergistic: domain-specific financial ontologies define what concepts mean and their business constraints, while general DQ ontologies provide how to assess quality of those concepts. For instance, FIBO defines that a “derivative contract” must reference an underlying asset, while DQAM provides the framework to check whether this referential integrity constraint is satisfied in actual data. This integration enables automated validation where domain semantics inform quality rules. Modern triple stores such as Stardog [19] support this integration by combining ontology reasoning (validating that data conforms to FIBO definitions) with constraint validation (checking SHACL rules derived from quality requirements), allowing both semantic consistency and data quality checks within a unified infrastructure.
These ontology-based approaches enable the following:
  • Formal, machine-readable definitions of quality dimensions;
  • Domain-specific constraint capture through logical axioms;
  • Automated validation via semantic reasoning;
  • Cross-institutional semantic interoperability.
The rise of financial ontologies has been primarily driven by the Bank of International Settlements’ future vision for a globally standard data infrastructure that operates in real-time with a higher granularity [20]. Many domestic regulators are also signalling future regulation to focus on standardisation, interoperability and data lineage as key pillars of data risk [21]. This gives rise to the potential of ontologies being used to address the challenges posed by the research gaps.
To address these gaps, the focus of this paper is on the following research questions:
  • How can contextual information be represented in datasets used by data consumers producing regulatory reports in a way that helps identifying DQ issues?
  • How can ontologies be leveraged in data processes and align with modern data quality methods in existing FI architectures?

2.2. EvaluationCriteria

The business problem that arises is that it is difficult to have a complete understanding of the transformations occurring in a big data process used for financial regulatory reporting and use this to inform the DQ issues expected throughout the data process. Data consumers will make data observations but lack the tools to contextualise these observations easily. While data architecture in banking focuses on transparent, streamlined data processes, there is a lack of focus on data architecture to support big DQ detection process. It is challenging for business teams who use data (i.e., non-technical data consumers) to understand, control, modify and maintain DQ detection processes used for financial regulatory reporting. This is further exacerbated in the context of big data. Therefore, three evaluation criteria will be used in this paper:
  • Context representation;
  • Scalability and maintainability;
  • Transparency to business users.
Based on academic and industry research, the evaluation criteria can be compared against several approaches an FI can take to alleviate these challenges, as summarised in Table 2.

3. Framework Design

3.1. Methodology

This research follows the Design Science Research Methodology (DSRM) framework [22], which is appropriate when research questions can be addressed through the creation and evaluation of innovative artifacts. DSRM proceeds through six activities: (1) problem identification and motivation, (2) definition of objectives, (3) design and development, (4) demonstration, (5) evaluation, and (6) communication. Importantly, DSRM is inherently iterative—each cycle involves design, demonstration, and evaluation, with findings informing subsequent refinements [23].
Iterative DSRM Process. This paper documents the first design iteration of CEDAR. Following established practices in design science [24,25], early iterations focus on the following:
  • Design novelty: Does the artifact represent a novel solution approach?
  • Technical feasibility: Can the proposed design be implemented with existing technologies?
  • Proof-of-concept: Does the artifact work for simplified scenarios?
Later iterations address:
  • Scalability: Does the artifact perform at realistic scales?
  • Comparative effectiveness: Does it outperform alternative approaches?
  • User utility: Does it provide value in real-world use?
  • Organizational impact: Does deployment improve outcomes?
This staged approach is methodologically sound—attempting comprehensive evaluation before establishing feasibility risks wasting resources on fundamentally flawed designs. As Hevner notes: “The rigor and completeness of design evaluation should be appropriate for the maturity and importance of the artifact” [26].

3.1.1. Methodology Application

The proposed artifact is in the form of a framework named CEDAR, which stands for Contextual Events and Domain-driven Associative Representation. CEDAR relies on different models (see Section 3.2) to contextualise temporal data and an ontology to represent the knowledge across the framework (see Section 3.3). The framework needs to be agile as it is intended to be applicable across any area within this architecture, and adaptable to the data quality requirements of the data processes in scope.
The motivating example used to design and explain the framework’s models will be based on trading an equity-based derivative, a financial instrument where trading patterns can be linked to an underlying equity (such as a stock).

3.1.2. Evaluation Approach

Given that this paper follows DSRM, it is important to contextualize evaluation approaches in related work to appropriately scope our validation strategy.
Evaluation in Ontology-Based Systems. Ontology evaluation typically proceeds through multiple stages [27]: (1) verification of technical correctness (syntax, consistency), (2) validation against competency questions, (3) application-based evaluation with real data, and (4) user studies assessing utility. Early-stage ontology research often focuses on stages 1–2, with stages 3–4 following in subsequent publications. For example, FIBO’s initial publications [14] focused on conceptual design and competency question validation before later work evaluated adoption and impact [28].
Evaluation in Data Quality Frameworks. DQ framework evaluation varies widely in rigor. Batini and Scannapieco [18] distinguish between (1) theoretical frameworks providing conceptual models without implementation, (2) prototypes demonstrating feasibility with small datasets, (3) validated systems with empirical testing on real data, and (4) deployed solutions with longitudinal case studies. The academic literature often presents stages 1–2, while commercial tools report stages 3–4 (though with limited methodological transparency). For instance, Luzzu [16] followed a progression: initial design paper (conceptual framework), implementation paper (prototype with 19 datasets), and adoption study (deployment case studies).
Design Science Evaluation Strategies. Venable et al.’s FEDS framework [25] recommends different evaluation strategies based on artifact maturity and risk. For novel technical designs with high technical risk, artificial evaluation (controlled experiments, synthetic data) should precede naturalistic evaluation (field deployment, real users). For early-stage artifacts, formative evaluation (identifying design improvements) is more appropriate than summative evaluation (proving effectiveness). This staged approach balances scientific rigor with practical constraints—comprehensive evaluation requires access to production systems, user populations, and time that may not be available for initial design publications.
Implications for CEDAR Evaluation. Following these precedents, our evaluation strategy adopts a staged approach: This paper (DSRM Iteration 1) focuses on conceptual design, technical feasibility demonstration, and competency question validation using synthetic data—consistent with early-stage ontology and DQ framework research. Subsequent iterations will progress to large-scale empirical validation, baseline comparisons, and user studies, following the evaluation progression documented in successful ontology-based systems.
Activity 2: Objectives of a Solution (Section 2.2)
  • Integrate ontological domain knowledge with event stream processing;
  • Enable detection of temporal and context-dependent quality issues;
  • Provide business-meaningful explanations for detected anomalies;
  • Maintain interoperability through standards-based design.
Activity 3: Design and Development (Section 3.2 and Section 3.3)
  • Conceptual framework: Four integrated models (contextualization, event, trend detection, domain);
  • Ontological design: Extension of FIBO with event and quality concepts;
  • Architecture: Integration of semantic web technologies with big data platforms.
Activity 4: Demonstration (Section 4 and Section 5)
  • Motivating example: Equity derivative trading scenario;
  • Worked example: Detection of unusual price jump with causal tracing;
  • Competency questions: Demonstration of ontological query capabilities.
Activity 5: Evaluation (Section 6)
  • Current iteration: Proof-of-concept implementation, controlled validation with synthetic data, competency question testing;
  • Future iterations: Large-scale empirical evaluation, baseline comparisons, user studies, deployment case studies (detailed roadmap in Section 2.2).
Activity 6: Communication
  • This paper: Design contribution for academic audience;
  • Planned: Practitioner-oriented publications, open-source prototype release, workshop presentations.

3.2. Models for Contextualising Temporal Data

3.2.1. Overview

These models are the key component of the framework, where a model represents a repeatable, systematic method that is applied to a dataset to carry out some processing task. An instance of a model type will exist for a specific application. There are four types of models required in the framework:
  • A domain model (DM)—the overarching model to represent the business knowledge and semantics used across the framework;
  • The trend detection model (TDM)—a model to detect trends (i.e., patterns of interest) for an input dataset guided by the domain model. The TDM is a repository of functions that applies to set of attributes to detect which trends have occurred;
  • The event pattern model (EPM)—a model to detect event patterns (i.e., combinations of events that, when related, reflect business semantics). For example, an observed change in an attribute for a data source is a pattern;
  • The contextualisation model (CM)—a model to provide business actions based on the derived context of the event pattern model.
These are described in the rest of this section. We will use as an example an application for analysing trading data for equity-based derivatives.

3.2.2. Domain Model

An essential requirement for heterogenous data quality management is defining the taxonomy and business information concepts. The domain model is used to define entities in a specific domain and capture the relationships between these entities, either within a dataset or across datasets, i.e., it defines the semantics of the datasets. The domain model is represented via an ontology that achieves the following:
  • Captures known entities, relationships, roles and resources across data sources within a defined scope;
  • Uses a standardised, defined taxonomy of entities and their relationships represented as classes and relations;
  • Provides axioms to define constraints, dependencies or logical relationships between entities.
In our example, a domain model would contain known entities and defined relationships between attributes (price and volume) of market transactions for equity-based derivatives and the underlying equities. These attributes would reflect trends in these financial instruments, as shown in Figure 1.

3.2.3. Trend Detection Model

The trend detection model is used to detect data of interest using statistical learning methods (e.g., analytics, mathematical/computation models, machine learning). The trend detection model is represented by a series of predicate functions given the specified domain model, where, when applied, may return additional information (e.g., a tag) that can be associated with the data point. Using the running example, a domain expert may have the following expectations:
  • Derivatives should only be priced below a defined threshold.
  • Derivatives should only see a price difference between two consecutive trading periods within a defined threshold.
A domain expert may then be expected to raise a potential DQ issue if any or both these expectations are not met. Therefore, two trend detection models can be applied to the attributes of the market transactions, detecting unusual/interesting changes in price for equities and derivatives. Example of trends and their functions are shown in Table 3. The last row represents when both conditions are met.

3.2.4. Event Pattern Model

The event pattern model is the primary contribution of this paper as it is the central component that connects the different models in the framework. An event is defined when ‘the action has taken place’. It describes an occurrence that has been captured and integrated into the system, carrying information about the action. In this framework, we adopt an event meta-model that is depicted in Figure 2 [29]. A simple event represents the occurrence of a single action and may be instantaneous or have a duration. A complex event represents a combination of events that, when related, reflect business semantics and are defined by an event pattern.
The framework considers the input data as simple events and uses event patterns to define hierarchies of more complex events. Let us continue with the previous example. Trading data, by nature, has a temporal component, and trading patterns in one domain may impact the trading pattern of another. This can be generically defined as an event-based pattern (where one action that has occurred can be linked to another action that has also occurred or is expected to occur). A simple event is the market transaction from the original data source, and a complex event may be a simple price jump, indicated by the occurrence of the respective event pattern. The resulting EPM can be detailed as per Table 4.
The highest order in the event pattern hierarchy would signify when the event model output is ready for business interpretation. In our example, a user wants to observe event-based patterns across related financial trades. Deviations from typical trading patterns can be detected and potentially described by the existence (or lack) of related changes. As a domain expert, the user has an expectation of the usual price/volume, data movements between transaction periods and understands the impact the relationship between equities and derivatives will have on trading volumes/prices. Based on checking for several trends, the domain expert can specify two types of events for derivative transactions:
  • The observed trading price of the derivative is not within expected thresholds. Therefore, the domain expert would label this event as unusual price.
  • The observed trading price jump across consecutive trading periods of the derivative is unusual and moves outside expectation. Therefore, the domain expert would label this context as a ‘simple price jump’.
The event hierarchy that would result from the events in Table 4 is shown in Figure 3.
The events (both simple and complex) are defined by the roles and properties of entities and their attributes from the domain model and overlayed with appropriate thresholds determined by subject matter experts in the trend detection model.

3.2.5. Contextualisation Model

The contextualisation model is the overarching model to determine the actions required by a business user. The CM takes on defined complex events based on the event hierarchy that would require business intervention. The business context is then represented as an action associated with the dataset, as shown in Table 5.
In our example, the business user may consider a dataset to require further action when unusual trading has been detected. Using the event hierarchy, this is represented by E3 which is defined as a business outcome in the CM. Therefore, action A1, or raise a DQ, from the CM would need to be taken by the business user.

3.2.6. Sample Application

In this section, the running example will be expanded to demonstrate the application of the models. Consider the trading ledger shown in Table 6, showing aggregated volumes and price per unit for trades for fictious equity ‘ABC’ at 30 min intervals as well as its equity-based derivative. Each row is a data point, representing the occurrence of an action, and can therefore be categorised as simple event of event type market transaction (E0 in the event model).
Using the event model specified in Table 4 and the event hierarchy in Figure 3, several complex events can be detected based on the event pattern occurrence as show in Table 7 and visualised in Figure 4. The resulting context, based on the CM (from Table 5) leads to an instance of A1; that is, a business user should raise a DQ issue for the specified trading log.

3.3. Knowledge Base Design

A second component of the framework to support the application of the models from Section 3.2 is the knowledge base. The terminology and its application used are derived from the RM-ODP standard [30] and extended into the context of event patterns via concepts presented in [31,32]. Other use cases include big data frameworks pertaining to DQ management and semantic models [33,34,35,36,37].
The key purpose of the knowledge base is to provide visibility and transparency to end-users as to how context was derived by the framework. These objectives are supported through the development of an ontology to be used in the knowledge base. The objectives are as follows:
  • Allowing users to track an instance through the framework’s application and investigate how the framework was applied and relate model outputs to their inputs;
  • Allowing users to understand the framework’s models and how information is related across the models.
The ontology was developed using the methodology outlined in [38]. The specified approach was a bottom-up approach to address a series of defined competency questions to meet the knowledge base objectives. Utilising the bottom- up approach, the ontology should be able to answer the example competency questions in Table 8 for the example in Section 3.2.6, and more generally, the generic competency questions that can be extrapolated.

3.3.1. Defining the Key Classes

To draw out the required classes and attributes of the ontology, the example competency questions in Table 8 can be abstracted as per Table 9.
As a result, the entities of the ontology are defined in Table 10. From the competency questions, two parent classes can be defined:
  • Reference–representing generalised information that defines expected behaviour and attributes of instances;
  • Instance–representing specific realisations of a reference entity.

3.3.2. The Ontology Framework and Logical Relationships

The CEDAR ontology is built on the following key logical predicates and relationships in Table 11. This presents the logical content and semantic justification for each relationship axiom in our ontology. These axioms establish the foundational constraints and dependencies that govern how concepts interact within the framework. Each relationship represents not merely a structural connection but a logical assertion about the necessary dependencies and semantic meanings that underpin visual analysis of software engineering data.
The resulting ontology is shown in Figure 5. The central component is the event type entity, linking the actions, trends and events in the datasets. The event type entity is used to define instances of the event (i.e., the event (instance) entity). Instances are detected through event pattern occurrence (instance), which, in turn, uses event instance(s) and datapoint instance(s) to determine if an event pattern has occurred. Datapoints are contained within a dataset (instance), conforming to the types of datasets defined in the dataset entity.
An event type is also used to link to an action type, where the action type entity includes the event type that would trigger the action. Instances of the action types are captured in the action (instance) entity. An event type is defined as per a trend, which is sourced from the trend detection model. The TDM is associated with the domain model entity, which also has a reflexive relationship defined to capture related domains.
This design integrates the subcomponents of the knowledge base to allow the framework to be applied to the dataset in question. The user can understand the derived context and resulting action through the ontology and how the framework and its models have been applied.

3.3.3. Transforming Raw Data for the Ontology

CEDAR’s data transformation pipeline converts operational data into queryable Resource Description Framework (RDF) through five stages. Extraction captures heterogeneous data (logs, feeds, reference data) from source systems. Transformation performs semantic enrichment by computing contextual attributes that enable quality assessment—price changes relative to historical values, time intervals between events, and derived metrics representing business patterns. Semantic mapping applies formal specifications defining correspondences between source data and ontology concepts: determining RDF class instantiation (e.g., mapping instrument type to cedar:DerivativeTradeEvent) and property assertions (e.g., trade id to cedar:hasTradeID). This mapping embodies domain expertise validated by experts to ensure accurate conceptual alignment. Serialization generates RDF triples expressing type assertions, data properties, object relationships, and provenance metadata. Loading and reasoning ingests triples where ontological inference derives implicit knowledge through transitive relationships and class hierarchies—transforming data points into semantically interconnected knowledge where business rules can be evaluated.
This design ensures scalability through separation of concerns (data-intensive transformations separated from semantic reasoning), timeliness through incremental processing enabling near-real-time assessment, and accuracy through formal mapping specifications that codify domain knowledge explicitly and maintain provenance tracking. The bidirectional pipeline is noteworthy: operational data populates ontology instances while ontological constraints derive quality rules, creating a feedback loop where encoded business knowledge governs how data are interpreted for quality assessment.

4. CEDAR System Architecture

This section details how the CEDAR framework will be integrated into an existing FI architecture and be used by data consumers to contextualise data.

4.1. CEDAR System Boundaries and Context

Based on Section 2, we assume an FI data architecture that follows up a bow-tie mechanism, from data production, to processing to consumption. The target users of the framework include data consumers and data roles aligned with the use of data within the enterprise, as shown in Figure 6.
There are three key use cases of CEDAR. They are as follows:
  • UC1—Implementing the framework’s components (models) and defining their interactions (in the knowledge base): this role is assumed by a knowledge engineer persona;
  • UC2—Executing the framework on an ongoing basis: this role is assumed by a data operations persona;
  • UC3—Analysing the outcome of the framework: this role is assumed by a data consumer persona.

4.2. CEDAR Architectural Layers

The CEDAR application architecture will have three layers, as shown in Figure 7.
Each layer is defined as follows:
  • Graphical user interface (GUI) layer: this layer allows consumers (data consumers and data risk personnel) to understand context that has been added to their data before usage and explore the framework and instances of its application.
  • Services: This layer is responsible for retrieving and applying the appropriate models. It also populates the ontology used to support framework transparency and user exploration. The ontology is populated from two different key sources: reference entities by domain experts, and instance entities by the datasets from the framework application.
  • Knowledge base components: this layer contains the datasets and models used by the framework, as well as instances of the framework application.
This architecture is designed to facilitate the CEDAR framework’s application in different FI environments. The three core services required in CEDAR are as follows:
  • Model Retrieval—which acts as the access layer to the Knowledge Base;
  • Model Application—which orchestrates the execution of models on datasets;
  • Framework Ontology Management—which maintains the structural organization of framework components.
The knowledge base itself comprises three components:
  • Producer Datasets—which represent the input data sources;
  • Knowledge Base Models—which contain the reusable structured artifacts that define processing logic;
  • Knowledge Base Instances—which store the outputs generated from applying models to datasets.

4.3. Use Cases and Sequence Diagrams

We will now explore how the use cases defined in Section 4.1 are implemented by the framework’s architecture. For these use cases, each service’s contribution to the workflows is defined in Table 12, and each knowledge base component’s contribution is defined in Table 13.

4.3.1. UC1—Implementing the Framework

Use case 1 focuses on how the framework in implemented initially within an FI’s architecture. This is shown in Figure 8. The knowledge engineer would start with using the GUI to build a model component of the framework. This would query the framework ontology management service return the model constraints that the knowledge engineer would then build within. For example, the knowledge engineer starts with wanting to define the domain model. The framework ontology management service would return the format for the knowledge engineer to define the domain model in, which would be the form of entities and relationships with defined attributes. If the knowledge engineer builds outside these constrains, the framework ontology management service would fail the verification and provide feedback to the knowledge engineer to resolve. If the model has been built within the defined constraints, it will be stored in the knowledge base. The knowledge engineer can then loop through this process to build more models for the framework, moving from the domain model to the trend detection model to the event pattern model to the contextualisation model. The knowledge engineer also needs to define the interactions between models, specifying how which models can be used together to ensure compatibility.
When defining the trend detection model, the knowledge engineer would convert business context into trends for the trend detection model and set which domain model is linked to which trend model. This will then be verified by the framework ontology management service and provide feedback to the engineer. The knowledge engineer would then abstract the trends into events to create the event pattern model, and set interactions between trends and event types. As part of the event pattern model, the event hierarchy would also be defined. The events (both simple and complex) would be defined by the roles and properties of entities and their attributes from the domain model, and overlaid with appropriate thresholds through the trend detection model, and informed by business experts. The final model for the knowledge engineer to configure would be the contextualisation model. Using the event hierarchy, the knowledge engineer would set the interactions between the event pattern model and event hierarchy to the contextualisation model.
Each of these models, once configured, would be stored in the knowledge base and their interactions defined in the framework ontology management service for use in the next use case.

4.3.2. UC2—Executing the Framework

Use case 2 focuses on how the framework is executed. This is shown in Figure 9. The data operations persona is required to execute the framework against a chosen dataset. The dataset will be validated and added to the knowledge base, before initialising a run of the framework. As defined by the knowledge engineer’s implementation in use case 1, the framework will look to apply the domain model first, and therefore retrieve the domain model pertaining to this dataset from the knowledge base. This will then be returned to the model retrieval service and sent back to the model application service. Based on this, and replicating the same sequence, the domain model will lead to the application of the trend detection model, which in turn will lead to the the application of the event pattern model. The event pattern model will be applied until the event hierarchy is triggered.
If the highest order complex event in the event hierarchy is found, the final model to be applied is the contextualisation model. This is requested by the model application service, and retrieved from the knowledge base by the model retrieval service. The model is then applied and the final result (i.e., context based on the CM) is taken from the defined model from the knowledge engineer. This final results are processed to comply based on framework ontology management service and stored in the knowledge base. Once stored, framework execution has completed. The data operations persona then sees the contextualised results on the GUI and is ready to share with the data consumer persona for use case 3.

4.3.3. UC3—Analysing the Framework Outcomes

Use case 3 focuses on how the framework outcomes are analysed. This is shown in Figure 10. The data consumer persona is interested in understanding the context provided and how the context was derived, as per Table 9. The user, via the GUI, requests to analysis an execution of the framework for the specific dataset. This leads to a query of the framework structure to the framework ontology management service. This fetches the output for each of the applied models from the knowledge base and organises the results based on the framework ontology. The structured framework results are returned to the user on the GUI.
The data consumer persona can now explore the results by selected components via the GUI. This will query the framework ontology management service to fetch detailed results from the knowledge base and return the output to the user. This will allow the data consumer to explore across the framework’s execution for the producer dataset and analyse the outcome.

5. Implementation

5.1. Prototype

The current prototype to implement the framework has utilised broker trading data sourced from the Securities Industry Research Centre of Asia-Pacific (SIRCA) and the Financial Industry Business Ontology (FIBO), and follows the example used in Section 3.2.6. As the initial prototype, the implementation has focused on the model development and application time-based data across related datasets.
We will focus on two scenarios defined in use case 3 as the key indicator of successful implementation. The following stack is used to demonstrate the case study:
  • CEDAR GUI—Stardog.
  • CEDAR Knowledge Base storage:
    Producer datasets—Stardog;
    Knowledge base models—Python 3.14;
    Knowledge base instances—Stardog.
  • CEDAR services:
    Model retrieval—Python;
    Model application—Python;
    Framework Ontology Management—Stardog.

5.2. Scenario 1: Anomaly Detection Knowledge Building

The first scenario evaluated is the detection of anomalous data, representing a potential DQ issue for a business user to raise. In this implementation, an unusual price with jump would represent such an anomaly. The CEDAR framework is applied to the derivatives dataset. Using the Stardog Explore interface, the application of the contextualisation model and event model are shown via Figure 11 or as a series of tuples via Table 14. Similarly, the application of the event model, trend detection model and domain model are shown in Figure 12.

5.3. Scenario 2: Anomaly Exploration

The second scenario evaluated is the framework users ability to self-service their use of the framework. To evaluate the prototype in the persona of a business user, the Stardog Explore GUI will be used to demonstrate outputs of the models and understand the framework application. As the framework ontology is the basis for business users to understand the framework, evaluation will focus on this ontology against the competency questions in Table 9, using the output data produced by the implementation in Section 5.1.
For example, competency question 1 stated the following: Why does a business user need to complete an [action instance]? Using the knowledge base instance and the framework ontology, we can start by querying which action instances were found (A1_1), what the reference action would be (A1), and then which event needs to occur for this action to be required (E3). This query, and output can be shown in Figure 13.
Similarly, we can evaluate another competency question: How many times [event instance] occured? Using market transaction (E0) as an example, we can count how many event instances of E0 occurred. This query, and the output can be show in Figure 14.

6. Discussion

This section interprets the validation findings, explores their significance and implications, positions results within the broader literature, and acknowledges limitations that contextualize the contribution’s scope.

6.1. Interpretation of Findings

This DSRM iteration focused on establishing technical feasibility for addressing the research questions: Can contextual information from event streams be integrated with ontological knowledge to identify context-dependent DQ issues (RQ1), and can ontology-based DQ assessment be operationalized in event-driven FI architectures (RQ2)?
Following the evaluation framework established in Section 2.2, we assess CEDAR against three key criteria from Table 2:

6.1.1. Criterion 1: Identification and Representation of Context-Dependent DQ Issues

Finding: The prototype successfully detected the unusual price jump anomaly by integrating temporal context with domain knowledge. This demonstrates advancement over traditional DQ methods and commercial tools that show “minimal” capability for context-dependent issues [7].
Significance: Unlike rule-based systems that detect violations of isolated constraints (e.g., “price is negative”), CEDAR identified an anomaly whose validity depends on temporal and domain context. The contextualization model enabled assessment of the derivative price change relative to the underlying equity’s behavior, addressing the challenge identified by Wang and Strong [3] that data quality is inherently context-dependent and user-specific.
Comparison to Alternatives: Supervised anomaly detection methods (Solution 2 in Table 2) require continuous human labeling and struggle with evolving contexts [9]. Unsupervised methods (Solution 3) lack interpretability—a limitation Nedelkoski et al. [39] quantified at 43% false positives in financial services due to missing domain context. CEDAR’s ontology-grounded approach provides both contextual awareness and interpretability, though at the cost of requiring explicit rule definition.
Relation to Literature: This finding extends Schneider’s [12] observation that ontologies are valuable additions to DQ management toolkits by demonstrating how ontologies can operationalize contextual assessment. While Debattista et al.’s Luzzu framework [40] applies ontologies to assess knowledge graph quality, CEDAR extends this to operational event streams—addressing Färber et al.’s [41] finding that 37% of quality issues in financial knowledge graphs are temporal and missed by static assessment.

6.1.2. Criterion 2: Scalability and Maintainability (Preliminary Assessment)

Finding: The prototype processed synthetic transactions with query response times under 200 ms. Rule addition requires ontology updates (estimated less than 1 day per new rule based on implementation experience) rather than coding in multiple systems.
Preliminary Interpretation: This suggests potential advantages over traditional methods where Loshin [42] documented 60–70% of effort spent maintaining rules across systems. However, with only minimal transactions, scalability claims are premature. The use of standards-based technologies (OWL, SPARQL, Spark) provides a foundation for scale, but performance at production volumes (millions of events) requires validation in DSRM Iteration 2.
Comparison to Alternatives: Commercial tools (Solution 4 in Table 2) offer proven scalability but with “high cost and low flexibility” due to vendor lock-in. CEDAR’s open standards approach trades proven performance for flexibility and interoperability—a trade-off that requires empirical validation to determine if the flexibility benefits outweigh potential performance costs.
Architectural Implications: CEDAR’s design aligns with FI trends toward centralized data management [1] while supporting the bow-tie architecture pattern [11] through semantic integration. The separation of computational models from semantic models enables independent scaling, though integration overhead needs measurement.

6.1.3. Criterion 3: Transparency to Data Consumers

Finding: Competency question validation demonstrated that the ontology enables querying from action (“Why must I raise this issue?”) to root cause (“Which market event triggered this?”) through semantic relationships. All competency questions were answerable via SPARQL or visual graph exploration.
Significance: This addresses a key limitation in existing approaches. Traditional DQ tools provide technical error codes; ML-based anomaly detection offers statistical scores; both lack business-meaningful explanations. CEDAR’s causal tracing capability—linking detected anomalies through event chains to violated business constraints—provides the “transparency to data consumers” that Munar et al. [10] identified as critical in complex FI architectures where data consumers lack visibility into upstream processing.
Important Caveat: Technical capability to answer questions does not validate actual usability for business users. Our validation used competency questions defined by researchers and executed by technical users. Whether data quality analysts and compliance officers can effectively use Stardog Explorer or formulate SPARQL queries requires user studies. The gap between technical capability and user utility is well-documented in ontology research [27].
Comparison to Alternatives: Solutions 2 and 3 (supervised/unsupervised anomaly detection) are “limited due to technical nature of implementation” [9]. CEDAR aims to bridge technical detection with business interpretation through ontological representation, but this bridge’s effectiveness for non-technical users remains unvalidated.

6.2. Theoretical Implications

6.2.1. For Data Quality Management Theory

CEDAR extends DQ management theory in three ways. First, it operationalizes the concept of semantic data quality—assessing quality not just against structural constraints but against rich domain knowledge encoded in ontologies. This addresses Wang and Strong’s [3] foundational observation that quality is fitness-for-use and context-dependent, providing a technical mechanism for representing and reasoning about context.
Second, it bridges the gap between static DQ frameworks [5,18] and dynamic event-driven environments. Existing frameworks assess quality at specific lifecycle points; CEDAR assesses quality across temporal event sequences, addressing Elouataoui et al.’s [2] critique that frameworks focus on specific lifecycle points rather than holistic management.
Third, it demonstrates that ontology-based approaches can address the context-dependency limitations that Zhang et al. [7] identified in automated DQ assessment—though our validation is too limited to claim this definitively.

6.2.2. For Ontology Application Research

This work contributes to ontology engineering practice by demonstrating how ontologies can move beyond static knowledge representation to drive operational processes. The bidirectional enrichment pattern—events populating the ontology, ontology enriching events—provides a model for integrating ontologies into dynamic systems, addressing a gap in ontology application literature that primarily focuses on static knowledge bases.
The extension of FIBO with event and quality concepts provides a reference model for operational financial ontology applications, complementing existing work on FIBO’s use for semantic consistency [14] and regulatory compliance [15].

6.2.3. For Financial Technology Research

CEDAR demonstrates practical application of semantic web standards in financial institutions—an industry traditionally resistant to academic technologies due to regulatory, security, and performance constraints. The successful integration with enterprise technologies (Databricks, Stardog) provides evidence that W3C standards can coexist with commercial platforms in regulated environments, addressing a barrier to semantic web adoption in finance.

6.3. Practical Implications

6.3.1. For Data Quality Practitioners

If validated at scale, CEDAR could offer several benefits:
Reduced Manual Effort: Ontological rule abstraction could reduce the 60–70% of time Loshin [42] found organizations spend maintaining rules. Rules defined once in the ontology automatically apply to related instruments through class inheritance.
Improved Root Cause Analysis: Automated causal tracing could reduce the 2–5 days domain experts in our partner institutions reported spending investigating cross-system DQ issues. However, these potential benefits require validation through user studies and deployment.
Enhanced Regulatory Alignment: Explicit constraint documentation in ontologies provides audit trails linking quality checks to regulatory requirements (e.g., MiFID II, Basel III)—addressing supervisor demands for data lineage transparency [21].
Adoption Barriers: Realizing these benefits requires: (1) skilled ontology engineers to create and maintain OWL ontologies, (2) organizational commitment to semantic approaches, (3) integration with existing DQ toolchains, and (4) training for business users. These barriers may limit adoption to larger FIs with mature data governance programs.

6.3.2. For Financial Institutions

CEDAR aligns with regulatory trends toward standardization and interoperability. The Bank for International Settlements’ vision for globally standardized data infrastructure [20] and domestic regulators’ focus on data lineage [21] suggest ontology-based approaches may become necessary rather than optional. Early adoption could provide competitive advantages in regulatory reporting efficiency and data risk management.
However, production deployment requires addressing: (1) performance at scale (billions of daily transactions), (2) integration with legacy systems, (3) governance of ontology evolution, and (4) cybersecurity for semantic technologies—all identified as future work requiring substantial investment.

6.3.3. For Technology Vendors

CEDAR’s standards-based design provides a blueprint for next-generation DQ tools. Commercial vendors could adopt the ontology-event integration pattern while leveraging their existing optimization, UI, and support capabilities. The framework is intentionally complementary to existing tools (e.g., consuming Informatica alerts and adding semantic interpretation) rather than a complete replacement, suggesting partnership opportunities.

6.4. Limitations

6.4.1. Validation Scope Limitations

Scale: Validation used only synthetic transactions—smaller than production FI environments processing millions of daily transactions. The perfect detection accuracy (precision = recall = 1.0) likely reflects the simplicity of the test case rather than real-world robustness.
Data Characteristics: Synthetic data with known ground truth eliminates confounding factors but does not represent real-world complexity: missing values, inconsistent formats, network latency, concurrent transactions, system errors. Real data will introduce edge cases and noise that may reveal design flaws not apparent in controlled tests.
Anomaly Diversity: Testing a single anomaly type (price jumps) does not validate generalization to the diverse quality issues FIs encounter: completeness violations, consistency errors, referential integrity failures, temporal sequence violations, regulatory constraint breaches. Each may require different detection approaches and ontological modeling patterns.
Domain Scope: The equity derivative scenario does not validate applicability to other financial instruments (fixed income, foreign exchange, commodities) or processes (risk aggregation, collateral management, regulatory reporting). While the models are designed generically, actual generalization requires empirical validation.

6.4.2. Evaluation Methodology Limitations

User Validation: Competency question validation demonstrates technical capability but not actual usability. We do not know if data quality analysts can effectively use Stardog Explorer, understand ontological representations, or trust system recommendations. User utility—the ultimate measure of success for a tool targeting business users—remains will be required in future work.
Longitudinal Assessment: Short-term technical validation does not reveal issues that emerge over time: rule maintenance burden as business logic evolves, ontology drift as concepts change, performance degradation as data accumulates, user adoption patterns. This also requires user validation.
Organizational Context: Laboratory validation ignores organizational factors critical to adoption: integration with existing workflows, training requirements, change management, political dynamics, budget constraints. This also requires user validation.

6.4.3. Technical Design Limitations

Ontology Maintenance Complexity: Creating and maintaining OWL ontologies requires specialized expertise that may not be available in all FIs. While we documented less than 1 day per new rule based on implementation experience, this assumes skilled ontology engineers—actual effort for typical data quality teams may be higher.
Rule-Based Detection Limitations: This iteration did not cover anomaly detection without ontology updates, unlike ML approaches that adapt automatically. Future iterations will focus on this, to reduce dependency on domain experts to anticipate and encode all relevant quality patterns—potentially missing emergent issues.

6.4.4. Threats to Validity Summary

Following Runeson and Höst’s [43] case study methodology taxonomy, the following can be determined:
Internal Validity: Synthetic data and controlled environment eliminate confounding factors but may not represent real-world causal relationships. Single anomaly type tested limits ability to infer general detection capability.
External Validity: Results from synthetic transactions in one scenario cannot be generalized to production environments, diverse instruments, or other FIs without replication studies.
Construct Validity: Measuring technical capability (competency question answerability) does not validate the construct of interest (user utility). This must be validated by a future user study.
Reliability: Detailed documentation of models, implementation, and validation procedures supports replication. However, dependencies on specific technologies and expert knowledge may limit reproducibility by other researchers.

6.5. Future Research Directions

The limitations above inform a structured research agenda across three DSRM iterations:
Iteration 2 (Technical Validation): Address scale, diversity, and comparison limitations through: 100,000+ real SIRCA transactions, 15–20 diverse anomaly types, statistical validation with significance testing, and performance profiling to identify bottlenecks.
Iteration 3 (User Validation): Address usability and utility limitations through: recruitment of FI practitioners (data quality analysts, compliance officers, data stewards), task-based usability studies measuring completion time and accuracy, system usability measurement, and qualitative interviews exploring perceived usefulness and adoption barriers.
Beyond CEDAR-specific validation, this work opens broader research questions: How should ontologies be governed in large organizations? What makes a “good” explanation for DQ issues? How to build appropriate user trust in ontology-driven systems? These questions extend to semantic systems and explainable AI generally.

6.6. Positioning Within Design Science Research

This paper represents an early-stage DSRM contribution focused on establishing design novelty and technical feasibility before comprehensive evaluation. This positioning is methodologically appropriate: attempting large-scale evaluation or user studies before confirming basic technical viability risks wasting resources on fundamentally flawed designs.
Following March and Smith’s [44] taxonomy, we have delivered a “construct” and “model” contribution (conceptual framework) with “instantiation” (working prototype) at the “proof-of-concept” stage. This is distinct from “proof-of-value” that requires comprehensive empirical evaluation demonstrating superiority over alternatives—the target of future iterations.
As Hevner [26] notes: “The rigor and completeness of design evaluation should be appropriate for the maturity and importance of the artifact.” For an initial design iteration, controlled artificial evaluation is appropriate; for production deployment decisions, naturalistic summative evaluation would be required.

7. Conclusions and Future Work

This paper presents CEDAR, a novel framework integrating ontological domain knowledge with event-driven data quality management for financial institutions. Following Design Science Research Methodology’s first iteration, we establish CEDAR’s design novelty and technical feasibility. CEDAR advances the state of the art through four innovations: (1) ontology-driven rule derivation automating constraint translation into detection logic, (2) temporal ontological reasoning extending static quality assessment with event dynamics, (3) explainable assessment through causal tracing, and (4) standards-based architecture using W3C technologies with FIBO. Comparative analysis demonstrates CEDAR addresses gaps at the intersection of rule-based DQ tools, log-based anomaly detection, knowledge graph quality frameworks, and CEP systems.
Controlled validation with synthetic equity derivative data confirms that CEDAR’s integrated models successfully detect context-dependent anomalies, trace causal chains, and support ontological exploration, answering both research questions on integrating contextual events with ontological knowledge and operationalizing ontology-based DQ assessment in event-driven architectures. While an early-stage prototype, it provides direction for future iterations that cover more complex models and integration, that can be comprehensively evaluated. This work demonstrates how semantic web technologies address operational DQ challenges in financial institutions, contributing to both theory and practice. As regulators demand standardization and data lineage transparency, ontology-based approaches like CEDAR may become essential infrastructure for modern financial data management.

Author Contributions

Conceptualization, A.T. and F.R.; methodology, A.T. and F.R.; software, A.T. and F.R.; validation, A.T. and F.R.; formal analysis, A.T. and F.R.; investigation, A.T. and F.R.; resources, A.T. and F.R.; data curation, A.T. and F.R.; writing—original draft preparation, A.T. and F.R.; writing—review and editing, A.T. and F.R.; visualization, A.T. and F.R.; supervision, F.R.; project administration, A.T. and F.R.; funding acquisition, F.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data from this research can be made available by the corresponding author following a justified request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
GFCGlobal Financial Crisis
CEDARContextual Events and Domain-driven Associative Representation
DQData Quality
FIFinancial institution
FIsFinancial institutions
FIBOFinancial Industry Business Ontology
DQOData Quality Ontology
DQAMData QUality Assessment Methodology
TDWMTotal Data Quality Management
DMDomain Model
TDMTrend Detection Model
EPMEvent Pattern Model
CMContextualisation Model
RDFResource Description Framework

References

  1. Cao, S.; Iansiti, M. Organizational Barriers to Transforming Large Finance Corporations: Cloud Adoption and the Importance of Technological Architecture; Working Paper 10142; CESifo: Munich, Germany, 2023. [Google Scholar]
  2. Elouataoui, W.; Alaoui, I.E.; Gahi, Y. Data Quality in the Era of Big Data: A Global Review. In Big Data: A Game Changer for Insurance Industry; Springer: Cham, Switzerland, 2022; pp. 1–25. [Google Scholar] [CrossRef]
  3. Wang, R.Y.; Strong, D.M. Beyond Accuracy: What Data Quality Means to Data Consumers. J. Manag. Inf. Syst. 1996, 12, 5–33. [Google Scholar] [CrossRef]
  4. Arolfo, F.; Vaisman, A. Data Quality in a Big Data Context. In Big Data: Concepts, Warehousing, and Analytics; Springer: Cham, Switzerland, 2018; pp. 159–172. [Google Scholar] [CrossRef]
  5. Cichy, C.; Rass, S. An overview of data quality frameworks. IEEE Access 2019, 7, 24634–24648. [Google Scholar] [CrossRef]
  6. Soni, S.; Singh, A. Improving Data Quality using Big Data Framework: A Proposed Approach. IOP Conf. Ser. Mater. Sci. Eng. 2021, 1022, 012092. [Google Scholar] [CrossRef]
  7. Zhang, P.; Xiong, F.; Gao, J.; Wang, J. Data quality in big data processing: Issues, solutions and open problems. In Proceedings of the 2017 IEEE SmartWorld, Ubiquitous Intelligence & Computing, Advanced & Trusted Computed, Scalable Computing & Communications, Cloud & Big Data Computing, Internet of People and Smart City Innovation, San Francisco, CA, USA, 4–8 August 2017; pp. 1–7. [Google Scholar] [CrossRef]
  8. Collibra. Collibra Data Quality & Observability. 2023. Available online: https://www.collibra.com/us/en/products/data-quality-and-observability (accessed on 20 June 2024).
  9. Thudumu, S.; Branch, P.; Jin, J.; Singh, J.J. A comprehensive survey of anomaly detection techniques for high dimensional big data. J. Big Data 2020, 7, 42. [Google Scholar] [CrossRef]
  10. Munar, A.; Chiner, E.; Sales, I. A Big Data Financial Information Management Architecture for Global Banking. In Proceedings of the 2014 International Conference on Future Internet of Things and Cloud, Barcelona, Spain, 27–29 August 2014; pp. 510–513. [Google Scholar] [CrossRef]
  11. Stockinger, K.; Bundi, N.; Heitz, J.; Breymann, W. Scalable architecture for Big Data financial analytics: User-defined functions vs. SQL. J. Big Data 2019, 6, 46. [Google Scholar] [CrossRef]
  12. Schneider, T.; Šimkus, M. Ontologies and Data Management: A Brief Survey. KI–Künstl. Intell. 2020, 34, 329–353. [Google Scholar] [CrossRef] [PubMed]
  13. FINOS. FINOS, Fintech Open Source Foundation. 2025. Available online: https://www.finos.org/ (accessed on 18 May 2025).
  14. Bennett, M. The financial industry business ontology: Best practice for big data. J. Bank. Regul. 2013, 14, 255–268. [Google Scholar] [CrossRef]
  15. Ryman-Tubb, N.F.; Krause, P. An ontology for Basel III capital requirements. In Proceedings of the 20th International Conference on Enterprise Information Systems (ICEIS), Funchal, Portugal, 21–24 March 2018; SCITEPRESS: Setúbal, Portugal, 2018; pp. 556–563. [Google Scholar]
  16. Debattista, J.; Auer, S.; Lange, C. daQ, an ontology for dataset quality information. In Proceedings of the LDOW@ WWW, Montreal, QC, Canada, 12 April 2016. [Google Scholar]
  17. Zaveri, A.; Rula, A.; Maurino, A.; Pietrobon, R.; Lehmann, J.; Auer, S. Quality assessment for linked data: A survey. Semant. Web 2016, 7, 63–93. [Google Scholar] [CrossRef]
  18. Batini, C.; Scannapieco, M. Data and Information Quality: Dimensions, Principles and Techniques; Springer: Cham, Switzerland, 2016. [Google Scholar]
  19. Stardog Union. Stardog: The Enterprise Knowledge Graph Platform. 2023. Available online: https://www.stardog.com/ (accessed on 13 December 2024).
  20. Colangelo, A.; Israël, J.M. How Integrated Reporting by Banks May Foster Sustainable Finance? Technical Report; European Central Bank: Frankfurt am Main, Germany, 2021. [Google Scholar]
  21. Moir, A.; Broadbent, B.; Woods, S. Transforming Data Collection from the UK Financial Sector; Technical Report; Bank of England: London, UK, 2020. [Google Scholar]
  22. Peffers, K.; Tuunanen, T.; Rothenberger, M.A.; Chatterjee, S. A Design Science Research Methodology for Information Systems Research. J. Manag. Inf. Syst. 2007, 24, 45–77. [Google Scholar] [CrossRef]
  23. Vom Brocke, J.; Winter, R.; Hevner, A.; Maedche, A. Design principles. In Design Science Research. Cases; Springer: Cham, Switzerland, 2020; pp. 15–31. [Google Scholar]
  24. Hevner, A.R.; March, S.T.; Park, J.; Ram, S. Design science in information systems research. MIS Q. 2004, 28, 75–105. [Google Scholar] [CrossRef]
  25. Venable, J.; Pries-Heje, J.; Baskerville, R. FEDS: A framework for evaluation in design science research. Eur. J. Inf. Syst. 2016, 25, 77–89. [Google Scholar] [CrossRef]
  26. Hevner, A.R. A three cycle view of design science research. Scand. J. Inf. Syst. 2007, 19, 4. [Google Scholar]
  27. Brank, J.; Grobelnik, M.; Mladenić, D. A survey of ontology evaluation techniques. In Proceedings of the Conference on Data Mining and Data Warehouses (SiKDD 2005), Ljubljana, Slovenia, 17 October 2005; pp. 166–170. [Google Scholar]
  28. Bennett, M.; Cray, S.; Hale, E.; Kendall, E. FIBO ontology: Adoption and application in the financial industry. In Proceedings of the ISWC (Posters/Demos/Industry), Online, 1–6 November 2020. [Google Scholar]
  29. Milosevic, Z.; Berry, A.; Chen, W.; Rabhi, F.A. An Event-Based Model to Support Distributed Real-Time Analytics: Finance Case Study. In Proceedings of the 2015 IEEE 19th International Enterprise Distributed Object Computing Conference, Adelaide, SA, Australia, 21–25 September 2015; pp. 122–127. [Google Scholar] [CrossRef]
  30. ITU-T Rec. X.902|ISO/IEC 10746-2; Information Technology–Open Distributed Processing–Reference Model: Foundations. Recommendation X.902. International Telecommunication Union: Geneva, Switzerland, 2010.
  31. Etzion, O.; Niblett, P. Event Processing in Action; Manning Publications: Shelter Island, NY, USA, 2010. [Google Scholar]
  32. Luckham, D. The power of events: An introduction to complex event processing in distributed enterprise systems. In Rule Representation, Interchange and Reasoning on the Web, Proceedings of the International Symposium, RuleML 2008, Orlando, FL, USA, 30–31 October 2008; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2008; Volume 5321, p. 3. [Google Scholar] [CrossRef]
  33. Barba-González, C.; Caballero, I.; Varela-Vaca, Á.J.; Cruz-Lemus, J.A.; Gómez-López, M.T.; Navas-Delgado, I. BIGOWL4DQ: Ontology-driven approach for Big Data quality meta-modelling, selection and reasoning. Inf. Softw. Technol. 2024, 167, 107378. [Google Scholar] [CrossRef]
  34. Benítez-Hidalgo, A.; Barba-González, C.; García-Nieto, J.; Gutiérrez-Moncayo, P.; Paneque, M.; Nebro, A.J.; Roldán-García, M.d.M.; Aldana-Montes, J.F.; Navas-Delgado, I. TITAN: A knowledge-based platform for Big Data workflow management. Knowl.-Based Syst. 2021, 232, 107489. [Google Scholar] [CrossRef]
  35. Dong, X.; He, H.; Li, C.; Liu, Y.; Xiong, H. Scene-based big data quality management framework. In Data Science, Proceedings of the 4th International Conference of Pioneering Computer Scientists, Engineers and Educators, ICPCSEE 2018, Zhengzhou, China, 21–23 September 2018; Communications in Computer and Information Science. Springer: Singapore, 2018; Volume 901, pp. 122–139. [Google Scholar] [CrossRef]
  36. Gu, R.; Qi, Y.; Wu, T.; Wang, Z.; Xu, X.; Yuan, C.; Huang, Y. SparkDQ: Efficient generic big data quality management on distributed data-parallel computation. J. Parallel Distrib. Comput. 2021, 156, 132–147. [Google Scholar] [CrossRef]
  37. Paneque, M.; Roldán-García, M.d.M.; Blanco, C.; Maté, A.; Rosado, D.G.; Trujillo, J. An ontology-based secure design framework for graph-based databases. Comput. Stand. Interfaces 2024, 88, 103801. [Google Scholar] [CrossRef]
  38. Noy, N.F.; McGuinness, D.L. Ontology Development 101: A Guide to Creating Your First Ontology; Technical Report KSL-01-05; Stanford University: Stanford, CA, USA, 2001. [Google Scholar]
  39. Nedelkoski, S.; Bogatinovski, J.; Acker, A.; Cardoso, J.; Kao, O. Anomaly detection from system tracing data using multimodal deep learning. In Proceedings of the 2020 IEEE 13th International Conference on Cloud Computing (CLOUD), Milan, Italy, 8–13 July 2020; pp. 179–186. [Google Scholar]
  40. Debattista, J.; Lange, C.; Auer, S. Luzzu—A methodology and framework for linked data quality assessment. J. Data Semant. 2016, 8, 159–175. [Google Scholar] [CrossRef]
  41. Färber, M.; Bartscherer, F.; Menne, C.; Rettinger, A. Linked data quality of DBpedia, Freebase, OpenCyc, Wikidata, and YAGO. Semant. Web 2018, 9, 77–129. [Google Scholar] [CrossRef]
  42. Loshin, D. Master Data Management; Morgan Kaufmann: Burlington, MA, USA, 2011. [Google Scholar]
  43. Runeson, P.; Höst, M. Guidelines for conducting and reporting case study research in software engineering. Empir. Softw. Eng. 2009, 14, 131–164. [Google Scholar] [CrossRef]
  44. March, S.T.; Smith, G.F. Design and natural science research on information technology. Decis. Support Syst. 1995, 15, 251–266. [Google Scholar] [CrossRef]
Figure 1. Example domain model.
Figure 1. Example domain model.
Electronics 15 00145 g001
Figure 2. Generic event model.
Figure 2. Generic event model.
Electronics 15 00145 g002
Figure 3. Event hierarchy.
Figure 3. Event hierarchy.
Electronics 15 00145 g003
Figure 4. Example of EPM application; Left: hierarchy of events detected; Right: events detected per time period.
Figure 4. Example of EPM application; Left: hierarchy of events detected; Right: events detected per time period.
Electronics 15 00145 g004
Figure 5. CEDAR framework ontology; parent class = reference in orange, parent class = instance in blue.
Figure 5. CEDAR framework ontology; parent class = reference in orange, parent class = instance in blue.
Electronics 15 00145 g005
Figure 6. Framework architecture combined with typical FI architecture. Arrows represent the flow of data.
Figure 6. Framework architecture combined with typical FI architecture. Arrows represent the flow of data.
Electronics 15 00145 g006
Figure 7. CEDAR application architecture.
Figure 7. CEDAR application architecture.
Electronics 15 00145 g007
Figure 8. CEDAR framework sequence diagram for implementing the framework in an FI’s architecture.
Figure 8. CEDAR framework sequence diagram for implementing the framework in an FI’s architecture.
Electronics 15 00145 g008
Figure 9. CEDAR framework sequence diagram for executing the framework on producer dataset.
Figure 9. CEDAR framework sequence diagram for executing the framework on producer dataset.
Electronics 15 00145 g009
Figure 10. CEDAR framework sequence diagram for analysing the outcome of the framework application.
Figure 10. CEDAR framework sequence diagram for analysing the outcome of the framework application.
Electronics 15 00145 g010
Figure 11. Contextualisation and event model. Parent class = reference in orange, parent class = instance in blue.
Figure 11. Contextualisation and event model. Parent class = reference in orange, parent class = instance in blue.
Electronics 15 00145 g011
Figure 12. Event, trend detection and domain model application.
Figure 12. Event, trend detection and domain model application.
Electronics 15 00145 g012
Figure 13. Evaluation of competency question 1.
Figure 13. Evaluation of competency question 1.
Electronics 15 00145 g013
Figure 14. Evaluation of competency question 6.
Figure 14. Evaluation of competency question 6.
Electronics 15 00145 g014
Table 1. Key regulatory reports.
Table 1. Key regulatory reports.
Report TypePurposeData Requirements
Basel III CapitalCapital adequacyTransaction-level position data
Liquidity CoverageShort-term liquidityDaily cash flow projections
Credit ExposureRisk concentrationCounterparty exposure data
Table 2. Evaluation criteria.
Table 2. Evaluation criteria.
No.Evaluation CriteriaSolution 1: Traditional DQ MethodsSolution 2: Anomaly Detection (Supervised)Solution 3: Anomaly Detection (Unsupervised)Solution 4: Commercial Tools
1Identification and representation of context dependent DQ issuesMinimalRelies on continual human inputLimitedLimited to application within tool
2Scalability and maintainability within existing architecture (incl. cost, flexibility, and timeliness) as datasets, processes and regulations evolveLimited application, high overhead to maintainHigh overhead to maintainHigh overhead to maintainHigh cost and low flexibility
3Transparency to data consumersComplexity dependent on complexity of data lineageLimited due to technical nature of implementationLimited due to technical nature of implementationLimited due to technical nature of implementation
Table 3. Example trends and their representation; where mt represents a market transaction.
Table 3. Example trends and their representation; where mt represents a market transaction.
TDM RefModel NameFunction DescriptionArgumentsFunction
UP U P ( m t ) Unusual price p = m t [ price ] p > threshold 2
SPJ S P J ( m t 1 , m t 2 ) Simple price jump t 1 = m t 1 [ time ] , t 2 = m t 2 [ time ] , p 1 = m t 1 [ price ] , p 2 = m t 2 [ price ] | t 1 t 2 | < time window 1 , | p 1 p 2 | > threshold 1
UPwJ U P w J ( S P J , U P ) Unusual price with jump t 1 = S P J [ time 1 ] , t 2 = U P [ time 2 ] | t 1 t 2 | < time window 2
Table 4. Event and event pattern types.
Table 4. Event and event pattern types.
Event Reference TypeEvent ClassEvent NameEvent Pattern ReferenceEvent Pattern TypeModel Ref (From TDM)
E0SimpleMarket transactionEP0AggregationMT
E1ComplexSimple price jumpEP1Filter over windowSPJ
E2ComplexUnusual priceEP2FilterUP
E3ComplexUnusual price with jumpEP3FilterUPwJ
Table 5. CM example.
Table 5. CM example.
Action ReferenceAction NameEvent Reference
A1Raise a DQ issueE3
Table 6. Trading log.
Table 6. Trading log.
Trade PeriodEquity Trading LogEquity-Based Derivative Trading Log
Simple Event ReferencePrice per Unit ($)Simple Event ReferencePrice per Unit ($)
1E0_11E0_54
2E0_22E0_65
3E0_33E0_715
4E0_41E0_85
Table 7. Sample of events detected and the related event pattern occurrences.
Table 7. Sample of events detected and the related event pattern occurrences.
Complex Event DetectedEvent Type NameEvent OccurrenceTrend Name
E1Simple price jumpE1_1 S P J ( E 0 _ 1 , E 0 _ 2 )
E1Simple price jumpE1_2 S P J ( E 0 _ 2 , E 0 _ 3 )
E1Simple price jumpE1_3 S P J ( E 0 _ 5 , E 0 _ 6 )
E1Simple price jumpE1_4 S P J ( E 0 _ 6 , E 0 _ 7 )
E2Unusual priceE2_1 U P ( E 0 _ 7 )
E3Unusual price with jumpE3_1 U P w J ( E 1 _ 4 , E 2 _ 1 )
Table 8. Example competency questions.
Table 8. Example competency questions.
Question TypeNo.Competency QuestionExpected Answer
Tracking an
instance
1Why does a business user need to complete action instance A1_1 ‘Raise a DQ issue’?Action A1_1 is action type A1 ‘Raise a DQ issue’ which is necessary when event type E3 ‘Unusual price with jump’ is detected. In this application, event instance E3_1 was detected.
2How was event instance E3_1 ‘Unusual price with jump’ detected?Event instance E2_1 ‘Unusual price’ and event instance E1_4 ‘Simple price jump’ were detected.
3Which datapoint instances is event instance E0_6 detected from?Event instance E0_6 is detected from datapoint instances DP_15 and DP_16.
4Which dataset instance does datapoint instance DP_15 come from?Dataset instance ‘Deriv_DS_01’, equity-based derivatives trading for trading period 1–4.
5How many datapoint instances were in the dataset instance ‘Deriv_DS_01’?8 datapoint instances.
Understanding
the framework
6Where is event type E3 ‘Unusual Price with Jump’ sourced from?Event type E3 corresponds to trend ‘Unusual price with jump’, which is sourced from the trend detection model ‘TradingMovements’.
7What domain model is associated with the trend detection model ‘Trading Movements’?It comes from the domain model ‘Market Transaction’.
8Is the domain model ‘Market Transaction’ connected to other domain models?Yes, the domain model ‘Tradable Instrument’.
9How many trends exist in the trend detection model ‘Trading Movements’?10 trends.
Table 9. Generic competency questions: where entities/attributes are shown in brackets.
Table 9. Generic competency questions: where entities/attributes are shown in brackets.
Question TypeNo.Competency QuestionExpected Answer
Tracking an
instance
1Why does a business user need to complete [action instance]?[Action instance] is [action type] which is necessary when [event type] is detected. In this application, [event instance] was detected.
2How was [event instance] detected?[Event instance(s)] was detected.
3Which [datapoint instances] is [event instance] detected from?[Event instance] is detected from [datapoint instance(s)].
4Which [dataset instance] does [datapoint instance] come from?[Dataset instance].
5How many [datapoint instance(s)] were in the [dataset instance]?count [datapoint instances] in [dataset instance].
Understanding
the framework
6Where is [event type] sourced from?[Event type] corresponds to [trend], which is sourced from [trend detection model].
7What [domain model] is associated with the [trend detection model]?It comes from [domain model].
8Is the [domain model] connected to other [domain model(s)]?Yes, [domain model(s)] with relationship “is associated with” to [domain model].
9How many [trend(s)] exist in the [trend detection model]?count [trend] in [trend detection model].
Table 10. Knowledge base classes.
Table 10. Knowledge base classes.
EntityDefinitionParent
Domain modelAn ontology defining business knowledge.Reference
Trend detection modelA mathematical or computing model to detect datapoints in the dataset that fit a specified behaviour based on the domain model.Reference
TrendA component of the trend detection model, where occurrence is captured as events.Reference
DatasetA predefined dataset for dataset instances, capturing consistent attributes for the dataset.Reference
Dataset (instance)A specific collection of data points that conforms to a reference dataset, representing real-world observations or transactions at a given time.Instance
Datapoint (instance)A value in a dataset representing an event.Instance
Event typeA predefined event, representing a trend.Reference
Event (instance)An occurrence of an event generated when an event pattern occurrence is detected.Instance
Event pattern
occurrence (instance)
A detected occurrence of a predefined event pattern.Instance
Action typeA predefined business process where occurrence is linked to an event type.Reference
Action (instance)An occurrence of an action type where the occurrence is linked to an event instance.Instance
Table 11. Logical content and semantic meaning of relationship axioms.
Table 11. Logical content and semantic meaning of relationship axioms.
AxiomConnectsLogical Content and Semantic Justification
is_contained_indatapoint_instance → dataset_instanceEstablishes compositional part-whole hierarchy where individual data points belong to dataset containers, ensuring data provenance and traceability for reproducibility.
conforms_todataset_instance → datasetEnforces structural consistency by requiring dataset instances to adhere to schema definitions, enabling validation and consistent interpretation.
detected_due_toevent_pattern_occurence → event_instanceEstablishes causal dependency between pattern occurrences and constituent events, enabling pattern mining traceability.
definesevent_type → event_instanceImplements type-instance pattern where event types serve as templates classifying individual instances, enabling generalization and filtering.
detected_throughevent_instance → event_pattern_occurenceRepresents inverse detection showing how events contribute to pattern recognition, enabling bidirectional graph navigation.
is_associated_withaction_type → event_typeModels action-event coupling where user actions trigger observable events, critical for understanding software behavior patterns.
is_associated_withtrend_detection_model → domain_modelContextualizes analytical models within domain models, providing semantic grounding for trend detection results.
is_associated_withdomain_model → domain_modelEnables modular ontology design where models reference or extend others, supporting cross-domain analysis and reusability.
is_instance_ofaction_instance → action_typeImplements type-instance distinction allowing specific actions to be classified by behavioral patterns.
sourced_fromevent_type → trend_detection_modelEstablishes analytical provenance showing which detection algorithms generated event type definitions.
is_detected_fromevent_instance → datapoint_instanceLinks data layer to event layer, showing transformation of raw measurements into meaningful events.
Table 12. Service definitions.
Table 12. Service definitions.
NameDefinitionUse Case 1: Building Framework Components (Knowledge Engineer)Use Case 2: Executing Framework Ongoing (Data Operations Persona)Use Case 3: Analysing Framework Outcomes (Data Consumer Persona)
Model RetrievalActs as a gateway for accessing and delivering model artifacts from the Knowledge Base, with caching, validation, and versioning.Used to verify and retrieve existing models for review or modification during setup.Relies on MR’s caching for fast model access during repeated executions.Indirectly supports by ensuring models are available for reference in analyses.
Model ApplicationOrchestrates the application of models to datasets, running pipelines to produce contextualised results.Not directly used; may be simulated for testing model interactions.Primary service for initiating and monitoring executions, applying models step-by-step.Not used; benefits from MA’s processed outputs stored in the KB.
Framework Ontology ManagementManages the structure and organization of the framework via an ontology, assembling models and results.Central for configuring ontology rules and defining model interactions in the KB.Supports by processing and storing results after executions.Key for requesting structured frameworks and exploring component details interactively.
Table 13. Knowledge base components.
Table 13. Knowledge base components.
NameDefinitionUse Case 1: Building Framework Components (Knowledge Engineer)Use Case 2: Executing Framework Ongoing (Data Operations Persona)Use Case 3: Analysing Framework Outcomes (Data Consumer Persona)
Producer DatasetsInput data sources (e.g., raw datasets from producers) that serve as the foundation for model application and analysis.Used as reference or test data to validate and build models, ensuring models align with real data structures.Selected and ingested as inputs for ongoing executions, where models are applied to process and transform the data.Referenced in contextualised results to understand the original data context behind analyses and outcomes.
Knowledge Base ModelsStructured artifacts (e.g., domain, trend detection, event pattern, contextualisation models) stored in the KB for reuse.Created, updated, and defined by the engineer, including their interactions and ontology rules within the KB.Retrieved and applied during executions to process datasets, with versioning ensuring consistency.Referenced in analyses to explain how results were derived, allowing consumers to explore model details.
Knowledge Base InstancesSpecific instances or results (e.g., contextualised outputs, execution snapshots) generated from applying models to datasets.Not directly used; instances are outcomes of testing or validation during building.Generated and stored as outputs of executions, capturing processed results for operational tracking.Queried and explored to interpret framework outcomes, providing actionable insights from processed data.
Table 14. Contextualisation and event model output.
Table 14. Contextualisation and event model output.
SubjectRelationshipTarget
A1_1is instance ofRaise a DQ issue
Raise a DQ issueresulted fromE3
E3corresponds toUnusual price with jump
E3definesE3_1
E3_1detected throughEP3_1
EP3_1detected due toE1_4
EP3_1detected due toE2_1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tafech, A.; Rabhi, F. CEDAR: An Ontology-Based Framework Using Event Abstractions to Contextualise Financial Data Processes. Electronics 2026, 15, 145. https://doi.org/10.3390/electronics15010145

AMA Style

Tafech A, Rabhi F. CEDAR: An Ontology-Based Framework Using Event Abstractions to Contextualise Financial Data Processes. Electronics. 2026; 15(1):145. https://doi.org/10.3390/electronics15010145

Chicago/Turabian Style

Tafech, Aya, and Fethi Rabhi. 2026. "CEDAR: An Ontology-Based Framework Using Event Abstractions to Contextualise Financial Data Processes" Electronics 15, no. 1: 145. https://doi.org/10.3390/electronics15010145

APA Style

Tafech, A., & Rabhi, F. (2026). CEDAR: An Ontology-Based Framework Using Event Abstractions to Contextualise Financial Data Processes. Electronics, 15(1), 145. https://doi.org/10.3390/electronics15010145

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop