Next Article in Journal
AI-Powered Trade Forecasting: A Data-Driven Approach to Saudi Arabia’s Non-Oil Exports
Previous Article in Journal
CME-YOLO: A Cross-Modal Enhanced YOLO Algorithm for Adverse Weather Object Detection in Autonomous Driving
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Comparison of Data Quality Frameworks: A Review

by
Russell Miller
1,
Sai Hin Matthew Chan
2,3,
Harvey Whelan
2,4 and
João Gregório
1,*
1
National Physical Laboratory, Informatics, Data Science Department, Glasgow G1 1RD, UK
2
National Physical Laboratory, Informatics, Data Science Department, Teddington TW11 0LW, UK
3
Department of Mathematics, University of Bath, Bath BA2 7AY, UK
4
Department of Natural Sciences, University of Bath, Bath BA2 7AX, UK
*
Author to whom correspondence should be addressed.
Big Data Cogn. Comput. 2025, 9(4), 93; https://doi.org/10.3390/bdcc9040093
Submission received: 22 January 2025 / Revised: 21 March 2025 / Accepted: 7 April 2025 / Published: 9 April 2025

Abstract

:
This study reviews various data quality frameworks that have some form of regulatory backing. The aim is to identify how these frameworks define, measure, and apply data quality dimensions. This review identified generalisable frameworks, such as TDQM, ISO 8000, and ISO 25012, and specialised frameworks, such as IMF’s DQAF, BCBS 239, WHO’s DQA, and ALCOA+. A standardised data quality model was employed to map the dimensions of the data from each framework to a common vocabulary. This mapping enabled a gap analysis that highlights the presence or absence of specific data quality dimensions across the examined frameworks. The analysis revealed that core data quality dimensions such as “accuracy”, “completeness”, “consistency”, and “timeliness” are equally and well represented across all frameworks. In contrast, dimensions such as “semantics” and “quantity” were found to be overlooked by most frameworks, despite their growing impact for data practitioners as tools such as knowledge graphs become more common. Frameworks tailored to specific domains were also found to include fewer overall data quality dimensions but contained dimensions that were absent from more general frameworks, highlighting the need for a standardised approach that incorporates both established and emerging data quality dimensions. This work condenses information on commonly used and regulation-backed data quality frameworks, allowing practitioners to develop tools and applications to apply these frameworks that are compliant with standards and regulations. The bibliometric analysis from this review emphasises the importance of adopting a comprehensive quality framework to enhance governance, ensure regulatory compliance, and improve decision-making processes in data-rich environments.

1. Introduction

Data quality (DQ) is defined by a set of values and attributes—often called dimensions—that can be qualitatively or quantitatively defined to describe the quality of datasets and other data structures [1,2]. The terminology used to describe these dimensions associated with data quality is complex, leading to standardisation efforts [2]. However, some commonly described dimensions include “accuracy”, “completeness”, “traceability”, and “timeliness”, which describe different aspects of a dataset. Given the increasing dependence on digital systems, primarily artificial intelligence (AI) and machine learning (ML) for informed decision making, having high-quality data is essential not only for ensuring operational efficiency but also to increase trust in these systems [3]. Standards, frameworks, guidelines, and regulations are commonly employed tools to ensure that high-quality data are used [1,4,5,6,7,8,9,10,11,12,13].
Data quality frameworks (DQFs) are structured methodologies used to assess, manage, and improve the quality of data. They can be directly built upon and supported by existing standards and regulations, or they can be specifically designed to address more tailored applications. These frameworks are essential for organisations to manage their data and demonstrate tangible evidence of the application of good data quality practices when communicating with stakeholders, both internal and external.
It is important to notice that this work is not the first attempt at providing an overview of existing DQFs, as there have been several attempts at doing so in recent years. Cichy et al. have published an overview of existing general-purpose data quality frameworks, where they offer a comprehensive systematic comparison between commonly known DQFs with the aim of informing data practitioners of the frameworks that better suit their needs [14]. While comprehensive, one limitation of this study that is recognised in their conclusions relates to the fact that several frameworks consider different dimensions of data quality, and no effort was done to map these using a common vocabulary or terminology. Additionally, the study also acknowledges the issue of regulatory compliance of the reviewed frameworks.
Similar studies of narrower scope have also been conducted, placing the focus on DQFs used only in specific domains such as healthcare [15,16] or finance [17]. While DQFs can be tailored for very specific applications, their wider applicability within each sector requires compliance with the established regulations of that sector. In many sectors, data management practices, and, by extension, data quality practices must be aligned with regulatory, governmental, or legislative standards [10,11,12,13]. This makes regulatory compliance a critical aspect of any DQF, ensuring that data management meets prespecified standards and increasing trust with stakeholders.
This paper provides a review of various DQFs that have received any type of regulatory backing across different sectors, offering a clear comparison of the sector-specific needs for high-quality data. Understanding how these regulated frameworks are used across different domains allows for the identification of common elements (i.e., DQ dimensions) between them. This compiled information can inform practitioners of the requirements of their specific domains, helping to avoid the mistakes associated with the use of unsuitable frameworks. This work builds upon previous work by the authors [2] that proposes a common terminology to describe data quality dimensions. This terminology is used to map dimensions between frameworks that have different nomenclatures, allowing a like-for-like comparison.
The significance of this review is its contribution to the understanding of DQFs and their regulatory compliance. It provides organisations and data practitioners with insights on how they can improve their data quality practices and ensure compliance with regulations. This information is equally relevant for emerging technologies, such as is the case with Large Language Model (LLM) AI systems [3]. This presents a rapidly evolving space that is cross-cutting over many different domains that regulators are struggling to keep up with. Leveraging existing regulated data quality frameworks provides a solid foundation for the development of AI-specific DQFs.
The frameworks covered in this paper include Total Data Quality Management (TDQM) [6]; Total Quality data Management (TQdM) [18]; ISO 8000 [1]; ISO 25012 [5]; Fair Information Practice Principles (FIPPS) [7]; the Quality Assurance Framework of the European Statistical System (ESS QAF) [8,19,20]; the IMF Data Quality Assessment Framework (DQAF) [10]; the UK Government Data Quality Framework [9,21]; the Data Management Body of Knowledge (DAMA DMBoK) [22,23]; the Basel Committee on Banking Supervision Standard (BCBS 239) [11]; the ALCOA+ Principles [24]; and the World Health Organization (WHO) Data Quality Assurance framework [13].
The remainder of this paper is structured as follows: Section 2 introduces general-purpose and foundational data quality frameworks; Section 3 presents data quality frameworks established as ISO standards; Section 4 discusses various governmental and international data quality frameworks; Section 5 highlights frameworks specifically used in the financial sector; Section 6 examines data quality frameworks employed in the healthcare sector; Section 7 collectively analyses and discusses all data quality frameworks reviewed by this work, comparing their common elements and identifying gaps in their assessments; and lastly, Section 8 summarises the findings of this review.

2. Data Management Frameworks

2.1. Total Data Quality Management (TDQM)

Total Data Quality Management (TDQM) is a holistic strategy for ensuring and monitoring the quality of data within organisations. TDQM was created in the 1980s by the MIT Sloan School of Management and was a pioneering research programme on data quality [6]. It has had a significant influence on the development of the data quality field [14,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42]. It views data as a commodity and employs methodologies and strategies to ensure their high quality. It focuses on different dimensions of data quality that correspond to data quality categories. These are accuracy, objectivity, believability, reputation, access, security, relevance, value-added, timeliness, completeness, amount of data, interpretability, ease of understanding, concise representation, and consistent representation.
The application process for TDQM comprises four stages: definition, measurement, analysis, and improvement [25]. This is also know as the DMAI (Define, Measure, Analyse, Improve) cycle and is shown here in Figure 1. The definition phase involves determining the relevant dimensions of data quality for both the organisation and the specific data being considered. The measurement phase involves assessing the existing condition of data quality, identifying any issues, and understanding their effects on the organisation. The analysis phase includes investigating the fundamental reasons behind data quality challenges. Finally, the improvement phase involves executing modifications that address the identified challenges to enhance the quality of the data.
This makes TDQM a method for executing the necessary cultural shift to establish a sustainable environment of ongoing data and information quality in an organisation. It offers a strong and all-encompassing framework for overseeing data quality. The framework acknowledges the critical importance of data in modern organisations and provides the necessary tools and methodologies to ensure the highest quality of this data.

2.2. Total Quality Data Management (TQdM)

Total Quality data Management (TQdM) shares the same holistic nature as TDQM. It consists of a comprehensive approach to managing data and information quality, with added emphasis on improving quality through detailed process analysis, and assumes the role of a value system within an organisation that integrates data quality management beliefs, principles, and methods into the organisational culture [18]. Given its holistic nature, it has served as the basis for the development of multiple data quality assessment tools [30,41,43,44,45,46,47,48,49,50,51,52], including its reference in the ISO 8000 series.
TQdM can be broken down into six sequential and cyclical processes: “Assess data definition and information architecture quality”, “Assess information quality”, “Measure nonquality information costs”, “Reengineer and cleanse data”, “Improve information process quality”, and “Establish the Information Quality Environment”.
The first two processes consist of measurements and aim to create an initial understanding of the current quality infrastructure and metrics being employed. The third process uses these measurements to develop a value proposition for improving data quality. The fourth process tackles the technical aspects of improving the quality of the data being generated and used. The fifth process deals with integrating the fourth process into the existing data pipeline. Lastly, the sixth process condenses the learning and benefits yielded by previous processes to establish a new and upgraded data quality environment.
These processes are then repeated in a continuous cycle of improvement. This makes TQdM a similar framework for continuous data and information quality improvement in organisations. It recognises the role that organisational culture plays in data and information management and establishes good practices for improving existing data and information quality infrastructure.

3. ISO Standards

3.1. ISO 8000

The ISO 8000 series of standards was developed to set the global benchmark for data quality and has seen widespread use [53,54,55,56]. This standards series issues frameworks for improving data quality specific to different data types, with a focus on the quality of enterprise master data [1]. These are data that are commonly used by organisations to manage information—which differs between organisations—that is critical to their operations. Master data can contain information on products, services, materials, clients, or financial transactions.
The series is organised into several parts that address specific data quality aspects and scenarios. These aspects and scenarios range from general principles of master data quality to specific applications, such as transactional data quality and product data quality. The data quality dimensions included in the ISO 8000 series are detailed in ISO 8000-8 and comprise accuracy, completeness, consistency, timeliness, uniqueness, and validity [57].
ISO 8000 also imports and incorporates notions such as the PDCA (Plan, Do, Check, Act) cycle, as shown here in Figure 2, which was outlined in ISO 9001 to improve data quality [4]. This cycle shares the same fundamentals as the DMAI cycle as described in Section 2.1 for the TDQM framework [58]. The planning stage aims to identify the relevant data quality dimensions for the organisation or task. The implementation stage entails the collection and processing of data. In the checking phase, the data quality dimensions considered in the planning stage are measured on the collected data. Lastly, the acting phase is implemented to continuously improve the processes of the full cycle. The practical application of the PDCA cycle is also described in ISO 8000-61 and provides a robust data quality management framework [59].

3.2. ISO 25012

The ISO 25012 standard is a part of the SQuaRE series of International Standards [5,60]. The latter establishes a general-purpose data quality model that can be applied to data stored within a structured computer system [61,62,63,64,65,66]. ISO 25012 can be specifically used to establish data quality requirements, define quality measures, and plan and perform data quality assessments. It performs a similar role to ISO 8000, but its application is more specific, primarily designed to be compatible with software development applications, processes, and pipelines.
Data quality dimensions in ISO 25012 are classified into fifteen unique characteristics: accuracy, completeness, consistency, credibility, currentness, accessibility, compliance, confidentiality, efficiency, precision, traceability, understandability, availability, portability, and recoverability. These dimensions are placed into a perspective spectrum that ranges from inherent to system-dependent. Inherent data quality dimensions are those that are intrinsic to the data regardless of the context of application and use, while system-dependent dimensions rely on the system and conditions of use to assess the quality of data. Some dimensions fall strictly into one of the two classes, while others need contextual consideration, balancing between inherent and system-dependent characteristics [5].
The comprehensive approach of ISO 25012 enables organisations to employ a single framework for consistently assessing the quality of their data under the assumption that the data are contained within a structured digital system. The dual perspective of inherent and system-dependent data quality dimensions offers a system for identifying the relevant data quality dimensions based on the application or use case while maintaining the sufficient granularity provided by the 15 individual dimensions for data quality assessment without redundancy [5]. However, the general approach of ISO 25012 may not suit organisations with highly specific and unique quality considerations.

4. Government and International Standards

4.1. Fair Information Practice Principles (FIPPS)

The Fair Information Practice Principles (FIPPS) are a set of guidelines established by the Federal Privacy Council (FPC), originally developed in 1973 by the United States Department of Health, Education, and Welfare, to address growing concerns about data privacy and the use of personal data, particularly when data are being used by automated systems [7]. These principles—transparency, individual participation, authority, purpose specification and use, limitation minimisation, access and amendment, quality and integrity, security, and accountability—see widespread use, acting as a foundational framework for ensuring data privacy and protection [67,68,69,70,71,72,73,74,75,76,77,78,79,80,81,82,83,84,85,86]. While not of regulatory status themselves, these principles have significantly influenced privacy legislation and policies, such as the General Data Protection Regulation (GDPR) and the Federal Information Processing Standards (FIPS) series developed by the National Institute of Standards and Technology (NIST) [87,88]. Notably, several FIPS standards are relevant for preserving data quality while addressing privacy and security concerns:
  • FIPS 180-4: Specifies secure hashing algorithms used to generate message digests, which help detect changes in messages and ensure data integrity during transfer and communication [89].
  • FIPS 199: Provides a framework for categorising federal information and information systems based on the level of confidentiality, integrity, and availability required. It helps in assessing the impact of data breaches and ensuring appropriate security measures [90].
  • FIPS 200: Outlines minimum security requirements for federal information and information systems, promoting consistent and repeatable security practices to protect data integrity and privacy [91].
Organisations wanting to use the FIPPS framework must assess the quality of personally identifiable information (PII) according to the principles it outlines for its effective use. This involves ensuring that data are accurate, relevant, and complete, as well as being collected and used transparently and with proper authority. By adhering to these principles, organisations can maintain high standards of data quality, which is essential for protecting privacy and ensuring the reliability of data-driven decisions.

4.2. Quality Assurance Framework of the European Statistical System (ESS QAF)

The Eurostat Quality Assurance Framework (QAF) is part of the Total Quality Management Framework that specifically addresses the quality of statistically generated outputs and data. It assesses the quality of these data according to five core principles, which are Relevance, Accuracy and Reliability, Timeliness and Punctuality, Coherence and Comparability, and Accessibility and Clarity.
The QAF itself is aligned with the European Statistics Code of Practice [8] and holds regulatory weight [19] under the Treaty on The Functioning European Union [20]. This alignment ensures that the statistical processes, outputs, and data adhere to high standards of quality, integrity, and reliability, which are essential for informed decision making and policy formulation within the European Union.
The QAF, as a data quality framework, focuses primarily on statistical processes and methods. This focus means it has reduced applicability compared to more general frameworks such as TDQM and TQdM, which were discussed previously. However, it shares similarities with these frameworks in its emphasis on key quality dimensions such as accuracy, timeliness, and relevance. This framework has thus been widely used and adapted both inside and outside of the European Union, demonstrating its robustness and flexibility in various contexts [92,93,94,95,96,97,98,99,100,101,102,103,104,105]. The QAF’s targeted approach ensures that European statistics are produced according to rigorous standards, making them reliable and useful for decision making and policy formulation within the European Union.

4.3. The UK Government Data Quality Framework

The UK Government Data Quality Framework, published in December 2020, addresses widespread concerns about data quality in the public sector. Motivated by the need to improve decision making, policy formation, and public services, it aims to standardise and enhance data quality practices across government organisations [9]. The framework consists of two parts: conceptual framework and practical guidance. The conceptual framework provides the structure for understanding and approaching data quality. It emphasises five data quality principles, describes the data lifecycle, shown here in Figure 3, and outlines the six core data quality dimensions used to evaluate the quality of data [9,21,106].
The five data quality principles are Commit to data quality, Know your users and their needs, Assess quality throughout the data lifecycle, Communicate data quality clearly and effectively, and Anticipate changes affecting data quality [9,21]. These principles are designed to create accountability and commit to ongoing assessment, improvement, and reporting of data quality. They promote understanding and prioritising user requirements to ensure that data are fit for purpose, focusing on quality measures and assurance at each stage of the data lifecycle. This ensures that end users understand data quality issues and their impact on data use, as well as helps them plan for and prevent future data quality issues through effective change management.
The conceptual framework also outlines a six-stage data lifecycle model consisting of the following activities: plan; collect, acquire, and ingest; prepare, store, and maintain; use and process; share and publish; and archive or destroy [9,21]. This lifecycle model helps identify potential quality issues at each stage and includes guidance on data management practices, quality considerations, and potential problems to address. Lastly, the conceptual framework defines six data quality dimensions as completeness, uniqueness, consistency, timeliness, validity, and accuracy [9,21]. It also provides examples on how to quantify them [106].
The second part of the framework provides practical tools and techniques for implementing the concepts introduced by the conceptual framework [21]. It includes guidance on data quality action planning, root cause analysis, metadata management, communicating data quality to users, and using data maturity models. These tools are designed to help organisations assess, improve, and communicate data quality effectively, supporting the principles and concepts outlined in the conceptual framework. The outputs from the practical guidance are then translated into a maturity assessment which places the data quality infrastructure of the assessed sector or organisation into a scale from 1 (unacceptable) to 5 (optimised) [21,106]. This scale acts as a guideline for improvement of data quality and its supporting infrastructure.

Data Management Body of Knowledge (DAMA DMBoK)

DAMA International is an organisation that specialises in advancing data and information best practices [22,23]. In 2009, it published the Data Management Body of Knowledge (DAMA DMBoK), which comprises a detailed set of guidelines for addressing data management challenges [22]. Among these guidelines, the DMBoK contains a functional framework and a common vocabulary for assessing data quality [107].
While not regulatory by itself, the DMBoK is heavily imparted in other frameworks, such as the UK Government Data Quality Framework. The DMBoK covers multiple data management knowledge areas that are leveraged by the UK Government in their own framework to understand the maturity levels of current data quality and management practices across different sectors. This allows the UK Government to identify areas for improvement and assess the requirements for carrying these improvements [9,108].
The knowledge areas covered by the DMBoK are Data Governance, Data Architecture, Data Modelling and Design, Data Storage and Operations, Data Security, Data Integration and Interoperability, Documents and Content, References and Master Data, Data Warehousing and Business Intelligence, Metadata, and Data Quality. The wide scope covered by these knowledge areas makes DMBoK a robust and adaptable framework for a diverse range of applications. The five data quality principles used by the UK Government in their own DQF, discussed in Section 4.3, were imported from DMBoK [108].

5. Financial Frameworks

5.1. IMF Data Quality Assessment Framework (DQAF)

The International Monetary Fund (IMF)’s Data Quality Assessment Framework (DQAF), first published in 2003 and updated in 2012, is a comprehensive methodology for assessing data quality specific to the financial sector and institutions. The framework is grounded in the Fundamental Principles of Official Statistics of the United Nations [10] and describes best practices accepted by the international community, such as the use of accepted methodologies, for assessing the quality of data.
The methodologies used by DQAF focus on the quality of statistical systems, processes, and products. As a result, the DQAF is defined by six quality dimensions: quality prerequisites, assurances of integrity, accuracy and reliability, serviceability, and accessibility [10].
Each dimension is further subdivided into elements that can be defined by specific indicators, such as legal and institutional support, adequate resources, relevance to user needs, professionalism, transparency, ethical standards, alignment with international standards, sound statistical techniques, timely and consistent data, and comprehensive metadata. These indicators can be quantified and used to describe the quality of data according to each dimension.
The main aim of the DQAF is to leverage these indicators to enhance the quality of data provided to the IMF by different countries in a standardised way [109]. This promotes financial institution transparency, which supports financial stability and aligns economic policies across participating countries. The DQAF is used and complemented by other IMF data dissemination standards, such as the General Data Dissemination System and the Special Data Dissemination Standard [110], to further advance the goal of using data in promoting transparency and supporting informed decision making.

5.2. Basel Committee on Banking Supervision Standard (BCBS 239)

The Basel Committee on Banking Supervision (BCBS 239)’s Principles for Effective Risk Data Aggregation and Risk Reporting, published in 2013, act as a framework for enhancing risk management practices in banking systems [11]. While BCBS 239 does not explicitly focus on data quality, it incorporates numerous terms and definitions relevant to data quality. This framework addresses the challenges many banks faced in aggregating and reporting risk data effectively during the 2008 financial crisis [111]. BCBS 239 comprises 14 principles organised into five key categories [112]. These categories are “Overarching Governance and Infrastructure”; “Risk Data Aggregation Capabilities”; “Risk Reporting Practices”; “Supervisory Review, Tools, and Cooperation”; and “Implementation Timeline and Transitional Arrangements”. The second category, “Risk Data Aggregation Capabilities”, focuses on the technical aspects of collecting, processing, and consolidating risk data, and it sets standards for data accuracy, adaptability, clarity, completeness, integrity, and timeliness [113,114].
The focus on these data quality dimensions is due to their relevance for effective risk management [115]. Accuracy reflects the closeness of data to the true values; adaptability refers to the ability to adjust to changing circumstances; clarity ensures that reports and data are easily understood; completeness is aimed at ensuring the availability of relevant data across all organisational units; integrity focuses on safeguarding data from unauthorised changes; and timeliness refers to data availability within the necessary timeframe for decision making. Together, these elements contribute to a robust framework for enhancing risk data aggregation and reporting practices.
By implementing the principles outlined in BCBS 239, banks are expected to enhance their risk management capabilities, increase transparency, and bolster their resilience to financial shocks [114,115,116]. This framework is particularly applicable to Global Systematically Important Banks (G-SIBs), which are subject to additional regulatory requirements, and encourages national authorities to extend their principles to Domestic Systematically Important Banks (D-SIBs) as well.

6. Healthcare Frameworks

6.1. ALCOA+ Principles

The ALCOA+ principles—attributable, legitimate, contemporaneous, original, accurate, plus complete, consistent, lasting, and available—establish a comprehensive framework for ensuring data integrity in regulated industries, with particular emphasis on medicine manufacturing [24]. These guidelines outline data quality aspects to be maintained throughout the data lifecycle [117]. The need for such a framework emerged from the growing complexity of data management in an increasingly digital landscape, where the risk of errors has risen in recent years [12,117,118].
The need for ALCOA+ arose from a multifaceted set of challenges in regulated industries [119,120]. Various factors contribute to poor data quality, including human and system errors, inadequate procedures, and inconsistencies across different platforms and processes. ALCOA+ addresses these challenges by offering clear data quality guidelines that enable organisations to implement robust processes, enhancing the overall trustworthiness of their data and data-driven decisions [12,117,118]. Additionally, ALCOA+ helps identify deliberate falsification of medical data, which, if undetected, can lead to severe consequences such as compromised patient safety, inaccurate clinical trial results, or regulatory noncompliance [24].
The U.S. Food and Drug Administration (FDA) has played a significant role in promoting and enforcing the ALCOA+ principles [24]. By adopting these guidelines, the FDA has established a clear standard for evaluating the reliability and trustworthiness of data submitted for regulatory purposes [24]. This adoption has far-reaching implications, as it guides FDA inspections and audits of regulated facilities and serves as a benchmark for compliance with good manufacturing practices (GMPs) and good laboratory practices (GLPs).

6.2. WHO Data Quality Assurance

The Data Quality Assurance (DQA) framework established by the World Health Organisation (WHO) provides a systematic approach for reviewing and enhancing data quality across healthcare facilities worldwide [13]. This framework is designed to identify weaknesses in data management systems and monitor data quality performance through structured processes. It encompasses routine data quality assurance activities, including regular reviews, discrete cross-sectional assessments, and periodic in-depth evaluations tailored to specific programs—all conducted through desk reviews or site assessments [121,122].
Desk reviews focus on analysing the completeness and consistency of existing aggregated data using established WHO metrics, such as completeness, internal consistency, external consistency with other data sources, and alignment with population data [121]. Site assessments evaluate the accuracy of health data through on-site evaluations guided by a checklist that examines data accuracy across reporting hierarchies and assesses the healthcare system’s capacity to generate quality data [122]. These assessments are routinely conducted by health facility staff on a monthly basis, with district-level staff performing periodic evaluations [123]. Both assurance methods use trace indicators as quantifiable measures to assess adherence to data quality standards; meeting established benchmarks indicates that data quality is satisfactory [124,125,126].
Implementing the DQA framework necessitates a coordinated effort involving all key stakeholders, reflecting the framework’s emphasis on data quality as a systemic issue influenced by interactions among various data quality activities. However, the framework does not primarily focus on overarching planning, technical support, funding, or promotional efforts.

7. Discussion

The frameworks presented in this review raise varying considerations for the data quality dimensions they include, depending on their respective domains of application. While differences among these frameworks were anticipated, identifying similarities and gaps in data quality coverage is a crucial aspect of data quality research. The regulatory nature of these frameworks significantly influences how quality data are described, communicated, and used by domain experts. Ultimately, any framework employed—whether developed in-house or adopted from existing frameworks—must align with the regulations set for their specific domains of application.
To facilitate a one-to-one comparison across all examined frameworks, standardised data terminology is needed. We employed a previously developed standardised data quality framework to achieve this comparison [2], which is found in Table 1. This approach uses the definitions of each individual data quality dimension, as outlined by their respective frameworks, and maps them to the corresponding definition in our framework to find the common name for that dimension. This standardisation is necessary because different frameworks describe the same data quality dimension using different terminology. Additionally, some dimensions, as specified by each framework, cover different aspects of the same overarching concept. For example,“enduring” (ALCOA+) [11,111,112,113,114] and “distribution” (BCBS-239) [12,24,117,118] both relate to governance but address different aspects of it. This highlights the importance of a standardised language to ensure clarity and usability across diverse frameworks.
Figure 4 offers a condensed view of the information provided by Table 1, which was made possible by the mapping exercise using our previously proposed data quality framework [2]. This figure presents a matrix grid that visually represents the presence or absence of data quality dimensions across the various frameworks reviewed in this study. The vertical axis has the data quality frameworks—in the same order they were introduced in this paper—while the horizontal axis contains all data quality dimensions considered. Blue cells indicate that a particular data quality dimension is included in the framework, while red cells indicate that it is not considered.
From Figure 4, it is noticeable that frameworks such as TDQM [6] and ISO 25012 [5] cover a broader range of dimensions—11 and 15 out of a total of 19, respectively—while other frameworks, such as FIPPS [7] and WHO’s DQA [13]—4 and 5 out of a total of 19, respectively—have notable gaps in their coverage. This shows a trend where frameworks designed for general or all-purpose applications tend to cover a larger range of data quality dimensions, while frameworks tailored for use in specific domains cover fewer dimensions. This observation is based in the sample size comprised of the frameworks reviewed, so it should be used with caution. However, the trend suggests that while general-purpose frameworks may offer broader coverage, specialised frameworks can provide targeted insights that are crucial for specific applications. While at first it can seem to indicate that specialised frameworks are more limited than broader frameworks, it is important to note that this can also make their use more streamlined, as they cover data quality dimensions more relevant for their specific domains of application.
Another relevant aspect to consider is that more specific frameworks often incorporate data quality dimensions that are not present in all-purpose frameworks. The IMF’s DQAF, BCBS 239, ALCOA+, and WHO’s DQA frameworks cover “governance”, “usefulness”, and “semantics” as dimensions of data quality [10,11,13,24,111]. All of these dimensions are absent from ISO 25012 [5], with the most comprehensive DQF reviewed in work using the number of data quality dimensions as a metric. The inclusion of these dimensions in frameworks such as ALCOA+ and WHO’s DQA demonstrates the need to address specific data quality concerns that can be overlooked in more generalisable frameworks. Additionally, the data quality concerns addressed by the inclusion of these dimensions in specialised frameworks are not consistent across frameworks. We previously discussed how “enduring” (ALCOA+) and “distribution” (BCBS-239) both relate to “governance” but address different facets pertinent to their respective domains: these being healthcare and finance [11,24,111]. While there is a risk of losing granularity in assessing data quality, aggregating data quality dimensions under a common terminology facilitates a comparison between frameworks that would otherwise not be feasible [2]. Notwithstanding this limitation, the presence of less common data quality dimensions in specialised frameworks highlights a potential understated value of these frameworks, which is that they can showcase the need for data quality dimensions that otherwise might not be included in all-purpose frameworks [2].
The only framework that accounted for “quantity” as a dimension of data quality was TDQM. This framework is the oldest covered in this review and one of the earliest examples of a structured DQF found [6]. The capacity of data storage systems has increased with time [127]; hence, the “quantity” of data played a larger role in overall data quality in the earlier days of computing, which can help explain its absence from more modern frameworks. However, with the growth of big data applications, data generation has been regularly outpacing data storage capacity [128]. As organisations increasingly deal with big data challenges, the relevance of “quantity” as a dimension of data quality is becoming more noticeable. This is exemplified by the increased generation of healthcare data, pushing forward the need to assess the quality of data in relation to available storage solutions and mechanisms [129,130].
One last noteworthy aspect is that core data quality dimensions, such as “accuracy”, “completeness”, “consistency”, and “timeliness” have a greater representation across all reviewed frameworks. Their prevalence reflects their historical weight, as these dimensions have been imported from older frameworks into modern ones. However, it is important to note that while these core dimensions are consistently represented, their definitions and metrics have evolved over time and across different domains. For instance, the definition of accuracy in early frameworks might have focused on the correctness of data entries, whereas modern interpretations could include aspects of data precision and reliability in complex datasets. Similarly, timeliness has evolved from simply ensuring data are up-to-date to encompassing real-time data processing and availability in dynamic environments [1]. Despite this, the rapid emergence of new technologies highlights the need for newer dimensions to be recognised and integrated into all-purpose data quality frameworks.
This is exemplified by the dimension of “semantics”, as semantic technologies, such as knowledge graphs and ontologies, have become more relevant to big data applications [131]. For instance, in healthcare, semantic frameworks can be used to design knowledge systems that are interoperable, allowing multiple stakeholders collaborating on the same processes to understand and validate each other [132]. This interoperability is crucial for ensuring that data are accurately interpreted and used across different systems and organisations. Another relevant example of the use of “semantics” comes from the finance sector, where tools such as ontologies and graph databases can support frameworks for bankruptcy prediction [133]. These applications require data to meet pre-established criteria such as detailed descriptions and nomenclature, further highlighting the need for a way to evaluate the quality of these data.
The rapid growth of AI technologies such as LLMs provides an example of why these less-represented dimensions are crucial. AI models rely on high-quality data to function effectively. Dimensions such as “quantity”, “availability”, “semantics”, “portability”, and “compliance” are very relevant in the AI domain. For instance, LLMs require large quantities of available and accessible data to learn from [134], and having data semantically linked is required for allowing these models to understand and process natural language [135,136]. Having data with a high degree of portability allows these data to be used across different systems and platforms, while “compliance” plays an inherent role in defining data quality specifically for AI systems, as it ensures that data meet legal and ethical standards [137].
This analysis of regulation-backed data quality frameworks reveals both strengths and limitations to their design and application. While core dimensions remain foundational across various frameworks, the emergence of new technologies imposes the inclusion of additional dimensions such as “semantics” to address the evolving landscape of data management and quality. The insights gained from this review highlight the importance of adopting a standardised approach to data quality that accommodates both established and emerging dimensions. By doing so, organisations can enhance their data management practices, ensure compliance with regulatory standards, and ultimately improve decision-making processes in an increasingly data-driven world.

8. Conclusions

The goal of this work was to review and compare different data quality frameworks that are underpinned by regulatory, statutory, or governmental standards and regulations. By examining frameworks such as TDQM, ISO 8000, ISO 25012, and others, this work sought to identify how these frameworks define, measure, and apply data quality dimensions. The focus on regulation-backed frameworks also allowed for the identification of frameworks used in heavily regulated industries, such as the IMF’s DQAF and BCBS 239 for the financial sector and WHO’s DQA and ALCOA+ for the healthcare sector. Understanding how data quality is applied across a varied landscape of domains, including their regulatory requirements, promotes the development of tools for applying these frameworks that comply with current requirements set by standards and regulations. Knowledge of how data quality is leveraged across specific domains is also valuable for informing regulators of emerging technologies such as AI systems [3].
A standardised data quality model, which was first suggested by the authors [2], was used to connect all the data quality dimensions found in each framework to a common vocabulary. This was done by looking at the different definitions that each framework gave to its dimensions and aligning them with dimensions sharing identical or similar definitions from other frameworks. The outcomes of this approach are shown in a matrix grid in Figure 4, with frameworks and dimensions on the vertical and horizontal axes, respectively. This made it possible to compare frameworks using a gap analysis.
Findings from the bibliometric analysis of all reviewed frameworks reveal that core data quality dimensions, such as “accuracy”, “completeness”, “consistency”, and “timeliness”, are well represented across all frameworks. Other dimensions, such as “semantics” and “quantity”, are overlooked by most frameworks; however, they highlight the need for modern dimensions to be considered. “Semantics” is a direct response to the increased use of knowledge graphs and ontologies, while “quantity” correlates with the capacity of data storage systems being outpaced by data generation. Additionally, frameworks tailored for specific domains, such as finance and healthcare, were found to often include data quality dimensions that were absent from more general frameworks due to specific industry needs. This highlights the importance of adopting a standardised approach to data quality that takes into account both established and emerging data quality dimensions and promotes data governance and compliance in data-rich environments.
This review provides insight into the data quality frameworks currently employed across a varied landscape of domains, including highly regulated industries. It fills the gaps of other modern reviews of this subject area by providing a like-for-like comparison of data quality frameworks used in regulated industries, without limiting the scope of these frameworks to framework types (all-purpose frameworks are reviewed alongside more specific frameworks) or to domains of applications (frameworks from multiple industries are reviewed). This study also highlights the need to develop and integrate more modern dimensions of data quality in existing frameworks to keep up with the needs of emerging technologies, such as is the case for AI systems based on LLM models. It also offers guidance for the creation of tools and applications that use these frameworks by highlighting the data quality dimensions that specific industries and domains have as set requirements. Lastly, this review also highlights the need to consider the inclusion of emerging data dimensions to enable established all-purpose frameworks to keep up with the rapidly evolving technological landscape.

Author Contributions

Conceptualisation: J.G. and R.M.; methodology: J.G. and R.M.; software: H.W. and S.H.M.C.; validation: J.G.; formal analysis: R.M. and S.H.M.C.; investigation: J.G., R.M., H.W. and S.H.M.C.; resources: J.G. writing—original draft preparation: J.G., R.M. and H.W.; writing—review and editing: J.G.; visualisation: S.H.M.C.; supervision: J.G.; project administration: J.G.; funding acquisition: J.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the UK Government Department for Science, Innovation, and Technology through the UK’s National Measurement System.

Data Availability Statement

No new data were created or analysed in this study. Data sharing is not applicable to this review.

Acknowledgments

Thanks to David Whittaker and Paul Duncan for providing feedback on the manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. ISO 8000-1:2022; Data Quality—Part 1: Overview. International Organization for Standardization: Geneva, Switzerland, 2022.
  2. Miller, R.; Whelan, H.; Chrubasik, M.; Whittaker, D.; Duncan, P.; Gregório, J. A Framework for Current and New Data Quality Dimensions: An Overview. Data 2024, 9, 151. [Google Scholar] [CrossRef]
  3. Levene, M.; Adel, T.; Alsuleman, M.; George, I.; Krishnadas, P.; Lines, K.; Luo, Y.; Smith, I.; Duncan, P. A Life Cycle for Trustworthy and Safe Artificial Intelligence Systems; Technical Report; NPL Publications: Teddington, UK, 2024. [Google Scholar]
  4. ISO 9001:2015; Quality Management Systems—Requirements. ISO: Geneva, Switzerland, 2015.
  5. ISO/IEC 25012:2008; Software Engineering—Software Product Quality Requirements and Evaluation (SQuaRE)—Data Quality Model. International Organization for Standardization: Geneva, Switzerland, 2008.
  6. MIT Information Quality Program. Total Data Quality Management (TDQM) Program. 2024. Available online: http://mitiq.mit.edu/ (accessed on 14 August 2024).
  7. Federal Privacy Council. Fair Information Practice Principles (FIPPS). 2024. Available online: https://www.fpc.gov/ (accessed on 14 August 2024).
  8. Eurostat. European Statistics Code of Practice—Revised Edition 2017; Publications Office of the European Union: Luxembourg, 2018. [Google Scholar] [CrossRef]
  9. Government Data Quality Hub. The Government Data Quality Framework. 2020. Available online: https://www.gov.uk/government/organisations/government-data-quality-hub (accessed on 1 October 2024).
  10. International Monetary Fund. Data Quality Assessment Framework (DQAF). 2003. Available online: https://www.imf.org/external/np/sta/dsbb/2003/eng/dqaf.htm (accessed on 1 October 2024).
  11. Basel Committee on Banking Supervision. Principles for Effective Risk Data Aggregation and Risk Reporting; Technical Report; Bank for International Settlements: Basel, Switzerland, 2013. [Google Scholar]
  12. Leach, C.D. Enhancing Data Governance Solutions to Optimize ALCOA+ Compliance for Life Sciences Cloud Service Providers. Ph.D. Thesis, Colorado Technical University, Colorado Springs, CO, USA, 2024. [Google Scholar]
  13. World Health Organization. Data Quality Assurance: Module 1: Framework and Metrics; World Health Organization: Geneva, Switzerland, 2022; p. vi, 30p. [Google Scholar]
  14. Cichy, C.; Rass, S. An overview of data quality frameworks. IEEE Access 2019, 7, 24634–24648. [Google Scholar] [CrossRef]
  15. Mashoufi, M.; Ayatollahi, H.; Khorasani-Zavareh, D.; Boni, T.T.A. Data quality in health care: Main concepts and assessment methodologies. Methods Inf. Med. 2023, 62, 005–018. [Google Scholar] [CrossRef]
  16. Fadahunsi, K.P.; O’Connor, S.; Akinlua, J.T.; Wark, P.A.; Gallagher, J.; Carroll, C.; Car, J.; Majeed, A.; O’Donoghue, J. Information quality frameworks for digital health technologies: Systematic review. J. Med. Internet Res. 2021, 23, e23479. [Google Scholar] [CrossRef]
  17. Landu, M.; Mota, J.H.; Moreira, A.C.; Bandeira, A.M. Factors influencing the quality of financial information: A systematic literature review. South Afr. J. Account. Res. 2024, 1–28. [Google Scholar] [CrossRef]
  18. English, L.P. Total quality data management (TQdM). In Information and Database Quality; Springer: Boston, MA, USA, 2002; pp. 85–109. [Google Scholar]
  19. European Parliament and Council of the European Union. Regulation (EC) No 223/2009 of the European Parliament and of the Council of 11 March 2009 on European Statistics; Technical Report; OJ L 87, 31.3.2009; European Union: Brussels, Belgium, 2009; pp. 164–173. [Google Scholar]
  20. European Union. Official Journal of the European Union, C 202; Technical Report; European Union: Maastricht, The Netherlands, 7 June 2016. [Google Scholar]
  21. Government Data Quality Hub. The Government Data Quality Framework: Guidance. 2020. Available online: https://www.gov.uk/government/publications/the-government-data-quality-framework/the-government-data-quality-framework-guidance (accessed on 1 October 2024).
  22. DAMA International. DAMA-DMBOK Data Management Body of Knowledge, 2nd ed.; Technics Publications: Sedona, AZ, USA, 2017; Available online: https://technicspub.com/dmbok/ (accessed on 1 October 2024).
  23. DAMA International. Body of Knowledge. 2024. Available online: https://www.dama.org/cpages/body-of-knowledge (accessed on 1 October 2024).
  24. Durá, M.; Sánchez-García, A.; Sáez, C.; Leal, F.; Chis, A.E.; González-Vélez, H.; García-Gómez, J.M. Towards a computational approach for the assessment of compliance of ALCOA+ Principles in pharma industry. In Challenges of Trustable AI and Added-Value on Health; IOS Press: Amsterdam, The Netherlands, 2022; pp. 755–759. [Google Scholar]
  25. Wang, R.Y. A product perspective on total data quality management. Commun. ACM 1998, 41, 58–65. [Google Scholar] [CrossRef]
  26. Bowo, W.A.; Suhanto, A.; Naisuty, M.; Ma’mun, S.; Hidayanto, A.N.; Habsari, I.C. Data quality assessment: A case study of PT JAS using TDQM Framework. In Proceedings of the 2019 Fourth International Conference on Informatics and Computing (ICIC), Semarang, Indonesia, 16–17 October 2019; pp. 1–6. [Google Scholar]
  27. Francisco, M.M.; Alves-Souza, S.N.; Campos, E.G.; De Souza, L.S. Total data quality management and total information quality management applied to costumer relationship management. In Proceedings of the 9th International Conference on Information Management and Engineering, Barcelona, Spain, 9–11 October 2017; pp. 40–45. [Google Scholar]
  28. Rahmawati, R.; Ruldeviyani, Y.; Abdullah, P.P.; Hudoarma, F.M. Strategies to Improve Data Quality Management Using Total Data Quality Management (TDQM) and Data Management Body of Knowledge (DMBOK): A Case Study of M-Passport Application. CommIT (Commun. Inf. Technol. J. 2023, 17, 27–42. [Google Scholar] [CrossRef]
  29. Wijnhoven, F.; Boelens, R.; Middel, R.; Louissen, K. Total data quality management: A study of bridging rigor and relevance. In Proceedings of the Fifteenth European Conference on Information Systems, ECIS 2007, St. Gallen, Switzerland, 7–9 June 2007. Number 15. [Google Scholar]
  30. Otto, B.; Wende, K.; Schmidt, A.; Osl, P. Towards a framework for corporate data quality management. In Proceedings of the Fifteenth European Conference on Information Systems, ECIS 2007, St. Gallen, Switzerland, 7–9 June 2007. Number 109. [Google Scholar]
  31. Wahyudi, T.; Isa, S.M. Data Quality Assessment Using Tdqm Framework: A Case Study of Pt Aid. J. Theor. Appl. Inf. Technol. 2023, 101, 3576–3589. [Google Scholar]
  32. Zhang, L.; Jeong, D.; Lee, S. Data quality management in the internet of things. Sensors 2021, 21, 5834. [Google Scholar] [CrossRef]
  33. Cao, J.; Diao, X.; Jiang, G.; Du, Y. Data lifecycle process model and quality improving framework for tdqm practices. In Proceedings of the 2010 International Conference on E-Product E-Service and E-Entertainment, Henan, China, 7–9 November 2010; pp. 1–6. [Google Scholar]
  34. Moges, H.T.; Dejaeger, K.; Lemahieu, W.; Baesens, B. A total data quality management for credit risk: New insights and challenges. Int. J. Inf. Qual. 2012, 3, 1–27. [Google Scholar] [CrossRef]
  35. Radziwill, N.M. Foundations for quality management of scientific data products. Qual. Manag. J. 2006, 13, 7–21. [Google Scholar] [CrossRef]
  36. Kovac, R.; Weickert, C. Starting with Quality: Using TDQM in a Start-Up Organization. In Proceedings of the ICIQ, Cambridge, MA, USA, 8–10 November 2002; pp. 69–78. [Google Scholar]
  37. Wilantika, N.; Wibowo, W.C. Data Quality Management in Educational Data: A Case Study of Statistics Polytechnic. J. Sist. Inf. J. Inf. Syst. 2019, 15, 52. [Google Scholar] [CrossRef]
  38. Shankaranarayanan, G.; Cai, Y. Supporting data quality management in decision-making. Decis. Support Syst. 2006, 42, 302–317. [Google Scholar] [CrossRef]
  39. Kovac, R.; Lee, Y.W.; Pipino, L. Total Data Quality Management: The Case of IRI. In Proceedings of the IQ; 1997; pp. 63–79. Available online: http://mitiq.mit.edu/documents/publications/TDQMpub/IRITDQMCaseOct97.pdf (accessed on 6 April 2025).
  40. Vaziri, R.; Mohsenzadeh, M. A questionnaire-based data quality methodology. Int. J. Database Manag. Syst. 2012, 4, 55. [Google Scholar] [CrossRef]
  41. Alhazmi, E.; Bajunaid, W.; Aziz, A. Important success aspects for total quality management in software development. Int. J. Comput. Appl. 2017, 157, 8–11. [Google Scholar]
  42. Shankaranarayanan, G. Towards implementing total data quality management in a data warehouse. J. Inf. Technol. Manag. 2005, 16, 21–30. [Google Scholar]
  43. Glowalla, P.; Sunyaev, A. Process-driven data quality management: A critical review on the application of process modeling languages. J. Data Inf. Qual. (JDIQ) 2014, 5, 1–30. [Google Scholar] [CrossRef]
  44. Otto, B. Quality management of corporate data assets. In Quality Management for IT Services: Perspectives on Business and Process Performance; IGI Global: Hershey, PA, USA, 2011; pp. 193–209. [Google Scholar]
  45. Otto, B. Enterprise-Wide Data Quality Management in Multinational Corporations. Ph.D. Thesis, Universität St. Gallen, St. Gallen, Switzerland, 2012. [Google Scholar]
  46. Caballero, I.; Vizcaíno, A.; Piattini, M. Optimal data quality in project management for global software developments. In Proceedings of the 2009 Fourth International Conference on Cooperation and Promotion of Information Resources in Science and Technology, Beijing, China, 21–23 November 2009; pp. 210–219. [Google Scholar]
  47. Siregar, D.Y.; Akbar, H.; Pranidhana, I.B.P.A.; Hidayanto, A.N.; Ruldeviyani, Y. The importance of data quality to reinforce COVID-19 vaccination scheduling system: Study case of Jakarta, Indonesia. In Proceedings of the 2022 2nd International Conference on Information Technology and Education (ICIT&E), Malang, Indonesia, 22 January 2022; pp. 262–268. [Google Scholar]
  48. Ofner, M.; Otto, B.; Österle, H. A maturity model for enterprise data quality management. Enterp. Model. Inf. Syst. Archit. (EMISAJ) 2013, 8, 4–24. [Google Scholar] [CrossRef]
  49. Fürber, C.; Fürber, C. Data quality. In Data Quality Management with Semantic Technologies; Springer Gabler: Wiesbaden, Germany, 2016; pp. 20–55. [Google Scholar]
  50. He, X.; Liu, R.; Anumba, C.J. Theoretical architecture for Data-Quality-Aware analytical applications in the construction firms. In Proceedings of the Construction Research Congress 2022, Arlington, VA, USA, 9–12 March 2022; pp. 335–343. [Google Scholar]
  51. Wende, K.; Otto, B. A Contingency Approach To Data Governance. In Proceedings of the ICIQ, Cambridge, MA, USA, 9–11 November 2007; pp. 163–176. [Google Scholar]
  52. Aljumaili, M.; Karim, R.; Tretten, P. Metadata-based data quality assessment. VINE J. Inf. Knowl. Manag. Syst. 2016, 46, 232–250. [Google Scholar] [CrossRef]
  53. Perez-Castillo, R.; Carretero, A.G.; Caballero, I.; Rodriguez, M.; Piattini, M.; Mate, A.; Kim, S.; Lee, D. DAQUA-MASS: An ISO 8000-61 based data quality management methodology for sensor data. Sensors 2018, 18, 3105. [Google Scholar] [CrossRef]
  54. Rivas, B.; Merino, J.; Caballero, I.; Serrano, M.; Piattini, M. Towards a service architecture for master data exchange based on ISO 8000 with support to process large datasets. Comput. Stand. Interfaces 2017, 54, 94–104. [Google Scholar] [CrossRef]
  55. Carretero, A.G.; Gualo, F.; Caballero, I.; Piattini, M. MAMD 2.0: Environment for data quality processes implantation based on ISO 8000-6X and ISO/IEC 33000. Comput. Stand. Interfaces 2017, 54, 139–151. [Google Scholar] [CrossRef]
  56. Carretero, A.G.; Caballero, I.; Piattini, M. MAMD: Towards a data improvement model based on ISO 8000-6X and ISO/IEC 33000. In Proceedings of the Software Process Improvement and Capability Determination: 16th International Conference, SPICE 2016, Dublin, Ireland, 9–10 June 2016; Proceedings 16. Springer: Cham, Switzerland, 2016; pp. 241–253. [Google Scholar]
  57. ISO 8000-8:2015; Data Quality—Part 8: Information and Data Quality: Concepts and Measuring. ISO: Geneva, Switzerland, 2015.
  58. Mohammed, A.G.; Eram, A.; Talburt, J.R. ISO 8000-61 Data Quality Management Standard, TDQM Compliance, IQ Principles. In Proceedings of the MIT International Conference on Information Quality, Little Rock, AR, USA, 6–7 October 2017. [Google Scholar]
  59. ISO 8000-61:2016; Data Quality—Part 61: Data Quality Management: Process Reference Model. ISO: Geneva, Switzerland, 2016.
  60. ISO/IEC 25000:2014; Systems and Software Engineering—Systems and Software Quality Requirements and Evaluation (SQuaRE)—Guide to SQuaRE. ISO: Geneva, Switzerland, 2014.
  61. Gualo, F.; Rodríguez, M.; Verdugo, J.; Caballero, I.; Piattini, M. Data quality certification using ISO/IEC 25012: Industrial experiences. J. Syst. Softw. 2021, 176, 110938. [Google Scholar] [CrossRef]
  62. Nwasra, N.; Basir, N.; Marhusin, M.F. A framework for evaluating QinU based on ISO/IEC 25010 and 25012 standards. In Proceedings of the 2015 9th Malaysian Software Engineering Conference (MySEC), Kuala Lumpur, Malaysia, 16–17 December 2015; pp. 70–75. [Google Scholar]
  63. Guerra-García, C.; Nikiforova, A.; Jiménez, S.; Perez-Gonzalez, H.G.; Ramírez-Torres, M.; Ontañon-García, L. ISO/IEC 25012-based methodology for managing data quality requirements in the development of information systems: Towards Data Quality by Design. Data Knowl. Eng. 2023, 145, 102152. [Google Scholar] [CrossRef]
  64. Verdugo, J.; Rodríguez, M. Assessing data cybersecurity using ISO/IEC 25012. Softw. Qual. J. 2020, 28, 965–985. [Google Scholar] [CrossRef]
  65. Pontes, L.; Albuquerque, A. Business Intelligence Development Process: An Approach with the Principles of Design Thinking, ISO 25012, and RUP. In Proceedings of the 2021 16th Iberian Conference on Information Systems and Technologies (CISTI), Chaves, Portugal, 23–26 June 2021; pp. 1–5. [Google Scholar]
  66. Galera, R.; Gualo, F.; Caballero, I.; Rodríguez, M. DQBR25K: Data Quality Business Rules Identification Based on ISO/IEC 25012. In Proceedings of the International Conference on the Quality of Information and Communications Technology, Aveiro, Portugal, 11–13 September 2023; pp. 178–190. [Google Scholar]
  67. Stamenkov, G. Genealogy of the fair information practice principles. Int. J. Law Manag. 2023, 65, 242–260. [Google Scholar] [CrossRef]
  68. Rasheed, A. Prioritizing Fair Information Practice Principles Based on Islamic Privacy Law. Berkeley J. Middle East. Islam. Law 2020, 11, 1. [Google Scholar]
  69. Paul, P.; Aithal, P.; Bhimali, A.; Kalishankar, T.; Rajesh, R. FIPPS & Information Assurance: The Root and Foundation. In Proceedings of the National Conference on Advances in Management, IT, Education, Social Sciences-Manegma, Mangalore, India, 27 April 2019; pp. 27–34. [Google Scholar]
  70. Klemovitch, J.; Sciabbarrasi, L.; Peslak, A. Current privacy policy attitudes and fair information practice principles: A macro and micro analysis. Issues Inf. Syst. 2021, 22, 145–159. [Google Scholar]
  71. Bruening, P.; Patterson, H. A Context-Driven Rethink of the Fair Information Practice Principles. SSRN 2016. [Google Scholar] [CrossRef]
  72. Gellman, R. Willis Ware’s Lasting Contribution to Privacy: Fair Information Practices. IEEE Secur. Priv. 2014, 12, 51–54. [Google Scholar] [CrossRef]
  73. Schwaig, K.S.; Kane, G.C.; Storey, V.C. Compliance to the fair information practices: How are the Fortune 500 handling online privacy disclosures? Inf. Manag. 2006, 43, 805–820. [Google Scholar] [CrossRef]
  74. Herath, S.; Gelman, H.; McKee, L. Privacy Harm and Non-Compliance from a Legal Perspective. J. Cybersecur. Educ. Res. Pract. 2023, 2023, 3. [Google Scholar] [CrossRef]
  75. Zeide, E. Student privacy principles for the age of big data: Moving beyond FERPA and FIPPS. Drexel Law Rev. 2015, 8, 339. [Google Scholar]
  76. Rotenberg, M. Fair information practices and the architecture of privacy (What Larry doesn’t get). Stan. Tech. Law Rev. 2001, 1, 1. [Google Scholar]
  77. Hartzog, W. The inadequate, invaluable fair information practices. Md. Law Rev. 2016, 76, 952. [Google Scholar]
  78. Proia, A.; Simshaw, D.; Hauser, K. Consumer cloud robotics and the fair information practice principles: Recognizing the challenges and opportunities ahead. Minn. J. Law Sci. Technol. 2015, 16, 145. [Google Scholar] [CrossRef]
  79. Karyda, M.; Gritzalis, S.; Hyuk Park, J.; Kokolakis, S. Privacy and fair information practices in ubiquitous environments: Research challenges and future directions. Internet Res. 2009, 19, 194–208. [Google Scholar] [CrossRef]
  80. Cavoukian, A. Evolving FIPPs: Proactive approaches to privacy, not privacy paternalism. In Reforming European Data Protection Law; Springer: Berlin/Heidelberg, Germany, 2014; pp. 293–309. [Google Scholar]
  81. Ohm, P. Changing the rules: General principles for data use and analysis. Privacy, Big Data, Public Good: Fram. Engagem. 2014, 1, 96–111. [Google Scholar]
  82. da Veiga, A. An online information privacy culture: A framework and validated instrument to measure consumer expectations and confidence. In Proceedings of the 2018 Conference on Information Communications Technology and Society (ICTAS), Durban, South Africa, 8–9 March 2018; pp. 1–6. [Google Scholar]
  83. Regan, P.M. A design for public trustee and privacy protection regulation. Seton Hall Legis. J. 2020, 44, 487. [Google Scholar]
  84. da Veiga, A. An Information Privacy Culture Index Framework and Instrument to Measure Privacy Perceptions across Nations: Results of an Empirical Study. In Proceedings of the HAISA, Adelaide, Australia, 28–30 November 2017; pp. 188–201. [Google Scholar]
  85. Da Veiga, A. An information privacy culture instrument to measure consumer privacy expectations and confidence. Inf. Comput. Secur. 2018, 26, 338–364. [Google Scholar] [CrossRef]
  86. Gillon, K.; Branz, L.; Culnan, M.; Dhillon, G.; Hodgkinson, R.; MacWillson, A. Information security and privacy—Rethinking governance models. Commun. Assoc. Inf. Syst. 2011, 28, 33. [Google Scholar] [CrossRef]
  87. European Parliament and Council. Regulation (EU) 2016/679 of the European Parliament and of the Council of 27 April 2016 on the Protection of Natural Persons with Regard to the Processing of Personal Data and on the Free Movement of Such Data, and Repealing Directive 95/46/EC (General Data Protection Regulation). 2016. Available online: https://eur-lex.europa.eu/eli/reg/2016/679/oj/eng (accessed on 14 August 2024).
  88. National Institute of Standards and Technology. Federal Information Processing Standards (FIPS) Publications. 2024. Available online: https://csrc.nist.gov/publications/fips (accessed on 14 August 2024).
  89. National Institute of Standards and Technology. Secure Hash Standard (SHS); Technical Report FIPS PUB 180-4; U.S. Department of Commerce: Washington, DC, USA, 2015. [Google Scholar]
  90. National Institute of Standards and Technology. Standards for Security Categorization of Federal Information and Information Systems; Technical Report FIPS PUB 199; U.S. Department of Commerce: Washington, DC, USA, 2004. [Google Scholar]
  91. National Institute of Standards and Technology. Minimum Security Requirements for Federal Information and Information Systems; Technical Report FIPS PUB 200; U.S. Department of Commerce: Washington, DC, USA, 2006. [Google Scholar]
  92. Sæbø, H.V. Quality in Statistics—From Q2001 to 2016. Stat. Stat. Econ. J. 2016, 96, 72–79. [Google Scholar]
  93. Revilla, P.; Piñán, A. Implementing a Quality Assurance Framework Based on the Code of Practice at the National Statistical Institute of Spain; Instituto Nacional de Estatistica (INE) Statistics Spain, Work. Pap.; Instituto Nacional de Estadística: Madrid, Spain, 2012. [Google Scholar]
  94. Nielsen, M.G.; Thygesen, L. Implementation of Eurostat Quality Declarations at Statistics Denmark with cost-effective use of standards. In Proceedings of the European Conference on Quality in Official Statistics, Vienna, Austria, 3–5 June 2014; pp. 2–5. [Google Scholar]
  95. Radermacher, W.J. The European statistics code of practice as a pillar to strengthen public trust and enhance quality in official statistics. J. Stat. Soc. Inq. Soc. Irel. 2013, 43, 27. [Google Scholar]
  96. Brancato, G.; D’Assisi Barbalace, F.; Signore, M.; Simeoni, G. Introducing a framework for process quality in National Statistical Institutes. Stat. J. IAOS 2017, 33, 441–446. [Google Scholar] [CrossRef]
  97. Stenström, C.; Söderholm, P. Applying Eurostat’s ESS handbook for quality reportson Railway Maintenance Data. In Proceedings of the International Heavy Haul STS Conference (IHHA 2019), Narvik, Norway, 12–14 June 2019; pp. 473–480. [Google Scholar]
  98. Mekbunditkul, T. The Development of a Code of Practice and Indicators for Official Statistics Quality Management in Thailand. In Proceedings of the 2017 International Conference on Economics, Finance and Statistics (ICEFS 2017), Hanoi, Vietnam, 25–27 February 2017; pp. 184–191. [Google Scholar]
  99. Radermacher, W.J.; Radermacher, W.J. Official Statistics—Public Informational Infrastructure. In Official Statistics 4.0: Verified Facts for People in the 21st Century; Springer: Cham, Switzerland, 2020; pp. 11–52. [Google Scholar]
  100. Sæbø, H.V.; Holmberg, A. Beyond code of practice: New quality challenges in official statistics. Stat. J. IAOS 2019, 35, 171–178. [Google Scholar] [CrossRef]
  101. Zschocke, T.; Beniest, J. Adapting a quality assurance framework for creating educational metadata in an agricultural learning repository. Electron. Libr. 2011, 29, 181–199. [Google Scholar] [CrossRef]
  102. Daraio, C.; Bruni, R.; Catalano, G.; Daraio, A.; Matteucci, G.; Scannapieco, M.; Wagner-Schuster, D.; Lepori, B. A tailor-made data quality approach for higher educational data. J. Data Inf. Sci. 2020, 5, 129–160. [Google Scholar] [CrossRef]
  103. Stagars, M. Data Quality in Southeast Asia: Analysis of Official Statistics and Their Institutional Framework as a Basis for Capacity Building and Policy Making in the ASEAN; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  104. Cox, N.; McLaren, C.H.; Shenton, C.; Tarling, T.; Davies, E.W. Developing Statistical Frameworks for Administrative Data and Integrating It into Business Statistics. Experiences from the UK and New Zealand. In Advances in Business Statistics, Methods and Data Collection; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2023; pp. 291–313. [Google Scholar]
  105. Ricciato, F.; Wirthmann, A.; Giannakouris, K.; Skaliotis, M. Trusted smart statistics: Motivations and principles. Stat. J. IAOS 2019, 35, 589–603. [Google Scholar] [CrossRef]
  106. Government Data Quality Hub. The Government Data Quality Framework: Case Studies. 2020. Available online: https://www.gov.uk/government/publications/the-government-data-quality-framework/the-government-data-quality-framework-case-studies (accessed on 1 October 2024).
  107. DAMA International. Mission, Vision, Purpose, and Goals. 2024. Available online: https://www.dama-belux.org/mission-vision-purpose-and-goals-2024/ (accessed on 1 October 2024).
  108. de Figueiredo, G.B.; Moreira, J.L.R.; de Faria Cordeiro, K.; Campos, M.L.M. Aligning DMBOK and Open Government with the FAIR Data Principles. In Proceedings of the Advances in Conceptual Modeling, Salvador, Brazil, 4–7 November 2019; pp. 13–22. [Google Scholar]
  109. Carson, C.S.; Laliberté, L.; Murray, T.; Neves, P. Toward a Framework for Assessing Data Quality. 2001. Available online: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=879374 (accessed on 1 October 2024).
  110. Kiatkajitmun, P.; Chanton, C.; Piboonrungroj, P.; Natwichai, J. Data Quality Assessment Framework and Economic Indicators. In Proceedings of the Advances in Networked-Based Information Systems, Chiang Mai, Thailand, 6–8 September 2023; pp. 97–105. [Google Scholar]
  111. Chakravorty, R. BCBS239: Reasons, impacts, framework and route to compliance. J. Secur. Oper. Custody 2015, 8, 65–81. [Google Scholar] [CrossRef]
  112. Prorokowski, L.; Prorokowski, H. Solutions for risk data compliance under BCBS 239. J. Invest. Compliance 2015, 16, 66–77. [Google Scholar] [CrossRef]
  113. Orgeldinger, J. The Implementation of Basel Committee BCBS 239: Short analysis of the new rules for Data Management. J. Cent. Bank. Theory Pract. 2018, 7, 57–72. [Google Scholar] [CrossRef]
  114. Harreis, H.; Tavakoli, A.; Ho, T.; Machado, J.; Rowshankish, K.; Merrath, P. Living with BCBS 239; McKinsey & Company: New York, NY, USA, 2017. [Google Scholar]
  115. Grody, A.D.; Hughes, P.J. Risk Accounting-Part 1: The risk data aggregation and risk reporting (BCBS 239) foundation of enterprise risk management (ERM) and risk governance. J. Risk Manag. Financ. Institutions 2016, 9, 130–146. [Google Scholar] [CrossRef]
  116. Elhassouni, J.; El Qadi, A.; El Madani El Alami, Y.; El Haziti, M. The implementation of credit risk scorecard using ontology design patterns and BCBS 239. Cybern. Inf. Technol. 2020, 20, 93–104. [Google Scholar] [CrossRef]
  117. Kavasidis, I.; Lallas, E.; Leligkou, H.C.; Oikonomidis, G.; Karydas, D.; Gerogiannis, V.C.; Karageorgos, A. Deep Transformers for Computing and Predicting ALCOA+ Data Integrity Compliance in the Pharmaceutical Industry. Appl. Sci. 2023, 13, 7616. [Google Scholar] [CrossRef]
  118. Sembiring, M.H.; Novagusda, F.N. Enhancing Data Security Resilience in AI-Driven Digital Transformation: Exploring Industry Challenges and Solutions Through ALCOA+ Principles. Acta Inform. Medica 2024, 32, 65. [Google Scholar] [CrossRef]
  119. Charitou, T.; Lallas, E.; Gerogiannis, V.C.; Karageorgos, A. A network modelling and analysis approach for pharma industry regulatory assessment. IEEE Access 2024, 12, 46470–46483. [Google Scholar] [CrossRef]
  120. Alosert, H.; Savery, J.; Rheaume, J.; Cheeks, M.; Turner, R.; Spencer, C.; Farid, S.S.; Goldrick, S. Data integrity within the biopharmaceutical sector in the era of Industry 4.0. Biotechnol. J. 2022, 17, 2100609. [Google Scholar] [CrossRef]
  121. World Health Organization. Data Quality Assurance: Module 2: Discrete Desk Review of Data Quality; World Health Organization: Geneva, Switzerland, 2022; p. vi, 47p. [Google Scholar]
  122. World Health Organization. Data Quality Assurance: Module 3: Site Assessment of Data Quality: Data Verification and System Assessment; World Health Organization: Geneva, Switzerland, 2022; p. viii, 80p. [Google Scholar]
  123. World Health Organization. Manual on Use of Routine Data Quality Assessment (RDQA) Tool for TB Monitoring; Technical report; World Health Organization: Geneva, Switzerland, 2011. [Google Scholar]
  124. World Health Organization. Data Quality Assessment of National and Partner HIV Treatment and Patient Monitoring Data and Systems: Implementation Tool; Technical report; World Health Organization: Geneva, Switzerland, 2018. [Google Scholar]
  125. World Health Organization. Preventive Chemotherapy: Tools for Improving the Quality of Reported Data and Information: A Field Manual for Implementation; Technical report; World Health Organization: Geneva, Switzerland, 2019. [Google Scholar]
  126. Yourkavitch, J.; Prosnitz, D.; Herrera, S. Data quality assessments stimulate improvements to health management information systems: Evidence from five African countries. J. Glob. Health 2019, 9, 010806. [Google Scholar] [CrossRef]
  127. Hilbert, M.; López, P. The world’s technological capacity to store, communicate, and compute information. Science 2011, 332, 60–65. [Google Scholar] [CrossRef]
  128. Bhat, W.A. Bridging data-capacity gap in big data storage. Future Gener. Comput. Syst. 2018, 87, 538–548. [Google Scholar] [CrossRef]
  129. Dash, S.; Shakyawar, S.K.; Sharma, M.; Kaushik, S. Big data in healthcare: Management, analysis and future prospects. J. Big Data 2019, 6, 1–25. [Google Scholar] [CrossRef]
  130. Abouelmehdi, K.; Beni-Hessane, A.; Khaloufi, H. Big healthcare data: Preserving security and privacy. J. Big Data 2018, 5, 1–18. [Google Scholar] [CrossRef]
  131. Janev, V.; Graux, D.; Jabeen, H.; Sallinger, E. Knowledge Graphs and Big Data Processing; Springer Nature: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
  132. Venkatasubramanian, V.; Zhao, C.; Joglekar, G.; Jain, A.; Hailemariam, L.; Suresh, P.; Akkisetty, P.; Morris, K.; Reklaitis, G.V. Ontological informatics infrastructure for pharmaceutical product development and manufacturing. Comput. Chem. Eng. 2006, 30, 1482–1496. [Google Scholar] [CrossRef]
  133. Yerashenia, N.; Bolotov, A. Computational modelling for bankruptcy prediction: Semantic data analysis integrating graph database and financial ontology. In Proceedings of the 2019 IEEE 21st Conference on Business Informatics (CBI), Moscow, Russia, 15–17 July 2019; Volume 1, pp. 84–93. [Google Scholar]
  134. Villalobos, P.; Ho, A.; Sevilla, J.; Besiroglu, T.; Heim, L.; Hobbhahn, M. Will we run out of data? Limits of LLM scaling based on human-generated data. arXiv 2022, arXiv:2211.04325. [Google Scholar]
  135. Hoseini, S.; Burgdorf, A.; Paulus, A.; Meisen, T.; Quix, C.; Pomp, A. Challenges and Opportunities of LLM-Augmented Semantic Model Creation for Dataspaces. In Proceedings of the European Semantic Web Conference, Crete, Greece, 26–30 May 2024; pp. 183–200. [Google Scholar]
  136. Cigliano, A.; Fallucchi, F. The Convergence of Open Data, Linked Data, Ontologies, and Large Language Models: Enabling Next-Generation Knowledge Systems. In Proceedings of the Research Conference on Metadata and Semantics Research, Athens, Greece, 19–22 November 2024; pp. 197–213. [Google Scholar]
  137. Hassani, S. Enhancing legal compliance and regulation analysis with large language models. In Proceedings of the 2024 IEEE 32nd International Requirements Engineering Conference (RE), Reykjavik, Iceland, 24–28 June 2024; pp. 507–511. [Google Scholar]
Figure 1. “Define, Measure, Analyse, Improve” cycle as outlined by TDQM [6,25] for refining adequate data management processes to implement.
Figure 1. “Define, Measure, Analyse, Improve” cycle as outlined by TDQM [6,25] for refining adequate data management processes to implement.
Bdcc 09 00093 g001
Figure 2. “Plan, Do, Check, Act” cycle, as outlined by the ISO 8000 and ISO 9000 series [4] The cycle follows an iterative process leading to incremental change and improvements and can be applied to design better data management processes.
Figure 2. “Plan, Do, Check, Act” cycle, as outlined by the ISO 8000 and ISO 9000 series [4] The cycle follows an iterative process leading to incremental change and improvements and can be applied to design better data management processes.
Bdcc 09 00093 g002
Figure 3. Data lifecycle as outlined by the the UK Government Data Quality Framework. It describes “the different stages the data will go through from design and collection to dissemination and archival/destruction” [9].
Figure 3. Data lifecycle as outlined by the the UK Government Data Quality Framework. It describes “the different stages the data will go through from design and collection to dissemination and archival/destruction” [9].
Bdcc 09 00093 g003
Figure 4. Gaps in data quality dimensions of the reviewed frameworks. Data quality dimensions present in specific frameworks are highlighted in blue, while absent dimensions are indicated in red.
Figure 4. Gaps in data quality dimensions of the reviewed frameworks. Data quality dimensions present in specific frameworks are highlighted in blue, while absent dimensions are indicated in red.
Bdcc 09 00093 g004
Table 1. Mapping data quality dimensions across frameworks: The first column contains the labels for the DQFs reviewed in this work. The second column lists only those dimensions explicitly described by the frameworks. The third column provides a standardised DQ nomenclature, as detailed by Miller et al. [2], for the data quality dimensions mentioned by the frameworks. Grey shading helps visualise instances where the mapping between the second and third column terms is not one-to-one.
Table 1. Mapping data quality dimensions across frameworks: The first column contains the labels for the DQFs reviewed in this work. The second column lists only those dimensions explicitly described by the frameworks. The third column provides a standardised DQ nomenclature, as detailed by Miller et al. [2], for the data quality dimensions mentioned by the frameworks. Grey shading helps visualise instances where the mapping between the second and third column terms is not one-to-one.
FrameworkDQ DimensionsDQ Common Nomenclature
TDQMAccuracyAccuracy
ObjectivityPrecision
BelievabilityCredibility
ReputationCredibility
AccessAccessibility
SecurityConfidentiality
RelevancyUsefulness
Value-added
TimelinessCurrenctness
CompletenessCompleteness
Amount of DataQuantity
InterpretabilityUnderstandability
Ease of Understanding
Concise RepresentationPrecision
Consistent RepresentationConsistency
ISO 8000AccuracyAccuracy
CompletenessCompleteness
ConsistencyConsistency
TimelinessCurrentness
UniquenessUsefulness
ValidityCredibility
ISO 25012AccuracyAccuracy
AccessibilityAccessibility
AvailabilityAvailability
CompletenessCompleteness
ComplianceCompliance
ConfidentialityConfidentiality
ConsistencyConsistency
CredibilityCredibility
CurrentnessCurrentness
EfficiencyEfficiency
PrecisionPrecision
PortabilityPortability
RecoverabilityRecoverability
TraceabilityTraceability
UnderstandabilityUnderstandability
FIPPSAccuracyAccuracy
RelevancyUsefulness
TimelinessCurrency
CompletenessCompleteness
ESS QAFStatistical Confidentiality and Data ProtectionConfidentiality
Accessibility and ClarityAccessibility
Availability
Understandability
Traceability
RelevanceUsefulness
Timeliness and PunctualityCurrentness
Accuracy and ReliabilityAccuracy
Impartiality and ObjectivityCredibility
Cost EffectivenessEfficiency
Coherence and ComparabilityConsistency
UK GOV DQF
(DAMA DMBoK)
CompletenessCompleteness
ConsistencyConsistency
TimelinessCurrentness
UniquenessUsefulness
ValidityCredibility
AccuracyAccuracy
IMFPrerequisites of qualityUsefulness
Assurance of IntegrityCredibility
Traceability
Methodological SoundnessSemantics
Accuracy and ReliabilityAccuracy
ServiceabilityConsistency
Currency
Traceability
AccessibilityAccessibility
Understandability
BCBS 239AccuracyAccuracy
Clarity and UsefulnessUnderstandability
ComprehensivenessCompleteness
FrequencyCurrentness
DistributionGovernance
ALCOA+AccurateAccuracy
AttributableTraceability
AvailableAvailability
CompleteCompleteness
ConsistentConsistency
EnduringGovernance
LegibleUnderstandability
OriginalTraceability
WHOCompletenessCompleteness
TimelinessCurrentness
Internal ConsistencyConsistency
Accuracy
External ConsistencyConsistency
Accuracy
Consistency of Population DataGovernance
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Miller, R.; Chan, S.H.M.; Whelan, H.; Gregório, J. A Comparison of Data Quality Frameworks: A Review. Big Data Cogn. Comput. 2025, 9, 93. https://doi.org/10.3390/bdcc9040093

AMA Style

Miller R, Chan SHM, Whelan H, Gregório J. A Comparison of Data Quality Frameworks: A Review. Big Data and Cognitive Computing. 2025; 9(4):93. https://doi.org/10.3390/bdcc9040093

Chicago/Turabian Style

Miller, Russell, Sai Hin Matthew Chan, Harvey Whelan, and João Gregório. 2025. "A Comparison of Data Quality Frameworks: A Review" Big Data and Cognitive Computing 9, no. 4: 93. https://doi.org/10.3390/bdcc9040093

APA Style

Miller, R., Chan, S. H. M., Whelan, H., & Gregório, J. (2025). A Comparison of Data Quality Frameworks: A Review. Big Data and Cognitive Computing, 9(4), 93. https://doi.org/10.3390/bdcc9040093

Article Metrics

Back to TopTop