Next Article in Journal
A Globally Optimal Alternative to MLP
Previous Article in Journal
Towards Accurate Thickness Recognition from Pulse Eddy Current Data Using the MRDC-BiLSE Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Mining User Perspectives: Multi Case Study Analysis of Data Quality Characteristics

1
Centre for Society and Policy, Indian Institute of Science, Bangalore 560012, India
2
Department of Management Studies, Indian Institute of Science, Bangalore 560012, India
*
Author to whom correspondence should be addressed.
Information 2025, 16(10), 920; https://doi.org/10.3390/info16100920
Submission received: 28 August 2025 / Revised: 7 October 2025 / Accepted: 15 October 2025 / Published: 21 October 2025

Abstract

With the growth of digital economies, data quality forms a key factor in enabling use and delivering value. Existing research defines quality through technical benchmarks or provider-led frameworks. Our study shifts the focus to actual users. Thirty-seven distinct data quality dimensions identified through a comprehensive review of the literature provide limited applicability for practitioners seeking actionable guidance. To address the gap, in-depth interviews of senior professionals from 25 organizations were conducted, representing sectors like computer science and technology, finance, environmental, social, and governance, and urban infrastructure. Data are analysed using content analysis methodology, with 2 level coding, supported by NVivo R1 software. Several newer perspectives emerged. Firstly, data quality is not simply about accuracy or completeness, rather it depends on suitability for real-world tasks. Secondly, trust grows with data transparency. Knowing where the data comes from and the nature of data processing matters as much as the data per se. Thirdly, users are open to paying for data, provided the data is clean, reliable, and ready to use. These and other results suggest data users focus on a narrower, more practical set of priorities, considered essential in actual workflows. Rethinking quality from a consumer’s perspective offers a practical path to building credible and accessible data ecosystems. This study is particularly useful for data platform designers, policymakers, and organisations aiming to strengthen data quality and trust in data exchange ecosystems.

Graphical Abstract

1. Introduction

Data quality (DQ) forms the foundational element for effective decision-making, innovation, and analytics. High-quality datasets streamline operations and provide organisations with a competitive advantage, particularly in rapidly evolving markets [1]. Historically, DQ has been conceptualised as a multidimensional construct encompassing intrinsic, contextual, and representational attributes. However, the exponential increase in global data generation, with International Data Corporation (IDC) predicting the Global Datasphere growth from 33 zettabytes (ZB) in 2018 to 175 ZB by 2025 [2], presents significant challenges in ensuring sustained DQ.
Data-sharing platforms have become crucial in managing the data surge. The growth of the data economy is driven by interactions between data providers, consumers, and platform intermediaries [3]. Data-sharing platforms form a central part of the ecosystem, facilitating access to large, diverse datasets [4]. However, most platforms focus on supply-side priorities such as volume, variety, and pricing, often neglecting consumer needs like traceability, usability, and relevance [5].
In India, the challenges of fragmented datasets, limited platform standardisation, and immature regulatory frameworks remain significant barriers to effective implementation [6]. For example, recent analyses of India’s Smart City Mission further highlights lack of unified data platforms hampering DQ interoperability [6]. Furthermore, smart city governance emphasizes the need for robust data frameworks, standardized data exchange platforms, and clear regulatory policies [6,7]. The Organization for Economic Cooperation and Development (OECD) report [6] notes retention or utilization of a small fraction of generated Internet of Things (IoT) data, underscoring the urgency of establishing national data strategies and governance mechanisms. The United Nations Habitat World Smart Cities Outlook [8] further stresses on fragmented data platforms and varying digital maturity levels impeding citizen centric services and limiting consumer influence. Moreover, recent evaluations of smart city projects in India reveals weak regulatory enforcement and immature data governance structures limiting the realization of real-time decision-making capabilities [6].
Furthermore, worldwide, most research on DQ has focused on theoretical frameworks, technical solutions, and provider-centric assessments, in the attempt to capture the multifaceted complexity of quality. The literature shows limited empirical exploration of how data consumers—those who utilise data in everyday decision-making—perceive, experience, and act upon DQ. The gap becomes particularly evident in the Indian context, where the data ecosystem is diverse and characterised by varying levels of digital maturity and limited standardisation.
While our review led to the identification of 37 distinct DQ dimensions (see Appendix A), such comprehensive listings can be overwhelming and do not inherently provide guidance on aspects more critical for real-world applications. These theoretical parameters raise a key practical question: out of all possible dimensions, which ones truly matter most to data users for everyday work? Addressing the gap, this study adopts a consumer-centric approach to explores how Indian data consumers interpret, experience, and act upon DQ in platform-based data exchange environments.
This study aims to move beyond theory by empirically investigating how data consumers prioritise and interpret DQ, to develop practical guidance to identify which aspects matter most for real-world applications. The 37 identified dimensions provide a conceptual baseline. Adopting a multiple-case study approach, following the methodology of Urbinati et al., 2019 [9], in-depth interviews were conducted with 25 senior data users across sectors such as artificial intelligence, financial services, environmental, social, and governance ESG consulting, and urban mobility to surface the lived experiences and practical definitions of DQ. Semi-structured interviews were conducted using Microsoft Teams between June and January 2024, with the interview protocol focusing on the following dimensions, namely, definitions and relevance of DQ for business goals; challenges in sourcing, cleaning, and integrating datasets; expectations from data providers and platforms; views on metadata, documentation, and standards; perceptions of pricing models, licensing, and monetisation strategies; and trust mechanisms, validation practices, and legal concerns in data sharing. The qualitative inquiry allows us to connect the general dimensions with actionable, context-driven priorities, and to determine consensus among users on what makes data “fit for use.”
Despite the extensive literature on DQ (see Appendix A), current frameworks remain provider-centric, privileging technical specifications over lived user needs. These approaches have limited capacity to explain why data may be “usable” yet not “used.” This creates a research gap: consumer perspectives on DQ remain underexplored, particularly in emerging economies such as India where platform ecosystems are rapidly evolving. Accordingly, the central research question guiding this study is: Which DQ dimensions matter most to Indian data consumers in everyday practice, and how do these priorities align with or diverge from existing theoretical frameworks?
This study offers four key contributions. First, this study reframes DQ assessments by shifting the focus from provider-led definitions to consumer-driven evaluations, highlighting the practical realities faced by Indian data consumers. Second, the paper provides sector-specific insights into the distinct usability, trust, and governance challenges encountered across industries. Third, we identify both social and technical barriers to effective use and integration of external datasets. Finally, this study presents actionable recommendations for the development of more responsive, consumer-informed, and trust-oriented data platforms.
While prior studies on consumer-centric governance and monetisation have largely focused on Europe and North America, this paper contributes originality by situating consumer perspectives within the Indian digital ecosystem. India is both a major hub for global data services and a representative Global South setting, where infrastructural constraints, regulatory frameworks, and market dynamics differ substantially from those in the Global North. By analysing how Indian data consumers evaluate DQ, usability, governance, and monetisation, this study extends platform governance debates beyond their dominant geographies and surfaces newer, context-specific quality concerns not captured in earlier work.
The remainder of this paper is structured as follows. Section 2 provides the background and reviews the relevant literature. Section 3 outlines the materials and methods, describing the case selection, participant profile, and procedures for data collection and analysis. Section 4 presents the results, highlighting consumer perspectives on fit-for-purpose quality, usability and format frictions, transparency and governance requirements, and monetisation concerns. Section 5 discusses the implications of these findings, reframing DQ as a consumer-driven construct and exploring platform frictions and governance challenges. Finally, Section 6 concludes this study, summarising the contributions and suggesting directions for future research.

2. Background

Table 1 summarizes the literature studies. Four dimensions are investigated, namely, DQ assessment frameworks and methods; consumer-centric governance in data platforms; monetisation models and consumer participation; and clinical and business protocols: lessons for DQ.
Table 1 provides a consolidated overview of the prior literature, highlighting the objectives, benefits, and persistent barriers across different domains. This summary underscores the fragmented nature of existing approaches and frames the need for a consumer-centric perspective adopted in this study.

2.1. DQ Assessment Frameworks and Methods

DQ research has evolved from conceptual frameworks to quantitative models and automated tools aiming to assess and improve DQ systematically. Historical contributions by Hoare (1975), Chapple (1976), Fox et al. (1994), and Ballou and Pazer (1985) [45,46,47,48] laid the foundation for defining and modelling data and process quality. De Amicis et al., 2006 [49] developed an analytical framework to assess interdependencies among DQ dimensions, which remains relevant for modern ecosystems. Wang and Strong, 1996 [1] introduced a multidimensional framework categorising DQ as intrinsic, contextual, representational, and accessibility-based. Later studies [11,12] expanded these models, though most remain provider-focused and insufficiently tested in complex, real-world environments.
Several studies have developed objective metrics and scalable tools to assess DQ [13,14,15,16,17]. Studies [18,19] further advanced automation for streaming and big data environments. However, these approaches often rely on clean, stable, and well-structured datasets, which are rarely available.
Emerging research advocates flexible, real-time assessment methods which can adapt to evolving data environments [2,4,20]. Sadiq and Indulska, 2017 [21] and Cai and Zhu, 2015 [22] have laid the groundwork for more dynamic quality assessment, a concept further advanced by recent studies. For instance, Abedjan et al., 2016 [23] has called for more emphasis on uniqueness and task-independent evaluation, critical for platform-based systems. This aligns with broader discussions on data governance in socio-technical systems [50].

2.2. Consumer-Centric Governance in Data Platforms

Platforms increasingly recognise the need to prioritise consumer needs such as transparency, traceability, privacy, and fairness [27] and DQ frameworks like Consumer-Oriented Data Control and Auditability (CODCA) [25] provide mechanisms for user-driven transparency and auditability, enhancing trust in data custodians. The automated governance framework [24] emphasizes how consumer perceptions and technical realities of DQ can be aligned through automated validation and governance processes, thereby supporting participatory and trustworthy data ecosystems. Complementing this, Acev et al., 2025 highlights foundational governance mechanisms such as provenance tracking, access control definitions, accountability structures, and interoperability standards underpinning consumer trust and elevate DQ within platforms facilitating shared data use [26].
Otto and Jarke (2019) [27] propose participatory ecosystems where consumers contribute directly to platform rules and data validation mechanisms. Veltri et al. (2020) [28] demonstrate greater platform transparency to significantly influence consumer trust and decision-making. Ducuing and Reich (2023) [29] explore governance tensions in emerging digital product passport frameworks, underscoring the need for balanced oversight and consumer influence. Agahari et al., 2022 [30] emphasise trust principles and participatory mechanisms as critical for sustainable data ecosystems. Dynamic consent models [44] offer practical governance solutions by enabling ongoing consumer control over data sharing, consent, and platform participation. These frameworks provide pathways for real-time consumer involvement and highlight need for participatory governance.

2.3. Monetisation Models and Consumer Participation

Understanding monetisation models and willingness to pay in the context of data platforms forms a significant gap in the literature. However, pricing opacity and the poor quality of available services frequently diminished the willingness. Yu and Zhang, 2017 [31] proposed a data-pricing strategy based on DQ, employing a bi-level programming model and a multi-version approach to enhance market segmentation, profitability, and consumer utility. Yang et al., 2019 [32] developed subscription models to jointly optimise DQ and consumer willingness to pay. Zhang et al., 2018 [33] introduced fairness-based pricing mechanisms, while Pei [34] and Miao et al., 2023 [35] offered comprehensive reviews of data-pricing strategies, including quality-driven, feature-based, and query-based pricing. Zhang et al., 2023 [4] provided a taxonomy of pricing models, further emphasising the importance of transparent, consumer-centric monetisation frameworks.
Malieckal et al., 2024 [37] extends the argument by demonstrating how pricing can be dynamically linked to concrete DQ dimensions such as timeliness, completeness, and contextual relevance through forecasting-based and AI-enabled mechanisms. The approach operationalises the conceptual link between quality and value, ensuring monetisation directly reflects consumers priorities. Despite these advances, limited empirical work has explored how consumer preferences shape pricing expectations in real-world data platforms. This study highlights the importance of tiered pricing, DQ-linked service level agreements (SLAs), and participatory pricing mechanisms.

2.4. Clinical and Business Protocols: Lessons for DQ

In clinical research, the SPIRIT 2013 guidelines [38] demonstrate how structured, detailed documentation and active stakeholder participation can significantly improve data usability and trust. Juddoo et al. (2018) [39] and Facile et al., 2022 [40] further reinforce the role of Clinical Data Interchange Standards Consortium (CDISC) protocols and key DQ dimensions such as accuracy, completeness, and timeliness in ensuring interoperability and traceability across clinical data systems.
Patient-centred studies [41] reveal strong preferences for transparent data sharing practices, semantic clarity, and participatory governance mechanisms. The dynamic consent model [44] and good clinical data management practices [40] reinforce the importance of traceability, auditability, and consumer governance in sensitive data environments. Etheredge (2007) [43] introduced the concept of a learning health system, which emphasises continuous improvement through consumer feedback and iterative data-driven development. These clinical governance structures provide useful frameworks to improve trust and usability in broader data-sharing platforms.

3. Materials and Methods

This study adopts a multiple-case study approach, following the methodology of Urbinati et al., 2019 [9], to explore how Indian data consumers experience and interpret DQ (DQ) in real-world settings.

3.1. Case Selection and Participant Profile

Participants are selected using purposive sampling to ensure they met the following criteria:
Regular interaction with third-party or external datasets.
Active use of data for decision-making, modelling, or product development.
Organisational responsibility for data-related strategy, integration, or evaluation.
A total of 25 cases are included, each representing an individual data consumer or a data-driven function within an organisation. The sample size of twenty-five cases was determined based on thematic saturation. Thematic saturation was reached after the seventeenth interview, when no substantially new insights emerged from subsequent data collection. At this stage, additional interviews yielded confirmatory rather than novel information. Nevertheless, further interviews were conducted to ensure adequate sectoral representation and to capture potential contextual variations across domains. This extended sampling strategy enhances the analytical generalisability and robustness of cross-case comparisons. Such an approach aligns with qualitative research conventions recognising that saturation is typically achieved between twelve and twenty interviews in focused studies, though additional cases are often collected in heterogeneous, multi-sector research to consolidate and verify emergent themes [51,52]. The sample included participants from the following sectors:
Computer Science and Technology
Financial Services and Banking
ESG Consulting and Workplace Strategy
Urban Infrastructure and Transport Analytics
Public Research and Statistical Modelling
Participants held senior positions such as CEOs, Product Owners, Data Architects, and Senior Managers (Table 2). Organisations ranged from startups and multinational firms to public sector bodies and consulting companies. The sectoral and organisational diversity of the sample allowed for rich cross-case comparison. The methodological choice is motivated by a desire to go beyond comprehensive but abstract DQ frameworks and to identify quality dimensions according to data users themselves.
While the sampling framework grouped participants into five broad sectors, the thematic analysis in the Results section further differentiated these into domain-specific categories. For example, “Financial Services and Banking” was analysed in greater detail as Finance and Analytics, Finance and Data Science, and Banking and FinTech. Similarly, “ESG Consulting and Workplace Strategy” corresponds to the Sustainability and General Business categories, while “Urban Infrastructure and Transport Analytics” aligns directly with Urban/Smart Cities. Additional categories such as Healthcare, Energy/Oil and Gas, and Retail/Manufacturing emerged inductively from participants’ job profiles and sector-specific discussions. This expanded classification allowed for richer sectoral comparisons in the results.

3.2. Data Collection

Semi-structured interviews were conducted using Microsoft Teams between June 2023 and January 2024. All interviews were conducted in English, which was the working language of participants, and lasted between 45 and 60 min and followed an interview protocol focusing on the following dimensions:
Definitions and relevance of DQ for business goals.
Challenges in sourcing, cleaning, and integrating datasets.
Expectations from data providers and platforms.
Views on metadata, documentation, and standards.
Perceptions of pricing models, licensing, and monetisation strategies.
Trust mechanisms, validation practices, and legal concerns in data sharing.
Participants provided consent prior to the interviews, and the interview guide was shared upon request. All interviews were audio-recorded, transcribed verbatim, anonymised, and manually cleaned for accuracy and completeness.

Informed Consent Statement and Approvals

All participants provided informed consent at the start of each interview. They were informed about the purpose of the research, the voluntary nature of participation, confidentiality assurances, and how the data would be used. Participants were explicitly told that, at any point, they could request the audio recording to be turned off and their response kept undocumented. They retained the right to withdraw their participation at any stage without consequence. No personally identifiable information (such as names or company details) has been disclosed without explicit permission.

3.3. Data Analysis

The cleaned transcripts are analysed using NVivo R1 (Release 1.7.2) software. A hybrid coding approach was applied. To ensure analytical reliability, the primary researcher conducted the coding of all transcripts. Coding decisions and emerging interpretations were periodically discussed with the co-author to review consistency and consolidate shared understanding. These discussions served as an informal reliability check and helped refine the thematic structure and definitions of key categories.
Deductive codes are developed based on established DQ frameworks [1,13] covering completeness, accuracy, timeliness, consistency, traceability, and documentation (Figure 1).
Inductive codes are derived from participant narratives, including terms and concerns specific to the Indian data ecosystem, such as, ‘data freshness’, ‘lineage visibility’, ‘manual cleaning burden’, ‘version control gaps’, and ‘APIs without schema’.
The dual coding approach allowed validation of established DQ dimensions while surfacing novel, practice-driven themes from the Indian context. Table 3 and Table 4 present a summary of the themes and the number of participants who raised them. The emerging thematic clusters are (1) Fit-for-Purpose Quality, (2) Usability and Format Frictions, (3) Transparency and Traceability and Governance, and (4) Monetisation and Sharing Models. In this paper, “extracts” refer to coded segments of interview transcripts, ranging from short phrases (e.g., “missing values issue”) to longer verbatim quotations that capture participants’ concerns in context. Each extract is a discrete passage linked to a theme or sub-theme during coding.
Illustrative quotes are extracted to support thematic validity and analytic depth. The approach follows methodological precedents from similar digital transformation studies [9], enabling both domain-specific insight and theory extension.
To systematically assess the strength and prominence of each domain’s engagement with identified DQ themes, we developed a theme engagement rating (Table 5). The coding tables (Table 3, Table 4 and Table 5) provide structured transparency by showing how raw interview data were systematically aggregated into first- and second-order codes and then synthesised into higher-level conceptual categories. These tables are not merely descriptive; they operationalise the methodological logic of the analysis by illustrating the progression from empirical excerpts to abstract themes. The theme engagement rating (Table 5) further strengthens methodological rigour by quantifying the relative emphasis of each theme across sectors, following the comparative approach outlined by Urbinati et al. (2019) [9]. The method classifies intensity levels on a four-point scale: High, Medium, Low, and None, based on:
The number of participants from each domain who discussed the theme,
The depth of the participant engagement (detail, multi-point extracts), and
The diversity of concerns (number of distinct sub-themes or sub-codes addressed).
While Appendix A catalogues 37 dimensions from decades of literature, our coding distilled them into four higher-order consumer-centric themes (see Table 3). This shows convergence around contextual usability and divergence on governance and monetization. The four emergent themes Fit-for-Purpose Quality, Usability Challenges, Governance/Trust, and Monetisation draw primarily from contextual (relevance, timeliness), representational (consistency, documentation), and accessibility (traceability, usability) dimensions, while placing limited emphasis on intrinsic measures like precision or minimality. This empirical condensation demonstrates the practical narrowing of theoretical constructs in lived contexts.

4. Results

Although four initial clusters emerged during first-level coding—Fit-for-Purpose Quality, Usability and Format Frictions, Transparency and Governance, and Monetisation—the final analysis concentrates on the first three dominant themes. These three capture the most frequent, conceptually rich, and cross-sectorally consistent concerns, representing over 95% of all coded extracts. The Monetisation category, while relevant, appeared sparsely discussed and is therefore referenced briefly within the discussion section. These categories emerged consistently across diverse sectors and organisational types. While the extracts illustrate sectoral diversity, the emphasis hereafter shifts from description to analysis, showing how recurring patterns across cases translate into theoretical constructs of consumer-centric data quality.

4.1. Fit-for-Purpose Quality: Needs Expressed by Data Consumers

The theme of fit-for-purpose quality captures the critical requirements data consumers have regarding the specific operational contexts. Across sectors, participants emphasise data must be corrected to align with particular business decisions. The diversity of application domains—from financial services to AI and urban mobility—further illustrated how DQ forms a nuanced and dynamic concept, contingent on domain-specific workflows and decision-making needs.
Across all sectors, participants consistently emphasise the importance of data being internally consistent, for timeliness, credibility, and completeness. Appropriate documentation was particularly stressed as foundational to data usability (Table 6, Table 7 and Table 8).
Table 6. Participant statements on Fit-for-Purpose Quality.
Table 6. Participant statements on Fit-for-Purpose Quality.
StatementParticipant
 “If even if, let’s say the accuracy is low, however, we measure the accuracy as long as they help in making business decisions.” AI Startup (P19)
“I think accuracy and newness of data matters because what happens is like there is data available, but those surveys have been done in 2020. Know how relevant they are in 2023 is questionable.”Mental Health Startup (P20)
“ There needs to be some reasonable level of accuracy, but the accuracy need not be 100% as long as the data is good enough to generate business decisions, that’s fine.”AI Startup (P19)
“In the open source systems, not really the paid version because the paid version didn’t have all the variety… when we want to buy data, we have to see the quality.”IT Consulting (P1)
“these are the basic things, to basically trust the data it has to be internally consistent.”AI Startup (P19)
“If you take the names of districts in India, it should be consistent within the same data sources and across data sources.”AI Startup (P19)
“Data validation is first thing… the context always changes with my digital footprintTechnology Private Firm (P6)
“If it has been used widely, all of that adds to credibility… if you cannot trust the data then you can’t do anything from it.”Oil and Gas Multinational (P24)
“Data comes in some file format which is not yet standardized. It could be Excel files without exact metadata about when it was collected or what standards it follows. There’s no description of the columns, inconsistent nomenclature, and the data is mostly static.”Urban Transport Analytics (P7)
“World Wide Web Consortium has given guidelines on metadata standards… all G20 countries adopted it except India.”Finance/Law Non-Profit (P18)
“The ability to look at how data has transformed from source till date… whole transformation lineage is critical.”Independent Consultant (P4)
“Publishing data methodology or data cookbook along with the dataset.”Finance/Law Non-Profit (P18)
Table 7. Sector-wise needs expressed.
Table 7. Sector-wise needs expressed.
SectorParticipants (IDs)Key Needs/Issues
IT/Tech and AIP1, P6, P10, P19, P22API documentation, accuracy for decisions, mission-critical data relevance
Finance and AnalyticsP4, P5, P8, P14Data freshness, lineage tracking, real-time updates
Urban/Smart CitiesP7, P12, P16Standardized formats, metadata completeness, context-aware data
HealthcareP17, P20, P25Historical data accuracy, medical subcode specificity
Energy/Oil and GasP11, P24Instrument reliability, credibility vs. real-world observations
Retail/ManufacturingP15, P21Large-volume processing, label consistency (e.g., product sizes)
SustainabilityP9, P23Cross-departmental data cleanliness, triangulation validation
Banking/FinTechP13Truthfulness verification, transaction data integrity
Table 8. Sector-wise engagement rating for Theme 1.
Table 8. Sector-wise engagement rating for Theme 1.
SectorParticipants (IDs)Depth/DiversityRating
IT/Tech and AIP1, P6, P10, P19, P22Detailed, multi-pointHigh
Finance and AnalyticsP4, P5, P8, P14Detailed, diverse issuesHigh
Urban/Smart CitiesP7, P12, P16ModerateHigh
HealthcareP17, P20, P25ModerateHigh
Energy/Oil and GasP11, P24BriefLow
Retail/Manufact.P15, P21BriefLow
SustainabilityP9, P23Some diversityMedium
Banking/FinTechP13BriefLow
The quotes reflect the pragmatic acceptance among some users on absolute technical precision being sometimes secondary to usability and decision support. The reflections underscore quality being judged in relation to the task rather than generic metrics, highlighting the importance of contextual relevance. The participants also stress core attributes of internal consistency and accuracy, demonstrating how these attributes build confidence in data-driven processes. The concern resonates strongly in sectors like finance and healthcare, where faulty decisions carry high stakes.
These insights emphasise the role of transparent documentation and lineage in accelerating user adoption and fostering trust. Particularly in advanced analytics and AI, understanding data provenance forms a prerequisite for reliable modelling. Overall, the analysis shows accuracy and freshness as paramount. Furthermore, participants evaluate DQ situationally, prioritizing task relevance and ease of evaluation through documentation and metadata. Thus, platform designs should focus on comprehensive metadata provision, clear lineage tracking, and domain-specific quality indicators to serve heterogeneous consumer needs effectively.
Participants from nearly all sectors described nuanced needs, relating to business relevance, timely access, and documentation. For example, those in IT/Tech and AI and Finance and Analytics domains spoke in detail about requirements for API documentation, data accuracy, and freshness, while sectors like Sustainability and Retail raised more limited or specialized concerns. Table 7 summarizes these patterns and captures the intensity of engagement for each domain, measured by the number of participants, the detail and diversity of the comments, and the overall prominence of the engagement.
The sectoral contrasts observed in Table 6, Table 7 and Table 8 illustrate how domain-specific priorities shape interpretations of fit-for-purpose quality. Finance and AI sectors emphasise accuracy, freshness, and lineage, driven by compliance obligations and automation-dependent decision systems. Urban infrastructure and sustainability domains foreground standardisation and metadata completeness to support interoperability and collaborative governance across multiple stakeholders and data custodians. In comparison, healthcare and ESG actors stress credibility and documentation, reflecting regulatory scrutiny and ethical accountability. These distinctions demonstrate that data quality is contextually enacted, contingent on sectoral workflows, risk exposure, and governance frameworks.

4.2. Usability and Format Frictions: Common Problems Encountered

Participants commonly reported significant usability challenges related to data integration, cleaning, and inconsistencies caused by multi-format inflows and missing values. These frictions not only increased operational overheads but also contributed to mistrust when data providers failed to ensure completeness or transparency. The burden of cleaning and organizing data forms a recurring concern, especially in sectors processing large volumes of heterogeneous inputs (Table 9, Table 10 and Table 11).
Table 9. Participant statements on Usability and Format Frictions.
Table 9. Participant statements on Usability and Format Frictions.
StatementParticipant
 “Eighty percent of my team’s effort is spent cleaning data and making sure it is all right—so much time wasted.” AI Startup (P19)
“Data cleaning takes up a lot of time.”Software (P3)
“The data cleaning process is definitely there… more than 50% [of time] goes into data cleaning only.”Financial Services (P8)
“Absence of data as well as missing information.”IT Consulting (P1)
“I do not know whether it is an applicable data source or not… there are missing values.”AI Startup (P19)
“Saving data in an organized format and handling multi-format data are some of the issues that we are facing when it comes to raw data handling.”Banking (P13)
“Standardization is needed to scale and explore and move into the market quickly .”Urban Transport Analytics (P7)
“The issue in most of the manufacturing area in India is that they’re not labelled the same way.”Banking/FinTech (P25)
Table 10. Sector-wise usability issues and format frictions.
Table 10. Sector-wise usability issues and format frictions.
SectorParticipants (IDs)Common Frictions
IT/Tech and AIP3, P10, P1980% of effort spent cleaning data, missing values, formatting overheads
Finance and Data ScienceP5, P8, P13, P14Multi-format chaos (CSV/Excel), manual noise reduction
Urban/Smart CitiesP7, P16Non-standardized files, static datasets without metadata
HealthcareP17, P25Unfilled fields, inconsistent medical labels (e.g., “Size L” ambiguity)
Energy/Oil and GasP11Sensor drift, instrument calibration issues
Retail/ManufacturingP15, P21Harmonization gaps (e.g., product IDs across vendors)
General BusinessP1, P6, P9Duplicate data, lack of documentation
Table 11. Sector-wise engagement rating for Theme 2.
Table 11. Sector-wise engagement rating for Theme 2.
SectorParticipants (IDs)Depth/DiversityRating
IT/Tech and AIP3, P10, P19Heavy cleaning, missing valuesHigh
Finance and Data ScienceP5, P8, P13, P14Multi-format issues, manual fixesHigh
Urban/Smart CitiesP7, P16Non-standardized files, metadataMedium
HealthcareP17, P25Unfilled fields, inconsistent labelsMedium
Energy/Oil and GasP11Sensor drift, calibrationLow
Retail/Manufact.P15, P21Harmonization gapsLow
General BusinessP1, P6, P9Duplication, lack of documentationHigh
The quotes highlight the heavy manual labour involved in preparing data for use, a critical barrier to scale and efficiency. Such extensive cleaning efforts reduce capacity for analysis and value extraction, impacting product timelines and innovation cycles. The participant underscores how even common DQ issues such as missing values propagate through workflows, undermining reliability and trust. The impact was felt acutely in modelling-intensive sectors like technology and finance.
The quotes also reflect a broad call for standardization to reduce format frictions, enabling interoperability and faster onboarding. Further, the analysis points to how fragmented data ecosystems increase costs and delay deployment. The analysis reveals usability and format challenges as central impediments to efficient data consumption, suggesting platform designs to prioritize automated cleaning aids, format standardization, and interoperability tools. Building trust also requires transparent communication about DQ issues from providers upfront.
Major usability pain points, such as extensive cleaning burdens and difficulties with multi-format data, were most vividly described by participants from IT, Finance, and General Business backgrounds, with others echoing similar but less frequent or less elaborate challenges. Table 9 presents domain-level engagement intensity.
The sectoral contrasts captured reveal that usability frictions intensify in ecosystems characterised by high data heterogeneity. Finance, technology and AI, and urban analytics sectors report substantial integration burdens due to multi-format inflows and variable schema quality, while retail and energy domains encountering more bounded operational challenges such as labelling inconsistencies or calibration issues. These differences affirm that cleaning overheads and interoperability gaps are structurally embedded within sectoral data architectures, reinforcing the importance of shared standards and collaborative governance mechanisms to streamline integration.

4.3. Platform Requirements: Transparency, Traceability, and Governance

Consumer expectations for data transparency and governance revolve around certification, authentication, standardization, and traceability guarantees from data providers. Participants expressed a need for verifiable quality credentials and common standards to reduce assessment burdens and increase trust. Governance mechanisms providing provenance and transformation histories become essential, particularly for sensitive or critical data (Table 12, Table 13 and Table 14).
The insights highlight the importance of quality metrics and standards, facilitating easier evaluation and comparability across providers. Insight accentuates the role of governance frameworks in operationalizing quality definitions. The participants emphasize the necessity for detailed lineage and governance artifacts to ensure data integrity over time, especially in domains with complex data flows or regulatory oversight.
Overall, transparency and governance emerged as essential for platform credibility and consumer trust. Implementing standardized certification, metadata to support lineage and version control, and clear governance policies will be indispensable design considerations for data marketplaces or platforms aiming to cater to complex consumer ecosystems.
Transparency-related needs such as demand for certification, transformation logs, and auditable processing was discussed in moderate detail by participants across several sectors. Table 14 captures these differences by domain.
The governance patterns summarised in Table 12, Table 13 and Table 14 reflect differentiated trust imperatives across sectors. Finance and legal domains prioritise version control and audit trails to align with stringent compliance regimes, whereas sustainability and public-sector actors emphasise methodology transparency and source validation to enable collaborative governance and stakeholder confidence. These variations highlight how governance mechanisms are selectively configured, shaped by regulatory maturity, institutional mandates, and the relational nature of trust within each domain.

5. Discussion and Implications

The discussion moves beyond descriptive reporting of user concerns to interpret findings through established theoretical lenses. Drawing on frameworks such as Wang and Strong’s (1996) [1] multidimensional model, Human–Data Interaction (HDI), and socio-technical systems theory, the analysis conceptualises data quality as a contextually negotiated construct shaped by both technical affordances and organisational governance. This theoretical integration clarifies how consumer-driven notions of “fit-for-purpose” quality reshape the meaning of data reliability and usability in platform contexts. Rather than a fixed attribute, quality is co-produced through users’ interactions with data, platforms, and governance mechanisms.

5.1. DQ as a Consumer-Driven Construct

This study contributes to the growing shift in the data governance literature from provider-led to consumer-led interpretations of DQ. Traditionally, DQ has been dominated by provider-centric frameworks with intrinsic measures such as correctness, consistency, and completeness. However, these approaches often fall short of capturing what actually matters to users in practice. Our findings challenge the orthodoxy: participants rarely described quality in abstract terms. Instead, quality was framed in relation to the decisions, workflows, and risks embedded in specific contexts of use.
Quality, from the consumer perspective, is judged by fitness for purpose and ability to integrate seamlessly into operational routines, Attributes such as contextual relevance, timeliness, and documented according to consumer needs emerged as essential. Across sectors such as artificial intelligence, oil and gas, and infrastructure, we observed varied tolerances for “imperfections,” contingent on the risk and stakes of the task. These findings echo recent critiques of static, universally applied quality metrics [27] and reinforce the idea of quality not being an inherent property of the dataset, but something constructed and enacted through use. The findings substantiate key tenets of Human–Data Interaction (HDI) [53,54] by demonstrating that data quality is experienced through usability, traceability, and contextual relevance rather than defined solely by intrinsic metrics. This experiential framing aligns with HDI’s principles of legibility, agency, and negotiability, wherein users actively interpret and shape data’s value in practice. In this view, participants’ emphasis on metadata and lineage reveals how data quality becomes a relational construct co-produced through ongoing human–data engagements rather than a static technical property.
Moreover, participants consistently called for richer, more customizable metadata, explicit lineage information, and dynamic documentation. These features allow consumers not only to assess but also to shape and co-produce quality standards in real time. The shift indicates data platforms should move beyond fixed metrics and instead design tools to empower consumers for negotiating quality for specific purposes.

5.2. Platform Frictions: Usability as a Neglected Priority

The persistent usability barriers reported in the analysis point to a significant disconnect between technical platform design and consumer priorities. While platforms often compete on volume, velocity, and cataloguing, participants from both technical and business backgrounds consistently described the bulk of effort going into cleaning, validating, and integrating external datasets before they became operationally useful.
Issues such as inconsistent formats, incomplete documentation, schema mismatches, and limited update transparency are seen as substantial productivity drains. Participants also expressed a recurring distrust of provider claims, due to poor interoperability and opaque preprocessing steps. These frictions reveal a critical gap in platform design thinking. From a socio-technical systems perspective, these frictions arise from the misalignment between human work practices and the technological infrastructures of data platforms. Data quality and usability are therefore emergent properties of interaction between people, processes, and tools rather than outcomes of system design alone. Documentation deficits, inconsistent schemas, and manual cleaning burdens highlight the need to re-engineer socio-technical configurations that support mutual adaptation between users and data technologies—an alignment central to sustainable data-quality management [50].
These findings echo prior research on the barriers posed by poor interoperability and fragmented governance [3]. For platforms to support scalable and trusted data exchange, usability must be treated as a fundamental design goal. Investments in schema standardization, automated validation tools, robust documentation practices, and clear communication of known limitations will evidently be required.
Platforms must enable rich metadata and lineage documentation to address fit-for-purpose needs while reducing cleaning and integration frictions through standardization and automation. Certification and transparent governance mechanisms establish trust and facilitate efficient data exchange. Addressing these interconnected challenges holistically will be essential to developing robust, user-centred data platforms capable of serving diverse sectors and evolving use cases.

5.3. The Central Role of Transparency and Governance

Trust in data platforms is strongly linked to DQ per se as well as governance mechanisms surrounding the platforms. Participants expected platforms to move beyond passive hosting and instead take active responsibility for ensuring data lineage, authentication, and accountability. Features such as detailed transformation logs, version control, certification, and complete provenance were repeatedly described as essential for building confidence, particularly in regulated or high-risk environments such as finance and energy.
This aligns with calls in the literature for participatory and consumer-centred data governance [5]. Frameworks such as the CODCA model provide conceptual grounding for these insights by emphasizing auditability, traceability, and user oversight as foundations of trust. The parallels are clear: participants’ demands for certification, lineage visibility, and verifiable quality credentials mirror CODCA’s design principles, suggesting that participatory governance is not merely normative but a functional requirement for trustworthy data-sharing ecosystems.
However, this study also draws attention to a deeper structural tension. While platforms face growing pressure for interoperability and integration across systems, expectations regarding transparency and certification remain sector-specific. The absence of shared standards and escalation mechanisms forces consumers to navigate fragmented and sometimes opaque governance environments. Participants voiced frustration at the lack of formal avenues to raise concerns or challenge provider claims—underscoring that trust is not given but actively sustained through visible, responsive, and participatory governance practices.
Collectively, these findings highlight the interdependence of data-quality assurance and governance design. Embedding audit trails, certification protocols, and user co-governance mechanisms would enable platforms to operationalize transparency and accountability while reinforcing consumer confidence across domains.

5.4. Toward Consumer-Centric Data Ecosystems

Taken together, the findings suggest a need to reimagine data platforms through a consumer-centric perspective. Rather than seeing users as passive recipients of data, platforms should recognize them as active stakeholders with distinct needs and the capacity to shape governance, standards, and value.
Realizing the shift will require several key changes:
Governance structures to allow users to shape DQ standards,
Platform features to promote traceability, offer certification, and provide clear avenues for raising concerns,
Transparent and flexible pricing systems to reflect demonstrable quality and relevance, and
Sustained investments in documentation, metadata tools, and automated usability enhancements.
The approach reframes DQ as something negotiated and dependent on context. Quality is shaped through everyday workflows, verification processes, and governance structures. Future innovations should focus on building participatory systems, such as user feedback mechanisms, stakeholder councils, or context-sensitive certification programs, involving consumers more directly in the lifecycle of data products.
These findings also converge with platform-governance research [27,28] that conceptualizes platforms as multi-sided socio-technical arenas where governance, value creation, and data quality are co-constituted. Users in our study articulated governance expectations such as traceability, certification, and grievance channels—that mirror the accountability mechanisms proposed in the literature. Embedding such user co-governance structures would operationalize the transition from compliance-oriented to participatory platform governance, where data quality, trust, and value evolve through iterative human–data and institutional interaction.
Evidently, DQ is not simply about accuracy or completeness, and rather depends on it being timely, well documented, and suitable for real-world tasks. Trust grows through data transparency. In other words, knowing where the data comes from, the nature of processing, and recent updates matter as much as the data per se. Users are open to paying for clean, reliable, and ready-to-use data. These and other results suggest rethinking quality from a consumer’s perspective offers a more practical path to building credible and accessible data ecosystems.

6. Conclusions

This study has explored how DQ is perceived, prioritised, and operationalised by Indian data consumers across diverse sectors. The applied data-centric perspective highlights a strong user consensus around certain core dimensions of DQ, particularly ones directly impacting efficiency and decision-making. Despite the theoretical proliferation of possible quality indicators, only a subset is consistently valued and used. Our findings underscore the essential need for user-informed prioritization in designing data platforms and quality assessment tools.
The findings indicate the limits of technical accuracy alone. Consumers value transparency about where the data comes from, how data changes, and whether the provider has taken steps to maintain data integrity. Documentation, version tracking, and clear governance structures become essential. Yet, many users still face routine struggles with messy formats, schema mismatches, and inconsistent updates, thus holding back the promise of data platforms.
More broadly, consumers continue to bear the burden of validating, cleaning, and adapting datasets to the workflows or use cases. The gap erodes trust and efficiency and highlights the limits of current data supply chains where responsibility for quality assurance is unevenly distributed. Sustainable data-sharing ecosystems must give consumers a stronger role in shaping quality checks and governance processes. Transparent pricing, visible quality improvements, and flexible pricing models become just as important as the data per se.
By reframing DQ as something built jointly by providers, platforms, and users, this study offers a path towards more reliable and consumer-friendly data markets. The analysis underscores the need for institutional mechanisms to formalise user input, whether through participatory standards-setting, routine feedback channels, or sector-specific certification regimes. Future work should explore ways to bring consumer voices directly into platform design and build shared quality practices that work across different sectors and data types. Policy interventions should likewise recognise data governance as not just a technical or economic issue, but a question of equity, accountability, and inclusion.
This study, however, has certain limitations. First, it draws on a sample of senior professionals in India, which may not fully represent the perspectives of data consumers in other regions or organizational levels. Second, the focus on sectors such as Computer Science and Technology, Finance, ESG, and Urban Infrastructure may overlook distinct challenges in other domains.
Future research should expand the scope by conducting comparative cross-country studies and large-scale quantitative surveys to test the prevalence of identified themes. Further exploration of automated usability tools, participatory governance models, and pricing mechanisms linked to data quality would provide actionable pathways for the design of next-generation data platforms.

Author Contributions

Conceptualization, A.G. and M.M.; methodology, M.M. and A.G.; software, M.M.; writing—original draft, M.M.; writing—review and editing, A.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by OMIDYAR NETWORK through funds (Grant No. SP/OMNI-20-0001.06) and Centre for Data for Public good (CDPG), India. The APC was funded by OMIDYAR NETWORK.

Institutional Review Board Statement

The study does not involve human biological materials, or human biological data. Therefore, it does not fall under the purview of ethics approval requirements as outlined by the Institutional Human Ethics Committee (IHEC) at the Indian Institute of Science (IISc).

Informed Consent Statement

All participants provided informed consent at the start of each interview. They were informed about the purpose of the research, the voluntary nature of participation, confidentiality assurances, and how the data would be used. Participants were explicitly told that, at any point, they could request the audio recording to be turned off and their response kept undocumented. They retained the right to withdraw their participation at any stage without consequence. No personally identifiable information (such as names or company details) has been disclosed.

Data Availability Statement

The data presented in this study are available on request from the corresponding author to maintain the privacy of the participants.

Acknowledgments

The authors would like to gratefully acknowledge Centre for Data for Public good (CDPG), India, for their support through multiple discussions on practical ways of pricing data. The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest. The funding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, and in the decision to publish the results.

Abbreviations

The following abbreviations are used in the manuscript:
AIArtificial Intelligence
APIApplication Programming Interface
CDISCClinical Data Interchange Standards Consortium
CODCAConsumer-Oriented Data Control and Auditability
CEOChief Executive Officer
CTOChief Technology Officer
DCATData Catalog Vocabulary
DQData Quality
ESGEnvironmental, Social, and Governance
IoTInternet of Things
ITInformation Technology
MLMachine Learning
OECDOrganisation for Economic Co-operation and Development
SLAService Level Agreement
SPIRIT 2013Standard Protocol Items: Recommendations for Interventional Trials 2013

Appendix A

Appendix A is retained to demonstrate the conceptual breadth of the data quality literature; however, this study operationalises only the most user-salient dimensions, evidencing a move from theoretical comprehensiveness to practical sufficiency.
Table A1. Data quality dimensions.
Table A1. Data quality dimensions.
Fisher and Kingma, 2001 [55]Pipino et al., 2002 [13]Parssian et al., 2004 [56]Shankarnarayanan and Chi, 2006 [57]Batini et al., 2009 [58]Sidi et al., 2009 [59]Even et al., 2010 [60]Moges et al., 2013 [61]Woodall et al., 2013 [62]Behkamal et al., 2014 [12]Becker et al., 2015 [63]Batini et al., 2015 [64]Rao et al.,2015 [65]Guo and Liu, 2015 [66]Taleb et al., 2016 [67]Vetro et al., 2016 [68]Taleb and Serhani 2017 [69]Zhang et al., 2018 [33]Jesiļevska 2017 [70]Färber et al., 2018 [71]Ardagna et al.,2018 [72]Cichy and Rass, 2019 [73]Byabazaire et al., 2020 [74]Gὀmez-Omella et al., 2022 [18]Makhoul 2022 [75]
 Part A 
Accessibility                
Accuracy        
Alignment                        
Believability                      
Completeness    
Compactness                        
Concise Representation                    
Correctness                        
Consistency             
Clarity                        
Cohesion                        
Confidentiality                        
Conformity                        
Distinctness                        
Ease of Manipulation                       
Free of Error                        
Interpretability                   
Minimality                        
Objectify                       
Part B
Precision                     
Pertinence                        
Pedigree                        
Relevance                  
Reliability                        
Reputation                     
Redundancy                        
Simplicity                         
Security                     
Timeliness            
Traceability                        
Trust                       
Understandability                     
Usability                        
Uniqueness                      
Value-Added                       
Validity                        
Volume                       
Note: Here, ✔ indicates that the corresponding literature considers the quality dimensions.

References

  1. Wang, R.Y.; Strong, D.M. Beyond Accuracy: What Data Quality Means to Data Consumers. J. Manag. Inf. Syst. 1996, 12, 5–33. [Google Scholar] [CrossRef]
  2. Reinsel, D.; Gantz, J.; Rydning, J. The Digitization of the World From Edge to Core. Framingham: International Data Corporation 2018, 16, 1–28. [Google Scholar]
  3. Richter, H.; Slowinski, P.R. The Data Sharing Economy: On the Emergence of New Intermediaries. IIC Int. Rev. Intellect. Prop. Compet. Law 2019, 50, 4–29. [Google Scholar] [CrossRef]
  4. Zhang, M.; Beltran, F.; Liu, J. A Survey of Data Pricing for Data Marketplaces. IEEE Trans. Big Data 2023, 9, 1038–1056. [Google Scholar] [CrossRef]
  5. Koutroumpis, P.; Leiponen, A.; Thomas, L.D.W. Markets for Data. Ind. Corp. Change 2020, 29, 645–660. [Google Scholar] [CrossRef]
  6. OECD. Smart City Data Governance; OECD: Paris, France, 2023; ISBN 9789264910201. [Google Scholar]
  7. Johnson, J.; Hevia, A.; Yergin, R.; Karbassi, S.; Levine, A. Data Governance Frameworks for Smart Cities: Key Considerations for Data Management and Use; Covington: Palo Alto, CA, USA, 2022; Volume 2022. [Google Scholar]
  8. Gerli, P.; Mora, L.; Neves, F.; Rocha, D.; Nguyen, H.; Maio, R.; Jansen, M.; Serale, F.; Albin, C.; Biri, F.; et al. World Smart Cities Outlook 2024 Acknowledgements Contributors (Data Collection); ESCAP: Bangkok, Thailand, 2014. [Google Scholar]
  9. Urbinati, A.; Bogers, M.; Chiesa, V.; Frattini, F. Creating and Capturing Value from Big Data: A Multiple-Case Study Analysis of Provider Companies. Technovation 2019, 84–85, 21–36. [Google Scholar] [CrossRef]
  10. Ghosh, D.; Garg, S. India’s Data Imperative: The Pivot towards Quality; 2005. Available online: https://www.niti.gov.in/sites/default/files/2025-06/FTH-Quaterly-Insight-june.pdf (accessed on 27 August 2025).
  11. Batini, C.; Scannapieca, M. Data Quality: Concepts, Methodologies and Techniques; Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  12. Behkamal, B.; Kahani, M.; Bagheri, E.; Jeremic, Z. A Metrics-Driven Approach for Quality Assessment of Linked Open Data. J. Theor. Appl. Electron. Commer. Res. 2014, 9, 64–79. [Google Scholar] [CrossRef]
  13. Pipino, L.L.; Lee, Y.W.; Wang, R.Y. Data Quality Assessment. Commun. ACM 2002, 45, 211–218. [Google Scholar] [CrossRef]
  14. Bovee, M.; Srivastava, R.P.; Mak, B. A Conceptual Framework and Belief-Function Approach to Assessing Overall Information Quality. Int. J. Intell. Syst. 2003, 18, 51–74. [Google Scholar] [CrossRef]
  15. Ehrlinger, L.; Werth, B.; Wöß, W. Automated Continuous Data Quality Measurement with QuaIIe. Int. J. Adv. Softw. 2018, 11, 400–417. [Google Scholar]
  16. Chug, S.; Kaushal, P.; Kumaraguru, P.; Sethi, T. Statistical Learning to Operationalize a Domain Agnostic Data Quality Scoring. arXiv 2021, arXiv:2108.08905. [Google Scholar] [CrossRef]
  17. Liu, Q.; Feng, G.; Tayi, G.K.; Tian, J. Managing Data Quality of the Data Warehouse: A Chance-Constrained Programming Approach. Inf. Syst. Front. 2021, 23, 375–389. [Google Scholar] [CrossRef]
  18. Meritxell, G.O.; Sierra, B.; Ferreiro, S. On the Evaluation, Management and Improvement of Data Quality in Streaming Time Series. IEEE Access 2022, 10, 81458–81475. [Google Scholar] [CrossRef]
  19. Taleb, I.; Serhani, M.A.; Bouhaddioui, C.; Dssouli, R. Big Data Quality Framework: A Holistic Approach to Continuous Quality Management. J. Big Data 2021, 8, 76. [Google Scholar] [CrossRef]
  20. Alwan, A.A.; Ciupala, M.A.; Brimicombe, A.J.; Ghorashi, S.A.; Baravalle, A.; Falcarin, P. Data Quality Challenges in Large-Scale Cyber-Physical Systems: A Systematic Review. Inf. Syst. 2022, 105, 101951. [Google Scholar] [CrossRef]
  21. Sadiq, S.; Indulska, M. Open Data: Quality over Quantity. Int. J. Inf. Manag. 2017, 37, 150–154. [Google Scholar] [CrossRef]
  22. Cai, L.; Zhu, Y. The Challenges of Data Quality and Data Quality Assessment in the Big Data Era. Data Sci. J. 2015, 14, 2. [Google Scholar] [CrossRef]
  23. Abedjan, Z.; Golab, L.; Naumann, F. Data Profiling. In Proceedings of the 2017 ACM International Conference on Management of Data, Chicago, IL, USA, 14–19 May 2017; ACM: New York, NY, USA, 2017; pp. 1747–1751. [Google Scholar]
  24. Podobnikar, T. Bridging Perceived and Actual Data Quality: Automating the Framework for Governance Reliability. Geosciences 2025, 15, 117. [Google Scholar] [CrossRef]
  25. Tapsell, J.; Akram, R.N.; Markantonakis, K. Consumer Centric Data Control, Tracking and Transparency—A Position Paper. In Proceedings of the 2018 17th IEEE International Conference on Trust, Security and Privacy in Computing and Communications/12th IEEE International Conference on Big Data Science and Engineering (TrustCom/BigDataSE), New York, NY, USA, 1–3 August 2018. [Google Scholar]
  26. Acev, D.; Biyani, S.; Rieder, F.; Aldenhoff, T.T.; Blazevic, M.; Riehle, D.M.; Wimmer, M.A. Systematic Analysis of Data Governance Frameworks and Their Relevance to Data Trusts. In Management Review Quarterly; Springer: Berlin/Heidelberg, Germany, 2025. [Google Scholar] [CrossRef]
  27. Otto, B.; Jarke, M. Designing a Multi-Sided Data Platform: Findings from the International Data Spaces Case. Electron. Mark. 2019, 29, 561–580. [Google Scholar] [CrossRef]
  28. Veltri, G.A.; Lupiáñez-Villanueva, F.; Folkvord, F.; Theben, A.; Gaskell, G. The Impact of Online Platform Transparency of Information on Consumers’ Choices. Behav. Public Policy 2023, 7, 55–82. [Google Scholar] [CrossRef]
  29. Ducuing, C.; Reich, R.H. Data Governance: Digital Product Passports as a Case Study. Compet. Regul. Netw. Ind. 2023, 24, 3–23. [Google Scholar] [CrossRef]
  30. Agahari, W.; Ofe, H.; de Reuver, M. It Is Not (Only) about Privacy: How Multi-Party Computation Redefines Control, Trust, and Risk in Data Sharing. Electron. Mark. 2022, 32, 1577–1602. [Google Scholar] [CrossRef]
  31. Yu, H.; Zhang, M. Data Pricing Strategy Based on Data Quality. Comput. Ind. Eng. 2017, 112, 1–10. [Google Scholar] [CrossRef]
  32. Yang, J.; Zhao, C.; Xing, C. Big Data Market Optimization Pricing Model Based on Data Quality. Complexity 2019, 2019, 1–10. [Google Scholar] [CrossRef]
  33. Zhang, D.; Wang, H.; Ding, X.; Zhang, Y.; Li, J.; Gao, H. On the Fairness of Quality-Based Data Markets. arXiv 2018, arXiv:1808.01624. [Google Scholar] [CrossRef]
  34. Pei, J. A Survey on Data Pricing: From Economics to Data Science. IEEE Trans. Knowl. Data Eng. 2022, 34, 4586–4608. [Google Scholar] [CrossRef]
  35. Miao, X.; Peng, H.; Huang, X.; Chen, L.; Gao, Y.; Yin, J. Modern Data Pricing Models: Taxonomy and Comprehensive Survey. arXiv 2023, arXiv:2306.04945. [Google Scholar] [CrossRef]
  36. Majumdar, R.; Gurtoo, A.; Maileckal, M. Developing a Data Pricing Framework for Data Exchange. Future Bus. J. 2025, 11, 4. [Google Scholar] [CrossRef]
  37. Malieckal, M.; Gurtoo, A.; Majumdar, R. Data Pricing for Data Exchange: Technology and AI. In Proceedings of the 2024 7th Artificial Intelligence and Cloud Computing Conference, Tokyo, Japan, 14–16 December 2024; Association for Computing Machinery: New York, NY, USA, 2025; pp. 392–401. [Google Scholar]
  38. Chan, A.W.; Tetzlaff, J.M.; Gøtzsche, P.C.; Altman, D.G.; Mann, H.; Berlin, J.A.; Dickersin, K.; Hróbjartsson, A.; Schulz, K.F.; Parulekar, W.R.; et al. SPIRIT 2013 Explanation and Elaboration: Guidance for Protocols of Clinical Trials. BMJ 2013, 346, e7586. [Google Scholar] [CrossRef]
  39. Juddoo, S.; George, C.; Duquenoy, P.; Windridge, D. Data Governance in the Health Industry: Investigating Data Quality Dimensions within a Big Data Context. Appl. Syst. Innov. 2018, 1, 43. [Google Scholar] [CrossRef]
  40. Facile, R.; Muhlbradt, E.E.; Gong, M.; Li, Q.; Popat, V.; Pétavy, F.; Cornet, R.; Ruan, Y.; Koide, D.; Saito, T.I.; et al. Use of Clinical Data Interchange Standards Consortium (CDISC) Standards for Real-World Data: Expert Perspectives From a Qualitative Delphi Survey. JMIR Med. Inform. 2022, 10, e30363. [Google Scholar] [CrossRef] [PubMed]
  41. Aggarwal, R.; Farag, S.; Martin, G.; Ashrafian, H.; Darzi, A. Patient Perceptions on Data Sharing and Applying Artificial Intelligence to Health Care Data: Cross-Sectional Survey. J. Med. Internet Res. 2021, 23, e26162. [Google Scholar] [CrossRef]
  42. Miller, R.; Whelan, H.; Chrubasik, M.; Whittaker, D.; Duncan, P.; Gregório, J. A Framework for Current and New Data Quality Dimensions: An Overview. Data 2024, 9, 151. [Google Scholar] [CrossRef]
  43. Etheredge, L.M. A Rapid-Learning Health System. Health Aff. 2007, 26, w107–w118. [Google Scholar] [CrossRef] [PubMed]
  44. Kaye, J.; Whitley, E.A.; Lund, D.; Morrison, M.; Teare, H.; Melham, K. Dynamic Consent: A Patient Interface for Twenty-First Century Research Networks. Eur. J. Hum. Genet. 2015, 23, 141–146. [Google Scholar] [CrossRef]
  45. Hoare, C.A.R. Data Reliability. ACM Sigplan Not. 1975, 10, 528–533. [Google Scholar] [CrossRef]
  46. Chapple, J.N. Business Systems Techniques: For the Systems Professional; Longman: London, UK, 1976. [Google Scholar]
  47. Fox, C.; Levitin, A.; Redman, T. The Notion of Data and Its Quality Dimensions. Inf. Process Manag. 1994, 30, 9–19. [Google Scholar] [CrossRef]
  48. Ballou, D.P.; Pazer, H.L. Modeling Data and Process Quality in Multi-Input, Multi-Output Information Systems. Manag. Sci. 1985, 31, 150–162. [Google Scholar] [CrossRef]
  49. Amicis, F.; Barone, D.; Batini, C. An Analytical Framework to Analyze Dependencies Among Data Quality Dimensions. In Proceedings of the 2006 International Conference on Information Quality, ICIQ 2006, Cambridge, MA, USA, 10–12 November 2006; pp. 369–383. [Google Scholar]
  50. Chatfield, A.T.; Reddick, C.G. A Framework for Internet of Things-Enabled Smart Government: A Case of IoT Cybersecurity Policies and Use Cases in U.S. Federal Government. Gov. Inf. Q. 2019, 36, 346–357. [Google Scholar] [CrossRef]
  51. Guest, G.; Bunce, A.; Johnson, L. How Many Interviews Are Enough? An Experiment with Data Saturation and Variability. Field Methods 2006, 18, 59–82. [Google Scholar] [CrossRef]
  52. Hennink, M.M.; Kaiser, B.N.; Marconi, V.C. Code Saturation Versus Meaning Saturation: How Many Interviews Are Enough? Qual. Health Res. 2016, 27, 591–608. [Google Scholar] [CrossRef]
  53. Mortier, R.; Haddadi, H.; Henderson, T.; McAuley, D.; Crowcroft, J. Human-data interaction: The human face of the data-driven society. arXiv 2014, arXiv:1412.6159. [Google Scholar] [CrossRef]
  54. Sailaja, N.; Jones, R.; McAuley, D. Designing for human data interaction in data-driven media experiences. In Proceedings of the Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems, Yokohama, Japan, 8–13 May 2021; pp. 1–7. [Google Scholar]
  55. Fisher, C.W.; Kingma, B.R. Criticality of Data Quality as Exempli®ed in Two Disasters. Inf. Manag. 2001, 39, 109–116. [Google Scholar] [CrossRef]
  56. Parssian, A.; Sarkar, S.; Jacob, V.S. Assessing Data Quality for Information Products: Impact of Selection, Projection, and Cartesian Product. Manag. Sci. 2004, 50, 967–982. [Google Scholar] [CrossRef]
  57. Shankaranarayanan, G.; Cai, Y. Supporting Data Quality Management in Decision-Making. Decis. Support Syst. 2006, 42, 302–317. [Google Scholar] [CrossRef]
  58. Batini, C.; Cappiello, C.; Francalanci, C.; Maurino, A. Methodologies for Data Quality Assessment and Improvement. ACM Comput. Surv. 2009, 41, 1–52. [Google Scholar] [CrossRef]
  59. Sidi, F.; Shariat Panahy, P.H.; Affendey, L.S.; Jabar, M.A.; Ibrahim, H.; Mustapha, A. Data Quality: A Survey of Data Quality Dimensions. In Proceedings of the 2012 International Conference on Information Retrieval & Knowledge Management, Kuala Lumpur, Malaysia, 13–15 March 2012; IEEE: New York, NY, USA, 2012; pp. 300–304. [Google Scholar]
  60. Even, A.; Shankaranarayanan, G.; Berger, P.D. Evaluating a Model for Cost-Effective Data Quality Management in a Real-World CRM Setting. Decis. Support. Syst. 2010, 50, 152–163. [Google Scholar] [CrossRef]
  61. Moges, H.-T.; Dejaeger, K.; Lemahieu, W.; Baesens, B. A Multidimensional Analysis of Data Quality for Credit Risk Management: New Insights and Challenges. Inf. Manag. 2013, 50, 43–58. [Google Scholar] [CrossRef]
  62. Woodall, P.; Koronios, A. An Investigation of How Data Quality Is Affected by Dataset Size in the Context of Big Data Analytics. In Proceedings of the International Conference on Information Quality, Xi’an, China, 1–3 August 2014. [Google Scholar]
  63. Becker, D.; King, T.D.; McMullen, B. Big Data, Big Data Quality Problem. In Proceedings of the 2015 IEEE International Conference on Big Data (Big Data), Santa Clara, CA, USA, 29 October–1 November 2015; IEEE: New York, NY, USA, 2015. [Google Scholar]
  64. Batini, C.; Rula, A.; Scannapieco, M.; Viscusi, G. From Data Quality to Big Data Quality. J. Database Manag. 2015, 26, 60–82. [Google Scholar] [CrossRef]
  65. Rao, D.; Gudivada, V.N.; Raghavan, V.V. Data Quality Issues in Big Data. In Proceedings of the 2015 IEEE International Conference on Big Data (Big Data), Santa Clara, CA, USA, 29 October–1 November 2015; IEEE: New York, NY, USA, 2015. [Google Scholar]
  66. Guo, J.; Liu, F. Automatic Data Quality Control of Observations in Wireless Sensor Network. IEEE Geosci. Remote Sens. Lett. 2015, 12, 716–720. [Google Scholar] [CrossRef]
  67. Taleb, I.; Kassabi, H.T.E.; Serhani, M.A.; Dssouli, R.; Bouhaddioui, C. Big Data Quality: A Quality Dimensions Evaluation. In Proceedings of the 2016 Intl IEEE Conferences on Ubiquitous Intelligence & Computing, Advanced and Trusted Computing, Scalable Computing and Communications, Cloud and Big Data Computing, Internet of People, and Smart World Congress (UIC/ATC/ScalCom/CBDCom/IoP/SmartWorld), Toulouse, France, 18–21 July 2016. [Google Scholar]
  68. Vetrò, A.; Canova, L.; Torchiano, M.; Minotas, C.O.; Iemma, R.; Morando, F. Open Data Quality Measurement Framework: Definition and Application to Open Government Data. Gov. Inf. Q. 2016, 33, 325–337. [Google Scholar] [CrossRef]
  69. Taleb, I.; Serhani, M.A. Big Data Pre-Processing: Closing the Data Quality Enforcement Loop. In Proceedings of the 2017 IEEE 6th International Congress on Big Data, BigData Congress 2017, Honolulu, HI, USA, 25–30 June 2017; Institute of Electrical and Electronics Engineers Inc.: New York, NY, USA, 2017; pp. 498–501. [Google Scholar]
  70. Jesiļevska, S. Data Quality Dimensions to Ensure Optimal Data Quality. Rom. Econ. J. 2017, 20, 89–103. [Google Scholar]
  71. Färber, M.; Bartscherer, F.; Menne, C.; Rettinger, A. Linked Data Quality of DBpedia, Freebase, OpenCyc, Wikidata, and YAGO. Semant. Web 2018, 9, 77–129. [Google Scholar] [CrossRef]
  72. Ardagna, D.; Cappiello, C.; Samá, W.; Vitali, M. Context-Aware Data Quality Assessment for Big Data. Future Gener. Comput. Syst. 2018, 89, 548–562. [Google Scholar] [CrossRef]
  73. Cichy, C.; Rass, S. An Overview of Data Quality Frameworks. IEEE Access 2019, 7, 24634–24648. [Google Scholar] [CrossRef]
  74. Byabazaire, J.; O’Hare, G.; Delaney, D. Using trust as a measure to derive data quality in data shared IoT deployments. In Proceedings of the 2020 29th International Conference on Computer Communications and Networks (ICCCN), Honolulu, HI, USA, 3–6 August 2020; pp. 1–9. [Google Scholar]
  75. Makhoul, N. Review of Data Quality Indicators and Metrics, and Suggestions for Indicators and Metrics for Structural Health Monitoring. Adv. Bridge Eng. 2022, 3, 17. [Google Scholar] [CrossRef]
Figure 1. Coding structure employed.
Figure 1. Coding structure employed.
Information 16 00920 g001
Table 1. Summary of the literature studied.
Table 1. Summary of the literature studied.
Domain of AnalysisObjectives and BenefitsBarriers and ProblemsReferences
DQ in Platforms in India (including Smart City and Real-Time Systems)Improved access, usability, traceability; real-time, high-quality data for decision-making; context-specific governance and consumer empowermentFragmented datasets, poor documentation, limited consumer input; complex real-time governance; regulatory immaturity and disempowermentRichter & Slowinski 2018 [3]; Koutroumpis et al., 2020 [5]; Ghosh & Garg 2025 [10]
DQ Assessment Frameworks and MethodsMultidimensional models; scalable, automated tools; real-time assessment adaptable to fragmented ecosystemsLimited real-world validation; provider-centric; requires structured datasets; consumer-friendly tools are largely missingWang & Strong 1996 [1] Batini & Scannapieco 2006 [11]; Behkamal et al., 2014 [12]; Pipino et al., 2002 [13]; Bovee et al., 2003 [14]; Ehrlinger et al., 2018 [15]; Chug et al., 2021 [16]; Liu et al., 2021 [17]; Gómez-Omella et al., 2022 [18]; Taleb et al., 2021 [19]; Zhang et al., 2023 [4]; Alwan et al., 2022 [20]; Sadiq and Indulska, 2017 [21]; Cai and Zhu, 2015 [22]; Abedjan et al., 2016 [23]
Consumer-Centric Governance in Data PlatformsTraceability, fairness, transparency, participatory governance, consumer trustWeak operationalisation, limited consumer influence, regulatory immaturityPodobnikar 2025 [24]; Tapsell et al., 2018 [25]; Acev et al., 2025 [26]; Otto & Jarke 2019 [27]; Veltri et al., 2020 [28]; Ducuing & Reich 2023 [29]; Agahari et al., 2022 [30]
Monetisation Models and Consumer ParticipationQuality-driven, tiered pricing linked to consumer willingness to payPricing opacity, inconsistent service quality, limited research on consumer preferencesYu and Zhang, 2017 [31]; Yang et al., 2019 [32]; Zhang et al., 2018 [33]; Pei 2020 [34]; Miao et al., 2023 [35]; Zhang et al., 2023 [4]; Majumdar et al., 2024 [36]; Malieckal et al., 2024 [37]
Clinical and Business ProtocolsInteroperability, continuous improvement, user-centred governance and consentRegulatory complexity, sector specificity, limited applicability to open platformsChan et al., 2017 [38]; Juddoo et al., 2018 [39]; Facile et al., 2022 [40]; Aggarwal et al., 2021 [41]; Miller et al., 2024 [42]; Etheredge 2007 [43]; Kaye et al., 2015 [44]
Table 2. Participant profiles.
Table 2. Participant profiles.
Case IDRoleSectorOrganization Type
P1Partner and Tech EvangelistIT ConsultingLarge Firm
P2DirectorPatent AnalyticsMid-Size Enterprise
P3Product OwnerSoftwareMultinational Corporation
P4Vice PresidentIndependent ConsultantIndependent
P5Senior Data ScientistCross-Industry (Retail, Manufacturing)Independent Consultant
P6CEO Technology Private Firm 
P7Data ArchitectUrban Transport AnalyticsGovernment-Affiliated
P8Quantitative AnalystFinancial ServicesMultinational Corporation
P9CEOESG and Workplace StrategyConsulting Firm
P10Founder DirectorComputer Science and TechnologyStartup
P11Statistical Modelling SpecialistOil and Gas ConsultingConsulting Firm
P12Knowledge Management ConsultantInfrastructureLarge Firm
P13Senior ManagerBankingMultinational Corporation
P14Consulting PartnerFinancial AnalyticsLarge IT Firm
P15Program ManagerRetail and Consumer GoodsIT Services Company
P16Product Head for Traffic SolutionsSmart Cities/Urban MobilityPrivate Firm
P17CEOHealthcareStartup
P18FounderFinance and LawNon-Profit
P19FounderArtificial IntelligenceStartup
P20FounderMental HealthStartup
P21FounderHuman ResourcesStartup
P22Head of Delivery ExcellenceCloud/TechMultinational Corporation
P23CEO, Co-founderSustainabilitySocial Enterprise
P24Opportunities Evaluation ManagerOil and GasMultinational Corporation
P25Co-Founder and CTOBanking/FinTechStartup
Table 3. First-level coding.
Table 3. First-level coding.
Quality ThemeKey Concepts and KeywordsParticipantsNumber of Extracts
Fit-for-Purpose Quality: Needs Expressed by Data ConsumersConsistency in Formats, Data Accuracy, Relevance to Task, Volume, Freshness, Frequency of Updating, Credibility, Completeness, Expert Validation, Documentation, API Documentation, Metadata, Data Handbook, Lineage23 (92%)92
Usability and Format Frictions: Common Problems EncounteredMissing data, Time Spent on Data Cleaning, Faulty Instruments, Data Duplication, Multi-Format Inflows (data/file), Lack of Documentation19 (76%)52
Transparency, Traceability, and Governance: Consumer Expectations and Platform RequirementsCertification, Authentication, Provider-Side Quality Checks, Establishing Common Standards, Version Control, Source Traceability, Transformation Logs10 (40%)22
Monetisation and Willingness to Pay: Pricing ConcernsWillingness to Pay Linked to Better-Quality, High-Quality Data Essential for Monetization3 (12%)3
Percentages represent the proportion of participants who mentioned each theme and do not sum to 100%, as a single participant may have contributed to multiple themes.
Table 4. Second-level Coding.
Table 4. Second-level Coding.
CodeSubcodeKeywords and ConceptsParticipantsNumber of Extracts
Fit-for-Purpose QualityContextual RelevanceAccuracy, Consistency, Data Hygiene, Usefulness, Timeliness, Completeness, Version Control, Metadata, Standards, Lineage, Transparency2380
Temporal DimensionsFreshness, Relevance, Timeliness, Real-Time Data, Historical Data, Mission-Critical912
Credibility MechanismsAuthenticity, Standardization, Traceability, Veracity, Certification, Validation716
Usability and Format FrictionsCleaning BurdenData Cleaning, Missing Values, Formatting Issues, Time-Consuming, Accuracy, Duplication, Multi-Format Data, Resource Drain, Value Addition.1021
Integration IssuesData Formats, Synchronization, Interoperability, Labelling, Vendor Lock-In, Standardization, Scalability, Harmonization, IoT Limitations.919
Documentation GapsMetadata, Standardization, Missing Information, Data Culture, Unstructured Data, Sampling Issues, Publication Details, Supplementary Fields.912
Transparency, Traceability, and GovernanceVerification NeedsAuthentication, Certification, Triangulation, Data Validation, Trust, Credentials.56
Traceability RequirementsVersion Control, Data Lineage, Processing History, Transparency, Documentation, Relevance.23
Standardization GapsMultiple Standards, Interoperability, DQ, Sector-Specific, Stakeholder Collaboration, Metadata Standards, Market Scalability.613
Table 5. Theme engagement rating criteria.
Table 5. Theme engagement rating criteria.
IntensityCriteria
High≥3 participants from the domain, with detailed or multi-point discussion (several sub-themes addressed)
Medium1–2 participants with moderate-to-deep engagement and moderate diversity
Low1 participant with a brief mention, or 2+ participants but only surface-level comments (minimal diversity/detail)
NoneNo participants from the domain discussed the theme
Table 12. Participant statements on Platform Transparency, Traceability, and Governance.
Table 12. Participant statements on Platform Transparency, Traceability, and Governance.
StatementParticipant
“I need the certification of the data authentication of the data.”Cloud/Tech (P22)
“standardization is needed to scale and explore and move into the market quickly or develop the market quickly.”Urban Transport Analytics (P7)
“the verification of the data is the most important thing.”ESG and Workplace Strategy (P9)
“Checking data quality becomes easier based on fixed parameters of data quality, because those can only be defined data.”Finance/Law Non-Profit (P18)
“Is not easy because if you take any domain, not just one standard, even if you see sensor for sensors only, at least I have come across more than three standards.”Software Multinational Corporation (P3)
“data standardization and the normalization are another challenge that you know I’ve been facing for in case of a data integration kind of thing”Smart Cities/Urban Mobility (P16)
“now that becomes challenging, maintaining that entire, uh, like the history of the version control of the data and understanding what are all the changes that that has gone through over the course in time.”Finance and Law Non-Profit (P18)
“One thing which when we’re talking about version control of the data, one thing which people do often tend to ignore is the sort of processing that has gone from the raw to the finalized data.”Finance and Law Non-Profit (P18)
Table 13. Sector-wise platform requirements.
Table 13. Sector-wise platform requirements.
SectorParticipants (IDs)Core Expectations
IT/Tech and ConsultingP1, P22Provider-side certification (e.g., API governance checks)
Finance and LegalP4, P14Version control, transformation logs for auditability
Urban/Smart CitiesP7, P12Adoption of DCAT standards, consortium-led quality parameters
SustainabilityP9, P23Methodology transparency, source validation protocols
HealthcareP10Raw-to-finalized data processing history
Cross-IndustryP5Lineage tracking for regulatory compliance
Table 14. Sector-wise engagement rating for Theme 3.
Table 14. Sector-wise engagement rating for Theme 3.
SectorParticipants (IDs)Depth/DiversityRating
IT/Tech and ConsultingP1, P22Certification, API governance checksMedium
Finance and LegalP4, P14Version control, audit logs, complianceMedium
Urban/Smart CitiesP7, P12Standards (DCAT), quality parametersMedium
SustainabilityP9, P23Methodology transparency, source validationMedium
HealthcareP10Data processing historyLow
Cross-IndustryP5Lineage trackingLow
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Malieckal, M.; Gurtoo, A. Mining User Perspectives: Multi Case Study Analysis of Data Quality Characteristics. Information 2025, 16, 920. https://doi.org/10.3390/info16100920

AMA Style

Malieckal M, Gurtoo A. Mining User Perspectives: Multi Case Study Analysis of Data Quality Characteristics. Information. 2025; 16(10):920. https://doi.org/10.3390/info16100920

Chicago/Turabian Style

Malieckal, Minnu, and Anjula Gurtoo. 2025. "Mining User Perspectives: Multi Case Study Analysis of Data Quality Characteristics" Information 16, no. 10: 920. https://doi.org/10.3390/info16100920

APA Style

Malieckal, M., & Gurtoo, A. (2025). Mining User Perspectives: Multi Case Study Analysis of Data Quality Characteristics. Information, 16(10), 920. https://doi.org/10.3390/info16100920

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop