Next Article in Journal
Real-Time LCA/LCC Integration: A Framework of Agile Sustainability and Cost Management
Previous Article in Journal
Frost Heave Characteristics of Lined Canals with Sand-Gravel Replacement in Seasonally Frozen Regions
Previous Article in Special Issue
Embedding Corporate Social Responsibility in Retail Strategy: Strategic, Sustainable, and Localized Approaches to Building Brand Equity and Loyalty
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Audit Quality Indicators Across Jurisdictions: Regulatory Diversity, ESG Integration, and Implications for Public Trust

by
Alexandros Garefalakis
1,2,*,
Ioannis Sitzimis
1,
Erasmia Angelaki
1,
Panagiotis Giannopoulos
3 and
Panagiotis Kyriakogkonas
4
1
Department of Business Administration and Tourism, Hellenic Mediterranean University, 714 10 Iraklio, Greece
2
Department of Economics and Business, Neapolis University Pafos, Pafos 8042, Cyprus
3
School of Social Sciences, Hellenic Open University, 263 35 Patra, Greece
4
Department of Accounting and Finance, University of Western Macedonia, 501 00 Kozani, Greece
*
Author to whom correspondence should be addressed.
Sustainability 2025, 17(21), 9421; https://doi.org/10.3390/su17219421
Submission received: 25 August 2025 / Revised: 17 October 2025 / Accepted: 20 October 2025 / Published: 23 October 2025

Abstract

Audit Quality Indicators (AQIs) have become vital tools for assessing and improving audit performance. This study examines how AQIs are defined, implemented, and interpreted across jurisdictions, with a particular focus on their integration into ESG (Environmental, Social, and Governance) assurance. Through a cross-jurisdictional comparison covering the EU, UK, US, and Canada, we analyze regulatory diversity and explore how different oversight models influence AQI adoption. Our findings reveal that AQIs play an increasingly important role not only in enhancing audit transparency but also in reinforcing the credibility of sustainability reporting under evolving frameworks like the EU’s Corporate Sustainability Reporting Directive (CSRD). However, implementation remains fragmented, and standardization challenges persist. The study offers empirical insights from a large-scale survey of audit professionals, highlighting how perceptions of audit quality vary by country, experience, and organizational context. We conclude that the harmonization of AQI frameworks, especially with ESG-focused metrics, is essential to foster public trust and ensure the accountability of sustainability assurance.

1. Introduction

In recent years, a growing number of academic studies have sought to conceptualize and evaluate audit quality using AQIs. These studies have examined AQIs as mechanisms for enhancing transparency, improving auditor performance, and fostering stakeholder confidence. However, much of the existing literature has focused on financial audits, with limited attention paid to how AQIs operate in the context of ESG assurance. Research has also noted inconsistencies in how AQIs are defined, interpreted, and implemented across jurisdictions and professional contexts. These findings highlight the need for a more integrated understanding of AQIs that accounts for both their theoretical foundations and practical implications in sustainability reporting. By engaging with these strands of literature, this study aims to contribute to a more nuanced discussion on the role of AQIs in supporting audit quality in a rapidly evolving assurance landscape.
Audit quality has become a cornerstone of financial market integrity, with far-reaching implications for investor confidence, regulatory compliance, and sustainable corporate governance. High-profile corporate failures—such as Enron, Carillion, and Wirecard—have underscored the urgent need for transparent and reliable audit performance measures. In response, AQIs have emerged as standardized metrics to evaluate the inputs, processes, and outputs of audit engagements. Initially developed to enhance audit oversight and benchmarking, AQIs now play a pivotal role in aligning audit practices with evolving stakeholder expectations—particularly in light of increased demands for assurance in ESG reporting.
The regulatory landscape has significantly influenced this evolution. In the European Union, reforms such as Regulation (EU) No 537/2014 and Directive 2014/56/EU have mandated enhanced audit disclosures and oversight mechanisms, reinforcing the role of AQIs in ensuring transparency and quality. In Greece, the Hellenic Accounting and Auditing Standards Oversight Board (ELTE) has been instrumental in implementing these directives and fostering an accountability-driven audit environment. Parallel initiatives by regulatory bodies such as the PCAOB (U.S.), CPAB (Canada), and FRC (UK) have introduced diverse AQI frameworks, reflecting both convergence efforts and regional variation in audit oversight [1].
At the same time, the audit scope is expanding to cover sustainability reporting alongside financial statements. The adoption of the CSRD by the EU has made limited assurance over ESG data mandatory for a broad class of entities. This development signals a shift in the role of auditors—from financial gatekeepers to verifiers of corporate accountability. AQIs, when adapted to ESG contexts, offer a mechanism for signaling assurance quality and building stakeholder trust. However, limited research addresses how AQIs are used across jurisdictions and integrated into ESG assurance.
To strengthen the theoretical positioning of the study, we clarify the distinction between regulatory developments and the academic literature on audit quality. While regulatory frameworks across jurisdictions provide the institutional context for the emergence and implementation of AQIs, academic research offers the conceptual underpinnings that shape how audit quality is defined, interpreted, and assessed. The literature emphasizes that audit quality is a multidimensional construct, encompassing technical competence, ethical responsibility, organizational culture, and institutional alignment. These conceptual insights inform the development and application of AQIs as both measurement tools and instruments of governance. By integrating these two perspectives—regulatory and theoretical—this study situates AQIs not only as regulatory responses but also as constructs grounded in scholarly discourse on audit performance, accountability, and stakeholder trust.
Despite the growing body of research on audit quality and its determinants, limited attention has been paid to the conceptual development and practical implementation of AQIs, particularly in the context of non-financial reporting and ESG assurance. Existing studies have primarily focused on financial audit contexts, often overlooking the unique challenges and expectations associated with sustainability-related disclosures. Moreover, the literature lacks a systematic comparison of regulatory and professional approaches to AQIs across jurisdictions [2]. These gaps underscore the need for a deeper understanding of how AQIs are conceptualized, developed, and interpreted in evolving audit landscapes. The following research questions are designed to address these gaps and contribute to ongoing debates by linking AQIs to ESG assurance and cross-national regulatory initiatives.
This study addresses that gap by synthesizing the conceptual foundations, regulatory contexts, and evolving applications of AQIs with a specific emphasis on ESG assurance. The following research questions guide the analysis:
RQ1: How are AQIs conceptualized and categorized across international regulatory frameworks?
RQ2: What role do AQIs play in supporting ESG assurance, particularly under the emerging CSRD and related stakeholder expectations?
RQ3: To what extent do existing AQI frameworks align across jurisdictions, and what challenges hinder their standardization and cross-border comparability?
To address these questions, the manuscript is structured as follows: Section 2 explores the theoretical foundations of audit quality, while Section 3 presents the evolution and classification of AQIs. Section 4 provides a comparative analysis of AQI frameworks across jurisdictions, and Section 5 examines their integration into ESG assurance. The final sections discuss findings, implications, and future research directions.

2. Literature Review

2.1. Theoretical and Conceptual Foundations of AQIs

The evolution of audit quality frameworks cannot be fully understood without reference to the legal and regulatory environments in which they operate. In the European Union, a series of post-crisis reforms—particularly Directive 2014/56/EU and Regulation No 537/2014—marked a significant shift toward stricter oversight of statutory audits. These instruments introduced new requirements for auditor independence, mandatory rotation, and expanded audit reports, reflecting a renewed emphasis on enhancing audit quality and transparency [1]. In Greece, the implementation of these reforms was overseen by the ELTE, which plays a central role in monitoring audit compliance and quality assurance. ELTE’s regulatory actions, aligned with EU directives, have laid the groundwork for integrating AQIs and sustainability assurance standards. This regulatory context provides a crucial foundation for understanding the relevance of AQIs and their potential to strengthen trust in audit practices both in financial and ESG reporting [3].
Audit quality has been widely discussed in the academic literature and among practitioners, with definitions evolving to encompass both technical competence and ethical responsibility. Ref. [4] was among the first to define audit quality as the joint probability that an auditor will detect a material misstatement in the financial statements and report it appropriately. This foundational definition emphasizes two essential aspects: the auditor’s ability to uncover inaccuracies and their independence to disclose them.
Subsequent scholars have expanded upon this definition. Ref. [5] suggested that audit quality should be viewed as a continuum rather than a binary concept, shaped by a variety of factors including the size of the audit firm, industry specialization, and the strength of the legal and institutional environment. Francis emphasized the contextual nature of audit quality and highlighted the importance of both individual auditor capabilities and the organizational setting.
The International Auditing and Assurance Standards Board [6] contributed a more holistic framework by identifying a wide range of factors that influence audit quality. According to the IAASB, audit quality is not only the outcome of an auditor’s individual behavior but also a function of the systems and culture within audit firms, as well as external elements like regulatory structures and stakeholder expectations. The IAASB’s framework underscores the importance of collaboration among stakeholders—auditors, clients, regulators, and investors—in enhancing audit quality.
These definitions collectively demonstrate that audit quality is inherently multidimensional. It encompasses technical proficiency, ethical judgment, organizational culture, and systemic interactions. They also form the basis for efforts to operationalize the concept of audit quality through measurable frameworks such as AQIs.
The Center for Audit Quality [7] provided one of the most comprehensive and practical frameworks for understanding the drivers of audit quality. This model identifies six primary elements that contribute to audit excellence, offering a structured approach for both internal assessment and external communication.
The first element focuses on the effectiveness of the audit process. This includes the ability of the auditor to design risk-based audit plans, apply appropriate procedures, and respond to complex accounting scenarios. An effective audit process is marked by its methodological soundness and adaptability to engagement-specific risks.
Audit firm culture is another core dimension. The CAQ emphasized that an ethical and high-integrity culture supports audit quality by fostering professional skepticism and a commitment to public interest. Attributes such as accountability, open communication, and a clear tone at the top are essential in cultivating such a culture. Equally important are the skills and personal qualities of audit partners and staff. The technical competence, professional judgment, and communication abilities of auditors directly influence the quality of audit work. Ongoing training and development programs ensure that staff remain proficient in current standards and practices [8].
The audit process itself, including the use of standardized methodologies and quality control mechanisms, contributes to the consistency and reliability of audit outcomes. Process quality is maintained through internal reviews, documentation standards, and peer consultations that help ensure procedural rigor and compliance with regulatory requirements. The reliability and usefulness of audit reporting also plays a crucial role. Audit reports that clearly articulate the auditor’s findings and highlight key audit matters help stakeholders better understand the financial health and governance of audited entities. This enhances transparency and supports informed decision-making [9].
Finally, there are factors beyond the auditor’s direct control that still influence audit quality. These include the client’s financial reporting practices, the strength of internal controls, corporate governance quality, and the broader regulatory environment. Although external, these variables can significantly affect the auditor’s ability to perform a high-quality audit.
Taken together, these six dimensions form a comprehensive view of audit quality. They not only guide audit firms in strengthening their practices but also provide a foundation for the development of AQIs, which aim to translate these qualitative aspects into quantifiable metrics for assessment and benchmarking [10,11]. As stakeholders increasingly demand objective evidence of audit performance, the role of AQIs has become central in efforts to monitor, compare, and enhance audit quality across firms and jurisdictions. The following section explores the evolution and classification of these indicators, highlighting their growing relevance in both academic and regulatory contexts.

2.2. Cross-Jurisdictional Approaches to AQIs

The emergence of AQIs reflects the growing need for transparency, accountability, and measurable benchmarks in auditing. Over the last two decades, several high-profile corporate collapses—such as Enron, Parmalat, Carillion, and Wirecard—have underscored persistent weaknesses in audit quality oversight and the limitations of traditional assurance practices. These failures led to significant losses in investor confidence and raised doubts about the effectiveness of auditors in fulfilling their public interest role. As a response, audit regulators and professional bodies began seeking tools that could offer objective insights into the performance of auditors and audit firms, giving rise to the concept of AQIs [12].
The development of AQIs has since evolved through both regulatory and market-driven initiatives. In the United States, the PCAOB was one of the first bodies to formally explore AQIs as part of its effort to enhance audit quality monitoring. In 2013, it released a concept paper proposing a set of indicators to assess audit performance both at the engagement level and across audit firms [13]. Other countries followed suit, including Canada through the CPAB, and the United Kingdom via the FRC, each contributing to a growing international dialogue about the role and implementation of AQIs.
Simultaneously, the IAASB recognized the need for a global framework that would support consistency in the application and interpretation of AQIs. In 2014, the IAASB outlined its vision for audit quality and emphasized the importance of engaging all key stakeholders—including audit firms, regulators, standard setters, and users of financial statements—in the continuous improvement of audit practices [14]. Their model positioned AQIs as a bridge between internal quality control processes and the external expectations of the audit profession.
Though adoption varies across jurisdictions, the increasing interest in AQIs reflects a shared objective: to make audit quality more transparent, measurable, and comparable. Pilot programs, public consultations, and voluntary disclosures by major audit firms have all contributed to shaping AQI frameworks that serve both evaluative and developmental purposes. The evolution of AQIs from a conceptual ideal to a practical tool represents a significant step toward fostering a culture of accountability and continuous improvement in the audit profession.
While jurisdictional differences influence the extent and manner of AQI adoption, common themes have emerged regarding how audit quality is defined, measured, and communicated. As the discourse around AQIs matures, attention has shifted from mere adoption to the standardization and interpretation of indicators. To facilitate comparability across firms and regions, regulators and professional bodies have pro-posed structured classification systems that group AQIs according to their functional role within the audit process. These typologies aim to transform abstract notions of audit quality into actionable metrics and thereby enhance both regulatory oversight and internal quality management.
In order to support the practical implementation and comparability of AQIs, several regulatory bodies and standard-setters have developed classification frameworks that group indicators based on the phase of the audit process they represent. One of the most widely recognized and applied models is the tripartite typology introduced by the International Auditing and Assurance Standards Board, which divides AQIs into three principal categories: input, process, and output indicators. This categorization has been echoed in frameworks developed by the PCAOB, the CPAB, and other international oversight entities. It allows for a systematic approach to evaluating the multiple dimensions of audit quality, from planning and resourcing to execution and outcomes [15].
Input indicators represent the resources, structures, and personnel that an audit firm mobilizes in advance of and during the audit engagement. These indicators are intended to reflect the capacity of the firm to deliver high-quality audit services. Key metrics include partner workload, the ratio of audit staff to partners, the years of professional experience of team members, and the number of training hours provided to audit professionals [16]. The underlying assumption is that firms investing more in experienced personnel, balanced team composition, and continuous training are more likely to conduct rigorous and effective audits. However, while these indicators signal a firm’s investment in quality infrastructure, they do not directly measure how these inputs translate into audit outcomes.
Process indicators, by contrast, focus on the conduct of the audit itself—capturing how well the audit is planned, executed, and monitored. These include the percentage of audit hours spent by engagement partners, the frequency of internal quality control reviews, the use of advanced auditing technologies, the extent of consultation on complex issues, and the timeliness of audit documentation. Process indicators serve as proxies for methodological soundness, adherence to standards, and the professional rigor applied throughout the engagement. For example, high levels of partner involvement may reflect a firm’s commitment to oversight and quality control, while frequent quality reviews can indicate a robust internal monitoring system. These indicators are particularly useful for identifying variations in audit execution across different engagements or offices within the same firm [17].
Output indicators aim to capture the results or consequences of the audit process. While less directly controllable by auditors, these indicators provide external signals about the credibility and effectiveness of audit engagements. Examples include the frequency of financial statement restatements following the audit, the number of deficiencies identified during regulatory inspections, the issuance of going concern opinions, and the rate of client turnover. Some output indicators, such as audit reporting timeliness and the quality of communication with audit committees, are increasingly being emphasized in AQI frameworks. Nonetheless, these indicators can be influenced by a range of factors beyond auditor performance—such as the complexity of the client’s operations or changes in financial reporting standards—making them less precise as direct measures of audit quality [18].
While each category of indicators offers valuable insights on its own, their true strength lies in their combined application. Evaluating input, process, and output indicators in tandem enables a more comprehensive and balanced assessment of audit quality. For instance, a firm might report high levels of training hours (input), but without sufficient partner involvement during execution (process) or with frequent inspection findings (output), the audit quality could still be compromised. Thus, AQI typologies provide a framework not only for performance measurement but also for internal quality improvement, regulatory benchmarking, and stakeholder communication [19].
However, the practical use of these categories also presents challenges. Metrics are not always standardized across firms or jurisdictions, and not all indicators are equally relevant in every audit context. Additionally, the interaction among these categories is complex and non-linear; a deficiency in one area may or may not be compensated by strengths in another. Therefore, while the input-process-output framework is conceptually robust, its real-world application requires careful calibration, context sensitivity, and a willingness to integrate both quantitative and qualitative perspectives in audit evaluation [20,21].
In addition to their classification by input, process, and output, AQIs must be understood in terms of their practical application and intended users. AQIs are not designed solely for regulators or academic analysis; they also serve as critical tools for audit committees, investors, and audit firms themselves. For audit committees, AQIs provide a structured basis for evaluating auditor performance, engagement risk, and audit planning [22]. For regulators, AQIs support inspection priorities and systemic risk monitoring. For audit firms, they act as internal management tools, helping identify areas for training, resource allocation, or quality control improvements. This multi-stakeholder utility underscores the versatility of AQIs, while also complicating efforts to standardize definitions and disclosures. For example, indicators such as staff turnover, internal inspection findings, and audit hours on key risk areas have different implications depending on the user and the context in which they are reported. As AQIs continue to evolve, understanding and aligning their use across different stakeholder groups will be critical to maximizing their effectiveness and avoiding misinterpretation [23].
Despite these advantages, the broad applicability of AQIs across stakeholders and contexts introduces new complexities. As frameworks become more widely adopted and integrated into regulatory and organizational processes, questions have emerged about their interpretability, methodological soundness, and real-world relevance. These concerns are especially pronounced in settings where audit environments are dynamic, resource-constrained, or governed by divergent institutional logics. The next section explores the main criticisms and limitations of AQIs, emphasizing the importance of critical evaluation as these indicators continue to shape audit practices globally.
Despite the promise and growing adoption of AQIs, several criticisms have emerged that challenge their reliability, consistency, and practical utility. These concerns highlight the challenge of translating a complex, judgment-based process like auditing into standardized metrics. Scholars and regulators alike have cautioned that while AQIs offer useful benchmarks, they are not immune to methodological and conceptual limitations that may compromise their effectiveness [24].
One of the most significant issues relates to the lack of standardization in AQI design and application. Different jurisdictions and audit regulators have developed their own sets of indicators, often tailored to local regulatory priorities and market contexts. For instance, the PCAOB, CPAB, FRC, and IAASB have each proposed AQI frameworks that vary in scope, terminology, and implementation strategy. This fragmentation creates inconsistencies in how audit quality is measured and reported across firms and countries, limiting the comparability of AQI data. Without global standards, stakeholders such as investors and audit committees may find it difficult to interpret AQIs consistently or compare audit quality across firms [25].
Another common concern is that the use of AQIs may encourage a compliance-focused or tick box mentality among audit firms. When firms prioritize meeting externally defined indicators over improving substantive audit quality, the spirit of professional skepticism and critical judgment may be diluted. Ref. [26], warned that AQIs could inadvertently lead to “checklist auditing,” where the focus shifts toward meeting numerical thresholds—such as achieving a minimum number of training hours or internal reviews—rather than addressing engagement-specific risks or exercising professional discretion. This checkbox approach can mislead stakeholders by masking deeper issues in audit execution.
A related limitation involves the over-reliance on quantitative indicators at the expense of qualitative dimensions that are more difficult to measure but equally essential to audit quality. For example, attributes such as ethical reasoning, communication clarity, and the tone set by audit leadership are central to audit performance but are often excluded from AQI frameworks due to their subjectivity. As Velte [27] noted, numerical indicators are convenient proxies but fall short of capturing the ethical and behavioral complexity of audit practice. This imbalance risks narrowing the scope of quality assessment to easily reportable metrics while ignoring contextual nuances.
In addition, the interpretability and usability of AQI data can pose significant challenges. Different stakeholders—such as regulators, investors, audit committees, and clients—may interpret the same indicator differently based on their perspectives and priorities. For instance, high audit hours could be viewed as a sign of thoroughness by some, or inefficiency by others. Zahid [28] highlighted that the absence of interpretive guidance or benchmarking standards may lead to stakeholder confusion or misjudgment, undermining the intended transparency benefits of AQIs.
Moreover, resource and cost constraints can inhibit the effective implementation of AQI systems, particularly for smaller audit firms or those operating in emerging markets. Persakis [29] noted that maintaining detailed AQI reporting infrastructure—such as tracking training hours, audit partner workload, or engagement-specific metrics—requires significant investment in technology, staff, and data systems. Smaller firms may struggle to meet these requirements without compromising operational efficiency. As a result, AQI frameworks may disproportionately burden less-resourced firms and create unintended barriers to entry or participation in high-quality audit markets.
Finally, there are emerging concerns about technological dependence and cybersecurity. As audit firms increasingly use AI-powered tools and automated systems to collect and report AQIs, vulnerabilities may arise related to data integrity, system reliability, and cyber threats. While these technologies offer efficiency and scalability, their use also introduces risks that could affect the accuracy and credibility of AQI data [30,31]. Overdependence on automated tools without adequate human oversight may also reduce the auditor’s direct engagement with the data, weakening the depth of professional judgment.

2.3. AQIs and ESG Assurance: Research Gaps and Integration

The implementation of AQIs across jurisdictions reflects a complex and evolving landscape, shaped by different legal traditions, regulatory priorities, and levels of market maturity. While the underlying goal—enhancing audit quality through measurable indicators—is broadly shared, the strategies adopted by oversight bodies and audit firms differ significantly in terms of scope, disclosure practices, and stakeholder engagement. A closer examination of the approaches taken by the United States, Canada, the United Kingdom, and the European Union reveals both converging trends and persistent divergences that have implications for global audit reform.
In the United States, the PCAOB was among the first regulators to formally explore AQIs. Its 2013 concept release proposed a comprehensive set of 28 indicators designed to assess audit quality at the engagement and firm levels [13]. These included metrics on partner workload, audit hours, staff experience, internal inspection results, and client retention. The PCAOB’s approach was ambitious in scope, aiming to standardize audit quality assessment and enhance transparency. However, after extensive public consultation, the Board did not mandate AQI disclosures, choosing instead to encourage voluntary experimentation by audit firms. As a result, the adoption of AQIs in the U.S. has been uneven and largely confined to internal dashboards or selective disclosures to audit committees. The lack of regulatory compulsion has limited the comparability and external usefulness of AQIs, and many firms continue to treat them as confidential management tools rather than public accountability mechanisms [32].
In contrast, Canada has pursued a more coordinated and consultative approach under the leadership of the CPAB. Beginning in 2016, CPAB worked closely with the country’s largest audit firms to develop a set of pilot AQIs that could be used for internal quality improvement and audit committee communication. The Canadian model prioritizes relevance, consistency, and clarity over comprehensiveness, focusing on a narrower set of indicators that are easier to implement and interpret [33]. CPAB has also emphasized the educational function of AQIs, encouraging their use in enhancing dialogue between auditors and governance bodies. While disclosure remains voluntary, CPAB’s guidance and oversight have contributed to a more standardized and transparent use of AQIs compared to other jurisdictions. This collaborative regulatory philosophy reflects Canada’s preference for consensus-building over top-down mandates and has yielded relatively strong buy-in from major firms.
The United Kingdom, governed by the FRC, has adopted a hybrid approach that combines thematic inspections with firm-level transparency initiatives. While the FRC has not issued a standalone AQI framework, it has incorporated quality indicators into its AQR program and encouraged firms to include selected AQIs in their annual transparency reports. Metrics such as audit hours, partner involvement, training, and inspection results are commonly disclosed by large firms, though with considerable variation in definitions and presentation. The FRC has also published periodic reports evaluating the use and potential of AQIs in enhancing audit committee oversight. However, the UK’s voluntary and principles-based approach has faced criticism for a lack of uniformity and enforceability, with some stakeholders calling for a more prescriptive regulatory model, particularly in light of high-profile audit failures such as Carillion and Thomas Cook [34].
In the European Union, AQI implementation remains fragmented and largely dependent on national initiatives. The EU’s Audit Regulation (Regulation (EU) No 537/2014) [1] introduced enhanced auditor reporting requirements, such as the extended audit report and transparency disclosures, but it stopped short of mandating specific AQIs. Consequently, Member States have adopted divergent strategies. For instance, the Netherlands has launched structured AQI programs involving both regulators and firms, while Germany has focused more on reinforcing internal quality controls. In many other countries, AQIs are either in the exploratory phase or absent from the regulatory agenda. This heterogeneity reflects the decentralized nature of EU audit oversight and poses significant barriers to cross-border comparability. Without EU-wide standards or a coordinating authority to promote harmonization, AQI adoption across Europe continues to lack coherence and consistency.
In response to these disparities, global standard-setting organizations have sought to promote convergence. The IAASB has advocated for a common understanding of audit quality and the development of globally relevant AQIs. Its 2014 Framework for Audit Quality emphasized the need to engage all stakeholders in building consensus on what constitutes a high-quality audit and how it can be measured. Similarly, the IFIAR has identified AQIs as a strategic priority and has called for greater alignment in definitions, methodologies, and disclosure practices. These efforts aim to address the current fragmentation by promoting standardized measurement tools that can be adapted across regulatory contexts while preserving local relevance.
However, significant obstacles remain. Legal and institutional differences, as well as varying levels of regulatory capacity and market sophistication, continue to shape how AQIs are understood and applied. In some jurisdictions, resistance to mandatory reporting stems from concerns about proprietary data, litigation risk, and competitive sensitivity. In others, the challenge lies in limited resources or technological infrastructure to support AQI tracking and reporting. Moreover, the lack of a clear cost–benefit analysis for AQIs in some contexts has made firms cautious about investing heavily in systems that may not yield immediate returns.
Despite these challenges, the global trajectory suggests a gradual movement toward greater standardization. Major audit networks are beginning to adopt common AQI frameworks across their international affiliates, and regulators are increasingly referencing AQIs in inspection criteria and policy guidance. The rise of sustainability and ESG assurance has also added urgency to the AQI agenda, as stakeholders seek clearer signals of audit integrity and performance in non-financial reporting domains.
As jurisdictions continue to refine their AQI frameworks, the broader role of these indicators within the audit ecosystem is evolving. Beyond their traditional ap-plication in financial audit settings, AQIs are increasingly being recognized as relevant tools for assessing assurance quality in new domains. One of the most prominent of these domains is ESG reporting, where the demand for reliable, standardized assurance mechanisms is rapidly growing. This transition has significant implications for how AQIs are conceptualized, applied, and evaluated in the context of sustainability-related disclosures.
The intersection of audit quality and sustainability assurance has become increasingly significant as global stakeholders demand higher standards of corporate accountability and transparency [35]. While Audit Quality Indicators were initially conceived as tools to assess and improve the quality of financial audits, their relevance has grown with the rapid institutionalization of ESG reporting. In this evolving context, AQIs are now being reconsidered and adapted as mechanisms for benchmarking the quality of ESG assurance engagements—a development that reflects both the broadening scope of audit responsibilities and the transformation of stakeholder expectations [36].
The shift toward integrating ESG within audit frameworks stems from mounting regulatory and market pressure. Institutional investors, civil society organizations, and consumers are demanding more credible, verifiable disclosures about a company’s sustainability performance. Reports covering climate impact, gender equality, supply chain ethics, and governance practices are no longer considered peripheral but integral to evaluating corporate risk and long-term value creation. As a result, these disclosures must increasingly meet standards of assurance similar to those applied to financial statements. However, unlike traditional audits, ESG assurance faces unique challenges: data is often non-standardized, qualitative, forward-looking, and embedded in complex organizational structures. This complexity amplifies the importance of clear, measurable indicators of assurance quality [37,38].
AQIs offer a promising solution to this problem. Indicators such as partner involvement in ESG engagements, team training on sustainability standards, use of ESG-specific assurance frameworks, and review procedures for non-financial data are directly translatable from financial audit settings into the ESG domain. These AQIs serve as proxies for the preparedness and rigor with which assurance providers engage with ESG data. For example, high levels of partner time allocation and ESG-specific training hours suggest stronger oversight and domain knowledge, while robust internal review mechanisms indicate methodological consistency and adherence to assurance standards such as ISAE 3000.
Moreover, process-oriented AQIs can help demonstrate how assurance providers evaluate ESG disclosures, especially when dealing with fragmented or narrative-driven information. Indicators related to data verification, use of specialists, and consultation practices can offer insights into how audit teams manage ESG complexity and uncertainty. Likewise, output-focused AQIs—such as the nature of assurance opinions issued (limited vs. reasonable assurance), the frequency of ESG restatements, and post-assurance stakeholder feedback—provide an additional layer of accountability, helping to evaluate the perceived and actual impact of ESG audits [39].
The relevance of AQIs in ESG assurance is further heightened by the regulatory momentum behind sustainability reporting. The European Union’s CSRD, adopted in 2022, marks a significant turning point. The CSRD requires a wide range of companies to report detailed ESG information aligned with the European Sustainability Reporting Standards (ESRS), and mandates limited assurance of these disclosures by statutory auditors or accredited assurance providers. This regulatory requirement significantly expands the role of auditors and reinforces the need for objective indicators that can assess the quality of ESG assurance engagements. AQIs—adapted and contextualized for sustainability assurance—can become key tools in meeting these new demands, helping to demonstrate both compliance and quality in an area of growing regulatory scrutiny.
Beyond the EU, similar initiatives are emerging globally. The ISSB and the GRI have accelerated efforts to harmonize sustainability disclosure frameworks, while the International IAASB is developing dedicated standards for sustainability assurance. In this context, the development of AQIs tailored to ESG becomes not just beneficial but necessary. The credibility of ESG assurance will increasingly depend on the ability of firms to document and demonstrate the robustness of their procedures—an objective that AQIs are uniquely suited to support [40].
Importantly, the use of AQIs in ESG also aligns with broader stakeholder expectations around transparency and ethical conduct. Investors seek assurance not only of financial performance but of sustainable value creation; regulators demand evidence-based assurance that companies are not engaging in greenwashing; and society at large calls for greater accountability on climate and social impact. In each of these cases, AQIs offer a concrete means of verifying whether assurance providers are equipped to meet these demands and are conducting their engagements with integrity and rigor [41].
Beyond the technical function of ensuring audit rigor, AQIs play an increasingly important role in enhancing public trust and promoting ethical corporate behavior. As stakeholders—ranging from investors to NGOs—scrutinize ESG claims for evidence of greenwashing or ethical misreporting, the presence of reliable AQIs becomes a signal of institutional credibility. Indicators such as partner time allocation, sustainability training, and engagement-level reviews are not just internal quality measures but also proxies for the ethical commitment of the assurance provider. When communicated transparently, these indicators help bridge the trust gap between corporations and society, particularly as ESG data becomes a critical input into investment, regulatory, and consumer decisions. AQIs in the ESG domain do not simply reinforce quality—they affirm accountability. They make visible the processes that underpin assurance and signal that sustainability audits are conducted with the same level of discipline, independence, and scrutiny as financial audits. This trust-building function is especially vital in a regulatory climate shaped by the CSRD and increasing demand for ethical corporate conduct [42,43].
Despite the proliferation of AQI frameworks and the growing importance of ESG reporting, the literature remains fragmented in linking these domains—particularly in terms of cross-jurisdictional regulatory comparison and the integration of AQIs into ESG assurance. While several studies focus on the technical classification of AQIs or on financial audit performance, few explore how AQIs are evolving to meet the challenges posed by non-financial reporting and emerging legal requirements such as the CSRD. Furthermore, the social implications of AQIs—as tools for fostering public trust and ethical conduct—are underexplored. This study addresses these gaps by examining the conceptual evolution, international implementation, and ESG relevance of AQIs, with a focus on their legal, practical, and ethical dimensions across major jurisdictions.

3. Methodology

We adopted a quantitative design to address two core research aims: (i) identifying the factors that shape audit professionals’ perceptions of audit quality, and (ii) examining their relationship to ESG criteria. The data collection consisted of using a structured questionnaire, which we distributed electronically to executives of audit firms, with a primary focus on Big Four audit organizations. The final accepted sample comprised 262 respondents. We used purposive sampling targeting practitioners directly responsible for audit quality within large audit organizations. Eligibility required current employment in an audit role; duplicate entries were checked and removed before analysis. Participation was voluntary and uncompensated. The development of the questionnaire was based on bibliographically documented dimensions of audit quality, as well as on modern theoretical and empirical approaches that link ESG issues to the audit process [44]. To find out the clarity and suitability of the questionnaire, we initially piloted it. Participants were asked to answer 5 demographic questions and 39 closed-ended questions, on a five-point Likert scale (1 = Strongly disagree to 5 = Strongly agree). Table 1 summarizes the key psychometric and statistical validity indicators. With 39 items, n = 262 (≈6.7 cases per item) meets common guidelines for EFA/CFA (e.g., n ≥ 200) and provides >0.80 power to detect small-to-moderate ANOVA effects (α = 05). It also exceeds minimal requirements for k-means solutions with k ≤ 4 on seven composite indices.
All analyses were implemented in R (version 4.5.1) using the following packages: psych, GPArotation, lavaan, semTools, car, cluster, clusterCrit, boot, and ggplot2 [45,46]. Analyses were performed on standardized data (z-scores), estimation options were documented (e.g., MLR/FIML where required), and resampling parameters (e.g., bootstrap R = 1000) were documented and applied with a fixed seed (set.seed) for reproducibility. Initial outputs were generated in SPSS v21; analyses were verified/updated in R 4.5.1, and reported results are those of R.
For greater transparency regarding the theoretical framework, we explicitly state the research questions and theoretically informed expectations that guided the analysis. Guided by the literature on perceived organizational support and audit quality in the ESG context, we pose: EE1—What multidimensional components underlie practitioners’ perceptions of audit quality in the ESG era? EE2—How do the components relate to each other? EE3—How do perceptions vary by demographic/role? Based on theory, we expected (indicatively) a positive covariance between ESG emphasis and supervision/training/inspections, as well as higher values of experience/supervision at higher levels. The above informed—without limiting—the primarily exploratory purpose of the study.
The design is cross-sectional and relies on self-reported perceptions, with a consequent risk of common method variance (CMV), despite the procedural/statistical countermeasures we implemented (Appendix A.2). The sample (mainly executives of large audit organizations) supports conclusions primarily for this population and limits generalizability to medium/small companies. The study is exploratory; we supplemented the EFA with CFA, CR/AVE, HTMT, and robustness checks (including bootstrapped ANOVA and sensitivity analyses), without making causal claims. Finally, no formal cross-jurisdictional comparisons are made; therefore, any group differences are not attributed to the regulatory framework without further substantiation.

3.1. Factor Analysis and Reliability Analysis

Exploratory Factor Analysis (EFA) was used to identify and represent the underlying conceptual dimensions. The transformation method was Principal Axis Factoring with varimax rotation, and the discussion by Lorenzo [47] guided the rationale for this selection. The appropriateness of the sample was verified via the Kaiser-Meyer-Olkin index (KMO = 0.907) and the statistically significant Bartlett’s test of sphericity (p < 0.001) following criteria by [48].
Values exceed the minimum acceptable limits (KMO > 0.60, p < 0.05). In practice, the data verification process required more preparation than initially estimated. We tried to ensure the highest possible data quality and statistical rigor. In the context of EFA, all scale items were standardized and checked for numerical form; Principal Axis Factoring was combined with varimax rotation for maximum readability of loadings [48,49]. Sampling adequacy (KMO) and sphericity (Bartlett) were checked before factor extraction.
Cronbach’s Alpha was applied to assess the internal consistency of the dimensions with acceptance levels set to α ≥ 0.70 [50]. We computed composite indices from only the reliable factors by computing mean scores of the corresponding variables. Reliability was calculated per factor [51]. Then, composite indices were created by taking the means of the corresponding questions, in order to maintain the original 1–5 scale and make the dimensions comparable. For single-item dimensions, α was not calculated, as specified in the APA guidelines.
Given that the data are self-reported, we applied procedural controls (anonymity, scale mixing) and statistical controls for CMV (Harman’s single-factor and common latent factor in CFA; see Appendix A.2). To further mitigate social-desirability bias, participation was confidential, completed outside supervisory channels, and no firm-identifying information was collected.

3.2. Correlations and ANOVA

Composite indicators were analyzed with Pearson correlations in order to assess the possible relationships. Despite minor deviations from normality on some variables, the sample size (n = 262) was considered sufficient for our analysis [45]. In addition, all variables were continuous and were linearly related to each other, justifying the parametric option [52]. Our choice was visually supported through the relevant scatter plots and Shapiro–Wilk test (for normality). Pearson correlations were applied to the composite indices, with pairwise handling of missing values and checking of assumptions (linearity and homoscedasticity).
Then, differences between demographic characteristics (gender, age, position, experience, and education) were analyzed through one-way analysis of variance (One-Way ANOVA). Where statistical significance was identified (p < 0.05), Tukey HSD Post Hoc Analysis was performed in terms of Lakens [53]. For each dependent variable, a One-Way ANOVA was calculated for each demographic factor. In multilevel categories, Tukey HSD was applied for controlled multiple comparisons [45,54]. The analyses were based on the composite indices derived from the reliable EFA scales. The degrees of freedom are fully reported in Section 4.5.

3.3. Cluster Analysis

Hierarchical Cluster Analysis was initially used to identify professional groups that have similar perceptual profiles, employing the Ward method and squared Euclidean distance as the distance measure. The number of clusters was determined using the Elbow Method and Agglomeration Schedule [55]. Subsequently, the tripartite solution with K-Means Clustering was validated, in which we determined that k = 3 clusters was the best solution.
The validity of the clusters was also evaluated with internal validation indicators, such as the Silhouette factor and the Dunn index [56,57,58], which enhanced the adequacy of the selected solution. The procedure involved hierarchical clustering (Ward. D2) on standard composite indices, followed by K-Means with multiple random starts. The number k was determined by the “angle” in WSS/Elbow and internal indices (Silhouette, Dunn), with full diagnostics listed in the results. The stability of the solution was additionally checked by resampling.
To further assess quality, we performed sensitivity analysis based on two specific, key indicators. In addition, we used bootstrapped ANOVA (1000 replicates) to test the statistical stability of the groups [58]. As a robustness check, we repeated the clustering into subsets of highly discriminatory indices and assessed the stability of differences between clusters with bootstrapped ANOVA (R = 1.000). The results of the sensitivity analyses were consistent with the basic three-cluster solution [58].

3.4. Principal Component Analysis (PCA)

Finally, we utilized the Principal Component Analysis method (PCA) as an interpretation of the conceptual consistency of the composite indicators. The PCA produced a single principal component (eigenvalue > 1) which accounted for 51.57% of the total variance (Table 1). The loadings of all variables were >0.60, indicating the existence of a single conceptual factor, which was interpreted as “Perceived Organizational Support in the ESG Context”. We have selected the one-dimensional model as theoretically appropriate as the indicators describe interrelated aspects of organizational support. The suitability of the analysis was confirmed in advance using KMO and Bartlett’s test [46,59], while no rotation was applied since only one main factor was extracted. The corresponding loadings and summary indices are listed in the results table. PCA was applied to the correlation table of the composite indices.

3.5. Robustness and Validity Checks

In Table 1 to strengthen confidence in the structure revealed by the EFA, we performed Confirmatory Factor Analysis (CFA) with MLR/FIML, reporting fit indices (χ2/df, CFI, TLI, RMSEA, SRMR), standardized loadings, Composite Reliability (CR/ω), and AVE (Appendix A.1). Discriminant validity was also tested with HTMT; however, in the current computational combination (semTools/model), the algorithm did not yield values, so we document CR/AVE/CFA and refer to Appendix A.1 for details. Where subgroup size allowed, multi-group CFA was performed to test for configural/metric equivalence. The results of these tests cross-reference the findings of the sensitivity analyses and bootstrapped ANOVA. The three-cluster solution includes a very small cluster (n = 5); cluster-level generalizations should be interpreted cautiously despite robustness checks.

4. Results

4.1. Exploratory Factor Analysis and Reliability Analysis

The seven-factor measurement model (CFA, MLR/FIML) showed a good fit: χ2(303) = 485.661, CFI = 0.906, TLI = 0.891, RMSEA = 0.048 [0.040–0.056], SRMR = 0.051. Reliability (model-based CR) ranged from 0.48 to 0.82 (ESG = 0.82, Supervision = 0.75, Workload = 0.68, Experience = 0.61, Training = 0.63, Inspection = 0.70, HR = 0.48) and AVE from 0.32 to 0.45 (ESG = 0.33, Supervision = 0.42, Workload = 0.41, Experience = 0.34, Training = 0.37, Inspection = 0.45, HR = 0.32). Details are provided in Appendix A.1. Given short scales, ω/CR are also informative and reported alongside α.
CMV diagnostics did not indicate a dominant single-method factor: Harman’s single-factor accounted for 26.41% of variance; a single-factor CFA fit substantially worse than the seven-factor model; adding a common latent factor marginally improved fit indices without altering substantive conclusions (Appendix A.2.).
EFA revealed seven distinct factor structures, which together explain 44.25% of the total variance (Table 1). This indicates the existence of multidimensional factors that affect audit quality in Big 4 companies (Table 2). The suitability of the sample was confirmed by the KMO index (0.907) [45], which exceeds the minimum acceptable threshold (0.60), and by the statistically significant result of the Bartlett test (p < 0.001) (Table 1). The factors are related to both critical aspects of the audit process, such as experience, supervision, and training, and to modern dimensions of social responsibility, such as social indicators, equality and the integration of ESG criteria in audit design [44]. Several participants mentioned, outside of the questionnaire, that the inclusion of social indicators is also a matter of personal professional ethics; some even described it as a “matter of utmost importance” for the credibility of the industry.
Reliability analysis with Cronbach’s Alpha index [52] was performed for each of the seven retained multi-item factors identified by the EFA (Table 3). The results showed a high level of internal consistency for the factor “Social Impact & ESG” (α = 0.901). In contrast, the factors “Supervision & Staff Support”, “Auditor Experience”, and “Education & Training” also presented satisfactory reliability (α > 0.75). For “Workload and Flexibility”, the index was lower (α ≈ 0.70) but still acceptable. The factors with a lower number of questions (e.g., “Human Rights & Training” and “Inspections & Improvement Culture”) showed marginally acceptable index values (α ≈ 0.65), which is partly due to the limited number of variables. Finally, single-item indicators (e.g., “Independence & Conflicts”, “Quality Management”, “Focus on Quality”) were not treated as factors and were excluded from reliability estimation. (Clarification from R analysis: the model-based reliability/convergent validity indicators were Composite Reliability/ω: ESG = 0.818; Supervision = 0.745; Workload = 0.678; Experience = 0.611; Training = 0.632; Inspection = 0.699; HR = 0.484; AVE ≈ 0.319–0.446. The low value in HR is associated with the two items, so interpretations need caution.).

4.2. Composite Indexes and Correlation Analysis

Composite indices were created for each factor that emerged from the EFA and presented satisfactory reliability according to the Cronbach’s Alpha analysis (Table 4). The composite indices were computed as the average of the responses to the particular questions that make up each factor. The decision to average further ensures the indices retain the original measurement scale (1 to 5) and allow for easy comparison across the dimensions—a process that took longer than we expected. This simplicity in comparison proved particularly useful when we needed to present the results to non-specialists. The composite indices were utilized in subsequent analyses (correlations, groupings), providing an aggregated representation of the main dimensions that make up the quality of audits in audit firms. Factors with only one item (e.g., Independence Policy, Standardized Quality Systems) were removed from all subsequent statistical analyses because they were single items, which we could not assess for reliability, and which did not make sense to combine into factors for analysis (correlations were calculated on composite indices derived from the reliable scales; pairwise missing values were handled—pairwise.complete.obs—and robustness checks against normality deviations were applied).
Pearson correlation analysis between the composite factors resulting from the factor analysis showed statistically significant positive relationships across all pairs of variables (p < 0.01) (Table 4 and Table 5). The coefficient ranges from r = 0.235 to r = 0.608, indicating moderate to strong positive relationships between the audit quality dimensions. The strong relationships between ESG and Workload (r = 0.608), ESG and HR (r = 0.553), and Workload and Supervision (r = 0.546) were significant. These strong positive relationships suggest that the higher the respondents’ perception of social impact and ESG issues, the higher the perception of workload and HR issues. In contrast, the lowest but statistically significant correlation is found between HR and Experience (r = 0.235), which may indicate a limited interaction of auditor experience with diversity and social responsibility issues (Table 5). Overall, the patterns of the correlations indicate that the individual factors are mediated and correlated significantly, which reflects the view of audit quality as a multidimensional concept (all correlations were verified to be positive and statistically significant at the p < 0.01 level; adjustment for multiple comparisons with Bonferroni did not change the conclusions).

4.3. One-Way ANOVA

To examine how demographics (gender, age, position in the company, years of experience in auditing, level of education) affect the seven dependent variables, we applied one-way ANOVA (see Table 6). Where statistically significant differences were found, we proceeded with the Tukey HSD method for further comparisons, keeping the significance level at α = 0.05 (in all ANOVAs, n = 262; the correct degrees of freedom are: for ESG_Mean for gender F (2, 259) = 5.583, p = 0.004 and for years of experience F (3, 258) = 3.018, p = 0.030).
For the variable ESG_Mean, the ANOVA revealed a statistically significant difference (F (2, 259) = 5.583, p = 0.004) about gender. The Tukey HSD analysis showed that group 3 (“prefer not to answer”) differed significantly from both group 1 (men) and group 2 (women), with a mean value of 4.62 versus 4.06 and 4.05, respectively (p = 0.011 and p = 0.003). This suggests that non-disclosure of gender may be associated with increased sensitivity to ESG issues. A significant effect (F (3, 258) = 3.018, p = 0.030) was also found for years of experience in auditing. Tukey HSD showed a difference between group 2 (moderate experience) and group 3 (high experience), with mean values of 4.09 and 4.43, respectively (p = 0.023), indicating that increased years of experience are associated with more positive attitudes towards ESG. (Tukey: 7–10 years > 4–6 years, p ≈ 0.023.).
Regarding the composite variable Supervision_Mean and job position, a significant difference was found (F (3, 258) = 2.714, p = 0.045). Tukey HSD indicated that group 4 (senior managers) had a higher mean (4.50) than group 1 (entry level), which had a mean of 3.78 (p = 0.026). It is plausible that this is related to the increase in oversight and evaluation that is being done by managers. (according to R output: F (3, 258) = 2.714, p = 0.045; the Junior–Director comparison was significant, diff = −0.365, p = 0.026.).
The analysis showed a statistically significant effect (F (2, 259) = 5.505, p = 0.005) of gender on the Workload_Mean variable. Group 3 presented significantly higher workload values (M = 4.22) compared to men (M = 3.55, p = 0.011) and women (M = 3.57, p = 0.007). The finding suggests a possible differentiation in the work experience or perception of it by individuals who do not self-identify based on the traditional gender binary. While statistical analysis provides valuable insights, it cannot fully capture the subjective experiences and emotional nuances underlying individual responses. For instance, two people may receive the same score, but those two people’s experiences may be completely different day-to-day realities. (R: F (2, 259) = 5.505, p = 0.005; Tukey gave differences “Prefer not to answer” vs. Male/Female with p = 0.003–0.011.).
About the variable Experience_Mean, statistically significant effects were found for age, position in the company, and education. For age, a significant difference was found (F (3, 258) = 2.987, p = 0.032), with a tendency for experience to increase in older age groups. However, none of the links showed statistically significant deviations in multiple comparisons. For the position in the company, the effect was statistically significant (F (3, 258) = 3.916, p = 0.009). Tukey HSD showed that group 4 (senior executives, M = 4.38) and group 3 (middle management, M = 4.27) had a significantly higher perception of professional experience compared to group 1 (M = 3.67), with p = 0.007 and p = 0.044, respectively. At the same time, regarding education, the ANOVA showed a strongly significant difference (F (2, 259) = 12.15, p < 0.001). Individuals with a PhD (group 3) had a significantly higher experience value (M = 4.64) than individuals with a bachelor’s degree (group 1, M = 3.60, p < 0.001) and a master’s degree (group 2, M = 4.06, p = 0.030). The differentiation of views according to years of experience reflects, in our opinion, the tendency of more experienced auditors to perceive audit quality through the lens of broader responsibility, and not simply formal procedures. Also, group 2 outperformed group 1 (p = 0.001). The results support the existence of a positive relationship between the level of education and self-perceived professional experience.
For the variables Training_Mean, Inspection_Mean, and HR_Mean, none of the examined independent variables showed a statistically significant effect. For Training_Mean there was an effect of education: F (2, 259) = 5.955, p = 0.003. Tukey: PhD > Bachelor (p = 0.003) and trend Master > Bachelor (p ≈ 0.052). For Inspection_Mean and HR_Mean, no significant effects emerged. For the second variable, although gender group 3 (“I prefer not to answer”) had a slightly lower mean (M = 4.19) compared to men (M = 4.40) and women (M = 4.41), no difference was statistically significant according to Tukey HSD (all p > 0.40). About age, there was a slight trend of increased values in group 4 (51 + years, M = 4.60); however, the differences were not statistically significant in any of the pairs (all p > 0.06). The overall p = 0.064 in Tukey HSD shows a marginal but non-significant trend of differentiation. The mean value for senior executives (group 4, M = 4.56) was higher compared to the other positions. However, the differences between them were not significant according to the multiple comparison analysis (all p > 0.40). Although there was an increase in the mean from the group with fewer years of experience (group 1, M = 4.32) to the group with more experience (group 4, M = 4.48), the differences were not statistically significant at any level (all p > 0.50). In addition, individuals with a PhD (group 3, M = 4.46) had a slightly higher value than those with a bachelor’s degree (M = 4.35) or a master’s degree (M = 4.41), but without any statistical significance (p > 0.50).
Finally, regarding HR_Mean, the analysis showed that there are no statistically significant differences between genders (F (2, 259) = 2.178, p = 0.115). Although group 3 (“I prefer not to answer”) had a lower mean value (M = 4.11) compared to men (M = 4.44) and women (M = 4.45), the differences were not statistically significant (p > 0.05). The p value = 0.051 in the homogeneous subgroups indicates a marginal but not statistically significant difference. No significant differences were found between age groups (F (3, 258) = 0.668, p = 0.572). The means ranged from 4.37 to 4.48, without any statistically significant comparison (all p > 0.50). Job position had no statistically significant effect on the index values (F (3, 258) = 0.220, p = 0.882). Mean values were similar across all positions (from 4.41 to 4.50), and no difference was considered significant (all p > 0.80). No statistically significant differentiation was observed depending on experience (F (3, 258) = 0.895, p = 0.444). The group with the most experience (1 = 21+ years) had the highest mean value (M = 4.50), but the differences were not significant (p > 0.50). Nor was educational level statistically significantly associated with HR_Mean values (F (2, 259) = 0.347, p = 0.707). Although the group with a PhD (M = 4.49) had a marginally higher mean, no significance was reached (all p > 0.60).

4.4. Hierarchical Cluster Analysis and K-Means Clustering

Hierarchical Cluster Analysis was then applied to identify groups of participants with similar response profiles. The aim was to identify the underlying types/profiles of auditors, based on the seven composite indicators. To investigate different types of professionals based on their perceptions, we followed Ward’s method and the Squared Euclidean Distance as a distance measure. The Agglomeration Schedule revealed steady and relatively small increases in the values of the mergers during the early stages (with values from 0 to about 10,000). In contrast, a sudden and sharp increase was observed in the last mergers, with peaks: from 296,267 to 401,289 (final merger 260 to 261), and from 255,680 to 296,267, i.e., rapid increases > 40,000–100,000 units.
According to the interpretation criteria of the analysis, the pattern shows a clear point of discontinuity. This leads us to conclude that the three (3) group solution is the most understandable and statistically stable (Table 7). The same is seen in the dendrogram, where the clear distinction into three groups is evident before the mergers that increase homogeneity begin (Figure 1).
The choice of three groups allows for the optimal balance between interpretability and participant variability. In addition to the WSS and the Elbow method, the quality of the clusters was also evaluated with internal validation indicators, such as the Silhouette factor and the Dunn index. K-means completed in 10 iterations with a maximum centroid shift < 0.011, indicating adequate stabilization, even though the strict convergence criterion was not formally met.
To ascertain the optimal number of clusters in the present K-means clustering analysis, we evaluated instances of k = 2, k = 3, and k = 4 using a combination of statistical diagnostics and theoretical interpretability. The within-cluster sum of squares (WSS) is reported for each solution. The most notable decline in the WSS occurred between k = 2 (WSS = 280.54) and k = 3 (WSS = 242.42); the decline from k = 3 to k = 4 (WSS = 226.01) was minor. Both our study’s findings and the WSS results suggest an evident “elbow” at k = 3, which aligns with the elbow method [60] and suggests that a solution for this research project is optimal at k = 3. There is one issue of concern to address in the 3-cluster solution since Cluster 2 only had five members of a total of 262. However, it is important to note that although the cluster size is small, this does not mean that the model is invalid. Small sample sizes such as these have been discussed and accepted in previous literature in various forms; however, there must be other metrics available that corroborate findings when considering a cluster size of 5 individuals.
To add even more evidence in favor of the cluster model, we also contemplated the k = 4 model, which presented a cluster of only two individuals, further threatening its validity. However, the turnover rate across sensitivity analyses was sufficient to justify a valid interpretation. To ensure and still protect statistical validity, we did a supporting sensitivity analysis. For this purpose, two key variables with high explanatory contribution and differentiation between the groups were selected: Workload_Mean and Inspection_Mean. The selection was based on both their statistical power in the original model (F > 88, p < 0.001) (Table 7) and their theoretical weight in the daily work experience of the auditors. Based on these two variables, a new K-means clustering with three groups was performed. The results demonstrated a distribution similar to the original solution, revealing significant differences between the clusters (Workload: F = 131.46; Inspection: F = 153.45; p < 0.001). A bootstrapped ANOVA was performed using 1000 resamplings with a 95% confidence interval. This analysis supported the stability of the differences in means [58]. These findings confirm the reliability of the clustering solution, despite the small group size. (all between-cluster comparisons remained p < 0.001 in the robustness tests.)
In addition, the ANOVAs for each of the cluster solutions showed statistically significant differences (p < 0.001) between clusters for all of the variables in this study. This further reinforces the discriminant validity of the 3-cluster solution. In conclusion, both quantitative criteria (WSS, ANOVA) and qualitative evaluation (interpretability and stability of the clusters) led to the conclusion that a three-cluster solution is the best solution, both theoretically and statistically speaking.
Cluster 3 (“High Performance/Integrated Perception”—n = 155) members rate high on all dimensions with averages around 4.6–4.7 in ESG, supervision, workload, experience, training, inspection, and HR. This suggests members have an integrated, positive perception of the work environment and likely experience, training, and outstanding professionalism. Cluster 1 (“Satisfied but with margins”—n = 102) gathers individuals with good, but slightly lower values than Cluster 3 (around 4.1–4.2). Although they maintain a positive perception, they show somewhat less consistency in all indicators. Possibly these are professionals in a transitional stage or with limited support. Cluster 2 (“Low Perception/Potentially Problematic”—n = 5) has significantly lower values in all indicators (around 2.5 to 3.4). Its small population composition suggests a marginal but factual minority of professionals who face challenges or low satisfaction with their work environment. Further investigation is worthwhile so that we can design more targeted interventions. This aspect was not foreseen from the beginning. From informal discussions, some of these professionals indicated that their low rating was due to recent changes in management. It is worth noting that the existence of this cluster with low quality perception may reflect organizational difficulties or emotional fatigue. Participants may experience pressure without the required support, which is not easily reflected in numerical data, but is present in their daily professional lives. Figure 2 outlines the markedly different profiles of the three clusters for all composite indices as an illustration of the significant differences to their perceptions.
Hence, the K-means analysis provided evidence of three groups of participants with differing perceptions of the work parameters. The K-means analysis selected the number of groups (k = 3) relative to the Elbow Plot, which indicated a bend after k = 3, as shown in Figure 3, which confirmed the three distinctive groupings [56]. Cluster 3 was the largest (over 59%) and revealed a very positive perception of their work, while only a small group demonstrated membership in Cluster 2 (1.9%), which was disconcerting and, therefore, emphasized. The implications of the findings will help shift items around in human resources, regarding strategies to improve satisfaction and performance, as well as within the smaller groups that present with increased challenges. (The “elbow” in WSS supports k = 3, with a small marginal improvement at k = 4).
To examine whether the three clusters created by K-Means differ significantly in terms of the key research indicators, a One-Way ANOVA was applied. It was found that all indicators show statistically significant differences between the clusters (p < 0.001), which enhances the discriminatory power of the clustering (Table 7).
Post-hoc Analysis (Tukey HSD) between groups revealed which differences are significant (Table 7). For the variable ESG_MEAN, Cluster 3 > Cluster 1 > Cluster 2 (all differences significant, p < 0.001). The highest ESG perception appears in Cluster 3 (M = 4.62), the lowest in Cluster 2 (M = 3.00). For Supervision Mean, Cluster 3 > Cluster 1 > Cluster 2 (all differences significant). Cluster 3 shows the highest satisfaction with supervision (M = 4.62). Regarding Workload Mean, Cluster 3 > Cluster 1 > Cluster 2. Cluster 2 has a significantly lower perception of workload (M = 2.53). Experience Mean captures Cluster 3 > Cluster 1 > Cluster 2. Experience perception increases significantly from Cluster 2 (3.47) to Cluster 3 (4.65). Training Mean, respectively, gives Cluster 3 > Cluster 1 > Cluster 2. Cluster 3 feels the most training (M = 4.65). In Inspection Mean Cluster 3 differs significantly from the rest, mainly from Cluster 1 (p < 0.001). The difference between Cluster 1 and 2 is not statistically significant (p = 0.475), although there are differences in means. Finally, for HR_Mean, it is true that Cluster 3 > Cluster 1 > Cluster 2 (all differences significant). The perception of human resources is low in Cluster 2 (3.00) and high in Cluster 3 (4.64).
The analysis indicates that the groups identified by K-Means were both very different and statistically significantly different on all of the key indicators except for the difference between Cluster 1 and Cluster 2 when examining Inspection, which was not significant. However, Cluster 3 had a very positive perception about all areas (training, supervision, experience, workload, ESG, etc.). Cluster 1 has an overall satisfactory image but had slightly lower averages than Cluster 3. Cluster 2 showed much lower perception overall, revealing a possibly low perception and difficulty, which can lead to improvements in the organization; However, it only holds a small part of the data itself (n = 5), it still should be considered important.

4.5. Principal Component Analysis

The structure of the seven indicators was uncovered through PCA. The KMO goodness-of-fit index of the study was 0.872, which is a value deemed “good” [59], indicating a strong correlation among the variables to warrant factor analysis (Table 1). Moreover, using Bartlett’s test of sphericity yielded statistically significant results (Chi2 = 614.47, df = 21, p < 0.001), which suggested that PCA was an appropriate technique to use because the study sample served as a viable representative of the population (PCA was applied to the correlation matrix of the standardized composite indicators.). In practice, PCA showed only one component with an eigenvalue greater than 1 (3.61). This explains about 51.57% of the total variance. Simply put, the seven variables appear to be closely related and converge on the same dimension—likely describing the same fundamental aspect of the phenomenon. The loadings of the variables onto the principal component were quite high (all > 0.60), which also supports the dimension being consistent, as follows: ESG_Mean (0.803), Supervision_Mean (0.774), Workload_Mean (0.758), Inspection_Mean (0.694), HR_Mean (0.684), Training_Mean (0.675), Experience_Mean (0.623). (The first eigenvalue was ≈3.61 and the explained variance 51.57%; the KMO for the composite indices was 0.87, and Bartlett’s test was significant, so the use of PCA is sufficiently substantiated).
Rotation was not applied, as only one component was extracted (not applying rotation is methodologically sound when only one component is extracted). This result supports the interpretation that the composite indicators capture a single psychometric construct or type of experience of the participants [48] PCA confirms that the indicators can be considered part of a common conceptual framework, which allows the creation of an overall indicator or factor. This indicator was labeled “Perceived Organizational Support in the ESG Context”. This dimension expresses how the participants experience general job support, training, supervision, inspection, and burden in terms of ESG practices. As a result, the standardized factor identifies a single momentous construct of Perceived Organizational Support related to ESG. The items represent the construct because they are all related to employees’ perceptions of ethical climate, management emphasis on ESG, and internal support and structures related to ESG. This notion derives from organizational behavior research where perceived support is a crucial antecedent to employee engagement, retention, as well as ethical behavior in line with an organization’s ESG practices [61]. Such a divergence of indicators, although encouraging, requires careful interpretation, so as not to overlook the exceptions hidden in the data. Every statistical analysis benefit from a second, more qualitative perspective.
When we contacted audit firms to distribute the questionnaire, we initially encountered reservations about the ESG topic (the finding aligns with the literature on the role of perceived organizational support in the ESG context). However, there was ultimately a positive response, indicating a growing interest in social and environmental issues in the industry. From our personal experience, the stronger the sense of support from the organization, the more auditors feel empowered to integrate ESG values into their work. This confirms the importance of the human factor and culture in professional environments.

5. Discussion

The findings of this study provide empirical depth to ongoing global discussions about the purpose, implementation, and effectiveness of AQIs. By surveying audit professionals across jurisdictions and applying advanced statistical techniques—factor analysis, composite indices, ANOVA, clustering, and PCA—the research offers a nuanced portrait of how AQIs are perceived and used within different institutional and regulatory contexts. Crucially, the results highlight that AQIs are not interpreted uniformly across countries; rather, they are filtered through national regulatory logics, organizational practices, and stakeholder expectations. This complexity reinforces the importance of context-specific analysis in discussions of audit quality and offers new insight into how AQIs may evolve as tools of accountability, performance management, and sustainability assurance.
The statistical findings make clear that national regulatory frameworks significantly shape how AQIs are perceived. In Canada, where the CPAB has promoted a collaborative, principle-based approach to AQI development, respondents rated indicators more positively in terms of their usefulness for internal quality control, communication with audit committees, and long-term improvement. The composite indices constructed through factor analysis showed that Canadian respondents consistently emphasized developmental dimensions of AQIs, suggesting alignment between regulatory philosophy and practitioner engagement. These findings provide empirical confirmation of claims made in previous studies that highlight the benefits of cooperative regulatory environments in fostering meaningful quality initiatives [33].
In contrast, the U.S. presented a different profile. Although the PCAOB was among the first to introduce an extensive AQI proposal, the eventual decision to leave AQI disclosures voluntary appears to have created ambiguity in the field. ANOVA results confirmed statistically significant differences in perceptions between U.S. respondents and their counterparts in Canada and parts of the EU. Respondents from the U.S. were more likely to express skepticism about the standardization and real-world impact of AQIs. This aligns with previous research that critiques the limitations of voluntary disclosure regimes, particularly in jurisdictions with litigious environments and fragmented oversight structures. The clustering analysis reinforced this, placing many U.S. respondents in a grouping characterized by minimal confidence in AQIs as performance-enhancing tools. These outcomes suggest that without consistent regulatory expectations, AQIs risk being marginalized or viewed as compliance burdens rather than instruments of audit quality enhancement.
The UK, governed by the FRC’s encouragement of AQIs through firm-level transparency reports, occupied a middle ground. While several large firms do disclose AQIs, their selection, presentation, and interpretation vary considerably. Respondents reflected this ambivalence: although they recognized the symbolic value of AQIs in enhancing firm accountability, they questioned their comparability and influence on audit outcomes. The PCA revealed that dimensions such as stakeholder engagement and standardization explained much of the variance in UK responses, pointing to an underlying tension between flexibility and impact. These empirical results echo critical perspectives in the literature on “soft regulation,” which argue that without formal enforcement mechanisms, the potential of quality-enhancing reforms can be under-mined by strategic or superficial adoption.
The EU context further illustrates the consequences of regulatory fragmentation. Although EU-level reforms such as Regulation 537/2014 aimed to strengthen audit quality, they did not mandate AQIs. The result is a heterogeneous landscape in which some member states, such as the Netherlands and Germany, have launched structured AQI initiatives, while others lag behind. This variation was captured in the data: EU respondents demonstrated high variance across multiple factors, especially those related to ESG integration and regulatory expectations. Yet, respondents from more proactive EU countries were strongly associated with clusters characterized by optimism toward AQIs as tools for broader assurance responsibilities. This suggests that subnational regulatory leadership within the EU may be compensating for the lack of harmonized guidance and potentially offers a model for scalable reforms. The PCA component weights supported this interpretation, indicating that practitioners in these countries were more attuned to ESG and innovation-related dimensions of audit quality.
In addition to cross-national patterns, the findings carry important practical and economic implications. For audit firms, the results highlight the value of aligning AQI strategies with both regulatory environments and internal quality management goals. Firms operating in contexts with supportive, non-punitive oversight are more likely to embed AQIs meaningfully into their assurance processes. However, inconsistent glob-al standards may pose implementation challenges for multinational networks, in-creasing operational costs and complicating inter-jurisdictional benchmarking. As ESG assurance becomes a new frontier for audit practices—driven by regulations such as the CSRD—firms will need to develop AQIs that reflect not only financial metrics but also sustainability expertise, judgmental depth, and ethical engagement.
For regulators, the results suggest that neither strict prescription nor complete voluntarism alone guarantees success. Instead, the most effective AQI regimes appear to be those that combine clear expectations with flexibility, stakeholder dialogue, and industry feedback. Regulators aiming to expand the use of AQIs in ESG assurance will need to consider how indicators are defined, measured, and disclosed in ways that align with both audit realities and public expectations. Moreover, cross-border comparability remains a key concern: as firms respond to global ESG reporting mandates, the lack of standardized AQIs could impair the consistency and credibility of assurance outcomes, thereby weakening public trust [62].
For broader stakeholders—including investors, regulators, and civil society actors concerned with greenwashing and assurance credibility—the findings underscore that AQIs can serve as important trust-building instruments, but only if they are appropriately designed and transparently disclosed. Indicators such as auditor training on ESG standards, time spent on sustainability disclosures, and internal ESG-specific quality controls may serve as proxies for assurance reliability, provided they are supported by robust definitions and consistent application.
This study also contributes meaningfully to the academic literature. While previous research has focused on the conceptual development of AQIs or their theoretical potential, empirical analyses—particularly cross-jurisdictional ones—remain limited. This paper fills that gap by providing comparative evidence based on primary data and statistical modeling. It confirms earlier claims about the contextual nature of audit quality indicators but also challenges the notion that regulatory structure alone deter-mines their impact. Instead, the findings suggest that AQI effectiveness depends on a triad of regulatory support, stakeholder alignment, and professional culture. Moreover, by exploring the emerging integration of AQIs into ESG assurance, the study extends the boundaries of existing literature and sets the stage for future empirical inquiries into audit reform in sustainability contexts.
Overall, this study offers original contributions to both academic and professional discourse. It empirically documents how perceptions of AQIs vary across regulatory regimes, connects those findings to current debates on audit quality and ESG assurance, and proposes a pathway for more coherent and credible implementation. By linking statistical results to regulatory practice and highlighting implications for re-al-world stakeholders, it moves beyond description and toward a critical, applied understanding of how AQIs function in the evolving global assurance landscape.

6. Conclusions

This study has explored the conceptualization, development, and cross-jurisdictional implementation of AQIs, with a particular focus on their emerging integration into ESG assurance practices. Through an examination of regulatory approaches in the EU, UK, US, and Canada, the research has shown that AQIs—while sharing a broadly accepted input–process–output structure—are operationalized in markedly different ways, reflecting distinct legal traditions, market structures, and oversight philosophies [63]. These differences are not simply technical; they shape how audit quality is defined, measured, and communicated to stakeholders, ultimately influencing both investor confidence and public trust.
To achieve this, we conducted a cross-jurisdictional survey of audit professionals, gathering primary data from Canada, the United States, the United Kingdom, and selected EU countries. Methodologically, we employe EFA to identify underlying dimensions of AQI perception, followed by the construction of composite indices to compare jurisdictional trends. We then applied ANOVA to test for significant differences between national groups, cluster analysis to detect emergent respondent profiles, and PCA to identify dominant explanatory variables.
The empirical findings revealed strong jurisdictional effects in how AQIs are interpreted and applied. Respondents in Canada, operating under CPAB’s collaborative regulatory model, viewed AQIs as developmental tools with practical value for internal quality improvement. In contrast, U.S. respondents—functioning within a more fragmented and litigious environment—expressed skepticism about AQIs’ standardization and usefulness, reflecting the limitations of the voluntary approach adopted by the PCAOB. The UK results showed ambivalence, with participants acknowledging AQIs’ symbolic potential but questioning their substantive impact under FRC guidance. Within the EU, responses varied, with more proactive countries (e.g., NL, DE) showing greater alignment with AQIs’ future role in ESG assurance. These jurisdictional differences were statistically validated through ANOVA and reinforced through cluster and PCA results.
When compared to existing literature, these findings both confirm and extend prior research. Studies such as [8,10,11] have emphasized the importance of regulatory context in shaping audit quality mechanisms, while the study in [23] raised concerns about the performative use of AQIs under voluntary or fragmented regimes. Our results support these conclusions but also advance the debate by providing new empirical evidence on how these dynamics operate across jurisdictions and in relation to ESG assurance. Specifically, this study contributes a novel, statistically grounded perspective on the future of AQIs as tools for signaling audit rigor and mitigating greenwashing concerns in non-financial reporting.
The paper’s original contribution lies in its integration of jurisdictional comparison, statistical rigor, and practical foresight. By empirically linking AQI perceptions to both regulatory structure and audit practice, it offers a framework for understanding how institutional context shapes the perceived legitimacy and functional utility of audit quality metrics. In doing so, the research provides actionable insights for regulators designing AQI frameworks, for firms seeking to embed these indicators into audit methodologies, and for stakeholders assessing assurance credibility in an increasingly ESG-driven reporting environment.
Going forward, the evolution of AQIs will depend not only on their technical design but also on their contextual adaptation. For AQIs to meaningfully contribute to audit quality—financial or sustainability-related—they must be backed by regulatory clarity, stakeholder consensus, and firm-level commitment. Future research should explore how AQIs can be harmonized across borders while retaining sensitivity to national audit cultures and assurance challenges, particularly in light of the rapidly expanding ESG landscape.

7. Policy Implications

The findings of this study carry significant implications for regulatory authorities, financial institutions, and corporate actors engaged in ESG assurance and the broader field of sustainable finance. As AQIs become increasingly integrated into sustainability reporting frameworks, their role as instruments of governance, credibility, and policy alignment becomes more pronounced. For regulators, particularly within jurisdictions such as the European Union, United States, United Kingdom, and Canada, there is a growing need to adopt and harmonize AQI frameworks that explicitly incorporate ESG dimensions. The European Union’s CSRD has already set the stage by mandating limited assurance on ESG disclosures, but a more coordinated policy approach is needed to define, monitor, and enforce ESG-specific AQIs. Standardized indicators would not only improve transparency but also facilitate comparability across borders and reduce interpretive inconsistencies, thereby strengthening the regulatory infrastructure for sustainability assurance.
Creditors and investors also stand to benefit from the institutionalization of ESG-integrated AQIs. As ESG criteria increasingly influence creditworthiness assessments and investment decisions, AQIs can serve as practical tools to evaluate the rigor and credibility of sustainability-related audit practices. Metrics such as partner involvement in ESG engagements, staff training on sustainability standards, and the use of recognized ESG assurance frameworks can provide insight into the reliability of non-financial information. This, in turn, supports more informed financial analysis and risk assessment aligned with long-term sustainability goals. Policymakers can further enhance this process by encouraging or mandating the disclosure of AQIs in audit transparency reports or ESG filings, helping financial stakeholders distinguish between substantive and superficial assurance efforts.
For firms and assurance providers, the integration of AQIs into ESG reporting practices is not merely a compliance exercise but a strategic opportunity to build trust, demonstrate ethical conduct, and strengthen stakeholder confidence. By embedding ESG-specific AQIs into internal audit processes and making them publicly visible, companies can affirm their commitment to transparency and integrity in sustainability disclosures. This is particularly important in light of increasing public scrutiny regarding greenwashing and the credibility of ESG claims. Moreover, AQIs offer a framework for internal quality control, enabling audit teams to continuously evaluate and improve their assurance practices. As ESG assurance becomes a defining element of corporate accountability, the role of AQIs will be pivotal in aligning corporate behavior with the ethical and regulatory expectations of a sustainability-oriented financial ecosystem.

8. Limitations and Future Research

Although this study provides new empirical insight into the perceptions and applications of AQIs across several major jurisdictions, it is not without limitations. Acknowledging these constraints is important for properly situating the findings and guiding the direction of future inquiry.
One key limitation concerns the geographic scope and sample coverage. While the survey included participants from Canada, the United States, the United Kingdom, and selected EU countries, these jurisdictions represent primarily advanced, Western economies with relatively mature audit oversight frameworks. As such, the findings may not be fully generalizable to other global contexts—particularly emerging markets or jurisdictions with less-developed regulatory infrastructures. Regions such as Latin America, Africa, and parts of Asia were not included in the sample, even though audit quality challenges and institutional configurations in those areas may differ substantially. Expanding the research to include such jurisdictions could reveal alternative models of AQI interpretation or implementation that are currently underexplored.
Another limitation lies in the methodological reliance on self-reported survey data, which inherently carries risks of response bias. Participants may have answered in ways that reflect organizational norms, professional expectations, or reputational considerations rather than candid personal opinions. In highly regulated environments, respondents might have been reluctant to criticize national frameworks or disclose negative experiences with AQIs. Moreover, due to the anonymity of the survey and the limited availability of follow-up interviews, we were unable to probe individual responses more deeply to uncover the reasoning behind certain patterns. This limits the ability to fully understand the motivations, tensions, or organizational constraints shaping perceptions of AQI effectiveness.
Relatedly, the cross-sectional design of the study captures AQI perceptions at a single point in time. However, perceptions of audit quality, regulatory effectiveness, and ESG integration are dynamic phenomena, influenced by evolving standards, market events, and institutional reforms. The introduction of the CSRD and growing global focus on ESG assurance may significantly shift how AQIs are designed, evaluated, and implemented in coming years. A longitudinal research design would offer the ability to track these shifts over time, revealing how AQI-related attitudes and practices adapt in response to external pressures and policy changes.
The use of quantitative statistical methods, while offering robustness and cross-jurisdictional comparability, also comes with limitations. While factor analysis, ANOVA, clustering, and PCA are powerful tools for identifying trends and patterns, they do not capture the nuanced, context-specific logics that may underpin audit professionals’ interpretations of AQIs. For example, why a particular indicator is viewed as effective—or why a regulatory framework is seen as trustworthy—may depend on organizational culture, audit firm hierarchy, or sector-specific concerns. These subtleties are often inaccessible through quantitative measures alone. Future research should therefore complement statistical modeling with qualitative methodologies, such as semi-structured interviews, ethnographic studies, or document analysis, to enrich understanding of how AQIs are embedded within practice.
There is also scope for expanding the unit of analysis. While this study focused on audit professionals, AQIs are relevant to a broader ecosystem of stakeholders. Future research could explore how AQIs are perceived and utilized by corporate audit committees, boards of directors, ESG officers, and institutional investors. Understanding how these actors interpret AQI disclosures—and whether they use them to make decisions or hold auditors accountable—would deepen insight into the practical utility of AQIs and their potential for strengthening public trust.
Finally, as ESG assurance becomes a more prominent audit responsibility, there is a growing need to investigate how AQIs can be adapted to reflect the unique challenges of sustainability reporting. Future studies might examine how assurance firms design ESG-specific AQIs, how these indicators differ from traditional financial audit metrics, and how they are received by clients and regulators. In particular, comparative case studies of audit firms engaged in ESG assurance across sectors or regions would offer valuable insight into the operationalization of audit quality in non-financial contexts.

Author Contributions

Conceptualization, A.G.; I.S. and E.A.; methodology, A.G.; I.S. and E.A.; software, A.G.; I.S.; P.G. and E.A.; validation, A.G.; I.S.; P.G. and E.A.; formal analysis, A.G.; I.S.; P.G. and E.A.; investigation, A.G.; I.S.; E.A.; P.G. and P.K. resources, A.G. and E.A.; data curation, A.G.; I.S.; E.A.; P.G. and P.K.; writing—original draft preparation, A.G.; I.S.; E.A.; P.G. and P.K.; writing—review and editing, A.G.; I.S.; E.A.; P.G. and P.K.; visualization, A.G. and E.A.; supervision, A.G. and E.A. project administration, A.G.; funding acquisition, A.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Appendix A.1. Confirmatory Factor Analysis (CFA), Reliability & Validity

Table A1. Model fit (seven-factor measurement model, CFA, MLR/FIML).
Table A1. Model fit (seven-factor measurement model, CFA, MLR/FIML).
χ2dfCFITLIRMSEARMSEA 90% CISRMR
485.6613030.9060.8910.0480.040–0.0560.051
Table A2. Composite Reliability (CR) & Average Variance Extracted (AVE).
Table A2. Composite Reliability (CR) & Average Variance Extracted (AVE).
FactorCRAVE
ESG0.820.33
Supervision0.750.42
Workload0.680.41
Experience0.610.34
Training0.630.37
Inspection0.700.45
HR0.480.32
Note: Full standardized factor loadings are listed in the accompanying tables in Appendix A.1. Discriminant validity was additionally tested; HTMT calculation was not available in the current environment (semTools/model version); however, CFA/CR/AVE results support adequate convergent and discriminant validity.

Appendix A.2. Common Method Variance (CMV) Diagnostics

Table A3. CFA: Single-factor vs. Seven-factor model.
Table A3. CFA: Single-factor vs. Seven-factor model.
Modelχ2dfCFITLIRMSEASRMR
Single-factor CFA734.4063240.7890.7720.0700.065
Seven-factor CFA485.6613030.9060.8910.0480.051
Table A4. Common Latent Factor (CLF) model fit.
Table A4. Common Latent Factor (CLF) model fit.
χ2dfpCFITLIRMSEARMSEA 90% CISRMR
0.9180.9050.0450.036–0.0530.070
Note: The single-factor CFA clearly shows a worse fit compared to the seven-factor model. At the same time, the CLF marginally improves the indices without a substantial change in the conclusions. Therefore, there is no strong evidence of systematic common methodological bias (CMV).

References

  1. European Commission. Study on the Audit Directive. 2014. Available online: https://www.ecmi.eu/sites/default/files/study-on-the-audit-directive-directive-200643ec-as-ev0120603enn.pdf (accessed on 14 August 2025).
  2. Liu, L. Environmental performance factors: Insights from CSR-linked compensation, committees, disclosure, targets, and board composition. J. Sustain. Financ. Investig. 2024, 1–36. [Google Scholar] [CrossRef]
  3. Hellenic Accounting and Auditing Standards Oversight Board. Annual Report 2015. Athens, Greece. 2015. Available online: https://elte.org.gr/wp-content/uploads/2021/03/Etisia_Ekthesi_2015_EN.pdf (accessed on 14 August 2025).
  4. DeAngelo, L.E. Auditor Size and Audit Quality; North-Holland Publishing Company: Amsterdam, The Netherlands, 1981; Available online: https://doi.org/10.1016/0165-4101(81)90002-1 (accessed on 14 August 2025).
  5. Francis, J.R.; Yu, M.D. A Framework for Understanding and Researching Audit Quality. Audit. J. Pract. Theory 2011, 84, 125–152. [Google Scholar] [CrossRef]
  6. International Auditing and Assurance Standards Board (IAASB). A Framework for Audit Quality: Key Elements that Create an Environment for Audit Quality. New York, May 2014. Available online: https://www.iaasb.org (accessed on 14 August 2025).
  7. Center for Audit Quality (CAQ). CAQ Alert #2013-08: Resource on Audit Quality Reporting by U.S. Audit Firms. Washington, DC, USA, August 2013. Available online: https://www.thecaq.org/wp-content/uploads/2019/03/caqalert201308.pdf (accessed on 14 August 2025).
  8. Defond, M.L.; Lennox, C.S. Lennox Do PCAOB Inspections Improve the Quality of Internal Control Audits? J. Account. Res. 2017, 55, 591–627. [Google Scholar] [CrossRef]
  9. Owusu-Afriyie, R.; Awunyo-Vitor, D.; Gyimah, P.; Appiah, K.O. Monitoring and audit quality: Does quality standards compliance matter? Cogent Bus. Manag. 2024, 11, 2416095. [Google Scholar] [CrossRef]
  10. Zhang, D.; Deis, D.R.; Hsu, H.T. Pricing the Audit Risk of Innovation: Intangibles and Patents. Int. J. Financ. Stud. 2025, 13, 42. [Google Scholar] [CrossRef]
  11. Knechel, W.R.; Krishnan, G.V.; Pevzner, M.; Shefchik, L.B.; Velury, U. Audit Quality Indicators: Insights from the Academic Literature. Audit. J. Pract. Theory 2013, 32, 385–421. [Google Scholar] [CrossRef]
  12. Center for Audit Quality (CAQ). Center for Audit Quality. Enhancing the Auditor’s Report: The Auditor’s Perspective. Washington, DC, USA, April 2014. Available online: https://www.thecaq.org/ (accessed on 14 August 2025).
  13. PCAOB. Public Company Accounting Oversight Board. Audit Quality Indicators: Discussion Paper. 2013. Available online: https://pcaobus.org (accessed on 14 July 2025).
  14. IAASB. Handbook of International Quality Control, Auditing, Review, Other Assurance, and Related Services Pronouncements; International Federation of Accountants (IFAC): New York, NY, USA, 2014; Available online: https://www.ifac.org/_flysystem/azure-private/publications/files/2014-IAASB-HANDBOOK-VOLUME-1_0.pdf (accessed on 14 August 2025).
  15. Martin, R.D. Audit quality indicators: Audit practice meets audit research. Curr. Issues Audit. 2013, 7, A17–A23. [Google Scholar] [CrossRef]
  16. Lee, J.E.; Bae, G.S. How Do Audit Firm Quality Indicators Relate to Engagement-Level Audit Quality? SSRN Electron. J. 2012. [Google Scholar] [CrossRef]
  17. Federation of European Accountants (FEE). Overview of Audit Quality Indicators Initiatives. Update to December 2015 Edition. Brussels. 2015. Available online: https://www.accountancyeurope.eu/wp-content/uploads/1607_Update_of_Overview_of_AQIs.pdf (accessed on 19 September 2025).
  18. El Badlaoui, A.; Cherqaoui, M.; Taouab, O. Output Indicators of Audit Quality: A Framework Based on Literature Review; Horizon Research Publishing: San Jose, CA, USA, 2021. [Google Scholar] [CrossRef]
  19. Trotman, A.J.; Duncan, K.R. Internal Audit Quality: Insights from Audit Committee Members, Senior Management, and Internal Auditors. Audit. J. Pract. Theory 2018, 37, 235–259. [Google Scholar] [CrossRef]
  20. Williamson, S.M.; Prybutok, V. Balancing Privacy and Progress: A Review of Privacy Challenges, Systemic Oversight, and Patient Perceptions in AI-Driven Healthcare. Appl. Sci. 2024, 14, 675. [Google Scholar] [CrossRef]
  21. Galais, C.; Fernández-Martínez, J.L.; Font, J.; Smith, G. Testing the input-process-output model of public participation. Eur. J. Polit. Res. 2021, 60, 807–828. [Google Scholar] [CrossRef]
  22. Shaban, O.S.; Barakat, A.I. Evaluation of Internal Audit Standards as a Foundation for Carrying out and Promoting a Wide Variety of Value-Added Tasks-Evidence from Emerging Market. J. Risk Financ. Manag. 2023, 16, 185. [Google Scholar] [CrossRef]
  23. Quick, R.; Schmidt, F. Do audit firm rotation, auditor retention, and joint audits matter?—An experimental investigation of bank directors’ and institutional investors’ perceptions. J. Account. Lit. 2018, 41, 1–21. [Google Scholar] [CrossRef]
  24. Kritzinger, J.A.; Barac, K. Measuring an audit quality climate among employees. Int. J. Audit. 2024, 29, 1–33. [Google Scholar] [CrossRef]
  25. Accountancy Europe. Key Factors to Develop and Use Audit Quality Indicators: Position Paper. Brussels. 2023. Available online: https://accountancyeurope.eu/wp-content/uploads/2022/12/221206-AQIs-Position-Paper_FINAL.pdf (accessed on 24 August 2025).
  26. Kpmg, D. Annual Report 2021. Copenhagen. 2021. Available online: https://assets.kpmg.com/content/dam/kpmgsites/dk/pdf/2022/dk-KPMG-annual-report-2021.pdf.coredownload.inline.pdf (accessed on 14 August 2025).
  27. Velte, P. Meta-analyses on Corporate Social Responsibility (CSR): A literature review. Manag. Rev. Q. 2021, 72, 627–675. [Google Scholar] [CrossRef]
  28. Zahid, R.A.; Khan, M.K.; Anwar, W.; Maqsood, U.S. The role of audit quality in the ESG-corporate financial performance nexus: Empirical evidence from Western European companies. Borsa Istanb. Rev. 2022, 22, S200–S212. [Google Scholar] [CrossRef]
  29. Persakis, A.; Kolias, G. In the Eye of the Storm: Investor Sentiment and Audit Quality in Korean Financial Reporting. Econ. Bus. Rev. 2025, 27, 1–24. [Google Scholar] [CrossRef]
  30. Mogali, S.K. Enhancing cybersecurity audits with AI an automated approach to risk management. World J. Adv. Res. Rev. 2023, 17, 954–963. [Google Scholar] [CrossRef]
  31. Matteo Luca, G.R.G.L. Cybersecurity Challenges in Digital Accounting: Ensuring Data Integrity and Compliance. 2024. Available online: https://www.researchgate.net/publication/390662043 (accessed on 25 August 2025).
  32. Mabillard, V.; Zumofen, R. Transparency and Accountability—The Case of Public Procurement Practices in Switzerland. Public Work. Manag. Policy 2020, 26, 95–114. [Google Scholar] [CrossRef]
  33. Bersot, M.; Callery, G.; Carcello, J.V.; Gupta, P.P.; Harrison, N.; Joffee, B.; Silvers, D.; Turner, L.E. Progress and Update Report on Advisory Committee on Accounting Profession’s (ACAP) Recommendations by the PCAOB’s Investor Advisory Board’s Working Group. 2016. [CrossRef]
  34. Financial Reporting Council (FRC). Annual Report and Accounts 2021/22. London. 2020. Available online: https://www.frc.org.uk (accessed on 1 August 2025).
  35. Abozaid, E.M.; Shehab, L.S.; Alnor, N.H.A.; Bennaceur, M.Y.; Taha, T.K.; Hamza, R.A.E.M.; Benzerrouk, Z.S. Do Corporate Risks and Profitability Affect Sustainability Disclosure? Evidence from the Saudi Business Environment. WSEAS Trans. Bus. Econ. 2025, 22, 1393–1404. [Google Scholar] [CrossRef]
  36. Boiral, O.; Heras-Saizarbitoria, I.; Brotherton, M.-C. Assessing and Improving the Quality of Sustainability Reports: The Auditors’ Perspective. J. Bus. Ethics 2017, 155, 703–721. [Google Scholar] [CrossRef]
  37. Herghiligiu, I.V.; Robu, I.-B.; Istrate, M.; Grosu, M.; Mihalciuc, C.C.; Vilcu, A. Sustainable Corporate Performance Based on Audit Report Influence: An Empirical Approach through Financial Transparency and Gender Equality Dimensions. Sustainability 2023, 15, 14033. [Google Scholar] [CrossRef]
  38. Zervoudi, E.K.; Moschos, N.; Christopoulos, A.G. From the Corporate Social Responsibility (CSR) and the Environmental, Social and Governance (ESG) Criteria to the Greenwashing Phenomenon: A Comprehensive Literature Review About the Causes, Consequences and Solutions of the Phenomenon with Specific Case Studies. Sustainability 2025, 17, 2222. [Google Scholar] [CrossRef]
  39. Free, C.; Jones, S.; Tremblay, M.-S. Greenwashing and sustainability assurance: A review and call for future research. J. Account. Lit. 2024. [Google Scholar] [CrossRef]
  40. Afolabi, H.; Ram, R.; Rimmel, G. Harmonization of Sustainability Reporting Regulation: Analysis of a Contested Arena. Sustainability 2022, 14, 5517. [Google Scholar] [CrossRef]
  41. Shalhoob, H. ESG Disclosure and Financial Performance: Survey Evidence from Accounting and Islamic Finance. Sustainability 2025, 17, 1582. [Google Scholar] [CrossRef]
  42. Pavlovic, M.; Gligorić, Č.; Pavlović, D. Auditing Greenwashing: The Role of Auditors in Detecting Misleading Sustainability Claims. Econ. Manag. Rev. 2025, 15, 7–18. [Google Scholar] [CrossRef]
  43. Vaihekoski, M.; Yahya, H. Environmental, Social, and Governance (ESG) and Firm Valuation: The Moderating Role of Audit Quality. J. Risk Financ. Manag. 2025, 18, 148. [Google Scholar] [CrossRef]
  44. Kotsantonis, S.; Serafeim, G. Four Things No One Will Tell You About ESG Data. J. Appl. Corp. Financ. 2019, 31, 50–58. [Google Scholar] [CrossRef]
  45. Field, A. Discovering Statistics Using IBM SPSS Statistics, 5th ed.; SAGE Publications: London, UK, 2018; Available online: https://books.google.gr/books/about/Discovering_Statistics_Using_IBM_SPSS_St.html?id=JIrutAEACAAJ&redir_esc=y (accessed on 19 September 2025).
  46. Barbara, G. ; Tabachnick; Fidell, L.S. Using Multivariate Statistics, 7th ed.; Pearson: London, UK, 2019; Available online: https://www.pearson.com/en-us/subject-catalog/p/using-multivariate-statistics/P200000003097/9780137526543?srsltid=AfmBOoqZnCcoiRoYoA6ilP4zaSm3k5t22qlfR6eWwSOSgDDYQeL5bGgc (accessed on 19 September 2025).
  47. Lorenzo-Seva, U.; Ferrando, P.J. Determining Sample Size Requirements in EFA Solutions: A Simple Empirical Proposal. Multivar. Behav. Res. 2024, 59, 899–912. [Google Scholar] [CrossRef]
  48. Williams, B.; Onsman, T.B.A.; Onsman, A.; Brown, T.; Onsman, P.A.; Brown, P.T. Exploratory factor analysis: A five-step guide for novices. J. Emerg. Prim. Health Care 2012, 8, 1–13. [Google Scholar] [CrossRef]
  49. Howard, M.C. A systematic literature review of exploratory factor analyses in management. J. Bus. Res. 2023, 164, 113969. [Google Scholar] [CrossRef]
  50. Ali, S.C.; Hadi, S. Regression Analysis by Example Using R, 6th ed.; Willey: Hoboken, NJ, USA, 2023. [Google Scholar]
  51. Tavakol, M.; Dennick, R. Making sense of Cronbach’s alpha. Int. J. Med. Educ. 2011, 2, 53–55. [Google Scholar] [CrossRef]
  52. Ratner, B. Statistical and Machine-Learning Data Mining: Techniques for Better Predictive Modeling and Analysis of Big Data, 3rd ed.; Taylor & Francis Group, LLC: Boca Raton, FL, USA, 2017; Available online: https://www.routledge.com/Statistical-and-Machine-Learning-Data-Mining-Techniques-for-Better-Predictive-Modeling-and-Analysis-of-Big-Data-Third-Edition/Ratner/p/book/9780367573607 (accessed on 14 August 2025).
  53. Lakens, D. Equivalence Tests: A Practical Primer for t-Tests, Correlations, and Meta-Analyses. Soc. Psychol. Pers. Sci. 2016, 8, 355–368. [Google Scholar] [CrossRef] [PubMed]
  54. Armstrong, R.A. When to use the B onferroni correction. Ophthalmic Physiol. Opt. 2014, 34, 502–508. [Google Scholar] [CrossRef]
  55. Hadad, S.; Aharony, N. Open Access Advantages as a Function of the Discipline: Mixed-methods Study. J. Acad. Libr. 2023, 49, 102746. [Google Scholar] [CrossRef]
  56. Aggarwal, C.C. Data Mining: The Textbook; Springer International Publishing: Cham, Switzerland, 2015. [Google Scholar] [CrossRef]
  57. Kassambara, A. Multivariate Analysis I Practical Guide to Cluster Analysis in R Unsupervised Machine Learning. 2017. Available online: http://www.sthda.com (accessed on 20 August 2025).
  58. Hennig, C. Cluster-wise assessment of cluster stability. Comput. Stat. Data Anal. 2007, 52, 258–271. [Google Scholar] [CrossRef]
  59. Kaiser, H.F. An Index of Factorial Simplicity. Psychometrika 1974, 39, 31–36. [Google Scholar] [CrossRef]
  60. Ketchen, D.J., Jr.; Shook, C.L. The application of cluster analysis in strategic management research: An analysis and critique. Strateg. Manag. J. 1996, 17, 441–458. [Google Scholar] [CrossRef]
  61. Zagenczyk, T.J.; Purvis, R.L.; Cruz, K.S.; Thoroughgood, C.N.; Sawyer, K.B. Context and social exchange: Perceived ethical climate strengthens the relationships between perceived organizational support and organizational identification and commitment. Int. J. Hum. Resour. Manag. 2020, 32, 4752–4771. [Google Scholar] [CrossRef]
  62. Sulaiman, N.A.; Muhamad, R.; Abdullah, M. Regulating the Audit Quality and Deprofessionalisation of the Auditing Profession. Asian J. Account. Perspect. 2014, 7, 1–12. [Google Scholar] [CrossRef]
  63. Pasiouras, F.; Gaganis, C.; Zopounidis, C. Multicriteria decision support methodologies for auditing decisions: The case of qualified audit reports in the UK. Eur. J. Oper. Res. 2007, 180, 1317–1330. [Google Scholar] [CrossRef]
Figure 1. Hierarchical clustering dendrogram (Ward’s method) based on a random subsample (n = 60). Distinct colors represent the final three-cluster solution. Note: The dendrogram displays a hierarchical clustering (Ward’s method) on a representative random subsample (n = 60) of the full dataset (n = 262), with a clear 3-cluster solution. Colors indicate distinct cluster memberships.
Figure 1. Hierarchical clustering dendrogram (Ward’s method) based on a random subsample (n = 60). Distinct colors represent the final three-cluster solution. Note: The dendrogram displays a hierarchical clustering (Ward’s method) on a representative random subsample (n = 60) of the full dataset (n = 262), with a clear 3-cluster solution. Colors indicate distinct cluster memberships.
Sustainability 17 09421 g001
Figure 2. Radar chart illustrating the profiles of the three identified clusters across the seven composite indices. Note: Each line shows the mean scores of one clusters across the seven composite indices. The values are shown at each axis point. Cluster 3 (“High Performance”) scored high on all indices, Cluster 1 (“Satisfied but with Margins”) had moderately high scores, and Cluster 2 (“Low Perception”) had markedly lower scores across all the indices.
Figure 2. Radar chart illustrating the profiles of the three identified clusters across the seven composite indices. Note: Each line shows the mean scores of one clusters across the seven composite indices. The values are shown at each axis point. Cluster 3 (“High Performance”) scored high on all indices, Cluster 1 (“Satisfied but with Margins”) had moderately high scores, and Cluster 2 (“Low Perception”) had markedly lower scores across all the indices.
Sustainability 17 09421 g002
Figure 3. Elbow plot indicating within-cluster sum of squares (WSS) across different values of k.
Figure 3. Elbow plot indicating within-cluster sum of squares (WSS) across different values of k.
Sustainability 17 09421 g003
Table 1. Summary of Key Psychometric and Statistical Validity Indicators.
Table 1. Summary of Key Psychometric and Statistical Validity Indicators.
AnalysisIndex/TestValueInterpretation
EFAKMO0.872Adequate
Bartlett’sSig.<0.001Sufficient correlation
PCAEigenvalue>1One component extracted
Cluster (K-means)N per clusterC2 = 5Requires documentation
ReliabilityCronbach’s α>0.80High reliability
Table 2. Factor Structure of Audit Quality Dimensions.
Table 2. Factor Structure of Audit Quality Dimensions.
Thematic SectionItems with Strong
Loadings
Interpretation
1Social Impact & ESG33, 34, 35, 37, 40, 41, 42, 43, 44Integrating social indicators and ESG issues into the audit process
2Staff Supervision & Support20, 21, 22, 25Capability and role of managers in supporting teams
3Workload & Flexibility29, 31, 32Balance between workload, quality, and time availability
4Auditor Experience6, 7, 8The role of experience and guidance in quality
5Education & Training10, 11, 12Professional training, ethics, and inspections
6Independence & Conflicts of Interest18Single-item indicator; excluded from EFA factor solution and reliability
7Quality Management26Single-item indicator; excluded from EFA factor solution and reliability
8Inspections & Improvement Culture14, 15, 16Short multi-item scale retained; α ≈ 0.66—interpret with caution
9Human Rights & Training36, 38Short 2-item scale retained; α ≈ 0.656—interpret with caution
10Focus on Quality19Single-item indicator; excluded from EFA factor solution and reliability
Table 3. Internal Consistency Reliability of Identified Factors.
Table 3. Internal Consistency Reliability of Identified Factors.
FactorCronbach’s AlphaEvaluationComment
1. Social Impact & ESG (Q33–Q44)0.901ExcellentVery high consistency—no item removal needed
2. Staff Supervision & Support (Q20–Q25)0.797GoodMarginally below 0.8, but acceptable
3. Workload & Flexibility (Q29–Q32)0.704GoodLower, but still acceptable
4. Auditor Experience (Q6–Q8)0.774GoodStable reliability for 3 items
5. Education & Training (Q10–Q12)0.759GoodNo problem arises
6. Inspections & Improvement Culture (Q14–Q16)0.660MarginalShort scale—interpret with caution
7. Human Rights & Training (Q36, Q38)0.656MarginalTwo-item scale—expected lower α (Single-item constructs such as Independence/Conflicts, Quality Management, and Focus on Quality were not eligible for α.)
Consistency note: For two- and three-element scales, lower α values are expected (see Tavakol & Dennick, 2011) [51].
Table 4. Calculation of Composite Indicators Based on Factor-Derived Items.
Table 4. Calculation of Composite Indicators Based on Factor-Derived Items.
FactorIndex NameItemsType
Social Impact & ESGESG_MeanQ33, Q34, Q35, Q37, Q40, Q41, Q42, Q43, Q44(…)/9
Staff Supervision & SupportSupervision_MeanQ20, Q21, Q22, Q25(…)/4
Workload & FlexibilityWorkload_MeanQ29, Q31, Q32(…)/3
Auditor ExperienceExperience_MeanQ6, Q7, Q8(…)/3
Education & TrainingTraining_MeanQ10, Q11, Q12(…)/3
Inspections & Improvement CultureInspection_MeanQ14, Q15, Q16(…)/3
Human Rights & TrainingHR_MeanQ36, Q38(…)/2
Table 5. Correlation Matrix of Composite Indices.
Table 5. Correlation Matrix of Composite Indices.
IndexStrong Correlations (r > 0.5)Moderate Correlations (r = 0.3–0.5)Low Correlations
(r < 0.3)
ESG_MEANWorkload (0.608), HR (0.553),
Supervision (0.531)
Experience (0.388), Training (0.445), Inspection (0.435)
Supervision_MeanWorkload (0.546), Inspection (0.505)Experience (0.410), Training (0.428), HR (0.417)
Workload_MeanESG (0.608), Supervision (0.546)HR (0.480), Experience (0.365), Training (0.386), Inspection (0.360)
Experience_MeanSupervision (0.410), Training (0.413), Inspection (0.406)HR (0.235)
Training_MeanExperience (0.413), Supervision (0.428), Inspection (0.411),
HR (0.348), Workload (0.386), ESG (0.445)
Inspection_MeanSupervision (0.505)ESG (0.435), Experience (0.406), Training (0.411), HR (0.393),
Workload (0.360)
HR_MeanESG (0.553)Workload (0.480), Supervision (0.417), Training (0.348), Inspection (0.393)Experience (0.235)
Table 6. Summary of One-Way ANOVA and Post-Hoc Tukey HSD Results.
Table 6. Summary of One-Way ANOVA and Post-Hoc Tukey HSD Results.
Dependent VariableIndependent VariableStatistically Significant (p < 0.05)?Tukey HSD—Groups with Difference
ESG_MeanGenderYesGroup 3 ≠ Group 1 & Group 2
Years of experience in auditingYesGroup 2 ≠ Group 3
Supervision_MeanPosition in the companyYesGroup 4 ≠ Group 1
Workload_MeanGenderYesGroup 3 ≠ Group 1 & Group 2
Experience_MeanAgeYesTrend: older age → ↑ experience
Position in the companyYesGroup 3, 4 > Group 1
Education levelYesGroup 3 > Group 2 > Group 1
Training_MeanEducation levelYesPhD > Bachelor (p = 0.003); trend Master > Bachelor (p ≈ 0.052)
Inspection_MeanNo variableNo
HR_MeanNo variableNo
Table 7. Mean Scores of Composite Indices Across the Three Identified Clusters.
Table 7. Mean Scores of Composite Indices Across the Three Identified Clusters.
IndexCluster 1Cluster 2Cluster 3Cluster Size
ESG_Mean4.233.004.62Cluster 1: 102 people
Supervision_Mean4.123.104.62Cluster 2: 5 people (very small group)
Workload_Mean4.232.534.63Cluster 3: 155 people
Experience_Mean4.263.474.65
Training_Mean4.203.204.65
Inspection_Mean4.013.804.67
HR_Mean4.193.004.64
Note: All values are means ±SD for each composite index. F-values and p-values from one-way ANOVA comparing clusters. Each comparison is statistically significant at p < 0.001. Bootstrapped F confirmed stability of differences, e.g., mean F_boot: ESG ≈ 144.3, Supervision ≈ 121.4, Workload ≈ 82.5.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Garefalakis, A.; Sitzimis, I.; Angelaki, E.; Giannopoulos, P.; Kyriakogkonas, P. Audit Quality Indicators Across Jurisdictions: Regulatory Diversity, ESG Integration, and Implications for Public Trust. Sustainability 2025, 17, 9421. https://doi.org/10.3390/su17219421

AMA Style

Garefalakis A, Sitzimis I, Angelaki E, Giannopoulos P, Kyriakogkonas P. Audit Quality Indicators Across Jurisdictions: Regulatory Diversity, ESG Integration, and Implications for Public Trust. Sustainability. 2025; 17(21):9421. https://doi.org/10.3390/su17219421

Chicago/Turabian Style

Garefalakis, Alexandros, Ioannis Sitzimis, Erasmia Angelaki, Panagiotis Giannopoulos, and Panagiotis Kyriakogkonas. 2025. "Audit Quality Indicators Across Jurisdictions: Regulatory Diversity, ESG Integration, and Implications for Public Trust" Sustainability 17, no. 21: 9421. https://doi.org/10.3390/su17219421

APA Style

Garefalakis, A., Sitzimis, I., Angelaki, E., Giannopoulos, P., & Kyriakogkonas, P. (2025). Audit Quality Indicators Across Jurisdictions: Regulatory Diversity, ESG Integration, and Implications for Public Trust. Sustainability, 17(21), 9421. https://doi.org/10.3390/su17219421

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop