1. Introduction
Digital transformation in the public sector reshapes how states design, deliver, and govern services, with direct implications for access, equity, accountability, and inclusion. We adopt a public-sector–specific view of digital transformation as a socio-technical, data-intensive reconfiguration of processes, capabilities, and citizen–state interactions aimed at creating public value and reducing exclusion. Consistent with Saeedikiya et al. (2025) [
1], we define digital transformation as a fundamental, technology-enabled change process that reconfigures capabilities and stakeholder interactions to produce public value; we also draw on Saeedikiya et al. (2024) [
2] to foreground dynamic capabilities as enablers of such change in public-service contexts. Within this lens, our three dimensions map as follows: D1 (Breadth of the Divide) ↔ foundational access/skills capabilities; D2 (Sectoral/Specific Divide) ↔ reconfigured service processes and workflows; D3 (GenAI Gap) ↔ emergent capabilities and user–system interactions specific to Generative AI (GenAI)—namely, access, task use, and competence.
Prior research on the digital divide has predominantly examined access and skills, with sectoral studies of electronic government (e-government) highlighting persistent gaps in usage, competence, and trust. Far less is known about how GenAI-enabled interfaces and workflows (e.g., conversational form completion, eligibility guidance, support for appeals) alter these gaps—who benefits, who is left behind, and under which conditions inclusion improves or deteriorates. This paper addresses that gap by proposing and testing a three-dimensional framework for digital inclusion in e-government: Breadth of the Divide (D1), Sectoral/Specific Divide (D2), and the GenAI Gap (D3), capturing access, task use, and competence related to GenAI.
Research problem: How does the integration of GenAI in e-government affect digital inclusion when assessed jointly across D1–D3, and what measurement design makes such assessment transparent, comparable, and policy-relevant?
Research questions and hypotheses:
- RQ1.
Can a composite indicator consistently quantify D1–D3 while preserving interpretability for policy?
- RQ2.
How do D1 and D2 relate to the emergent D3 in an upper-middle-income setting?
H1. Higher GenAI access and task-specific competence are associated with narrower e-government usage and competence gaps (ceteris paribus).
H2. In the short run, insufficient GenAI literacy and uneven access widen disparities among vulnerable groups, even where basic access (D1) is high.
H3. The relationship between D3 and D2 is moderated by trust and identification frictions (e.g., electronic identification, eID), such that GenAI benefits concentrate among digitally confident users.
Note: The literature remains divided: one stream expects GenAI to lower interaction costs and barriers via assistive interfaces [
3,
4]. At the same time, another anticipates a higher competence and trust threshold, risking a new axis of inequality [
5,
6]. Some reviews explicitly frame this tension as a “double-edged sword” [
7,
8]. Our design explicitly tests both possibilities.
Contributions:
We introduce the E-Government Divide Measurement Indicator (EGDMI) with explicit indicator lists, sources, units, transformations, and uniform normalization across D1–D3.
We replace ad hoc weights with a documented weighting strategy and provide sensitivity and uncertainty analyses.
We formally position D3 (GenAI Gap) as exploratory, given current data constraints, outline direct measures (access, task use, competence), and avoid its aggregation with D1–D2 where data do not warrant summation.
We deliver actionable implications by mapping inclusion barriers to concrete e-government workflows (identity proofing, eligibility assessment, form completion, and appeals). These contributions align with the journal’s scope on AI methods and impacts in public-service contexts, speaking directly to AI for digital government and governance.
Empirical setting and originality: Using official statistics and qualitative evidence for Serbia, we demonstrate how EGDMI reveals multi-layered exclusion in e-government and how GenAI may both lower interaction costs and raise competence/ethics thresholds. The proposed architecture, transparency checks, and reporting templates aim to standardize monitoring and enable cross-country learning.
Structure of the paper: Section 2 develops the theoretical foundation and defines D1–D3.
Section 3 details the index architecture (indicators, sources, transformations, normalization, weighting, and sensitivity).
Section 4 presents the methodology and data.
Section 5 reports results for Serbia and a robustness suite.
Section 6 discusses findings, limitations, and implications (ethical, practical, and theoretical).
Section 7 concludes with policy pathways and a roadmap for future data collection on the GenAI Gap.
Principal conclusions (preview): Our preliminary evidence indicates layered exclusion in e-government; D3 (GenAI Gap) is presently best reported separately from D1–D2 due to data immaturity, with immediate implications for inclusive design, literacy programs, and staged policy adoption.
2. Theoretical Background
2.1. Digital Transformation in Public Services: A Socio-Technical Lens
Digital transformation (DT) in the public sector is not a technology roll-out but a socio-technical reconfiguration of processes, capabilities, and actor relationships that aims to create public value and reduce exclusion. Consistent with Saeedikiya et al. (2025) [
1], we view DT as a technology-enabled, capability-driven change process that reshapes organizational routines and stakeholder interactions; Saeedikiya et al. (2024) [
2] emphasize dynamic capabilities (sensing, seizing, reconfiguring) as enablers of sustained transformation in service settings. Applying this lens to public administration implies that inclusion outcomes depend jointly on (i) foundational capabilities (access, basic skills, affordability), (ii) service-specific process design (identification, eligibility, form completion, appeals, redress), and (iii) emergent AI-mediated interaction capabilities (the ability of users—and institutions—to work productively and safely with GenAI interfaces).
This framing establishes clear theoretical anchors for the three dimensions we study: D1 (Breadth) reflects foundational capabilities; D2 (Sectoral/Specific) captures service-process reconfiguration; and D3 (GenAI Gap) captures emergent interaction capabilities at the human–AI boundary.
2.2. Revisiting the Digital Divide: From Access to Outcomes
The digital divide literature has evolved from a first-level focus on access (infrastructure, devices) to second-level skills and uses, and third-level outcomes (who benefits, how much, and in what ways). This evolution from first-level (access) to second-level (skills/uses) and third-level (outcomes) divides is well documented in the digital divide literature [
9]. In public services, these divides manifest as persistent gaps in usage, competence, and trust, even where connectivity is high. Recent e-government studies document that channel usage (online vs. phone/front-desk) and competence (e.g., eID use, completing multi-step transactions) remain uneven, with subjective non-use (perceived complexity, low need, mistrust) acting as an additional barrier. This literature motivates distinguishing baseline conditions (D1) from sector-specific adoption frictions (D2) that arise in concrete service journeys.
2.3. The Sector-Specific (E-Government) Divide
E-government embeds the general divide in process-intensive workflows (such as identity proofing, eligibility checks, form completion, payment, tracking, and appeals). Sectoral context matters: health, education, taxation, or licensing each imposes its own cognitive and procedural burdens, with varying requirements for identification, documentation, and temporal coordination. The literature consistently shows that even among the online population, a sizable share does not transact digitally with the government—due to skill gaps, low perceived utility, or trust and identification frictions (e.g., reluctance to use eID). This supports modeling a distinct D2 that (a) stratifies the online population into users vs. non-users of e-government, and (b) distinguishes competence (e.g., eID usage) from subjective non-use (e.g., “no need”).
2.4. Generative AI and the Emergence of a New Divide
Generative AI (GenAI) introduces conversational, assistance-oriented interfaces that can lower interaction costs (natural-language guidance, summarization, translation, form pre-fill), but also raise capability thresholds (prompting skill, verification literacy, understanding of data ethics, and provenance). The literature remains divided: some studies expect GenAI to lower interaction costs via assistive interfaces, while others anticipate higher competence and trust thresholds, potentially widening inequalities [
3,
4,
6]. Two competing mechanisms therefore co-exist:
Barrier-reducing mechanism: GenAI can simplify navigation, comprehension, and completion of forms; support eligibility reasoning; and provide multilingual, accessibility-friendly assistance (audio/text, plain-language explanations).
Barrier-raising mechanism: Effective and safe use requires functional GenAI literacy (task framing, checking model outputs, handling identity data responsibly), domain knowledge (to validate content), and ethical competence (privacy, bias, attribution). Institutions must also implement guardrails.
Given these countervailing forces—and the current scarcity of direct measures of GenAI access, use, and competence in official statistics—D3 (GenAI Gap) should presently be treated as exploratory, reported and analyzed separately from D1–D2 unless commensurate measures are available. This stance aligns with the precautionary framing in recent AI governance and e-government literatures and minimizes over-claiming in composite aggregation [
10,
11,
12,
13].
2.5. Mapping D1–D3 to Digital Transformation Theory
Grounding the three dimensions in DT theory clarifies their distinct roles and testable mechanisms:
D1—Breadth of the Divide (Foundational capabilities).
DT mapping: Sensing infrastructural gaps and seizing access/affordability interventions (devices, connectivity, training).
Mechanism: Without baseline access and basic digital skills, subsequent service reforms will not translate into inclusion.
D2—Sectoral/Specific Divide (Service-process design).
DT mapping: Reconfiguring processes and user journeys (identity proofing, eligibility, submissions, appeals), reducing procedural complexity and transaction costs.
Mechanism: Even at high D1, poor service design (e.g., eID hurdles, opaque instructions) generates usage and competence gaps.
D3—GenAI Gap (Emergent interaction capabilities).
DT mapping: New dynamic capabilities at the human–AI boundary—on both the citizen side (GenAI literacy, verification, ethics) and the administration side (safe orchestration, auditability).
Mechanism: GenAI can reduce cognitive/linguistic barriers but increase verification and ethical burdens, potentially widening disparities if literacy programs and safeguards lag.
Building on Saeedikiya et al. (2024; 2025) [
1,
2], the three dimensions are mapped to the core digital transformation capabilities to clarify their conceptual boundaries and mechanisms (see
Table 1).
Implementation note: Given current data limitations, we classify D3 as exploratory and recommend transparent reporting (indicator lists, sources, units, transformations) and uniform normalization across dimensions, while postponing aggregation of D3 with D1–D2 until direct, stable measures of GenAI access/use/competence become available.
2.6. Synthesis and Implications for Measurement (EGDMI)
The theoretical synthesis suggests three design imperatives for a measurement architecture:
Dimensional clarity and commensurability
Report explicit indicator lists with sources, units, and transformations; apply uniform normalization across D1–D3 to support interpretability and comparability.
Transparent weighting with robustness checks
Replace ad hoc weights with a documented scheme (expert elicitation, MCDA, or data-driven alternatives) and provide sensitivity/uncertainty analysis (e.g., ±10% weight shifts; bootstrapped confidence bands).
Precautionary treatment of D3
Until direct measures are available, label D3 as exploratory and avoid summing with D1–D2; provide scenario- or dashboard-based reporting for D3 (access, task use, competence, verification, and ethics sub-domains).
This theoretical consolidation directly informs the design of the E-Government Divide Measurement Indicator (EGDMI) used in this study. It motivates our empirical choices in the next section (data, normalization, weighting, robustness).
3. Methodology
This study adopts a mixed-methods research design to examine the E-Government Divide in the context of Generative Artificial Intelligence (GenAI). The methodology integrates quantitative and qualitative components to ensure comprehensive measurement, interpretation, and validation of the E-Government Divide Measurement Indicator (EGDMI). The approach aligns with the best international practices in composite index development, including the European Commission’s DESI [
14], the UN E-Government Survey (UN-EGDI) [
15], the ITU IDI framework [
16], and the OECD digital measurement standards [
17]. It is adapted to the specific context of GenAI-enabled public services.
3.1. Basic Indicator Setting
Building on the conceptual framework presented in
Section 2 and its mapping to Digital Transformation theory (see
Table 1), the EGDMI serves as a structured measurement architecture composed of three dimensions:
D1—Breadth of the Divide (Basic Digital Divide): foundational access, affordability, and basic digital skills enabling digital participation.
D2—Sectoral/Specific E-Government Divide: actual use and competence within e-government workflows (eID, identity proofing, eligibility, submissions, appeals) and reasons for non-use.
D3—GenAI Gap (Exploratory): emerging disparities in access to, task use of, and competence with GenAI tools in public service contexts, including verification and ethical literacy.
EGDMI is operationalized into two composite sub-indices (D1 and D2), while D3 is computed and reported separately as an exploratory layer until direct, stable measures of GenAI access and competence become available. Qualitative insights from focus groups inform indicator selection, interpretation, and policy implications but are not numerically aggregated with quantitative scores. The overall research design follows established mixed-methods guidance to integrate quantitative measurement with qualitative interpretation [
18].
3.2. Algorithms and Computation of the EGDMI
The algorithm for constructing the EGDMI follows established guidelines for composite indicators (OECD/JRC Handbook) and ensures consistency across dimensions through uniform normalization, expert-derived weights, and robustness testing [
12].
Step 1—Indicator definition and mapping:
Identify candidate indicators
for each dimension
{1, 2, 3}, mapped to conceptual constructs in
Table 1 with directionality adjusted so that higher values denote greater inclusion.
Step 2—Data screening:
Check coverage, timeliness, and reliability; exclude indicators with insufficient quality or missing data.
Step 3—Pre-processing:
Standardize units, apply necessary transformations (e.g., reverse coding for negative indicators), and document all processing steps [
12].
Step 4—Normalization (uniform):
All indicators are scaled to [0,1] using min–max normalization:
(Uniform application per OECD/JRC guidance) [
12].
Step 5—Weighting for D1 and D2 (expert elicitation):
Weights
are derived via expert elicitation from five specialists in digital governance, statistics, and AI policy. Experts rated the relevance of each indicator on a 1–5 scale; the average ratings were normalized so that
per dimension. A Delphi-style consensus process enhances transparency and avoids ad hoc choices [
19].
Step 6—Computation of dimension scores:
For
and
, sub-indices are calculated as weighted sums of normalized indicators:
Each dimension is then rescaled to a 0–100 range for interpretability [
12].
Step 7—GenAI Gap (D3):
D3 is treated as an exploratory index capturing preliminary measures of GenAI access, usage, and competence. Its indicators are normalized and summarized descriptively, but not aggregated with D1 and D2. Interpretation draws on emerging AI literacy guidance to contextualize capabilities at the human–AI boundary [
20].
Step 8—Robustness and uncertainty analysis:
To test stability, local sensitivity checks vary each
(with renormalization) and recompute D1 and D2. Absence of rank reversals or material score shifts indicates robustness. The approach follows good practice in global sensitivity analysis [
21].
Step 9—Validity and reliability assessment:
Three tests were conducted:
Content validity: expert review of indicator coverage and conceptual relevance [
12].
Construct validity: Spearman correlations among D1 and D2 sub-indices to confirm expected relationships [
12].
Reliability: Cronbach’s alpha for multi-item constructs, with psychometric benchmarks from measurement literature [
22,
23].
Given its exploratory status, D3 was excluded from reliability testing.
Step 10—User Segmentation (Cluster Analysis): To move beyond generalized findings and identify distinct citizen profiles based on their inclusion levels, a cluster analysis was performed. We employed a K-means clustering algorithm, a non-hierarchical partitioning method, using the ten normalized sub-dimension scores (D11–D33, as presented in
Table 2) as input variables. The optimal number of clusters was determined using the ‘Elbow’ method, which identifies the point of diminishing returns in the within-cluster sum of squares (WCSS). This four-cluster solution was further validated for its robustness by assessing its Silhouette coefficient, which confirmed good cluster cohesion and separation. The final clusters, presented in
Section 4.5, were confirmed to be statistically distinct.
3.3. Data Sources and Qualitative Component
The foundation for the quantitative analysis is data from the official survey “Usage of Information and Communication Technologies in the Republic of Serbia, 2023” (SORS 2023), conducted by the Statistical Office of the Republic of Serbia (SORS). This research was conducted as a CATI (telephone) survey on a two-phase, stratified sample comprising 2800 households and 2800 individuals. The target population included all individuals aged 16 to 74 and households with at least one member in that age range. The reference period for household and individual data was the three months preceding the interview.
This dataset provided the primary indicators for the following:
D1 (Breadth of the Divide): data on device access, connectivity, affordability, and basic digital skills.
D2 (E-Government Divide): behavioral data on service use, barriers, and eID (electronic identification) adoption.
Data for D3 (GenAI Gap), which is exploratory in nature, were derived from pilot CATI and online surveys on GenAI use, providing proxy measures for access, task use, and verification literacy.
To complement the quantitative findings, test the proposed framework, and clarify methodological approaches, qualitative research was conducted from October to November 2023. This research involved two focus groups (N = 8 and N = 9) with participants representing diverse demographic profiles (age, gender, education level, employment status, and living environment). Each session lasted 90 min.
3.4. Computation and Reporting
Composite scores for D1 and D2 were aggregated as weighted averages of normalized indicators and rescaled to a 0–100 range [
12]. D3 is presented side-by-side as a separate layer to highlight emerging AI inequalities without inflating composite values [
20]. This presentation supports policy relevance while maintaining methodological transparency and comparability over time.
3.5. Process Overview
The overall methodological process, summarized in
Figure 1, followed a structured seven-step sequence. This approach integrates conceptualization (mapping D1–D3 to DT theory) with empirical validation and transparent reporting.
The process was guided by established international best practices for measuring digital development [
14,
15,
16,
17] and relies on the OECD/JRC Handbook [
12] for index construction, normalization, and expert-based weighting [
19]. The validation steps align with standards for reliability [
22,
23], and the sensitivity analysis [
21], while the exploratory treatment of D3 is based on AI guidelines [
20].
The following section further clarifies the key terminology used within this methodological framework.
3.6. Note on Terminology
Throughout the paper,
EGDMI refers to the overall measurement architecture. In the current release, the index produces two composite scores (D1 and D2) and an exploratory GenAI layer (D3), reflecting data availability and theoretical maturity. Future iterations will enable complete aggregation once reliable GenAI measures become available, ensuring continuity with global digital inclusion metrics [
12,
20].
4. Results
This section presents the empirical findings of the E-Government Divide Measurement Indicator (EGDMI). Results are structured as follows:
- (i)
Descriptive findings across sub-dimensions of D1–D2;
- (ii)
Composite index scores for D1 and D2;
- (iii)
Exploratory results for D3 (GenAI Gap);
- (iv)
Validation and robustness checks;
- (v)
User segmentation based on cluster analysis.
4.1. Descriptive Results Across D1–D2
Table 2 provides an overview of scores for all ten sub-dimensions. The results indicate relatively strong foundational digital conditions (D1), contrasted with significantly lower engagement with e-government services (D2), and early-stage, uneven adoption of GenAI tools (D3).
Policy implication: Despite strong digital access, e-government engagement remains low, suggesting that the primary barriers are not connectivity but service design, motivation, and trust.
To better illustrate the distribution of performance across the ten sub-dimensions,
Figure 2 visualizes the relative strengths and weaknesses across D1, D2, and D3.
As shown in
Figure 1, the scores for foundational conditions (D1) are high, particularly in
Internet access (D12 = 85.6) and
Access to Technology (D11 = 73.4). Conversely, the Support & Training score
(D14 = 10.7) is critically low. The E-Government (D2) dimensions confirm this gap: while barriers like
Subjective Non-use (D2.4 = 53.7) are high, actual uptake, such as
E-Government Users (D2.1 = 34.2) and
eID Competence (D2.3 = 29.0), is very low. This confirms that the core bottleneck is no longer access, but skills, support, and motivation.
4.2. Composite Index Results for D1 and D2
Composite index scores show a notable divergence between general digital readiness (D1) and the actual adoption of e-government services (D2). D1 reaches a relatively high level (73.6), while D2 remains critically low (19.9), indicating a significant conversion gap between capability and usage (
Table 3).
Policy implication: The EGDMI core confirms a structural shift: the digital divide has moved from “basic access” toward service interaction, usability, and trust barriers. Policy should now focus on human-centered service redesign, rather than infrastructure alone.
To compare overall digital readiness with the actual adoption of e-government services,
Figure 3 presents the composite scores for D1 and D2.
Figure 3 clearly shows a significant gap between general digital readiness (D1) and e-government usage (D2). Despite strong foundational conditions (D1 = 73.6), user uptake of e-government remains critically low (D2 = 19.9). This confirms that digital inclusion efforts must now shift from infrastructure and skills toward improving service usability, perceived value, and trust.
4.3. GenAI Gap (D3)—Exploratory Results
D3 results are presented separately due to their exploratory status and the limited availability of standardized indicators. The GenAI Gap average score is
43.6, reflecting early adoption patterns and high inequality across socio-demographic groups (
Table 4).
Policy implication: Without early intervention, GenAI may widen existing divides, reinforcing inequalities in access, comprehension, and benefit from digital public services. Early investment in GenAI literacy and verification skills is essential.
To provide deeper insights into the exploratory GenAI Gap dimension,
Figure 4 displays the scores for its three sub-dimensions.
As illustrated in
Figure 3, GenAI adoption is at an early stage, with modest levels of access (
D31 = 47.8) and limited task diversity (
D32 = 39.4). Competence and verification literacy (
D33 = 43.6) remain underdeveloped, indicating that many users lack the skills to assess AI-generated content critically. These results highlight the importance of early investments in GenAI literacy to prevent the deepening of existing disparities.
4.4. Validity, Reliability, and Robustness Checks
The results of validity and robustness tests are provided in
Table 5. All metrics meet accepted thresholds, indicating that the core index (D1 and D2) is methodologically sound and stable for policy use.
Interpretation: EGDMI (core) demonstrates robust internal structure. D3 remains exploratory and should not yet be aggregated with D1 and D2 until direct measures of GenAI are standardized.
4.5. User Segmentation—Cluster Analysis
Cluster analysis was used to segment the population into digital & GenAI inclusion profiles. Four clusters emerged, enabling targeted intervention design (
Table 6).
Policy implication: Interventions must be tailored:
C1 → inclusion & access;
C2 → motivation & awareness;
C3 → trust and service redesign;
C4 → co-creation of AI-assisted public services.
To support targeted policy and intervention design,
Figure 5 maps four citizen clusters across the three main dimensions (D1–D3).
Figure 5 reveals four distinct digital inclusion profiles, with substantial variation in readiness and in the benefits derived from digital and GenAI-assisted services. While “GenAI-Augmented Citizens” (C4) are positioned to benefit most, “Digitally Excluded” (C1) and “Basic Digital Users” (C2) remain at risk of being left behind. This segmentation highlights the need for differentiated policy approaches tailored to each cluster rather than one-size-fits-all interventions.
4.6. Limitations of Interpretation
These results should be interpreted with caution due to the exploratory nature of GenAI indicators (D3) and limited longitudinal data. While D1 and D2 are statistically reliable, D3 requires further empirical development and should be monitored over time.
6. Conclusions
This study examined the evolving nature of the digital divide in the era of Generative AI (GenAI), with a focus on the e-government context. By developing and applying the E-Government Divide Measurement Indicator (EGDMI), the research provided a comprehensive assessment across three dimensions: the Basic Digital Divide (D1), the E-Government Divide (D2), and the GenAI Gap (D3). The findings show that while digital access and basic skills are relatively strong, e-government adoption remains low, and GenAI-related capabilities are still emerging and unevenly distributed. Based on the evidence, the initial hypothesis—that the introduction of GenAI may amplify the digital divide—was partially confirmed: GenAI holds significant potential to support inclusion, but also carries a clear risk of widening inequalities if safeguards and literacy efforts are not prioritized.
6.1. Key Contributions
This study offers three key contributions. First, it updates the conceptual understanding of the digital divide by expanding it beyond access and skills to include service interaction and GenAI-assisted capabilities. Second, it develops the EGDMI as a multidimensional measurement framework, enabling more apparent distinctions among foundational readiness (D1), behavioral use of public digital services (D2), and emerging AI-related capabilities (D3). Third, it provides empirical evidence from Serbia that illustrates a structural shift—from digital inequality based on access to one based on meaningful use, agency, and algorithmic literacy.
6.2. Practical and Policy Implications
The results highlight a need for a strategic shift in digital inclusion policies. Rather than focusing primarily on infrastructure and basic digital skills, governments should prioritize human-centered redesign of public services to strengthen usability, trust, and perceived value. GenAI should be introduced gradually, supported by targeted GenAI literacy programs that build citizens’ ability to assess and use AI-generated outputs safely and critically. Furthermore, differentiated policy approaches are required: initiatives should be tailored to distinct user groups—from digitally excluded citizens to advanced GenAI users—to prevent further stratification and unlock inclusive public value.
6.3. Limitations and Directions for Future Research
This research has several limitations that should be considered when interpreting the findings. The GenAI dimension (D3) remains exploratory due to the limited availability of standardized indicators and behavioral data. The study is cross-sectional, capturing a snapshot in time; future research should adopt a longitudinal approach to observe changes as GenAI integration in public services evolves. Expanding the model to other sectors (e.g., health, education, justice) and comparing results across countries would further validate the EGDMI and support the development of international benchmarks. Finally, future work should integrate qualitative user research to better understand motivational, cultural, and behavioral factors shaping meaningful digital participation.