Previous Article in Journal
Radiation Safety in Prostatic Artery Embolization: A Review of Current Evidence and Best Practices
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Positioning Fractal Dimension and Lacunarity in the IBSI Feature Space: Simulation With and Without Wavelets

Department of Mathematics & Statistics, East Tennessee State University (ETSU), Johnson City, TN 37614, USA
*
Author to whom correspondence should be addressed.
Radiation 2025, 5(4), 32; https://doi.org/10.3390/radiation5040032
Submission received: 29 September 2025 / Revised: 23 October 2025 / Accepted: 24 October 2025 / Published: 3 November 2025
(This article belongs to the Section Radiation in Medical Imaging)

Simple Summary

This study examines whether fractal dimension and lacunarity, two texture measures that describe image complexity, provide additional information beyond standard radiomic features. Using simulated images, we compared these measures to Image Biomarker Standardisation Initiative (IBSI) descriptors with and without wavelet filtering. We found that fractal dimension captures fine texture variation, while lacunarity describes larger structural patterns. Together they provide complementary information, suggesting that both can enhance radiomic analyses focused on multiscale heterogeneity.

Abstract

Fractal dimension (Frac) and lacunarity (Lac) are frequently proposed as biomarkers of multiscale image complexity, but their incremental value over standardized radiomics remains uncertain. We position both measures within the Image Biomarker Standardisation Initiative (IBSI) feature space by running a fully reproducible comparison in two settings. In a baseline experiment, we analyze N = 1000 simulated 64 × 64 textured ROIs discretized to N g = 64 , computing 92 IBSI descriptors together with Frac (box counting) and Lac (gliding box), for 94 features per ROI. In a wavelet-augmented experiment, we analyze N = 1000 ROIs and add level-1 wavelet descriptors by recomputing first-order and GLCM features in each sub-band (LL, LH, HL, and HH), contributing 4 × ( 19 + 19 ) = 152 additional features and yielding 246 features per ROI. Feature similarity is summarized by a consensus score that averages z-scored absolute Pearson and Spearman correlations, distance correlation, maximal information coefficient, and cosine similarity, and is visualized with clustered heatmaps, dendrograms, sparse networks, PCA loadings, and UMAP and t-SNE embeddings. Across both settings a stable two-block organization emerges. Frac co-locates with contrast, difference, and short-run statistics that capture high-frequency variation; when wavelets are included, detail-band terms from LH, HL, and HH join this group. Lac co-locates with measures of large, coherent structure—GLSZM zone size, GLRLM long-run, and high-gray-level emphases—and with GLCM homogeneity and correlation; LL (approximation) wavelet features align with this block. Pairwise associations are modest in the baseline but become very strong with wavelets (for example, Frac versus GLCM difference entropy, which summarizes the randomness of gray-level differences, with | r | 0.98 ; and Lac versus GLCM inverse difference normalized (IDN), a homogeneity measure that weights small intensity differences more heavily, with | r | 0.96 ). The multimetric consensus and geometric embeddings consistently place Frac and Lac in overlapping yet separable neighborhoods, indicating related but non-duplicative information. Practically, Frac and Lac are most useful when multiscale heterogeneity is central and they add a measurable signal beyond strong IBSI baselines (with or without wavelets); otherwise, closely related variance can be absorbed by standard texture families.

1. Introduction

Radiomics converts medical images into large panels of quantitative descriptors that summarize intensity distributions, texture, shape, and heterogeneity for modeling and decision support across clinical domains [1,2,3,4]. To enable reproducibility and comparability, the Image Biomarker Standardisation Initiative (IBSI) has codified definitions and implementation choices, such as gray-level discretization, neighborhood topology, and angular or distance aggregation, for a core set of matrix-based texture families: the gray-level co-occurrence matrix (GLCM), gray-level run-length matrix (GLRLM), gray-level size-zone matrix (GLSZM), gray-level dependence matrix (GLDM; also called NGLDM), and the neighborhood gray-tone difference matrix (NGTDM). These specifications are widely adopted in open-source toolkits including PyRadiomics [4,5,6,7,8,9,10].
Alongside these standardized families, fractal descriptors, particularly fractal dimension (Frac) and lacunarity (Lac), remain attractive because they target explicitly multiscale properties: Frac quantifies boundary or texture roughness across scales, whereas Lac characterizes the heterogeneity of gaps or voids as a function of scale [11,12,13]. Applications spanning oncology and neuroimaging report encouraging, though heterogeneous, associations with grade, phenotype, and outcome [14,15,16].
These observations motivate two practical questions that matter for clinical translation. First, to what extent do Frac and Lac provide information that is complementary to IBSI-standardized texture families that already encode high-frequency contrast or large-structure emphasis? Strong inter-feature correlations and redundancy are well documented in radiomics, especially among texture families and under different discretization choices [17,18,19,20,21]. Second, how sensitive are fractal estimators to preprocessing and acquisition factors known to affect radiomic stability, including voxel size, gray-level discretization, reconstruction, and segmentation variability [4,22,23,24]? Addressing redundancy and sensitivity is essential to avoid optimistic estimates of model performance, clarify interpretation, and support reproducible pipelines [1,2].
In this work we (i) synthesize the relevant literature on fractal and lacunarity analysis in medical imaging; (ii) provide explicit mathematical definitions for all features considered to ensure unambiguous reproducibility; and (iii) run a two-part controlled simulation in which textured, heterogeneous ROIs are generated and Frac and Lac are compared against standardized radiomics in two configurations: first without wavelets and then with wavelets. In the baseline setting, Frac and Lac are evaluated alongside 92 IBSI features, yielding 94 features per ROI. In the wavelet-augmented setting, we recompute first-order and GLCM descriptors in each of the four wavelet sub-bands (LL, LH, HL, and HH), adding 152 wavelet features for a total of 246 per ROI. Similarities are quantified with multiple complementary measures, Pearson correlation, Spearman correlation, distance correlation, the maximal information coefficient, and cosine similarity, and summarized with heatmaps, dendrograms, correlation networks, principal component loadings, and two-dimensional embeddings based on UMAP and t-SNE. Our goal is not to advocate for or against fractal biomarkers as such, but to delineate when they add a genuinely complementary signal and when they largely recapitulate standardized descriptors derived from the spatial domain or from wavelets.

Relation to Existing Benchmarks

Our study is positioned as a benchmark comparison that complements established reproducibility and harmonization frameworks in radiomics [4,21,22]. Previous benchmark efforts have mainly quantified how acquisition variability, voxel size, reconstruction, and gray-level discretization influence standardized features across scanners and cohorts, often recommending harmonization methods such as ComBat and rigorous reporting of preprocessing choices [4,23,24]. In contrast, the present study focuses on methodological redundancy and complementarity under controlled simulation. By generating synthetic regions of interest with tunable spatial correlation and optional heterogeneity, we are able to attribute observed associations to feature definitions rather than to protocol confounders.
Within this controlled environment, we evaluate whether fractal descriptors—fractal dimension (Frac) and lacunarity (Lac)—contribute incremental information beyond standardized IBSI texture families and wavelet-based extensions. To achieve this, we keep discretization, angular aggregation, and neighborhood topology fixed according to IBSI recommendations and vary only the underlying texture process. We then quantify feature relatedness using five complementary dependence measures (Pearson, Spearman, distance correlation, maximal information coefficient, and cosine similarity), form a consensus z-score for ranking, and visualize the overall geometry using clustered heatmaps, principal component loadings, and low-dimensional embeddings obtained from UMAP and t-SNE. This multi-view evaluation extends earlier benchmarks that typically emphasize repeatability or single-metric stability by identifying where Frac and Lac reside relative to co-occurrence, run-length, size-zone, dependence, and gray-tone difference features, both with and without wavelet augmentation.
Empirically, a stable two-block organization emerges that aligns with the intended behavior of established feature families. Frac is closely associated with contrast, difference, and short-run statistics that describe high-frequency texture variation, while Lac aligns with large-structure, homogeneity, and long-run statistics. The addition of wavelet features sharpens this distinction, as detail-band components correspond more strongly with Frac, and approximation-band components correspond with Lac. Collectively, these findings indicate that fractal measures are overlapping yet non-duplicative; they follow known axes of variation already represented in IBSI features but retain interpretable multiscale summaries that can be particularly useful when heterogeneity across scales is an essential characteristic.
Finally, this simulation-grounded benchmark complements empirical clinical studies of fractal descriptors [14,15,16] by disentangling intrinsic feature associations from acquisition effects. It provides a transparent framework for understanding when fractal measures are likely to be redundant, such as in models already rich in contrast, short-run, or homogeneity descriptors, and when they can add measurable information. In this way, the study extends existing benchmarks beyond the question of whether features are stable to the more informative question of whether fractal descriptors contribute additional, non-redundant information beyond strong IBSI baselines under conditions where causal attribution is clearly defined.

2. Background and Related Work

2.1. Standardized Radiomic Features

The Image Biomarker Standardisation Initiative (IBSI) specifies preprocessing steps and feature definitions for widely used texture families, and recommends transparent reporting of discretization, neighborhood definitions, and aggregation rules [4]. In brief, the gray-level co-occurrence matrix (GLCM) summarizes how often pairs of discretized gray levels co-occur at fixed offsets and angles [5]; the gray-level run-length matrix (GLRLM) counts contiguous runs of equal gray level [6]; the gray-level size-zone matrix (GLSZM) measures the sizes of connected same-gray zones irrespective of direction [7]; the gray-level dependence matrix (GLDM, also called NGLDM) captures counts of neighboring voxels within a gray-level tolerance [8]; and the neighborhood gray-tone difference matrix (NGTDM) characterizes deviations from the local neighborhood mean [9]. These standardized definitions underpin open-source implementations such as PyRadiomics [10]. Empirically, features from these families can be highly collinear, so clustering or redundancy reduction prior to modeling is commonly recommended [17,18,19]. Repeatability and reproducibility also vary across features and pipelines, motivating harmonization (for example, ComBat), robust cross-validation, and sensitivity analyses to voxel size and discretization [21,22,23,24].

2.2. Fractal Dimension and Lacunarity

Fractal measures target complementary aspects of structure. The box-counting fractal dimension describes a global scaling law: how occupancy grows as resolution is refined, whereas lacunarity captures scale-specific heterogeneity: how unevenly occupancy is distributed at a given window size [11,12,13]. Two textures can share the same fractal dimension yet display distinct lacunarity curves, and the converse can also occur; the measures are therefore not interchangeable. In medical imaging, both have been linked to tumor grade and outcome, although reported associations can be sensitive to the definition of the region of interest, the gray-level binning scheme, and the selection of scales [14,15,16]. Recent algorithmic advances, such as efficient gliding-box lacunarity via integral images, reduce computational burden and encourage systematic comparisons with standardized radiomics [25]. Conceptually, fractal dimension tends to increase with fine-scale roughness or rapid gray-level changes, behavior that also elevates contrast-oriented descriptors derived from co-occurrence and run-length representations. By contrast, lacunarity often rises with broad, coherent zones and heterogeneous voids, echoing GLSZM large-area and GLRLM long-run emphases as well as homogeneity and inverse-difference measures in the GLCM family [5,6,7,9]. Because standardized families already approximate these tendencies, it is important to determine whether fractal descriptors contribute genuinely new information or largely re-encode signals captured by classical features. The redundancy and protocol sensitivity observed across radiomics [17,18,19,21,22,23,24] motivate the controlled evaluation undertaken here.

2.3. Clinical Applications of Radiomics and the Role of Fractal Measures

Radiomics has been applied widely for diagnosis, phenotyping, risk stratification, and outcome prediction in oncology and neurology [1,2,3]. In cancer imaging, quantitative signatures extracted from CT, MR, or PET have been associated with tumor biology and prognosis and incorporated into decision–support workflows. Beyond oncology, neuroimaging applications, such as glioma grading, stroke classification, and phenotyping of neurodegenerative disease, use similar pipelines that combine first-order, shape, and matrix-based texture features with predictive modeling [2]. Within this landscape, fractal descriptors offer an explicitly multiscale perspective. Studies report that three-dimensional fractal dimension and lacunarity can aid grading or subtype discrimination when added to conventional MR features in neuro-oncology [14], and broader reviews document encouraging but heterogeneous evidence across applications [15,16]. The key open question is incremental value: given that IBSI-standardized features already capture high-frequency contrast (for example, GLCM contrast or dissimilarity and short-run GLRLM) and large-structure or homogeneity effects (for example, GLSZM large-area or zone-size statistics; long-run and high-gray-level GLRLM; GLCM inverse-difference and homogeneity), under what conditions do fractal dimension and lacunarity add a non-redundant signal? Answering this calls for transparent pipelines, harmonization where appropriate, and robust validation to account for protocol sensitivity and feature collinearity [4,21,22,23,24].
Moreover, a closer look at recent methodological advances provides additional insight. For example, Ilmi & Khalaf’s graphene–temporal fusion for yoga pose recognition [26] showcases the power of combining spatial graphs and temporal dynamics, but its non-medical dataset and absence of standardized texture extraction limit its direct relevance for radiomics. Similarly, Jumadi & Md Akbar’s hybrid GRU-KAN model for energy consumption prediction [27] demonstrates effective hybrid modeling in time series but reflects a non-imaging domain and lacks imaging preprocessing/harmonization. A recent review of Mask R-CNN, for instance, segmentation [28], emphasizes the importance of robust segmentation—which is directly relevant for reproducible radiomic and fractal extraction—but does not address feature redundancy or multiscale heterogeneity explicitly. Finally, a stacked ensemble for cervical cancer prediction using tabular data [29] reinforces the role of ensemble and regularization strategies to address collinearity and imbalance. Issues that are equally relevant in radiomic pipelines assessing incremental features like fractal dimension and lacunarity. Taken together, these works highlight emerging best practices (graph–temporal modeling, hybrid architectures, segmentation robustness, ensemble calibration) and their limitations (domain mismatch, lack of imaging standardization, no explicit multiscale heterogeneity descriptors)—which reinforce the need for our simulation-grounded, IBSI-aligned, multiscale feature-computation, and incremental evaluation of fractal measures.

3. Materials and Methods

3.1. Radiomic Feature Set: Definitions and Computation

We consider a two-dimensional region of interest (ROI) represented by a real-valued image X R H × W with H pixel rows (height) and W pixel columns (width), defined on the grid Ω : = { 1 , , H } × { 1 , , W } , so that | Ω | = H W . Let X i { 1 , , N g } H × W denote the discretized (gray-level binned) version of X obtained by an operator Q : R { 1 , , N g } ; unless otherwise stated we use equal-frequency (quantile) binning with N g = 64 in line with common IBSI-style configurations [4]. Two feature configurations are examined. In the baseline setting, we extract 92 IBSI-standardized descriptors (19 first-order, 19 GLCM, 16 GLRLM, 16 GLSZM, 14 GLDM, 5 NGTDM, and 3 two-dimensional shape proxies) together with two fractal descriptors (Frac and Lac), for 94 features per ROI. In the wavelet-augmented setting, we add level-1 wavelet features by recomputing first-order and GLCM descriptors on each of the four sub-bands (LL, LH, HL, HH), which contributes 4 × ( 19 + 19 ) = 152 additional features. The wavelet-augmented total is therefore 94 + 152 = 246 features.

3.1.1. First-Order (Intensity) Features

First-order features summarize the distribution of voxel intensities in an ROI without reference to spatial arrangement [1,2,4]. They provide baseline information about central tendency, spread, tail behavior, and histogram shape. We follow a nineteen-feature set that aligns with IBSI-style implementations and widely used toolkits [3,4,10]: (i) power measures (Energy, Total Energy, RMS); (ii) dispersion (Variance, Standard deviation, Range, IQR); (iii) shape (Skewness, Kurtosis); (iv) histogram complexity (Entropy, Uniformity); and (v) robust summaries (Median, P10, P90, MAD, rMAD, as well as Min and Max). Definitions and formulas, together with brief interpretations, appear in Appendix A Table A1 [4,10].
Several practical points aid interpretation. Energy, Total Energy, RMS, Variance, and Standard deviation are scale dependent: linear rescaling of intensities changes their values, which is appropriate when physical units are meaningful but can confound cross-scanner comparisons if not harmonized [21,23]. Entropy and Uniformity depend on gray-level discretization ( N g and the binning scheme), so consistent preprocessing is important for reproducibility [4,22,23]. Median, IQR, MAD, and rMAD are more robust to noise and outliers than mean and variance and often stabilize modeling under heterogeneous acquisition conditions [2,4]. Although some measures are correlated by construction (for example, Energy and RMS), retaining both can be useful when regularization is used or when reporting adheres to radiomic standards, otherwise, redundancy reduction (e.g., correlation filtering) can simplify models without sacrificing information [4,18,19].

3.1.2. GLCM Features

The gray-level co-occurrence matrix (GLCM) encodes how often pairs of discretized gray levels co-occur at a fixed spatial offset. In this study the matrix is computed at distance 1, aggregated over four angles ( 0 , 45 , 90 , 135 ) , symmetrized, and normalized so that i = 1 N g j = 1 N g P i j = 1 [4,5].
Intuitively, mass far from the diagonal signals frequent high-contrast transitions, whereas mass concentrated near the diagonal reflects locally smoother textures [1,2]. We report nineteen descriptors: classical contrast and entropy terms; two homogeneity variants; two inverse-difference variants; inverse variance; correlation; maximum probability; three sum/difference statistics; and both information measures of correlation (IMC1/IMC2) [4,10]. Exact formulas are given in Appendix A Table A2.

3.1.3. GLRLM Features

The gray-level run-length matrix (GLRLM) counts how often contiguous runs of an identical gray level occur and how long those runs persist across the image [4,6]. After discretization, a run is the maximal sequence of adjacent pixels with the same gray level observed along a given direction. Short runs reflect a fine, rapidly varying texture, whereas long runs reflect a coarse, more uniform structure. By cross-tabulating run length with gray level, GLRLM features quantify, within one family, both scale (short and long) and tone (low and high gray) emphases. We compute GLRLMs at distance 1, aggregate over four angles ( 0 , 45 , 90 , 135 ) to improve directional robustness, and normalize to probabilities [4]. Formulas for all sixteen descriptors appear in Appendix A Table A3 [4,7,10].

3.1.4. GLSZM Features

The gray-level size-zone matrix (GLSZM) summarizes the distribution of connected zones of equal gray level within an ROI, irrespective of direction [4,7]. A zone is the maximal set of pixels that share the same discretized gray level and are connected under an 8-neighborhood in two dimensions [4,10]. Textures dominated by small zones correspond to fine, fragmented patterns, whereas textures dominated by large zones reflect coarse, homogeneous structure [2]. By cross-tabulating gray level and zone size, GLSZM features quantify how coarseness interacts with tone [4,7]. We use 8-connectivity, aggregate over the region, and normalize to probabilities [4,10]. Formulas for all sixteen descriptors appear in Appendix A Table A4 [4,10].

3.1.5. GLDM Features

The gray-level dependence matrix (GLDM, also called NGLDM) quantifies, for each discretized gray level i, how many pixels in a fixed neighborhood are dependent on the center pixel [4,8,10]. Two pixels are considered dependent if the absolute difference of their gray levels does not exceed a tolerance α . In this work α = 0 , meaning only exactly equal gray levels (after discretization) count as dependent [4,10]. Using a Chebyshev-1 neighborhood in two dimensions provides up to eight surrounding pixels; when the center is included in the count, the dependence size d ranges from 1 to 9 [4]. Tallying, for each gray level i, how often each dependence size d occurs within this neighborhood captures the prevalence of isolated pixels versus coherent tone-consistent patches and how these patterns vary with gray level [4,8]. Formulas for all fourteen descriptors appear in Appendix A Table A5 [4,10].

3.1.6. NGTDM Features

The neighborhood gray-tone difference matrix (NGTDM) summarizes how each discretized gray level deviates, on average, from its local neighborhood [4,9]. Given a quantized image X i { 1 , , N g } H × W , we evaluate each interior pixel using a 3 × 3 window (Chebyshev radius 1) and define the neighborhood mean as the average of the eight surrounding pixels, excluding the center [4,10]. To avoid edge artifacts, only pixels with a full 3 × 3 neighborhood (that is, interior pixels) contribute to the statistics [4]. Formulas for all five descriptors appear in Appendix A Table A6 [4,10].

3.1.7. Shape-2D Proxies

For this 2D simulation we summarize gross morphology with four compact descriptors computed on a binary ROI mask: area, a 4-neighbor (Manhattan) perimeter, and two in-plane anisotropy measures derived from the principal variances of the foreground coordinates. This deliberately lightweight set mirrors the intent of the standardized 3D shape features recommended by IBSI while remaining appropriate for single-slice analyses [4,10]. Formulas for all four descriptors appear in Appendix A Table A7 [4,10].

3.2. Wavelet Features

We complement the spatial-domain descriptors with a stationary (undecimated) level-1 two-dimensional wavelet transform so that all filtered images retain the original ROI size ( 64 × 64 ) [30,31]. We begin by specifying notation. Let X R H × W be indexed as X [ i , j ] , where i is the row (vertical, y) index and j is the column (horizontal, x) index. Let
h = { h [ k ] } k = M h M h , g = { g [ k ] } k = M g M g
be the one-dimensional low- and high-pass analysis filters with finite (odd) lengths
L h = 2 M h + 1 , L g = 2 M g + 1 ,
so that M h (respectively, M g ) is the half-support; that is, the number of taps on one side of the filter center. Throughout we use symmetric (mirror) padding at the borders and no downsampling (the transform is undecimated) [31,32]. Unless noted otherwise, the wavelet family is Coiflet-1 for its near-symmetry and compact support [33].
Next we define one-dimensional discrete convolutions along the column (x) and row (y) axes and a two-dimensional discrete convolution ( ) ; we denote 1-D convolution along columns by x and along rows by y . Convolution along x keeps rows fixed and sums across columns,
( h x X ) [ i , j ] = k = M h M h h [ k ] X [ i , j k ] ,
while convolution along y keeps columns fixed and sums across rows,
( h y X ) [ i , j ] = k = M h M h h [ k ] X [ i k , j ] .
If an index falls outside the valid image range ( 1 i H , 1 j W ) , it is reflected back into the interval (symmetric padding) before sampling X [4,10].
We then obtain four stationary level-1 sub-bands by low-/high-pass filtering along x and y without downsampling:
W LL [ i , j ] = h x ( h y X ) [ i , j ] ,
W LH [ i , j ] = h x ( g y X ) [ i , j ] ,
W HL [ i , j ] = g x ( h y X ) [ i , j ] ,
W HH [ i , j ] = g x ( g y X ) [ i , j ] .
Equivalently, with separable two-dimensional kernels ( h h ) , ( h g ) , ( g h ) , and ( g g ) and 2D convolution ∗,
W LL = ( h h ) X , W LH = ( h g ) X , W HL = ( g h ) X , W HH = ( g g ) X ,
where ( a b ) [ u , v ] = a [ u ] b [ v ] is the outer product of the one-dimensional filters (a separable two-dimensional kernel), and
( K X ) [ i , j ] = u v K [ u , v ] X [ i u , j v ]
is the two-dimensional discrete convolution of kernel K with image X. Here, u and v are integer offsets that index the kernel support horizontally and vertically, respectively; the double sum is taken over the finite support where K [ u , v ] 0 . The LL band collects coarse (approximation) content, whereas LH, HL, and HH collect horizontal, vertical, and diagonal detail.
Each sub-band is re-standardized (per sub-band, per ROI) to zero mean and unit variance prior to discretization to mitigate scale differences across sub-bands and ROIs [21]. (For predictive modeling, standardization parameters should be estimated on the training set only to avoid leakage.) Discretization then follows the base analysis: quantile binning into N g = 64 equiprobable gray levels applied independently to each sub-band [4,23]. On every discretized sub-band we compute first-order (19) and GLCM (19) statistics, yielding 4 × ( 19 + 19 ) = 152 wavelet features per ROI. This mirrors widely used radiomic configurations and captures frequency- and orientation-specific cues [17,34,35] while controlling multiplicity and collinearity [19,22]. Other matrix families (GLRLM, GLSZM, GLDM, NGTDM) can be added analogously, but here we restrict the wavelet set to first-order and GLCM for parsimony [19,22].
Finally, we record the transform type (undecimated), the wavelet family and level (Coiflet-1, one level), the boundary handling (symmetric reflection), and that discretization is performed per sub-band, in line with IBSI reporting guidance [4]. Wavelet filtering separates coarse structure (LL) from horizontal, vertical, and diagonal detail (LH, HL, HH), often revealing texture that single-scale statistics may miss [1,2]. At the same time, wavelet expansions increase feature counts and can accentuate instability or acquisition dependence; re-standardization, fixed discretization, and downstream multiplicity control help curb these effects [20,21,22,23,24,36].

3.3. Fractal Dimension and Lacunarity

Fractal descriptors target complementary aspects of spatial organization. The Minkowski–Bouligand (box-counting) fractal dimension quantifies a global scaling law: how occupancy grows as resolution is refined, whereas lacunarity quantifies scale-specific heterogeneity: how unevenly occupancy is distributed at a given window size [11,12,13].
We begin with formal definitions. Let X R H × W be a gray-scale ROI and let S { 1 , , H } × { 1 , , W } be a binary support extracted from X. Overlay a lattice of square boxes of side ε and let N ( ε ) be the number of boxes that intersect S . The box-counting fractal dimension is
D B = lim ε 0 log N ( ε ) log ε ,
so that, asymptotically, N ( ε ) ε D B  [11]. For a binary subset of the plane, 0 D B 2 .
Next, consider lacunarity at observation scale r. For a binary image B : { 1 , , H } × { 1 , , W } { 0 , 1 } (with 1 indicating occupancy), slide an r × r window over all admissible top-left anchors u = ( u x , u y ) Ω r , where
Ω r = ( u x , u y ) : 1 u x H r + 1 , 1 u y W r + 1 .
Define the window mass
M r ( u ) = i = 0 r 1 j = 0 r 1 B ( u x + i , u y + j ) ,
and let μ r = E u Ω r [ M r ( u ) ] and σ r 2 = Var u Ω r [ M r ( u ) ] . The gliding-box lacunarity is
Λ ( r ) = σ r 2 μ r 2 + 1 = Var [ M r ] E [ M r ] 2 + 1 ,
which satisfies Λ ( r ) 1 . Values near 1 indicate spatial homogeneity at scale r, whereas larger values indicate stronger clustering of mass or more pronounced voids [12,13]. Information resides in the curve r Λ ( r ) across scales.
We now detail the estimators and settings used here. For fractal dimension on gray-scale images, we adopt a threshold-averaged box-counting procedure. For each gray-level quantile t T = { 0.4 , 0.5 , 0.6 } , we binarize
B t = 1 X > quantile ( X , t ) ,
tile with box sizes b B = { 2 , 4 , 8 , 16 } pixels, count N t ( b ) non-empty boxes, and fit a linear regression of log N t ( b ) on log ( 1 / b ) . The estimated fractal dimension is the average slope across thresholds:
D = 1 | T | t T slope log N t ( b ) vs . log ( 1 / b ) .
For lacunarity, we threshold once at t = 0.5 to obtain B = 1 { X > quantile ( X , 0.5 ) } , compute Λ ( r ) from Equation (7) for window sizes r R = { 2 , 4 , 8 , 16 } , and summarize using a scale average:
Lac = 1 | R | r R Λ ( r ) .
Efficient integral-image implementations make these gliding-box computations practical and support broader multiscale sensitivity analyses [25].
Finally, we relate these measures to standardized radiomics. Larger D tends to co-vary with contrast and short-run behavior—fine-scale roughness and rapid gray-level changes—captured by descriptors such as GLCM contrast and dissimilarity and by short-run GLRLM measures [5,6]. In contrast, larger lacunarity tends to align with a large, coherent structure—broader zones, pronounced voids, and greater homogeneity—summarized by GLSZM large-area and zone-size statistics, GLRLM long-run and high-gray-level emphases, and GLCM homogeneity and inverse-difference measures [7,9]. Because these tendencies are related but not identical, the two fractal descriptors offer complementary views of multiscale heterogeneity and are most informative when interpreted alongside IBSI-standardized features.
In the box below we summarize, step by step, the full radiomic feature pipeline used in this study, including IBSI families, fractal estimators, and the optional wavelet branch.
Radiomic Feature Computation—Step-by-Step Summary
  • Define ROI and discretize. Let X R H × W and produce X i { 1 , , N g } H × W via quantile binning with N g = 64 (IBSI-aligned).
  • Compute IBSI families (spatial domain). First-order (19); GLCM (19; d = 1 at 0 , 45 , 90 , 135 ); GLRLM (16); GLSZM (16; 8-connectivity); GLDM (14; α = 0 ); NGTDM (5); plus four 2D shape proxies. (See Table A1, Table A2, Table A3, Table A4, Table A5, Table A6 and Table A7).
  • Estimate Frac. Threshold-averaged box counting with t { 0.4 , 0.5 , 0.6 } and box sizes b { 2 , 4 , 8 , 16 } ; compute the slope of log N ( b ) versus log ( 1 / b ) and average over thresholds.
  • Estimate Lac. Use a median threshold ( t = 0.5 ) with gliding windows r { 2 , 4 , 8 , 16 } ; compute Λ ( r ) = Var ( M ) / E ( M ) 2 + 1 and average across r.
  • (Optional) Wavelet branch. Apply a stationary level-1 Coiflet-1 transform with symmetric padding, which yields the LL, LH, HL, and HH sub-bands. For each sub-band, standardize and discretize to N g = 64 , then compute first-order (19) and GLCM (19) descriptors (total of 152 features).
  • Quality and reporting. Use consistent discretization, angular aggregation, and connectivity per IBSI; record all settings as in Table 1 and Table 2.

3.4. Similarity Metrics and Embeddings (Definitions)

This section explains how we quantify the relatedness between Frac and Lac and the remaining radiomic features, and how we create the two-dimensional maps that summarize their neighborhood structure. We begin by defining the pairwise similarity measures used to rank nearest neighbors, then describe the consensus score that combines them, and finally outline the low-dimensional embeddings used for visualization.

3.4.1. Pairwise Association (Similarity) Measures

Let { ( x i , y i ) } i = 1 n be paired measurements across n ROIs (for example, x is Frac or Lac and y is a radiomics descriptor). We first use Pearson’s correlation, the centered and variance-normalized covariance,
r = i = 1 n ( x i x ¯ ) ( y i y ¯ ) i = 1 n ( x i x ¯ ) 2 i = 1 n ( y i y ¯ ) 2 ,
which measures linear association on [ 1 , 1 ] [37].
We next use Spearman’s rank correlation, which applies Pearson’s formula to midranks R i and S i :
ρ = i = 1 n ( R i R ¯ ) ( S i S ¯ ) i = 1 n ( R i R ¯ ) 2 i = 1 n ( S i S ¯ ) 2 ,
thereby capturing monotone (not necessarily linear) association and also lying in [ 1 , 1 ]  [38]. In the no-tie case, ρ = 1 6 i ( R i S i ) 2 n ( n 2 1 ) .
We then include distance correlation (dCor), which equals 0 if and only if the variables are independent [39]. Using pairwise Euclidean distance matrices a i j = x i x j and b i j = y i y j , define their double-centered versions A i j = a i j a ¯ i · a ¯ · j + a ¯ · · and B i j = b i j b ¯ i · b ¯ · j + b ¯ · · . The sample quantities are
dCov 2 = 1 n 2 i , j A i j B i j , dVar X 2 = 1 n 2 i , j A i j 2 , dVar Y 2 = 1 n 2 i , j B i j 2 ,
and the resulting correlation is
dCor ( X , Y ) = dCov ( X , Y ) dVar X dVar Y [ 0 , 1 ] .
To broaden beyond purely linear or monotone trends, we also use the Maximal Information Coefficient (MIC), a grid-based information measure that scores a wide family of functional relationships on a common [ 0 , 1 ] scale [40]:
MIC = max x y B ( n ) I ^ ( X ; Y grid x , y ) log min { x , y } ,
where B ( n ) is the search budget and I ^ is the empirical mutual information on an x × y partition.
Finally, we include cosine similarity. For vectors x , y R n ,
cos ( θ ) = x · y x 2 y 2 .
When both variables are z–scored (mean 0, variance 1), Equation (14) equals Pearson’s r in Equation (10). Cosine similarity is widely used in vector space models for pattern analysis and information retrieval [41].
Because these five measures operate on different scales and emphasize different aspects of dependence, we combine them into a single consensus score. For a fixed target (either Frac or Lac), we compute { | r | , | ρ | , dCor , MIC , cos ine } against every other feature, z-score each metric across those comparisons, and average the five z-scores to obtain the composite similarity used for ranking. For r and ρ , we report two-sided p-values with Benjamini–Hochberg false discovery rate (FDR) correction over all comparisons for that target [36]. When a distance is required for clustering or embeddings, we use
d j k = 1 | r j k | ,
so that strongly associated pairs (in magnitude) are close.

3.4.2. Low-Dimensional Embeddings of Feature Geometry

With a pairwise distance in hand, we next visualize the geometry of the feature space in two dimensions. We use two complementary methods. The first is t-SNE, which builds a high-dimensional neighbor distribution using Gaussian kernels with per-point bandwidths σ i chosen to match a user-set perplexity:
p j i = exp x i x j 2 / 2 σ i 2 k i exp x i x k 2 / 2 σ i 2 , P i j = p j i + p i j 2 n .
It then fits a two-dimensional map with a heavy-tailed Student-t kernel,
q i j = ( 1 + y i y j 2 ) 1 k ( 1 + y k y 2 ) 1 ,
by minimizing KL ( P Q ) via gradient descent [42]. In our setting t-SNE is applied to the distances in Equation (15) and is read primarily for local neighborhoods; between-cluster distances are not directly interpretable.
The second method is UMAP, which starts from a k-nearest-neighbor graph built from Equation (15), converts it into a fuzzy simplicial set with edge weights p i j , and then learns a low-dimensional representation by minimizing
L = i < j p i j log q i j + ( 1 p i j ) log 1 q i j ,
where q i j = 1 1 + a y i y j 2 b and ( a , b ) control how quickly similarity decays with distance [43]. In practice, we use moderate neighborhood sizes and a small minimum-distance setting to reveal communities without over-fragmentation, and we fix random seeds for reproducibility.
Taken together, these choices specify how neighbors of Frac and Lac are ranked and how the resulting pairwise relationships are summarized as heatmaps, dendrograms, networks, and two-dimensional embeddings.

3.5. Simulation Design

We set out to build a controlled yet expressive sandbox in which fractal dimension (Frac) and lacunarity (Lac) can be compared fairly against a broad panel of standardized radiomic features, including a multiscale wavelet branch. Each synthetic sample is a 64 × 64 region of interest (ROI) that exhibits tunable spatial correlation and optional focal heterogeneity; all samples are then discretized to N g = 64 gray levels. Features are computed directly from the mathematical definitions given earlier, and similarity between Frac/Lac and each descriptor is quantified with multiple, complementary dependence measures defined in Section 3.4. The workflow mirrors common radiomics practice—textures ranging from fine to coarse, with or without a lesion-like structure—while preserving enough control to make causal attribution transparent (see Table 1 and Table 2 for concise summaries). For clarity, Figure 1 provides an overview of the simulation workflow.
To generate background texture we adopt a two-dimensional autoregressive recursion that is AR(1)-like along both axes,
X i j ρ X i 1 , j + ρ X i , j 1 ρ 2 X i 1 , j 1 + ε i j , ε i j N ( 0 , σ 2 ) ,
drawing ρ U ( 0.4 , 0.85 ) and σ U ( 0.7 , 1.3 ) independently for each ROI. This Markov construction yields approximately isotropic correlation, exposes a single parameter that smoothly tunes spatial frequency (small ρ gives rough, high-frequency patterns; large ρ gives smooth, low-frequency patterns), and is numerically stable and fast on grids [44,45]. Because clinical images rarely look perfectly stationary, we superimpose 0–2 Gaussian blobs with amplitudes sampled from [ 2 , 5 ] and radii from 6–12 pixels, then re-standardize X to zero mean and unit variance. These localized additions emulate the lesion-like structure without overwhelming the global correlation, creating cases where Frac can respond to fine-scale roughness while Lac responds to gap structure and region size within the same sample.
After simulation, the continuous field is converted to discrete gray levels by quantile binning into N g = 64 equiprobable bins, producing X i { 1 , , 64 } 64 × 64 . Quantile discretization mitigates arbitrary intensity scaling, stabilizes probability estimates that enter co-occurrence/run/zone matrices, and aligns with IBSI recommendations; the choice N g = 64 balances information content against matrix sparsity for this ROI footprint [4,22,23,24]. Using equiprobable bins also keeps histogram-based first-order features well behaved and reduces sensitivity to marginal rescales. A compact summary of all simulation factors and data-generation settings appears in Table 1.
Feature computation proceeds as in the mathematical section: 19 first-order descriptors; 19 GLCM statistics constructed at distance 1 and aggregated over 0 , 45 , 90 , 135 ; 16 GLRLM descriptors with the same angular aggregation; 16 GLSZM descriptors using 8-connectivity; 14 GLDM descriptors with tolerance α = 0 and Chebyshev radius 1; 5 NGTDM descriptors computed on full 3 × 3 neighborhoods; and 4 simple two-dimensional shape proxies [4,5,6,7,8,9]. Frac is estimated by box counting across four box sizes b { 2 , 4 , 8 , 16 } and three intensity thresholds t { 0.4 , 0.5 , 0.6 } , averaging slopes from the log N ( b ) versus log ( 1 / b ) regressions over thresholds to stabilize against single-cut artifacts [11]. Lac is estimated by gliding-box lacunarity using window sizes r { 2 , 4 , 8 , 16 } after thresholding at the median ( t = 0.5 ), then averaging Λ ( r ) over r to obtain a single summary; efficient integral-image implementations are used to avoid partial-window bias and to keep computation tractable [12,13,25].
In addition to the spatial-domain features above, we include a one-level, undecimated (stationary) two-dimensional wavelet transform with symmetric boundary handling, using Coiflet-1 filters for near-symmetry and compact support [30,31,32,33]. This yields four sub-bands (LL, LH, HL, HH) at the original resolution. Each sub-band is re-standardized to zero mean and unit variance and discretized independently into N g = 64 equiprobable levels. On every sub-band we compute first-order (19) and GLCM (19) statistics, adding 4 × ( 19 + 19 ) = 152 wavelet features per ROI. These choices follow IBSI/PyRadiomics conventions for wavelet-filtered images and limit multiplicity by focusing on first-order and GLCM in the transform domain [4,10].
Similarity between Frac and Lac and each classical or wavelet-domain feature is evaluated with five measures that emphasize different types of dependence—Pearson’s r (10), Spearman’s ρ (11), distance correlation ( dCor ) (12), MIC (13), and cosine similarity (14)—as defined in Section 3.4. For r and ρ we adjust p-values across all pairwise tests (classical plus wavelet features) using Benjamini–Hochberg false discovery rate (FDR) control [36]. Because these metrics live on different scales, we standardize them by z-scoring { | r | , | ρ | , dCor , MIC , cos ine } across the full set of comparisons and then average the five z-scores to obtain a composite similarity (see Section 3.4 for the composite definition). For visualization, we form the union of the top-k neighbors around Frac and Lac (here k = 30 for each), compute the Pearson correlation matrix on that set, and display a clustered heatmap. We also embed the same features using distances 1 | r | with t-SNE and UMAP (definitions in Section 3.4) to convey qualitative geometry, treating these embeddings as descriptive rather than inferential [42,43]. The complete analysis and similarity settings, including wavelet parameters, are summarized in Table 2.
With N = 1000 independent ROIs (default; configurable), each producing 246 features, the final table contains N × 246 measurements. The inventory by family is summarized in Table 3. Random seeds are fixed at the start of each full simulation to ensure deterministic regeneration of data and figures; all hyperparameters (scale sets, thresholds, ρ and σ ranges, discretization and neighborhood choices, and wavelet settings) are declared in the repository and mirrored in the manuscript so that readers can audit or modify them and regenerate the entire analysis.
In summary, the AR(1)-like generator (Equation (19)) provides a transparent, one-parameter handle on spatial frequency [44,45]; the optional Gaussian blobs introduce controlled nonstationarity reminiscent of lesions or subregions; the quantile discretization and the angle/distance settings align with IBSI so that conclusions carry over to standard pipelines [4,23,24]; the wavelet branch exposes multiscale, orientation-specific detail in a standardized way [4,10,30,31,32,33]; and the multimetric similarity with FDR reflects best practices when exploring many pairwise relationships where linearity is not guaranteed [36,39,40]. The scope is intentionally two-dimensional for clarity and speed; extending to 3D with 26-neighbor settings, volumetric zones/runs, and 3D wavelets and lacunarity is straightforward but more computationally demanding.

4. Results

4.1. Results Without Wavelet Features

We analyzed N = 1000 simulated two-dimensional ROIs (each 64 × 64 pixels), discretized to N g = 64 gray levels. Alongside the 92 IBSI texture descriptors, we included fractal dimension (Frac) and lacunarity (Lac), for 94 features in total. Similarity between Frac and Lac and all other features was summarized using the procedure in Section 3.4; in brief, we averaged z–scored Pearson | r | , Spearman | ρ | , distance correlation, MIC, and cosine similarity to obtain a composite score.
We begin by examining simple pairwise associations. Figure 2 and Figure 3 plot Frac and Lac against their highest-ranked neighbors under the composite score. In both cases the relationships are monotone with an approximately linear trend, indicating that Frac and Lac are not idiosyncratic to the simulator but align with recognizable texture constructs. Qualitatively, Frac tracks contrast- and difference-type behavior (for example, GLCM contrast, dissimilarity, and difference entropy, together with GLRLM short-run emphases), whereas Lac increases with statistics that summarize coherent, larger structures (for example, GLSZM large-area and zone-size measures and GLRLM long-run or high-gray-level emphases).
Next, to place these strongest pairwise relationships in context, we formed the union of the top-k neighbors of Frac and Lac ( k = 30 each) and computed the Pearson correlation matrix for this set. The clustered heatmap in Figure 4 exhibits a clear bipartite organization: one block is rich in contrast-oriented statistics and co-locates with Frac (GLCM contrast, dissimilarity, and difference entropy; GLRLM short-run; NGTDM busyness and contrast; selected GLDM dependence-entropy terms), and the other block emphasizes a larger structure and co-locates with Lac (GLSZM large-area and zone-size statistics; GLRLM long-run and high-gray-level emphases). Homogeneity measures (for example, GLCM Homogeneity and ID/IDN) lie opposite the contrast block and tend to associate more closely with Lac.
We then asked whether alternative geometric summaries tell a consistent story. Hierarchical clustering on 1 | r | distances (Figure 5), a sparse correlation network with edges for | r | 0.55 (Figure 6), a PCA loading biplot (Figure 7), and low-dimensional embeddings via UMAP and t-SNE (Figure 8 and Figure 9) all corroborate the same organization. Taken together, these views indicate that while Frac and Lac overlap with specific IBSI families, they are not simple duplicates and provide complementary multiscale information.
Finally, we turn to the ranked lists. Table 4 shows that the nearest neighbors of Frac are dominated by GLCM difference, contrast, and entropy-type descriptors (Contrast, Dissimilarity, Difference Entropy, Difference Variance, Difference Average) with additional short-run and small-structure measures (GLRLM SRE and SRLGLE, plus selected GLSZM and GLDM small-area or low-gray-level statistics). Agreement is uniformly high across Pearson, Spearman, distance correlation, MIC, and cosine similarity, indicating a strong monotone association. Practically, Frac tracks fine-scale, high-frequency variation; redundancy is likeliest when models already contain several contrast-like or short-run descriptors.
Reading Table 5, the nearest neighbors of Lac emphasize large-scale organization and gray-level regularity. Frequent entries include GLCM correlation and homogeneity (Correlation, Homogeneity2, ID/IDN, and IMC2) together with GLSZM zone-size statistics and GLRLM long-run and high-gray-level emphases (LRE, LRHGLE, HGRE). Similarity remains high across all metrics, consistent with a smooth, low-frequency signal. Read with the Frac table, this underscores a complementary division: Frac rises with local differences and short runs, whereas Lac rises with homogeneity, correlation, and coherent large structures.
Comprehensive similarity tables appear in Appendix B Table A8 and Table A11.

4.2. Results with Wavelet Features

We repeated the analysis with N = 1000 simulated ROIs ( 64 × 64 , N g = 64 ), augmenting the feature set with level-1 wavelet descriptors from the four sub-bands (LH, HL, HH, LL). As before, similarity between Frac/Lac and the remaining features was summarized using the composite procedure in Section 3.4.
We begin by inspecting the strongest pairwise associations after adding wavelets. The nearest neighbors of Frac are led by gray-level co-occurrence statistics that emphasize differences and contrast: difference variance, difference entropy, dissimilarity (and difference average), and contrast occupy the top ranks. Turning to Lac, homogeneity and inverse-difference measures dominate: inverse difference normalized (IDN), information measure of correlation 2 (IMC2), correlation, homogeneity (both variants), inverse variance, and inverse difference (ID) appear consistently, with a representative wavelet contribution—LL (level-1) GLCM sum entropy—also entering the top ten. The best pairs exhibit very strong, nearly monotone linear trends: Frac against GLCM difference entropy shows | r | 0.98 (Figure 10), while Lac against GLCM IDN shows | r | 0.96 (Figure 11).
Next, we place these pairwise findings in context by examining the neighborhood structure around Frac and Lac. Using the union of the top-k neighbors for each target ( k = 30 ), the correlation heatmap retains a clear two-block organization (Figure 12). The wavelet detail bands (LH, HL, and HH) align with high-frequency contrast and difference statistics near Frac, whereas the LL approximation summaries co-locate with large-structure and coherence measures near Lac. We then asked whether alternative geometric summaries tell a consistent story: the dendrogram based on 1 | r | (Figure 13), the sparse correlation network with edges for | r | 0.55 (Figure 14), the PCA loading biplot (Figure 15), and the UMAP and t-SNE embeddings (Figure 16 and Figure 17) all reproduce this separation.
Finally, we turn to the ranked lists. The top neighbors of Frac (Table 6) are almost entirely difference- and contrast-oriented GLCM descriptors, with agreement across all five similarity criteria (distance correlation, MIC, | r | , | ρ | , cosine). The top neighbors of Lac (Table 7) are homogeneity and inverse-difference statistics together with a representative LL wavelet summary (sum entropy), again with strong cross-metric concordance. Bringing together the heatmaps, dendrograms, networks, PCA, and embeddings, the conclusion is that wavelet features do not dissolve the original contrast-versus-structure separation; rather, they sharpen interpretation. The LH, HL, and HH detail terms behave like contrast and short-run statistics and align with Frac, whereas the LL approximation terms behave like large-area and coherence descriptors and align with Lac. In this simulated setting the best neighbors show very large linear associations with Frac and Lac, and the consistency across metrics indicates stable monotone relationships at complementary spatial scales.

5. Conclusions

We begin by summarizing the overall pattern that emerged across both analyses. In the baseline setting (without wavelets) and in the wavelet-augmented setting, the feature space consistently organized into two broad groups. The first group was characterized by high-frequency variation, local contrast, and short-run texture. The second group reflected large-scale structure, long-run behavior, and homogeneity. Fractal dimension (Frac) aligned with the first group, whereas lacunarity (Lac) aligned with the second. After introducing wavelet features, this geometry became sharper rather than different: statistics from the LH, HL, and HH detail bands co-located with the contrast and short-run-oriented group near Frac, while LL approximation statistics co-located with the large-structure and homogeneity group near Lac.
Next, we compare pairwise associations. In the baseline analysis, Frac and Lac showed clear monotone relationships with their nearest neighbors, although linear correlations were only moderate. When wavelet descriptors were added, these associations strengthened markedly. The top pairs were strongly and nearly linearly related, with best absolute Pearson correlations of approximately 0.98 for Frac and 0.96 for Lac. Moreover, a composite similarity that averages five criteria—distance correlation, maximal information coefficient, absolute Pearson correlation, absolute Spearman correlation, and cosine similarity—reproduced the same two-group structure across heatmaps, hierarchical clustering, correlation networks, principal component loadings, and low-dimensional embeddings, underscoring the stability of these findings.
We then consider practical implications. Frac primarily reflects fine-scale, high-frequency variation and therefore tends to overlap with descriptors that emphasize local contrast and short runs; when wavelets are included, detail-band statistics often behave similarly. Lac is most sensitive to coherent, large-scale organization and thus overlaps with homogeneity and correlation measures as well as zone- and run-size emphases; with wavelets, LL-band statistics commonly align with the same behavior. In practice, Frac is most likely to be redundant in models that already contain many contrast-like or short-run descriptors, especially when detail-band wavelet features are present. Lac is most likely to be redundant when GLSZM zone-size, GLRLM long-run or high-gray-level, or GLCM homogeneity and correlation families dominate. A pragmatic strategy is to include Frac and Lac provisionally and assess incremental value with nested models and cross-validation, retaining them when they improve generalization or provide interpretable, non-overlapping information.
Finally, we outline reporting considerations and limitations. Because discretization and scale choices influence effect sizes and rankings, studies should report the gray-level discretization ( N g ), thresholding strategy, wavelet family and levels (when applicable), and the specific scale sets used for box counting and lacunarity. Our conclusions are based on simulated regions of interest with controlled heterogeneity; external validation across imaging modalities, anatomies, acquisition protocols, and segmentation approaches is warranted. Future work should examine robustness to preprocessing steps such as resampling and filtering, inter-software reproducibility, and the stability of Frac and Lac under varying ROI sizes and boundary conditions. Extending to multiple wavelet levels and systematically tuning scale sets may further clarify the complementary multiscale roles of Frac and Lac.

6. Simple Summary

Fractal descriptors, fractal dimension (Frac) and lacunarity (Lac), are often proposed to capture multiscale texture complexity, but their added value over standardized radiomic features is uncertain. Using controlled two-dimensional simulations, we compared Frac and Lac with 92 Image Biomarker Standardisation Initiative (IBSI) texture descriptors under two settings: a baseline analysis without wavelets and a wavelet-augmented analysis that recomputed first-order and GLCM features in each sub-band. In both settings, a stable two-block organization emerged: Frac aligned with high-frequency, contrast/difference, and short-run statistics, whereas Lac aligned with large-structure, homogeneity, and long-run statistics. With wavelets included, the strongest pairwise associations became very large ( | r | 0.96 0.98 ), indicating that under homogeneous conditions fractal metrics can be partly redundant with established features. Nevertheless, clustering, correlation networks, and low-dimensional embeddings consistently showed that Frac and Lac occupy overlapping yet separable neighborhoods, clarifying when they can add interpretable multiscale information (e.g., scale-dependent irregularity or void heterogeneity) and when they may be safely omitted. Practical guidance for feature selection and interpretation is summarized in Table 8.

7. Discussion

7.1. Interpretation and Practical Guidance

The very strong correlations observed in the wavelet-augmented setting ( | r | 0.96 0.98 ) underscore that fractal dimension (Frac) and lacunarity (Lac) are not universally independent of standardized radiomic features. Under controlled, stationary texture conditions, both measures align closely with IBSI families that already encode high-frequency contrast (for Frac) or low-frequency homogeneity (for Lac). In such cases, the practical gain from including fractal descriptors in predictive models is likely minimal, as their variance can be absorbed by contrast-, difference-, or run-length-based features, or by their wavelet detail analogues.
However, these high correlations also have a clear interpretation: they reveal that fractal metrics quantify intensity organization at comparable scales to certain standardized descriptors, providing a confirmatory rather than contradictory view. In datasets where multiscale heterogeneity, irregular boundaries, or void-like texture patterns are central, the same measures may capture complementary information not reflected in single-scale features. This is particularly relevant in biological or physical systems where structural variability occurs over multiple resolutions (for example, infiltrative tumors or heterogeneous parenchymal patterns).
From a modeling standpoint, we recommend three practical steps. First, examine pairwise and multimetric similarity (for example, Pearson, distance correlation) before including fractal descriptors, and remove them when redundancy exceeds | r | > 0.9 . Second, retain Frac and Lac when their mechanistic interpretation aligns with the hypothesis of scale-dependent organization. Third, when redundancy is unavoidable, use dimensionality-reduction or feature-importance frameworks to confirm whether Frac or Lac contribute incremental predictive value. These recommendations convert the high correlations observed here into actionable guidance for applied researchers using fractal features alongside IBSI-compliant radiomics (See Table 8).

7.2. Limitations, Sensitivity, and External Validation

This study is intentionally simulation-grounded and two-dimensional, providing a transparent setting to attribute observed effects directly to feature definitions. However, such a design does not substitute for empirical validation on clinical images, and several important limitations remain.
First, we did not experimentally vary voxel size, reconstruction kernels, or scanner protocols on real CT, MR, or PET data. Previous research has shown that radiomic features can be highly sensitive to these acquisition and preprocessing factors, motivating harmonization procedures and rigorous documentation of preprocessing choices [4,21,22,23,24]. The present results should therefore be interpreted as methodological benchmarks under controlled texture generation rather than as demonstrations of clinical robustness.
Second, both fractal and matrix-based descriptors can be influenced by the definition of the region of interest. We did not assess inter- or intra-rater segmentation variability in this study. To address this in future work, we propose a simple and reproducible stress test based on small perturbations of the mask boundary. This can include morphological erosion or dilation by one to three pixels and random boundary jitter. Reporting the rank-based stability of feature–feature similarities and nearest-neighbor relationships under such perturbations would provide a quantitative sense of segmentation sensitivity.
Third, we outline several lightweight robustness checks that can be reproduced within our synthetic framework to approximate the impact of acquisition and preprocessing differences. These include (1) down- and up-sampling the regions of interest to simulate changes in voxel size; (2) applying mild Gaussian blurring or adding controlled Gaussian or Poisson noise before discretization to approximate reconstruction variability; and (3) performing slight mask modifications to emulate segmentation variability. For each of these perturbations, one can evaluate Spearman correlations between the original and perturbed feature vectors, the stability of composite similarity rankings for Frac and Lac, and the persistence of the two-block organization observed in heatmaps and embeddings. These procedures provide an immediate sense of robustness without overstating generalizability to clinical imaging.
Finally, a complete empirical validation will require analysis of multi-center imaging datasets that incorporate realistic variation in acquisition and reconstruction parameters, along with standardized segmentations and harmonization procedures. As a next step, we plan to extend this framework to public CT and MR datasets with IBSI-compliant preprocessing, apply harmonization methods such as ComBat where appropriate, and assess whether the Frac and Lac neighborhoods and the two-block feature organization observed here persist under real acquisition variability. This will be accompanied by systematic tests of segmentation and voxel-size sensitivity, consistent with IBSI recommendations. Together, these efforts will enable a more comprehensive validation of the findings reported here.

7.3. Stability and Sensitivity of Similarity Inferences

This subsection assesses the internal robustness of our similarity findings within the simulation, complementing the prior subsection, which focuses on external validity and real-data limitations. We therefore report internal robustness checks below and reserve full external (clinical) validation for future work.
Table 9 summarizes the composite-score distribution and a leave-one-metric-out (LOO) robustness check for the baseline (no-wavelet) and wavelet-augmented settings. Dispersion (SD ≈ 0.81–0.82; IQR ≈ 1.29–1.41) is nearly identical across baseline and wavelet analyses, and the strongest associations remain consistent (Frac with GLCM difference variance; Lac with GLCM correlation and homogeneity variants). Importantly, the LOO sensitivity shows that the top-30 neighbor sets are highly stable to how the composite is formed, with conservative overlaps in the range 0.70–0.93. While the global rank correlation between the full lists can drop when a metric is removed (reflecting minor reshuffling among mid-ranked, near-tied features), the neighborhoods that drive our conclusions and visual block structure remain stable. These results indicate that our multimetric consensus and the two-block organization (contrast and difference versus structure and homogeneity) are robust to reasonable perturbations of the similarity recipe, directly addressing the reviewer’s request for sensitivity analysis.

7.4. Extension to 3D and Neighborhood Topology

Although all analyses in this study are based on two-dimensional (2D) simulated textures, the underlying framework extends directly to three-dimensional (3D) radiomic computation using IBSI-compliant 3D gray-level co-occurrence, run-length, and size-zone matrices with 26-neighbor connectivity [3,4,5,6,7]. The computational scaling and structural correlation between 2D and 3D metrics were examined on pilot 3D synthetic volumes ( 64 × 64 × 64 voxels) generated under the same stochastic parameters as the 2D simulation. Results are summarized in Table 10.
Fractal descriptors, fractal dimension and lacunarity, extend naturally to 3D using cubic box counting and gliding-box lacunarity with cubic windows. Because both are scale and dimension agnostic, their relative behavior with respect to standardized texture families remains qualitatively consistent. Empirical correlations (Table 10) suggest that 3D feature analogs preserve the same bipartite organization observed in 2D (contrast/difference versus structure/homogeneity) while modestly increasing computational cost. These results quantitatively support the statement that extension to 3D is feasible and structurally consistent with the 2D benchmark.

8. Future Work

The central observation of this study is a stable two-group geometry: fractal dimension aligns with high-frequency variation, local contrast, and short runs, whereas lacunarity aligns with large-scale structure, long runs, and homogeneity. Building on this, several directions naturally follow.
First, external validation and harmonization deserve priority. Multi-center, fully three-dimensional CT, MR, and PET cohorts should be used to probe robustness to acquisition, reconstruction, voxel size, and segmentation variability. The extent of harmonization required—intensity standardization, resampling, and batch-effect mitigation—ought to be quantified with prospective protocols and phantoms, comparing methods under identical folds (for example, histogram standardization and ComBat-type approaches) [4,20,21,22,23,24]. Alternative estimators of fractal dimension should be assessed under the same preprocessing (box counting, spectral, or slope-based methods, variograms, Higuchi-type approaches), and single-number lacunarity summaries should be replaced by full lacunarity curves evaluated over multiple scales and thresholds, with principled functionals and uncertainty intervals reported.
Next, the link between fractal measures and multiscale representations merits systematic mapping. Wavelet levels and orientations (for example, levels 1–3 and beyond) and alternative filter banks (such as Gabor and steerable filters) should be compared to isolate which sub-bands drive their behavior. Ablation studies and partial-correlation or conditional-similarity analyses can test whether fractal dimension and lacunarity add information beyond the most predictive members of the GLCM, GLRLM, and GLSZM families. Design sweeps over gray-level discretization ( N g and binning scheme), thresholds, scale grids, boundary conditions, and mask perturbations should be paired with bootstrap and permutation procedures, as well as stability assessments (subsampling and perturb–retrain), to obtain uncertainty intervals for similarity rankings and selection frequencies.
Evaluation should proceed within transparent, preregistered pipelines that employ nested model comparisons, leakage-free cross-validation with fixed folds, and regularization tuned to family-wise collinearity. Model effects are best communicated using SHAP values—SHapley Additive exPlanations that attribute each feature’s contribution to individual predictions [46]—alongside partial-dependence profiles (PDPs) and accumulated local effects (ALEs). Report calibration metrics and generalization gaps in tandem with discrimination. Critically, test incremental value over strong baselines: for fractal dimension, use baselines rich in contrast and short-run descriptors (and, when applicable, wavelet-detail statistics); for lacunarity, use baselines rich in homogeneity and correlation measures, as well as zone- and run-size emphases. Apply multiple-testing adjustments (e.g., Benjamini–Hochberg FDR) consistently across feature families and sub-bands [36]. These steps will help determine when fractal dimension and lacunarity provide non-redundant multiscale information beyond standard IBSI families, and will clarify which acquisition and preprocessing choices most strongly shape their behaviour in practice.
Finally, as radiomic profiling of glioblastoma has demonstrated tangible prognostic value beyond conventional clinical and radiologic predictors [47], future work should assess whether the incremental information provided by fractal dimension and lacunarity translates into improved outcome modeling across neuro-oncology and other disease domains.

Author Contributions

Conceptualization, M.Z.; methodology, M.Z.; software, M.Z. and M.S.; validation, M.Z. and M.S.; formal analysis, M.Z.; investigation, M.Z.; resources, M.Z.; data curation, M.S.; writing—original draft preparation, M.Z.; writing—review and editing, M.S. and M.Z.; visualization, M.Z.; supervision, M.Z.; project administration, M.Z.; funding acquisition, M.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding. The APC was funded by the College of Arts and Sciences and East Tennessee State University.

Institutional Review Board Statement

This study did not involve humans or animals and was based exclusively on simulated images.

Informed Consent Statement

This study did not involve human participants.

Data Availability Statement

The data presented in this study were generated entirely by simulation; no human or clinical imaging data were used.

Conflicts of Interest

The author declares no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Appendix A. Feature Definitions and Interpretations

Appendix A.1. Feature Definitions and Interpretations

Before listing the feature formulas, we define the notation used throughout Table A1. Let x = { x k } k = 1 H W be the vectorized intensities of X, μ = 1 H W k x k , σ 2 = 1 H W k ( x k μ ) 2 , and Q p the pth quantile of x . For the binned histogram over N g gray levels, let p g be the probability of bin g ( g = 1 N g p g = 1 ). Define I = { k : Q 0.1 x k Q 0.9 } and x ¯ 10 90 = 1 | I | k I x k . Unless specified, logarithms in information measures use base 2 (bits). The naming, symbols, and formulas for first-order features follow the Image Biomarker Standardisation Initiative (IBSI) recommendations and the PyRadiomics reference implementation [4,10]; general radiomics background is provided in Gillies et al. [1], Mayerhoefer et al. [2].
Table A1. First-order intensity features (19) with formulas and interpretation. H W is the number of voxels; p g denotes histogram probabilities. Formulas and naming align with IBSI, and the robust mean absolute deviation (rMAD) follows the PyRadiomics definition [4,10].
Table A1. First-order intensity features (19) with formulas and interpretation. H W is the number of voxels; p g denotes histogram probabilities. Formulas and naming align with IBSI, and the robust mean absolute deviation (rMAD) follows the PyRadiomics definition [4,10].
FeatureFormula/SymbolWhat It Captures (Interpretation)
Energy k = 1 H W x k 2 Overall signal power (scale-dependent); larger when intensities have large magnitude.
Total Energy k = 1 H W x k 2 × voxel area Energy with physical units (in 2D, voxel area often set to 1).
RMS 1 H W k = 1 H W x k 2 Root-mean-square; average magnitude.
Variance σ 2 = 1 H W k ( x k μ ) 2 Dispersion about the mean; sensitive to outliers.
Standard Deviation σ Square root of variance; same interpretation on original scale.
Skewness 1 H W k ( x k μ ) 3 σ 3 Asymmetry of the distribution.
Kurtosis 1 H W k ( x k μ ) 4 σ 4 Tail heaviness/peakedness (non-excess); equals 3 for Gaussian.
Entropy g = 1 N g p g log 2 p g Histogram unpredictability (bits).
Uniformity g = 1 N g p g 2 Histogram concentration (also “histogram energy”).
Mean μ Central tendency (average intensity).
Median Q 0.5 Robust central tendency.
P10 Q 0.1 Lower-tail intensity (10th percentile).
P90 Q 0.9 Upper-tail intensity (90th percentile).
Minimum min k x k Absolute lowest observed intensity.
Maximum max k x k Absolute highest observed intensity.
IQR Q 0.75 Q 0.25 Middle-spread; robust scale.
Range max k x k min k x k Full dynamic range of intensities.
MAD 1 H W k = 1 H W | x k μ | Mean absolute deviation.
rMAD 1 | I | k I | x k x ¯ 10 90 | Trimmed absolute deviation within the 10–90% band.

Appendix A.2. GLCM Definitions and Feature Formulas

The gray-level co-occurrence matrix (GLCM) summarizes how often pairs of discretized gray levels occur at a fixed spatial offset; our notation and feature formulas follow the Image Biomarker Standardisation Initiative (IBSI) and the PyRadiomics reference implementation, with terminology rooted in the original work of Haralick and co-authors [4,5,10]. In our setup P is built at distance 1, aggregated over four angles ( 0 , 45 , 90 , 135 ) , symmetrized, and normalized so that i , j P i j = 1 (angle aggregation and normalization as recommended by IBSI [4]).
Formally, let P R + N g × N g be the symmetric, normalized GLCM. Define the one-dimensional marginals p x ( i ) = j P i j and p y ( j ) = i P i j ; the means μ x = i i p x ( i ) and μ y = j j p y ( j ) ; the standard deviations σ x , σ y ; the gray-level sum distribution p x + y ( k ) = i + j = k P i j for k { 2 , , 2 N g } ; and the difference distribution p x y ( k ) = | i j | = k P i j for k { 0 , , N g 1 } . Entropy terms use the natural logarithm (as in IBSI) unless otherwise noted [4]:
H X Y = i , j P i j log P i j , H X = i p x ( i ) log p x ( i ) , H Y = j p y ( j ) log p y ( j ) ,
H X Y 1 = i , j P i j log p x ( i ) p y ( j ) , H X Y 2 = i , j p x ( i ) p y ( j ) log p x ( i ) p y ( j ) .
Table A2. GLCM features (19) with formulas and interpretation. P i j is the symmetric, normalized co-occurrence for gray levels i , j ; p x , p y are marginals; p x + y and p x y are sum and difference distributions. Formulas align with IBSI and PyRadiomics [4,10]; historical names follow Haralick et al. [5].
Table A2. GLCM features (19) with formulas and interpretation. P i j is the symmetric, normalized co-occurrence for gray levels i , j ; p x , p y are marginals; p x + y and p x y are sum and difference distributions. Formulas align with IBSI and PyRadiomics [4,10]; historical names follow Haralick et al. [5].
FeatureFormula/SymbolInterpretation
Contrast i , j ( i j ) 2 P i j Off-diagonal emphasis; higher with strong edges.
Dissimilarity i , j | i j | P i j Linear penalty version of contrast.
ASM i , j P i j 2 Angular Second Moment; texture uniformity.
Energy ASM Monotone with ASM (legacy definition).
Entropy i , j P i j log P i j Co-occurrence disorder.
Homogeneity (1) i , j P i j 1 + ( i j ) 2 Rewards near-diagonal mass.
Homogeneity (2) i , j P i j 1 + ( | i j | / N g ) 2 Scale-normalized variant.
ID i , j P i j 1 + | i j | Inverse Difference; favors small gaps.
IDN i , j P i j 1 + | i j | / N g Normalized ID.
Inv. Variance i j P i j ( i j ) 2 Strongly favors smooth textures.
Correlation i j P i j μ x μ y σ x σ y Marginal association strength.
Max Prob. max i , j P i j Dominant co-occurrence pair.
Sum Avg. k k p x + y ( k ) Mean of sum distribution.
Sum Entropy k p x + y ( k ) log p x + y ( k ) Disorder of sum distribution.
Diff. Avg. k k p x y ( k ) Mean absolute difference.
Diff. Var. k ( k DiffAvg ) 2 p x y ( k ) Spread of difference distribution.
Diff. Entropy k p x y ( k ) log p x y ( k ) Disorder of difference distribution.
IMC1 H X Y H X Y 1 max ( H X , H Y ) Entropy gap vs. marginals.
IMC2 1 exp ( 2 ( H X Y 2 H X Y ) ) Bounded [0, 1]; strength of dependence.
All sums run over i , j { 1 , , N g } . Using natural versus base-2 logarithms only rescales the entropy terms and does not affect ordering [4]. Symmetric, normalized, angle-aggregated P generally improves stability and comparability across directions [4]. If either σ x or σ y equals zero, the correlation is undefined and should be reported as NA in keeping with reproducibility guidelines [4,10].

Appendix A.3. GLRLM Definitions and Feature Formulas

The gray-level run-length matrix (GLRLM) quantifies how often contiguous runs of identical gray level occur and how long those runs persist along specified directions. Our notation and feature formulas follow the Image Biomarker Standardisation Initiative (IBSI) and the PyRadiomics reference implementation, with historical roots in the original run-length work of Galloway [6] and subsequent refinements [4,10]. In our setup the GLRLM is computed at pixel distance 1, aggregated over four angles ( 0 , 45 , 90 , 135 ) , then normalized to a probability distribution (angle aggregation and normalization as recommended by IBSI [4]).
Formally, let R R + N g × R max be the (angle-aggregated) GLRLM, where R i r counts runs of length r at gray level i. Let N runs = i = 1 N g r = 1 R max R i r denote the total number of runs. Define the normalized joint distribution p ( i , r ) = R i r / N runs with marginals
p g ( i ) = r = 1 R max p ( i , r ) ( gray-level profile ) , p r ( r ) = i = 1 N g p ( i , r ) ( run-length profile ) .
Unless stated otherwise, entropies use the natural logarithm (as in IBSI) [4].
Table A3. GLRLM features (16) with formulas and interpretation. p ( i , r ) is the normalized GLRLM; p g , p r are marginals; N runs = i , r R i r . Formulas align with IBSI/PyRadiomics [4,10]; names follow Galloway [6].
Table A3. GLRLM features (16) with formulas and interpretation. p ( i , r ) is the normalized GLRLM; p g , p r are marginals; N runs = i , r R i r . Formulas align with IBSI/PyRadiomics [4,10]; names follow Galloway [6].
FeatureFormula/SymbolInterpretation
SRE i , r p ( i , r ) r 2 Short-run emphasis; larger for fine, rapidly varying texture.
LRE i , r p ( i , r ) r 2 Long-run emphasis; larger for coarse, extended uniform regions.
GLN (counts) i r R i r 2 Run mass concentrated in few gray levels (tone nonuniformity).
GLNN i p g ( i ) 2 = GLN N runs 2 Scale-normalized GLN; reduces dependence on run count.
RLN (counts) r i R i r 2 Run mass concentrated in few lengths (scale nonuniformity).
RLNN r p r ( r ) 2 = RLN N runs 2 Scale-normalized RLN; reduces dependence on run count.
RP (IBSI) N runs N p Run density per pixel; higher when runs are more numerous.
LGRE i , r p ( i , r ) i 2 Emphasizes runs at low gray levels (darker tones).
HGRE i , r p ( i , r ) i 2 Emphasizes runs at high gray levels (brighter tones).
SRLGLE i , r p ( i , r ) i 2 r 2 Short runs at low gray levels (fine, dark texture).
SRHGLE i , r p ( i , r ) i 2 r 2 Short runs at high gray levels (fine, bright texture).
LRLGLE i , r p ( i , r ) r 2 i 2 Long runs at low gray levels (coarse, dark texture).
LRHGLE i , r p ( i , r ) i 2 r 2 Long runs at high gray levels (coarse, bright texture).
Run Entropy i , r p ( i , r ) log p ( i , r ) Disorder of the joint run distribution.
RLV (over p r ) r r r r p r ( r ) 2 p r ( r ) Spread of run lengths; multiple scales ⇒ large.
GLV (over p g ) i i i i p g ( i ) 2 p g ( i ) Spread across gray levels among runs.
Angle aggregation stabilizes estimates across directions, and a pixel distance of one aligns with common IBSI defaults [4]. If entropies are reported in bits, replace log with log 2 ; this only rescales values and does not affect rankings. As with other texture matrices, GLRLM features are sensitive to gray-level discretization and to pre-filtering that affects run continuity (e.g., smoothing or denoising); documenting these choices is essential for reproducibility and fair comparison across studies [4,10].

Appendix A.4. GLSZM Definitions and Feature Formulas

The gray-level size-zone matrix (GLSZM) summarizes the distribution of connected zones of equal gray level within a region of interest, independently of direction. It was introduced for image texture analysis by Thibault et al. [7] and subsequently standardized by the Image Biomarker Standardisation Initiative (IBSI) and implemented in reference toolkits such as PyRadiomics [4,10]. In two dimensions we use 8-connectivity to define zones, as recommended in IBSI for enhanced stability across orientations [4].
Formally, let Z R + N g × S max be the matrix where Z i s counts the number of 8-connected zones of size s observed at gray level i in the discretized image. Let N z = i = 1 N g s = 1 S max Z i s denote the total number of zones, and let N p be the number of pixels (voxels in 2D) in the ROI. Define the normalized joint distribution p ( i , s ) = Z i s / N z with marginals
p g ( i ) = s = 1 S max p ( i , s ) ( gray-level profile ) , p s ( s ) = i = 1 N g p ( i , s ) ( size profile ) .
Unless stated otherwise, entropies use the natural logarithm (as in IBSI) [4].
Table A4. GLSZM features (16) with formulas and interpretation. p ( i , s ) is the normalized GLSZM; p g , p s are marginals; N z = i , s Z i s (zones); N p is pixel count. Formulas align with IBSI/PyRadiomics [4,7,10].
Table A4. GLSZM features (16) with formulas and interpretation. p ( i , s ) is the normalized GLSZM; p g , p s are marginals; N z = i , s Z i s (zones); N p is pixel count. Formulas align with IBSI/PyRadiomics [4,7,10].
FeatureFormula/SymbolInterpretation
SAE i , s p ( i , s ) s 2 Emphasizes small zones (fine/fragmented textures).
LAE i , s p ( i , s ) s 2 Emphasizes large zones (coarse/homogeneous regions).
GLN (counts) i s Z i s 2 Zone mass concentrated in few gray levels (tone nonuniformity).
GLNN i p g ( i ) 2 = GLN N z 2 Scale-normalized GLN; reduces dependence on zone count.
SZN (counts) s i Z i s 2 Zone mass concentrated in few sizes (size nonuniformity).
SZNN s p s ( s ) 2 = SZN N z 2 Scale-normalized SZN; reduces dependence on zone count.
ZP (IBSI) N z N p Zone density per pixel; higher when zones are numerous.
LGZE i , s p ( i , s ) i 2 Emphasizes low gray-level zones (darker tones).
HGZE i , s p ( i , s ) i 2 Emphasizes high gray-level zones (brighter tones).
SA_LGLE i , s p ( i , s ) i 2 s 2 Small zones at low gray levels (fine, dark patterns).
SA_HGLE i , s p ( i , s ) i 2 s 2 Small zones at high gray levels (fine, bright patterns).
LA_LGLE i , s p ( i , s ) s 2 i 2 Large zones at low gray levels (coarse, dark regions).
LA_HGLE i , s p ( i , s ) i 2 s 2 Large zones at high gray levels (coarse, bright regions).
Zone Entropy i , s p ( i , s ) log p ( i , s ) Disorder of gray-level/size distribution.
Zone Size Var. (over p s ) s s s s p s ( s ) 2 p s ( s ) Spread of zone sizes; multiscale structures ⇒ large.
Zone Size Mean (over p s ) s s p s ( s ) Average zone size; complements SAE/LAE.
Using 8-connectivity reduces artificial fragmentation of zones and aligns with IBSI-compliant choices [4]. Reporting entropies in bits is possible by replacing log with log 2 ; this only rescales values and does not affect rank-based comparisons. As with other matrix families, GLSZM features are sensitive to gray-level discretization and to any pre-filtering that modifies connectivity (e.g., smoothing or morphological operations). Thorough documentation of these settings is essential for reproducibility and fair cross-study comparison [4,7,10].

Appendix A.5. GLDM Definitions and Feature Formulas

The gray-level dependence matrix (GLDM) quantifies, for each discretized gray level, how many neighbors in a fixed window are “dependent” on the center pixel, i.e., how many neighbors have a gray-level difference within a user-set tolerance. The concept traces back to the neighboring gray-level dependence framework of Sun & Wee [8] and has been standardized by the Image Biomarker Standardisation Initiative (IBSI) and implemented in reference toolkits such as PyRadiomics [4,10]. In two dimensions we adopt a Chebyshev-1 neighborhood (up to eight neighbors); unless stated otherwise, the dependence tolerance is α = 0 , so only exactly equal gray levels (after discretization) are considered dependent.
Let D R + N g × D max be the matrix whose entry D i d is the number of pixels of gray level i having exactly d dependent neighbors within the Chebyshev-1 neighborhood including the center. In 2D with the center included, D max = 9 (eight neighbors plus the center); implementations that exclude the center use D max = 8 with analogous formulas. Define the total event count N d = i = 1 N g d = 1 D max D i d and the normalized joint distribution
p ( i , d ) = D i d N d , p g ( i ) = d = 1 D max p ( i , d ) , p d ( d ) = i = 1 N g p ( i , d ) .
Unless otherwise specified, entropies use the natural logarithm (as in IBSI) [4].
Table A5. GLDM features (14) with formulas and interpretation. p ( i , d ) is the normalized GLDM; p g , p d are marginals; N d = i , d D i d . Formulas follow IBSI/PyRadiomics [4,8,10].
Table A5. GLDM features (14) with formulas and interpretation. p ( i , d ) is the normalized GLDM; p g , p d are marginals; N d = i , d D i d . Formulas follow IBSI/PyRadiomics [4,8,10].
FeatureFormula/SymbolInterpretation
SDE i , d p ( i , d ) d 2 Small-dependence emphasis; larger for fine/fragmented textures.
LDE i , d p ( i , d ) d 2 Large-dependence emphasis; larger for smooth/homogeneous regions.
GLN i p g ( i ) 2 Events concentrated in few gray levels (tone nonuniformity).
GLNN GLN N d 2 Scale-normalized GLN; reduces dependence on total events.
DN d p d ( d ) 2 Events concentrated in few dependence sizes (scale nonuniformity).
DNN DN N d 2 Scale-normalized DN; reduces dependence on total events.
DP N d N g D max Dependence event density (implementation-consistent).
LGLE i , d p ( i , d ) i 2 Emphasizes low gray-level events (darker tones).
HGLE i , d p ( i , d ) i 2 Emphasizes high gray-level events (brighter tones).
SDLGLE i , d p ( i , d ) i 2 d 2 Small dependence at low gray levels (fine, dark patterns).
SDHGLE i , d p ( i , d ) i 2 d 2 Small dependence at high gray levels (fine, bright patterns).
LDLGLE i , d p ( i , d ) d 2 i 2 Large dependence at low gray levels (coherent, dark regions).
LDHGLE i , d p ( i , d ) i 2 d 2 Large dependence at high gray levels (coherent, bright regions).
DEntropy i , d p ( i , d ) log p ( i , d ) Disorder of the joint dependence distribution.
Several implementation details guide interpretation and reproducibility. With α = 0 , only exactly equal neighbors contribute to dependence; enlarging α makes the criterion more permissive and shifts probability mass toward larger d, thereby increasing LDE-type terms and decreasing SDE-type terms [4,10]. In two dimensions, a Chebyshev-1 neighborhood provides up to eight neighbors; including the center yields D max = 9 . Entropy can be reported in bits by replacing log with log 2 ; the choice of base does not affect ranking or correlation analyses. As with other texture matrices, GLDM features are sensitive to gray-level discretization ( N g ) and any pre-filtering that alters local agreement, so these preprocessing choices should be documented for comparability [4]. Finally, the DP denominator ( N g D max ) follows our implementation to provide a consistent density proxy within a fixed configuration; absolute scaling may differ across software, but relative comparisons within a study remain interpretable.

Appendix A.6. NGTDM Definitions and Feature Formulas

For every gray level i { 1 , , N g } , let n i be the number of interior pixels with value i. For those pixels, aggregate neighborhood deviation via
s i = x : X i ( x ) = i i g ¯ neigh ( x ) ,
where g ¯ neigh ( x ) is the mean of the eight neighbors around location x (center excluded). Let N c = i = 1 N g n i be the total counted pixels and define p ( i ) = n i / N c . The average gray level is μ = i = 1 N g i p ( i ) and the weighted average absolute neighborhood deviation is S = i = 1 N g p ( i ) s i . A small constant ϵ (e.g., 10 12 ) is added in denominators to prevent division by zero in nearly homogeneous images.
Table A6. NGTDM features (5) with formulas and interpretation. p ( i ) = n i / N c , s i is the average absolute deviation from the neighborhood mean at gray level i, μ = i i p ( i ) , and S = i p ( i ) s i .
Table A6. NGTDM features (5) with formulas and interpretation. p ( i ) = n i / N c , s i is the average absolute deviation from the neighborhood mean at gray level i, μ = i i p ( i ) , and S = i p ( i ) s i .
FeatureFormula/SymbolInterpretation
Coarseness 1 S + ϵ Larger for smooth/uniform neighborhoods (small S).
Contrast i p ( i ) ( i μ ) 2 N g ( N g 1 ) S Global tone spread modulated by local dissimilarity S.
Busyness S i   j   i   p ( i ) j p ( j ) + ϵ Rate of local change relative to tone separation.
Complexity i j | i j | p ( i ) s i + p ( j ) s j S + ϵ Pairwise tone gaps weighted by local deviations.
Strength i j p ( i ) p ( j ) ( i j ) 2 S + ϵ Quadratic tone separation moderated by local roughness.
In practice, the NGTDM family is driven by the magnitude of S. When S is small, the image is locally smooth and Coarseness and Strength tend to be high, while Busyness becomes low. As S increases—reflecting greater local departures from neighborhood means—Busyness and, often, Contrast increase. Because p ( i ) and μ depend on gray-level discretization, these features are sensitive to the choice of N g and to any pre-filtering that changes neighborhood averages; reporting these preprocessing decisions is essential for reproducibility [4]. Restricting the computation to interior pixels accords with the original definition and helps avoid boundary bias where neighborhoods are incomplete [9].

Appendix A.7. Shape-2D Definitions and Feature Formulas

Let M { 0 , 1 } H × W denote the binary ROI mask (1 = foreground) and ( Δ x , Δ y ) the in-plane pixel spacings. The in-plane area is
A = i = 1 H j = 1 W M i j , A phys = A ( Δ x Δ y ) ,
reported as pixel count or, after scaling, in mm2. Boundary length is approximated by a 4-neighbor edge count. Writing N 4 = { ( 1 , 0 ) , ( 1 , 0 ) , ( 0 , 1 ) , ( 0 , 1 ) } and treating out-of-bounds neighbors as background,
P = i = 1 H j = 1 W ( δ x , δ y ) N 4 1 M i j = 1 , M i + δ x , j + δ y = 0 .
With isotropic spacing ( Δ x = Δ y = Δ ), a first-order physical estimate is P phys Δ P ; if spacings differ, horizontal and vertical edge contributions can be weighted by Δ x and Δ y , respectively. To characterize anisotropy, collect foreground coordinates C = { ( i , j ) : M i j = 1 } , compute c ¯ = 1 | C | ( i , j ) C ( i , j ) , and
Σ = 1 | C | ( i , j ) C i j i j c ¯ c ¯ .
Let λ 1 λ 2 0 be the eigenvalues of Σ . In 2D, IBSI’s elongation ( λ 2 / λ 1 ) and flatness coincide; guard divisions with a small ϵ when λ 1 is near zero, and report undefined values rather than coercing.
Table A7. Shape-2D proxies (4): formulas and interpretation. Convert to physical units using ( Δ x , Δ y ) . In 2D, elongation and flatness are identical; in 3D, IBSI defines distinct measures [4].
Table A7. Shape-2D proxies (4): formulas and interpretation. Convert to physical units using ( Δ x , Δ y ) . In 2D, elongation and flatness are identical; in 3D, IBSI defines distinct measures [4].
FeatureFormula/SymbolInterpretation
Area (A) A = i = 1 H j = 1 W M i j , A phys = A ( Δ x Δ y ) In-plane size of ROI (pixels or mm2).
Perimeter ( P , 4 neigh.) P = i , j ( δ x , δ y ) N 4 1 { M i j = 1 , M i + δ x , j + δ y = 0 } Boundary length on a Manhattan grid; reproducible contour measure.
Elongation (2D) Elong = λ 2 λ 1 + ϵ [ 0 , 1 ] In-plane anisotropy; ≈0 for stretched, ≈1 for compact.
Flatness (2D) Flatness = λ 2 λ 1 + ϵ [ 0 , 1 ] Same expression in 2D; distinct from elongation only in 3D.
Together these four descriptors provide a compact picture of size (area), boundary complexity at the grid scale (perimeter), and anisotropy (elongation/flatness). For volumetric cohorts and standardized reporting, these proxies should be replaced by the full IBSI 3D shape family and computed on the native voxel grid with recorded spacings and resampling details [4].

Appendix B. Associations of Frac and Lac with IBSI Features (Excluding Wavelets)

Table A8. Associations of fractal dimension (Frac) with IBSI features.
Table A8. Associations of fractal dimension (Frac) with IBSI features.
Featurer p r ρ p ρ dCor MIC cos p r adj p ρ adj Composite
GLCM_DiffEntropy0.9811.7 × 10 107 0.97900.9741.0000.9811.4 × 10 105 01.267
GLCM_Contrast0.9692.8 × 10 92 0.98000.9701.0000.9693.4 × 10 91 01.256
GLCM_DiffVariance0.9817.1 × 10 107 0.98200.9780.9680.9812.9 × 10 105 01.250
GLCM_DiffAverage0.9752.0 × 10 98 0.97800.9710.9740.9753.3 × 10 97 01.243
GLCM_Dissimilarity0.9752.0 × 10 98 0.97800.9710.9740.9753.3 × 10 97 01.243
GLCM_IMC10.9493.2 × 10 76 0.94600.9320.9160.9492.5 × 10 75 01.144
GLCM_Entropy0.9493.3 × 10 76 0.94500.9320.9020.9492.5 × 10 75 01.135
GLSZM_SAE0.9471.2 × 10 74 0.94200.9340.8760.9478.2 × 10 74 01.115
GLSZM_SZN0.9463.8 × 10 74 0.94100.9330.8770.9462.2 × 10 73 01.114
GLDM_SDE0.9385.5 × 10 70 0.93400.9220.8670.9382.7 × 10 69 01.090
GLDM_DN0.9351.3 × 10 68 0.93400.9210.8600.9355.1 × 10 68 01.084
GLDM_DNN0.9351.3 × 10 68 0.93400.9210.8600.9355.1 × 10 68 01.084
GLRLM_RLN0.9261.4 × 10 64 0.92400.9050.8420.9264.6 × 10 64 01.049
GLRLM_SRE0.9248.5 × 10 64 0.92200.9020.8320.9242.7 × 10 63 01.039
GLDM_SDLGLE0.8571.7 × 10 44 0.86700.8450.7460.8574.5 × 10 44 00.863
GLSZM_SA_LGLE0.8277.6 × 10 39 0.83500.8180.7320.8271.9 × 10 38 00.795
GLCM_Homogeneity2−0.9746.6 × 10 97 −0.98000.9720.982−0.9749.2 × 10 96 00.681
GLCM_Correlation−0.9695.4 × 10 92 −0.98000.9700.982−0.9695.6 × 10 91 00.679
GLSZM_LGZE0.7861.1 × 10 32 0.79800.7840.6560.7862.5 × 10 32 00.671
GLDM_SDHGLE0.7861.1 × 10 32 0.79200.7760.6590.7862.5 × 10 32 00.665
GLCM_SumEntropy−0.9331.3 × 10 67 −0.98300.9600.982−0.9334.7 × 10 67 00.664
GLCM_IDN−0.9762.5 × 10 99 −0.97700.9700.946−0.9766.8 × 10 98 00.655
GLRLM_SRLGLE0.7781.3 × 10 31 0.80100.7780.6020.7782.8 × 10 31 00.628
GLCM_ID−0.9543.6 × 10 79 −0.94900.9380.933−0.9543.3 × 10 78 00.608
GLCM_IMC2−0.9351.8 × 10 68 −0.94600.9270.916−0.9356.9 × 10 68 00.583
GLRLM_GLVariance0.7663.4 × 10 30 0.77000.7600.5850.7667.2 × 10 30 00.579
GLRLM_LGRE0.7441.0 × 10 27 0.77400.7560.5980.7442.1 × 10 27 00.569
GLRLM_RunEntropy−0.9384.9 × 10 70 −0.93600.9210.895−0.9382.6 × 10 69 00.562
GLSZM_ZoneSizeMean−0.9283.0 × 10 65 −0.9241.5 × 10 63 0.9080.912−0.9281.0 × 10 64 2.6 × 10 63 0.556
GLCM_Homogeneity1−0.9365.1 × 10 69 −0.93100.9160.882−0.9362.2 × 10 68 00.548
GLDM_DEntropy−0.9372.0 × 10 69 −0.93600.9220.870−0.9379.2 × 10 69 00.547
GLSZM_ZoneEntropy−0.9463.5 × 10 74 −0.94300.9330.847−0.9462.2 × 10 73 00.544
GLCM_InverseVariance−0.9406.2 × 10 71 −0.93300.9210.842−0.9403.4 × 10 70 00.527
GLSZM_GLNN−0.9152.4 × 10 60 −0.90900.8930.825−0.9157.5 × 10 60 00.480
GLRLM_GLNN−0.8816.9 × 10 50 −0.87700.8560.808−0.8811.9 × 10 49 00.421
GLCM_Energy−0.8841.1 × 10 50 −0.87700.8580.775−0.8843.1 × 10 50 00.402
GLCM_ASM−0.8832.0 × 10 50 −0.87700.8560.775−0.8835.8 × 10 50 00.400
GLRLM_LRE−0.8455.1 × 10 42 −0.84200.8230.701−0.8451.3 × 10 41 00.305
GLSZM_SA_HGLE0.6137.9 × 10 17 0.64100.6280.5490.6131.5 × 10 16 00.276
GLRLM_RP0.5571.4 × 10 13 0.5921.4 × 10 15 0.5510.7030.5572.3 × 10 13 2.3 × 10 15 0.255
GLDM_LDE−0.8242.2 × 10 38 −0.8261.0 × 10 38 0.8040.658−0.8245.4 × 10 38 1.7 × 10 38 0.252
GLRLM_LRLGLE−0.7384.7 × 10 27 −0.83200.7880.651−0.7389.5 × 10 27 00.218
GLSZM_SZNN−0.7743.4 × 10 31 −0.77600.7590.600−0.7747.5 × 10 31 00.148
GLDM_LDLGLE−0.7312.6 × 10 26 −0.79500.7590.598−0.7315.1 × 10 26 00.146
GLSZM_LA_LGLE−0.6324.0 × 10 18 −0.80500.7280.637−0.6327.5 × 10 18 00.131
GLSZM_GLN−0.7212.4 × 10 25 −0.73700.7320.603−0.7214.5 × 10 25 00.097
GLRLM_SRHGLE0.4902.0 × 10 10 0.5111.3 × 10 11 0.5180.5090.4903.1 × 10 10 2.0 × 10 11 0.009
GLSZM_LAE−0.5571.4 × 10 13 −0.65800.6400.643−0.5572.3 × 10 13 0−0.020
GLRLM_LRHGLE−0.5901.9 × 10 15 −0.62600.6240.622−0.5903.4 × 10 15 0−0.051
GLSZM_ZP0.4997.9 × 10 11 0.3587.1 × 10 6 0.5130.4640.4991.3 × 10 10 1.0 × 10 5 −0.102
GLRLM_RLNN−0.5702.8 × 10 14 −0.63900.6240.503−0.5705.0 × 10 14 0−0.125
GLRLM_GLN−0.5581.2 × 10 13 −0.64600.6170.491−0.5582.0 × 10 13 0−0.136
GLDM_LDHGLE−0.5063.9 × 10 11 −0.54200.5520.548−0.5066.3 × 10 11 0−0.211
GLRLM_RunVariance−0.3889.3 × 10 7 −0.5702.8 × 10 14 0.4410.649−0.3881.4 × 10 6 4.5 × 10 14 −0.225
GLCM_MaxProbability−0.4921.6 × 10 10 −0.4941.3 × 10 10 0.5350.432−0.4922.6 × 10 10 2.0 × 10 10 −0.325
FO_Range0.3752.2 × 10 6 0.3491.4 × 10 5 0.3560.2990.3753.3 × 10 6 2.0 × 10 5 −0.406
FO_Maximum0.3481.3 × 10 5 0.3131.0 × 10 4 0.3400.3110.3481.9 × 10 5 1.5 × 10 4 −0.452
GLSZM_LA_HGLE−0.2236.2 × 10 3 −0.4402.5 × 10 8 0.4080.511−0.2237.9 × 10 3 3.7 × 10 8 −0.452
GLSZM_HGZE0.2245.8 × 10 3 0.2766.6 × 10 4 0.2880.2930.2247.5 × 10 3 9.0 × 10 4 −0.620
FO_Kurtosis0.2679.7 × 10 4 0.2344.0 × 10 3 0.2710.2590.2671.4 × 10 3 5.3 × 10 3 −0.638
NGTDM_Busyness0.2285.0 × 10 3 0.2452.6 × 10 3 0.2570.2690.2286.6 × 10 3 3.5 × 10 3 −0.667
NGTDM_Contrast−0.2541.7 × 10 3 −0.3041.7 × 10 4 0.3380.325−0.2542.4 × 10 3 2.3 × 10 4 −0.678
GLSZM_ZoneVariance−0.0317.0 × 10 1 −0.3159.2 × 10 5 0.2500.458−0.0317.1 × 10 1 1.3 × 10 4 −0.699
GLRLM_HGRE0.1585.3 × 10 2 0.2334.3 × 10 3 0.2650.2990.1586.3 × 10 2 5.5 × 10 3 −0.711
GLCM_SumAverage−0.2363.6 × 10 3 −0.2061.1 × 10 2 0.2550.294−0.2364.8 × 10 3 1.4 × 10 2 −0.805
NGTDM_Complexity0.1951.7 × 10 2 0.0723.8 × 10 1 0.2260.2580.1952.0 × 10 2 4.3 × 10 1 −0.819
FO_MAD−0.2679.5 × 10 4 −0.2197.1 × 10 3 0.2530.235−0.2671.4 × 10 3 9.1 × 10 3 −0.827
FO_rMAD−0.2502.0 × 10 3 −0.2119.6 × 10 3 0.2420.250−0.2502.7 × 10 3 1.2 × 10 2 −0.834
NGTDM_Coarseness0.1012.2 × 10 1 0.1506.7 × 10 2 0.2400.2450.1012.5 × 10 1 7.9 × 10 2 −0.855
FO_P90.90%−0.2197.2 × 10 3 −0.1595.2 × 10 2 0.2330.235−0.2199.1 × 10 3 6.2 × 10 2 −0.887
FO_IQR.75%−0.2091.0 × 10 2 −0.1743.4 × 10 2 0.2070.249−0.2091.3 × 10 2 4.1 × 10 2 −0.887
FO_Skewness0.1191.5 × 10 1 0.0674.1 × 10 1 0.1970.2770.1191.7 × 10 1 4.5 × 10 1 −0.891
FO_P10.10%0.0952.5 × 10 1 0.0823.2 × 10 1 0.1840.2360.0952.8 × 10 1 3.6 × 10 1 −0.937
NGTDM_Strength−0.0694.0 × 10 1 −0.0713.8 × 10 1 0.2430.285−0.0694.2 × 10 1 4.3 × 10 1 −0.942
FO_Mean0.1379.4 × 10 2 0.0703.9 × 10 1 0.0000.2940.1371.1 × 10 1 4.4 × 10 1 −0.973
FO_Median.50%−0.0029.8 × 10 1 −0.0416.2 × 10 1 0.1860.306−0.0029.8 × 10 1 6.2 × 10 1 −0.997
FO_Minimum−0.0803.3 × 10 1 −0.0872.9 × 10 1 0.1670.243−0.0803.6 × 10 1 3.4 × 10 1 −0.999
FO_Energy−0.0803.3 × 10 1 −0.0565.0 × 10 1 0.1170.218−0.0803.6 × 10 1 5.3 × 10 1 −1.060
FO_TotalEnergy−0.0803.3 × 10 1 −0.0565.0 × 10 1 0.1170.218−0.0803.6 × 10 1 5.3 × 10 1 −1.060
GLDM_LGLE0.0084.6 × 10 1 0.0555.0 × 10 1 0.0000.2440.0084.7 × 10 1 5.3 × 10 1 −1.125
FO_RMS−0.0694.0 × 10 1 −0.0277.4 × 10 1 0.0000.239−0.0694.2 × 10 1 7.4 × 10 1 −1.132
FO_Variance−0.0773.5 × 10 1 −0.0555.1 × 10 1 0.0000.189−0.0773.7 × 10 1 5.3 × 10 1 −1.145
FO_StdDev−0.0783.4 × 10 1 −0.0535.2 × 10 1 0.0000.185−0.0783.7 × 10 1 5.3 × 10 1 −1.148
FO_Entropy0.0001.00.0001.00.0000.0000.0001.01.0−1.000
FO_Uniformity0.0001.00.0001.00.0000.0000.0001.01.0−1.000
Shape2D_Area0.0001.00.0001.00.0000.0000.0001.01.0−1.000
Shape2D_Perimeter0.0001.00.0001.00.0000.0000.0001.01.0−1.000
Shape2D_Elongation0.0001.00.0001.00.0000.0000.0001.01.0−1.000
Shape2D_Flatness0.0001.00.0001.00.0000.0000.0001.01.0−1.000
GLDM_GLN0.0001.00.0001.00.0000.0000.0001.01.0−1.000
GLDM_GLNN0.0001.00.0001.00.0000.0000.0001.01.0−1.000
GLDM_HGLE0.0001.00.0001.00.0000.0000.0001.01.0−1.000
GLDM_DP−0.0001.00.0001.00.0000.000−0.0001.01.0−1.000
Table A9. Associations of lacunarity (Lac) with IBSI features.
Table A9. Associations of lacunarity (Lac) with IBSI features.
Featurer p r ρ p ρ dCor MIC cos p r adj p ρ adj Composite
GLCM_Correlation0.9611.4 × 10 84 0.96300.9510.9240.9611.1 × 10 83 01.195
GLCM_Homogeneity20.9634.8 × 10 86 0.96200.9520.9130.9635.7 × 10 85 01.189
GLCM_IDN0.9652.7 × 10 88 0.96200.9530.8960.9652.2 × 10 86 01.180
GLCM_IMC20.9611.3 × 10 84 0.96400.9530.8870.9611.1 × 10 83 01.171
GLCM_ID0.9571.3 × 10 81 0.95400.9420.8660.9579.1 × 10 81 01.142
GLCM_InverseVariance0.9524.1 × 10 78 0.95000.9360.8590.9522.6 × 10 77 01.126
GLCM_Homogeneity10.9475.3 × 10 75 0.94300.9290.8460.9473.1 × 10 74 01.105
GLSZM_ZoneSizeMean0.9363.9 × 10 69 0.9311.6 × 10 66 0.9160.8200.9361.8 × 10 68 2.7 × 10 66 1.063
GLSZM_GLNN0.9304.5 × 10 66 0.92400.9090.8130.9301.8 × 10 65 01.043
GLRLM_RunEntropy0.9365.5 × 10 69 0.93000.9160.7850.9362.4 × 10 68 01.037
GLDM_DEntropy0.9317.0 × 10 67 0.92600.9110.7980.9312.9 × 10 66 01.037
GLCM_Energy0.9253.7 × 10 64 0.91800.9030.7950.9251.2 × 10 63 01.020
GLCM_ASM0.9228.5 × 10 63 0.91800.9000.7950.9222.6 × 10 62 01.015
GLSZM_ZoneEntropy0.9175.4 × 10 61 0.91400.8940.8040.9171.6 × 10 60 01.012
GLCM_SumEntropy0.8994.7 × 10 55 0.90600.8950.7830.8991.3 × 10 54 00.977
GLRLM_GLNN0.8981.5 × 10 54 0.89100.8760.7410.8984.1 × 10 54 00.927
GLRLM_LRE0.8702.6 × 10 47 0.86100.8510.7210.8706.7 × 10 47 00.856
GLDM_LDE0.8499.2 × 10 43 0.8431.1 × 10 41 0.8310.6390.8492.3 × 10 42 1.7 × 10 41 0.759
GLSZM_SZNN0.8362.4 × 10 40 0.83300.8250.6450.8365.8 × 10 40 00.742
GLCM_Contrast−0.9611.4 × 10 84 −0.96200.9510.924−0.9611.1 × 10 83 00.640
GLCM_DiffEntropy−0.9633.8 × 10 86 −0.96300.9520.915−0.9635.3 × 10 85 00.634
GLCM_DiffAverage−0.9659.5 × 10 88 −0.96200.9530.905−0.9652.6 × 10 86 00.629
GLCM_Dissimilarity−0.9659.5 × 10 88 −0.96200.9530.905−0.9652.6 × 10 86 00.629
GLSZM_GLN0.7886.2 × 10 33 0.79200.7950.5970.7881.4 × 10 32 00.625
GLCM_IMC1−0.9641.2 × 10 86 −0.96400.9520.887−0.9642.0 × 10 85 00.616
GLCM_Entropy−0.9641.2 × 10 86 −0.96400.9520.887−0.9642.0 × 10 85 00.616
GLCM_DiffVariance−0.9621.9 × 10 85 −0.96000.9490.880−0.9621.9 × 10 84 00.607
GLSZM_SAE−0.9261.3 × 10 64 −0.92100.9060.860−0.9264.6 × 10 64 00.534
GLSZM_SZN−0.9253.2 × 10 64 −0.92000.9040.860−0.9251.1 × 10 63 00.532
GLRLM_LRHGLE0.7093.4 × 10 24 0.73000.7380.6630.7097.2 × 10 24 00.531
GLDM_DN−0.9384.5 × 10 70 −0.93200.9190.820−0.9382.2 × 10 69 00.525
GLDM_DNN−0.9384.5 × 10 70 −0.93200.9190.820−0.9382.2 × 10 69 00.525
GLDM_SDE−0.9384.0 × 10 70 −0.93100.9180.820−0.9382.2 × 10 69 00.524
GLRLM_RLN−0.9283.7 × 10 65 −0.92300.9060.792−0.9281.4 × 10 64 00.489
GLRLM_SRE−0.9279.9 × 10 65 −0.92000.9050.779−0.9273.6 × 10 64 00.478
GLSZM_LAE0.6482.9 × 10 19 0.72800.7160.6510.6485.8 × 10 19 00.456
GLDM_SDHGLE−0.8895.0 × 10 52 −0.88700.8810.791−0.8891.3 × 10 51 00.441
GLDM_LDHGLE0.6437.4 × 10 19 0.66000.6800.5590.6431.4 × 10 18 00.326
GLRLM_GLN0.6164.7 × 10 17 0.68600.6740.5400.6168.1 × 10 17 00.302
GLRLM_RLNN0.6324.4 × 10 18 0.68100.6800.5180.6328.2 × 10 18 00.300
GLRLM_LRLGLE0.6042.7 × 10 16 0.67200.6290.4930.6044.6 × 10 16 00.225
GLSZM_SA_HGLE−0.7743.7 × 10 31 −0.79200.7800.655−0.7748.3 × 10 31 00.197
GLRLM_GLVariance−0.8018.4 × 10 35 −0.79400.8010.618−0.8012.0 × 10 34 00.193
GLCM_MaxProbability0.5892.2 × 10 15 0.5636.2 × 10 14 0.6280.4780.5893.5 × 10 15 9.3 × 10 14 0.134
GLDM_LDLGLE0.5892.2 × 10 15 0.62400.5980.4150.5893.6 × 10 15 00.110
GLSZM_LA_LGLE0.5141.7 × 10 11 0.65100.5830.4640.5142.6 × 10 11 00.086
GLRLM_RunVariance0.4401.8 × 10 8 0.6129.4 × 10 17 0.4970.5570.4402.7 × 10 8 1.4 × 10 16 0.012
GLDM_SDLGLE−0.7184.3 × 10 25 −0.71700.6790.498−0.7189.4 × 10 25 0−0.031
GLSZM_LA_HGLE0.3772.0 × 10 6 0.57200.5260.5760.3772.9 × 10 6 0−0.038
GLRLM_SRHGLE−0.6707.1 × 10 21 −0.67400.6800.543−0.6701.4 × 10 20 0−0.040
GLRLM_RP−0.6305.5 × 10 18 −0.6362.3 × 10 18 0.6220.582−0.6301.0 × 10 17 3.5 × 10 18 −0.082
GLSZM_SA_LGLE−0.6762.2 × 10 21 −0.67100.6430.498−0.6764.5 × 10 21 0−0.092
GLSZM_LGZE−0.6193.0 × 10 17 −0.62000.5990.457−0.6195.4 × 10 17 0−0.193
GLRLM_SRLGLE−0.6193.3 × 10 17 −0.62300.6030.417−0.6195.8 × 10 17 0−0.218
GLRLM_LGRE−0.5816.2 × 10 15 −0.58400.5780.387−0.5819.7 × 10 15 0−0.287
GLSZM_ZP−0.6042.9 × 10 16 −0.4461.1 × 10 8 0.6060.459−0.6044.9 × 10 16 1.6 × 10 8 −0.298
GLSZM_HGZE−0.4622.7 × 10 9 −0.4912.4 × 10 10 0.4830.394−0.4624.0 × 10 9 3.6 × 10 10 −0.428
GLCM_SumAverage0.3683.7 × 10 6 0.3382.4 × 10 5 0.3640.3300.3685.2 × 10 6 3.2 × 10 5 −0.450
FO_P90.90%0.3431.8 × 10 5 0.3041.7 × 10 4 0.3650.3340.3432.3 × 10 5 2.2 × 10 4 −0.488
GLRLM_HGRE−0.3841.2 × 10 6 −0.4267.2 × 10 8 0.4430.375−0.3841.7 × 10 6 1.0 × 10 7 −0.528
GLSZM_ZoneVariance0.1585.4 × 10 2 0.4043.8 × 10 7 0.3220.4520.1586.5 × 10 2 5.4 × 10 7 −0.535
FO_Range−0.3664.0 × 10 6 −0.3821.8 × 10 6 0.3890.396−0.3665.6 × 10 6 2.4 × 10 6 −0.576
NGTDM_Contrast0.2412.9 × 10 3 0.2883.7 × 10 4 0.3360.3390.2413.7 × 10 3 4.8 × 10 4 −0.601
NGTDM_Busyness−0.3491.2 × 10 5 −0.3521.1 × 10 5 0.3760.384−0.3491.6 × 10 5 1.6 × 10 5 −0.615
FO_Minimum0.2422.8 × 10 3 0.2442.7 × 10 3 0.2720.3130.2423.6 × 10 3 3.5 × 10 3 −0.681
NGTDM_Complexity−0.3491.2 × 10 5 −0.2393.3 × 10 3 0.3390.264−0.3491.6 × 10 5 4.1 × 10 3 −0.787
FO_Skewness0.1477.2 × 10 2 0.1713.7 × 10 2 0.2610.3050.1478.6 × 10 2 4.4 × 10 2 −0.820
FO_P10.10%0.1595.1 × 10 2 0.1379.4 × 10 2 0.2500.3130.1596.3 × 10 2 1.1 × 10 1 −0.831
FO_Median.50%−0.2601.3 × 10 3 −0.1971.6 × 10 2 0.3120.286−0.2601.7 × 10 3 1.9 × 10 2 −0.840
NGTDM_Strength0.0932.6 × 10 1 0.0773.5 × 10 1 0.3050.3380.0932.9 × 10 1 3.9 × 10 1 −0.876
FO_Maximum−0.2129.1 × 10 3 −0.2119.6 × 10 3 0.2460.231−0.2121.1 × 10 2 1.2 × 10 2 −0.921
NGTDM_Coarseness−0.0813.3 × 10 1 −0.1341.0 × 10 1 0.2760.309−0.0813.6 × 10 1 1.2 × 10 1 −0.936
FO_MAD0.1141.7 × 10 1 0.0962.4 × 10 1 0.2020.2330.1141.9 × 10 1 2.8 × 10 1 −0.978
FO_rMAD0.0882.8 × 10 1 0.0743.7 × 10 1 0.2010.2120.0883.1 × 10 1 4.1 × 10 1 −1.029
FO_Kurtosis−0.1071.9 × 10 1 −0.0942.5 × 10 1 0.1960.227−0.1072.2 × 10 1 2.9 × 10 1 −1.055
FO_Energy0.0793.3 × 10 1 0.0674.2 × 10 1 0.1170.2480.0793.6 × 10 1 4.5 × 10 1 −1.065
FO_TotalEnergy0.0793.3 × 10 1 0.0674.2 × 10 1 0.1170.2480.0793.6 × 10 1 4.5 × 10 1 −1.065
FO_IQR.75%0.0356.7 × 10 1 0.0267.5 × 10 1 0.1690.2740.0356.7 × 10 1 7.7 × 10 1 −1.080
FO_Mean−0.1211.4 × 10 1 −0.0555.1 × 10 1 0.0000.316−0.1211.7 × 10 1 5.3 × 10 1 −1.125
FO_Variance0.0753.6 × 10 1 0.0634.4 × 10 1 0.0000.2450.0753.7 × 10 1 4.7 × 10 1 −1.139
FO_StdDev0.0753.6 × 10 1 0.0614.6 × 10 1 0.0000.2250.0753.7 × 10 1 4.8 × 10 1 −1.154
GLDM_LGLE−0.0056.7 × 10 1 −0.0168.5 × 10 1 0.0000.218−0.0056.7 × 10 1 8.6 × 10 1 −1.251
FO_RMS0.0356.7 × 10 1 0.0039.7 × 10 1 0.0000.1620.0356.7 × 10 1 9.7 × 10 1 −1.269
FO_Entropy0.0001.00.0001.00.0000.0000.0001.01.0−1.000
FO_Uniformity0.0001.00.0001.00.0000.0000.0001.01.0−1.000
Shape2D_Area0.0001.00.0001.00.0000.0000.0001.01.0−1.000
Shape2D_Perimeter0.0001.00.0001.00.0000.0000.0001.01.0−1.000
Shape2D_Elongation0.0001.00.0001.00.0000.0000.0001.01.0−1.000
Shape2D_Flatness0.0001.00.0001.00.0000.0000.0001.01.0−1.000
GLDM_GLN0.0001.00.0001.00.0000.0000.0001.01.0−1.000
GLDM_GLNN0.0001.00.0001.00.0000.0000.0001.01.0−1.000
GLDM_HGLE0.0001.00.0001.00.0000.0000.0001.01.0−1.000
GLDM_DP0.0001.00.0001.00.0000.0000.0001.01.0−1.000

References

  1. Gillies, R.J.; Kinahan, P.E.; Hricak, H. Radiomics: Images are more than pictures, they are data. Radiology 2016, 278, 563–577. [Google Scholar] [CrossRef]
  2. Mayerhoefer, M.E.; Materka, A.; Langs, G.; Häggström, I.; Szczypińska, A.; Gibbs, P.; Cook, G.J.R. Introduction to radiomics. J. Nucl. Med. 2020, 61, 488–495. [Google Scholar] [CrossRef]
  3. Tian, J.; Dong, D.; Liu, Z.; Wei, J. Radiomics and Its Clinical Application: Artificial Intelligence and Medical Big Data; Academic Press: Cambridge, MA, USA; Elsevier: Amsterdam, The Netherlands, 2021; ISBN 978-0-12-818101-0/978-0-12-818102-7. Available online: https://www.sciencedirect.com/book/9780128181010/radiomics-and-its-clinical-application#book-description (accessed on 28 September 2025).
  4. Zwanenburg, A.; Vallières, M.; Abdalah, M.A.; Aerts, H.J.W.L.; Andrearczyk, V.; Apte, A.; Ashrafinia, S.; Bakas, S.; Beukinga, R.J.; Boellaard, R.; et al. The Image Biomarker Standardisation Initiative: Standardized quantitative radiomics. Radiology 2020, 295, 328–338. [Google Scholar] [CrossRef]
  5. Haralick, R.M.; Shanmugam, K.; Dinstein, I. Textural features for image classification. IEEE Trans. Syst. Man Cybern. 1973, 3, 610–621. [Google Scholar] [CrossRef]
  6. Galloway, M.M. Texture analysis using gray level run lengths. Comput. Graph. Image Process. 1975, 4, 172–179. [Google Scholar] [CrossRef]
  7. Thibault, G.; Angulo, J.; Meyer, F. Advanced statistical matrices for texture characterization: Application to cell nuclei classification. In Proceedings of the Pattern Recognition and Information Processing (PRIP), Minsk, Belarus, 2009; Available online: https://www.thibault.biz/Doc/Publications/AdvancedStatisticalMatrices_ICIP_2011.pdf (accessed on 28 September 2025).
  8. Sun, C.; Wee, W.G. Neighboring gray level dependence matrix for texture classification. Comput. Vis. Graph. Image Process. 1983, 23, 341–352. [Google Scholar] [CrossRef]
  9. Amadasun, M.; King, R. Textural features corresponding to textural properties. IEEE Trans. Syst. Man Cybern. 1989, 19, 1264–1274. [Google Scholar] [CrossRef]
  10. van Griethuysen, J.J.M.; Fedorov, A.; Parmar, C.; Hosny, A.; Aucoin, N.; Narayan, V.; Beets-Tan, R.G.H.; Fillion-Robin, J.-C.; Pieper, S.; Aerts, H.J.W.L. PyRadiomics: Feature Definitions and Implementation Notes. Online Documentation. Available online: https://pyradiomics.readthedocs.io/en/latest/features.html (accessed on 28 September 2025).
  11. Mandelbrot, B.B. The Fractal Geometry of Nature; W.H. Freeman: New York, NY, USA, 1982; Available online: https://en.wikipedia.org/wiki/The_Fractal_Geometry_of_Nature (accessed on 28 September 2025).
  12. Allain, C.; Cloitre, M. Characterizing the lacunarity of random and deterministic fractal sets. Phys. Rev. A 1991, 44, 3552–3558. [Google Scholar] [CrossRef]
  13. Plotnick, R.E.; Gardner, R.H.; O’Neill, R.V. Lacunarity indices as measures of landscape texture. Landsc. Ecol. 1993, 8, 201–211. [Google Scholar] [CrossRef]
  14. Kim, S.; Park, Y.W.; Ahn, S.S.; Choi, H.S.; Kim, E.H.; Kang, S.-G.; Kim, S.H.; Chang, J.H.; Kim, D.W.; Lee, S.-K.; et al. Magnetic resonance imaging–based 3D fractal dimension and lacunarity analyses for meningioma grading. Brain Tumor Res. Treat. 2020, 8, e3. [Google Scholar] [CrossRef]
  15. Battalapalli, D.; Siar, H.; Gutman, D.; Villanueva-Meyer, J.; Cha, S.; Daniels, D.; Hess, C.; Sugrue, L.; Jain, R.; Bilello, M.; et al. Fractal dimension as an imaging biomarker in neuro-oncology: Concepts, methods, and emerging evidence. Front. Physiol. 2023, 14, 1201617. [Google Scholar] [CrossRef]
  16. Paun, M.-A.; Moldoveanu, F.; Dogaru, F.; Stamate, C. Fractal analysis in medical imaging quantification: A review. Front. Biosci. 2022, 27, 66–90. [Google Scholar] [CrossRef]
  17. Yip, S.S.F.; Aerts, H.J.W.L. Applications and limitations of radiomics. Phys. Med. Biol. 2016, 61, R150. [Google Scholar] [CrossRef]
  18. Parmar, C.; Grossmann, P.; Bussink, J.; Lambin, P.; Aerts, H.J.W.L. Machine learning methods for quantitative radiomic biomarkers. Sci. Rep. 2015, 5, 13087. [Google Scholar] [CrossRef]
  19. Scrivener, M.; de Jong, E.E.C.; van Timmeren, J.E.; Kooi, T.; Leijenaar, R.T.H.; Lambin, P.; van Griethuysen, J.J.M.; Aerts, H.J.W.L. Radiomic feature redundancy and collinearity in oncologic imaging: A review. Insights Imaging 2016, 7, 1023–1036. [Google Scholar] [CrossRef]
  20. Qiu, Q.; Duan, J.; Gong, X.; Wang, X.; Xie, H.; Zhang, L.; Sun, Y.; Wang, W.; Xie, Z.; He, J.; et al. Reproducibility and generalizability of radiomic features: A multi-center study on lung cancer CT. Med. Phys. 2019, 46, 4630–4648. [Google Scholar] [CrossRef]
  21. Berenguer, R.; Pastor-Juan, M.d.R.; Canales-Vázquez, J.; Castro-García, M.; Villas, M.V.; Legorburo, F.M.; Sabater, S. Radiomics of CT Features May Be Nonreproducible and Redundant: Influence of CT Acquisition Parameters. Radiology 2018, 288, 407–415. [Google Scholar] [CrossRef]
  22. Traverso, A.; Wee, L.; Dekker, A.; Gillies, R. Repeatability and reproducibility of radiomic features: A systematic review. Eur. Radiol. 2018, 28, 5451–5464. [Google Scholar] [CrossRef]
  23. Shafiq-ul-Hassan, M.; Zhang, G.G.; Latifi, K.; Ullah, G.; Hunt, D.C.; Balagurunathan, Y.; Abdalah, M.A.; Schabath, M.B.; Goldgof, D.G.; Mackin, D.; et al. Intrinsic dependencies of CT radiomic features on voxel size and number of gray levels. Med. Phys. 2017, 44, 1050–1062. [Google Scholar] [CrossRef]
  24. Orlhac, F.; Frouin, V.; Nioche, C.; Ayache, N.; Buvat, I. Validation of a method to compensate multicenter effects affecting CT radiomics. J. Nucl. Med. 2018, 59, 1321–1328. [Google Scholar] [CrossRef]
  25. Kovács, B.B.H. Efficient gliding-box lacunarity. Pattern Anal. Appl. 2024, 27, 68. [Google Scholar] [CrossRef]
  26. Ilmi, H.R.; Khalaf, E.T. Blaze Pose Graph Neural Networks and Long Short-Term Memory for Yoga Posture Recognition. Int. J. Adv. Comput. Inform. 2025, 1, 79–88. [Google Scholar] [CrossRef]
  27. Jumadi, J.; Md Akbar, J.U. Hybrid GRU–KAN Model for Energy Consumption Prediction in Commercial Building Cooling. Int. J. Adv. Comput. Inform. 2025, 1, 69–78. [Google Scholar] [CrossRef]
  28. Erniwati, S.; Afifah, V.; Imran, B. Mask Region-Based Convolutional Neural Network in Object Detection: A Review. Int. J. Adv. Comput. Inform. 2025, 1, 106–117. [Google Scholar] [CrossRef]
  29. Rickisastra, R.; Hariyanto, D.; Apriyansyah, B. Stacked Ensemble of TabNet, Random Forest, and LightGBM for Enhanced Cervical Cancer Prediction. Int. J. Adv. Comput. Inform. 2025, 1, 56–68. [Google Scholar] [CrossRef]
  30. Mallat, S. A Wavelet Tour of Signal Processing, 2nd ed.; Academic Press: Cambridge, MA, USA, 1999; Available online: https://www.sciencedirect.com/book/9780124666061/a-wavelet-tour-of-signal-processing (accessed on 28 September 2025).
  31. Nason, G.P.; Silverman, B.W. The stationary wavelet transform and some statistical applications. In Wavelets and Statistics; Antoniadis, A., Oppenheim, G., Eds.; Lecture Notes in Statistics; Springer: Berlin/Heidelberg, Germany, 1995; Volume 103, pp. 281–299. [Google Scholar] [CrossRef]
  32. Shensa, M.J. The discrete wavelet transform: Wedding the à trous and Mallat algorithms. IEEE Trans. Signal Process. 1992, 40, 2464–2482. [Google Scholar] [CrossRef]
  33. Daubechies, I. Ten Lectures on Wavelets; SIAM: Philadelphia, PA, USA, 1992. [Google Scholar] [CrossRef]
  34. Lambin, P.; Leijenaar, R.T.; Deist, T.M.; Peerlings, J.; De Jong, E.E.; Van Timmeren, J.; Sanduleanu, S.; Larue, R.T.H.M.; Even, A.J.G.; Jochems, A.; et al. Radiomics: The bridge between medical imaging and personalized medicine. Nat. Rev. Clin. Oncol. 2017, 14, 749–762. [Google Scholar] [CrossRef]
  35. Aerts, H.J.; Velazquez, E.R.; Leijenaar, R.T.; Parmar, C.; Grossmann, P.; Carvalho, S.; Bussink, J.; Monshouwer, R.; Haibe-Kains, B.; Rietveld, D. Decoding tumour phenotype by noninvasive imaging using a quantitative radiomics approach. Nat. Commun. 2014, 5, 4006. [Google Scholar] [CrossRef]
  36. Benjamini, Y.; Hochberg, Y. Controlling the false discovery rate: A practical and powerful approach to multiple testing. J. R. Stat. Soc. B 1995, 57, 289–300. [Google Scholar] [CrossRef]
  37. Pearson, K. Note on regression and inheritance in the case of two parents. Proc. R. Soc. Lond. 1895, 58, 240–242. [Google Scholar] [CrossRef]
  38. Spearman, C. The proof and measurement of association between two things. Am. J. Psychol. 1904, 15, 72–101. [Google Scholar] [CrossRef]
  39. Székely, G.J.; Rizzo, M.L.; Bakirov, N.K. Measuring and testing dependence by correlation of distances. Ann. Stat. 2007, 35, 2769–2794. [Google Scholar] [CrossRef]
  40. Reshef, D.N.; Reshef, Y.A.; Finucane, H.K.; Grossman, S.R.; McVean, G.; Turnbaugh, P.J.; Lander, E.S.; Mitzenmacher, M.; Sabeti, P.C. Detecting novel associations in large data sets. Science 2011, 334, 1518–1524. [Google Scholar] [CrossRef]
  41. Salton, G.; McGill, M.J. Introduction to Modern Information Retrieval; McGraw–Hill: New York, NY, USA, 1983. [Google Scholar]
  42. van der Maaten, L.J.P.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  43. McInnes, L.; Healy, J.; Melville, J. UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction. arXiv 2018, arXiv:1802.03426. [Google Scholar] [CrossRef]
  44. Besag, J. Spatial interaction and the statistical analysis of lattice systems. J. R. Stat. Soc. B 1974, 36, 192–236. [Google Scholar] [CrossRef]
  45. Rue, H.; Held, L. Gaussian Markov Random Fields: Theory and Applications; Chapman & Hall/CRC: Boca Raton, FL, USA, 2005. [Google Scholar] [CrossRef]
  46. Lundberg, S.M.; Lee, S.-I. A unified approach to interpreting model predictions. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  47. Kickingereder, P.; Burth, S.; Wick, A.; Götz, M.; Eidel, O.; Schlemmer, H.-P.; Maier-Hein, K.H.; Wick, W.; Bendszus, M.; Radbruch, A.; et al. Radiomic Profiling of Glioblastoma: Identifying an Imaging Predictor of Patient Survival with Improved Performance over Established Clinical and Radiologic Risk Models. Radiology 2016, 280, 880–889. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Overview of the simulation workflow. The compact version summarizes all stages of the pipeline from synthetic texture generation to feature computation, similarity analysis, and visualization.
Figure 1. Overview of the simulation workflow. The compact version summarizes all stages of the pipeline from synthetic texture generation to feature computation, similarity analysis, and visualization.
Radiation 05 00032 g001
Figure 2. Frac versus its top neighbor under the composite score (baseline, no wavelets). The association is monotone with a shallow linear trend.
Figure 2. Frac versus its top neighbor under the composite score (baseline, no wavelets). The association is monotone with a shallow linear trend.
Radiation 05 00032 g002
Figure 3. Lac versus its top neighbor under the composite score (baseline, no wavelets). The association is monotone with a shallow linear trend.
Figure 3. Lac versus its top neighbor under the composite score (baseline, no wavelets). The association is monotone with a shallow linear trend.
Radiation 05 00032 g003
Figure 4. Correlation heatmap for the union of the top neighbors of Frac and Lac (baseline, no wavelets). Red and blue denote positive and negative correlations. The set partitions into a contrast-oriented block (Frac-aligned) and a large-structure block (Lac-aligned).
Figure 4. Correlation heatmap for the union of the top neighbors of Frac and Lac (baseline, no wavelets). Red and blue denote positive and negative correlations. The set partitions into a contrast-oriented block (Frac-aligned) and a large-structure block (Lac-aligned).
Radiation 05 00032 g004
Figure 5. Feature dendrogram based on 1 | r | distances (baseline, no wavelets). Frac clusters with contrast- and difference-type descriptors; Lac clusters with zone/run size and homogeneity descriptors.
Figure 5. Feature dendrogram based on 1 | r | distances (baseline, no wavelets). Frac clusters with contrast- and difference-type descriptors; Lac clusters with zone/run size and homogeneity descriptors.
Radiation 05 00032 g005
Figure 6. Correlation network for the same neighborhood set ( | r | 0.55 ; baseline, no wavelets). Communities reflect the contrast-heavy (Frac-adjacent) and large-structure (Lac-adjacent) groups.
Figure 6. Correlation network for the same neighborhood set ( | r | 0.55 ; baseline, no wavelets). Communities reflect the contrast-heavy (Frac-adjacent) and large-structure (Lac-adjacent) groups.
Radiation 05 00032 g006
Figure 7. PCA biplot of correlation-based loadings (baseline, no wavelets). PC1 spans high-frequency variation (contrast and short run; Frac side) to low-frequency coherence (large area, long run, and homogeneity; Lac side).
Figure 7. PCA biplot of correlation-based loadings (baseline, no wavelets). PC1 spans high-frequency variation (contrast and short run; Frac side) to low-frequency coherence (large area, long run, and homogeneity; Lac side).
Radiation 05 00032 g007
Figure 8. UMAP of feature similarity ( 1 | r | distance; baseline, no wavelets). Frac sits within the contrast-heavy region; Lac anchors a neighboring large-structure region.
Figure 8. UMAP of feature similarity ( 1 | r | distance; baseline, no wavelets). Frac sits within the contrast-heavy region; Lac anchors a neighboring large-structure region.
Radiation 05 00032 g008
Figure 9. t-SNE of feature similarity ( 1 | r | distance; baseline, no wavelets). The contrast-versus-structure separation mirrors the heatmap and PCA.
Figure 9. t-SNE of feature similarity ( 1 | r | distance; baseline, no wavelets). The contrast-versus-structure separation mirrors the heatmap and PCA.
Radiation 05 00032 g009
Figure 10. Frac versus its composite top neighbor (GLCM difference entropy); tight, nearly linear trend ( | r | 0.98 ).
Figure 10. Frac versus its composite top neighbor (GLCM difference entropy); tight, nearly linear trend ( | r | 0.98 ).
Radiation 05 00032 g010
Figure 11. Lac versus its composite top neighbor (GLCM inverse difference normalized, IDN); strong monotone trend ( | r | 0.96 ).
Figure 11. Lac versus its composite top neighbor (GLCM inverse difference normalized, IDN); strong monotone trend ( | r | 0.96 ).
Radiation 05 00032 g011
Figure 12. Correlation heatmap for the union of top neighbors (with wavelets). LH, HL, and HH detail terms join the contrast/difference block; LL approximation terms align with the large-structure block.
Figure 12. Correlation heatmap for the union of top neighbors (with wavelets). LH, HL, and HH detail terms join the contrast/difference block; LL approximation terms align with the large-structure block.
Radiation 05 00032 g012
Figure 13. Dendrogram on 1 | r | distances (with wavelets). Frac clusters with contrast/difference features; Lac clusters with homogeneity/inverse-difference and large-structure statistics.
Figure 13. Dendrogram on 1 | r | distances (with wavelets). Frac clusters with contrast/difference features; Lac clusters with homogeneity/inverse-difference and large-structure statistics.
Radiation 05 00032 g013
Figure 14. Correlation network ( | r | 0.55 ; with wavelets). Communities echo the contrast versus large-structure split.
Figure 14. Correlation network ( | r | 0.55 ; with wavelets). Communities echo the contrast versus large-structure split.
Radiation 05 00032 g014
Figure 15. PCA loading biplot (with wavelets). PC1 spans high-frequency/contrast (Frac side) to coherent/large-structure (Lac side); wavelet details load with the former, LL with the latter.
Figure 15. PCA loading biplot (with wavelets). PC1 spans high-frequency/contrast (Frac side) to coherent/large-structure (Lac side); wavelet details load with the former, LL with the latter.
Radiation 05 00032 g015
Figure 16. UMAP on 1 | r | distances (with wavelets). Frac is embedded in the contrast/wavelet-detail region; Lac anchors the large-structure/homogeneity region.
Figure 16. UMAP on 1 | r | distances (with wavelets). Frac is embedded in the contrast/wavelet-detail region; Lac anchors the large-structure/homogeneity region.
Radiation 05 00032 g016
Figure 17. t-SNE on 1 | r | distances (with wavelets). The two blocks remain visible after adding wavelet features.
Figure 17. t-SNE on 1 | r | distances (with wavelets). The two blocks remain visible after adding wavelet features.
Radiation 05 00032 g017
Table 1. Simulation factors and data-generation settings.
Table 1. Simulation factors and data-generation settings.
ComponentSettingValues/LevelsNotes
ROISize 64 × 64 Fixed across all samples
ReplicatesN1000 (default; configurable)Independent ROIs
Gray levels N g 64IBSI-aligned; mitigates sparsity
DiscretizationMethodQuantile (fixed-N)Equiprobable bins; jitter if needed
BackgroundModelAR(1)-like 2D recursion (Equation (19))Approximately isotropic; fast/stable
Texture ρ U ( 0.4 , 0.85 ) Per-ROI draw
Noise σ U ( 0.7 , 1.3 ) Per-ROI draw
Heterogeneity# blobs 0 , 1 , 2 with probs ( 0.3 , 0.5 , 0.2 ) Interior centers (avoid edges)
(lesion-like)Amplitude A U ( 2 , 5 ) Per blob
Radius r { 6 , , 12 } pxPer blob
StandardizationScaleZero mean, unit SDAfter blob addition
MaskBinaryFull ROIFor shape proxies
SoftwareRuntimeR 4.4.2 (GUI 1.81; Big Sur ARM build 8462)All simulations executed in R
RandomnessSeed20250915Reproducible end-to-end
Table 2. Feature families, wavelet settings, and similarity analysis choices.
Table 2. Feature families, wavelet settings, and similarity analysis choices.
CategorySettingValues/LevelsNotes
TargetsFractal dimension (Frac)Thresholds t { 0.4 , 0.5 , 0.6 } ; boxes b { 2 , 4 , 8 , 16 } Slope of log N vs. log ( 1 / b ) ; mean over t
Lacunarity (Lac)Median threshold ( t = 0.5 ); windows r { 2 , 4 , 8 , 16 } Λ ( r ) = Var ( M ) / E ( M ) 2 + 1 ; mean over r
WaveletTransformUndecimated 2D; Coiflet-1; level 1Symmetric padding; LL/LH/HL/HH
DiscretizationPer sub-band; N g = 64 quantile binsRe-standardize each sub-band
Sub-band featuresFirst-order (19) + GLCM (19) 4 × ( 19 + 19 ) = 152 features/ROI
EfficiencyIntegral imageUsed for FD and Lac O ( 1 ) window sums; avoids partial-window bias
First-order19 featuresEnergy, entropy, quantiles, …From discretized image
GLCM19 features d = 1 ; 0 , 45 , 90 , 135 Symmetric, normalized, aggregated
GLRLM16 featuresSame four anglesAggregated across angles
GLSZM16 features8-connectivityZone counts across sizes
GLDM14 features α = 0 ; Chebyshev radius 18-neighborhood
NGTDM5 features 3 × 3 neighborhoodsCoarseness, contrast, …
Shape-2D4 proxiesArea, perimeter, elongation, flatnessFull-ROI mask
SimilarityMetricsPearson r, Spearman ρ , dCor, MIC, cosineDefinitions in Section 3.4
Multiple testingBH-FDR on r and ρ  p-valuesAcross all comparisons
Composite scoreMean of z-scores of { | r | , | ρ | , dCor , MIC , cos ine } Standardized per metric across features
VisualizationHeatmap setUnion of top-k neighbors ( k = 30 each for Frac/Lac)Pearson correlation; clustered
EmbeddingsDistances 1 | r | ; UMAP and t-SNEDescriptive geometry (not inferential)
Table 3. Feature counts per family used in the simulation.
Table 3. Feature counts per family used in the simulation.
FamilyCount
First-order (spatial)19
GLCM (spatial)19
GLRLM16
GLSZM16
GLDM14
NGTDM5
Shape-2D proxies3
Wavelet sub-bands (LL, LH, HL, HH): FO + GLCM 4 × ( 19 + 19 ) = 152
Fractal dimension (Frac)1
Lacunarity (Lac)1
Total (wavelet-augmented)246
Table 4. Top-10 most similar IBSI descriptors to Frac under the composite similarity score.
Table 4. Top-10 most similar IBSI descriptors to Frac under the composite similarity score.
FeaturePearsonSpearmanDcorMICCosineSimilarity_Composite
GLCM_DiffVariance0.9800.9840.9760.9270.9801.245
GLCM_DiffEntropy0.9800.9800.9720.9190.9801.235
GLCM_DiffAverage0.9750.9790.9690.9150.9751.227
GLCM_Dissimilarity0.9750.9790.9690.9150.9751.227
GLCM_Contrast0.9700.9820.9690.9170.9701.225
GLCM_Entropy0.9490.9490.9330.8030.9491.095
GLCM_IMC10.9490.9490.9330.8030.9491.095
GLSZM_SZN0.9420.9440.9280.8060.9421.086
GLSZM_SAE0.9430.9450.9290.7960.9431.082
GLDM_SDE0.9360.9370.9200.7930.9361.065
Composite score is the z-average of | r | , | ρ | , distance correlation, MIC, and cosine similarity (when available).
Table 5. Top-10 most similar IBSI descriptors to Lac under the composite similarity score.
Table 5. Top-10 most similar IBSI descriptors to Lac under the composite similarity score.
FeaturePearsonSpearmanDcorMICCosineSimilarity_Composite
GLCM_IDN0.9630.9650.9550.8660.9631.159
GLCM_Correlation0.9590.9650.9530.8670.9591.155
GLCM_Homogeneity20.9610.9650.9530.8630.9611.154
GLCM_IMC20.9590.9660.9550.8390.9591.137
GLCM_InverseVariance0.9490.9560.9420.8370.9491.115
GLCM_ID0.9530.9580.9450.8220.9531.112
GLCM_Homogeneity10.9430.9490.9350.8060.9431.081
GLSZM_ZoneSizeMean0.9290.9360.9190.7850.9291.037
GLRLM_RunEntropy0.9330.9370.9210.7690.9331.031
GLDM_DEntropy0.9310.9340.9170.7630.9311.022
Composite score is the z-average of | r | , | ρ | , distance correlation, MIC, and cosine similarity (when available).
Table 6. Top-10 most similar IBSI descriptors to Frac under the composite similarity score considering wavelet features.
Table 6. Top-10 most similar IBSI descriptors to Frac under the composite similarity score considering wavelet features.
FeaturePearsonSpearmanDcorMICCosineSimilarity_Composite
GLCM_DiffVariance0.9780.9820.9740.9250.9781.226
GLCM_Contrast0.9680.9790.9670.9260.9681.213
GLCM_DiffEntropy0.9780.9770.9700.9090.9781.210
GLCM_DiffAverage0.9740.9770.9680.9050.9741.202
GLCM_Dissimilarity0.9740.9770.9680.9050.9741.202
GLCM_IMC10.9460.9460.9310.8380.9461.091
GLCM_Entropy0.9460.9460.9310.8380.9461.091
GLSZM_SAE0.9400.9410.9240.8020.9401.052
GLSZM_SZN0.9400.9390.9230.7890.9401.041
GLDM_DNN0.9330.9330.9160.7760.9331.019
Composite score is the z-average of | r | , | ρ | , distance correlation, MIC, and cosine similarity (when available).
Table 7. Top-10 most similar IBSI descriptors to Lac under the composite similarity score considering wavelet features.
Table 7. Top-10 most similar IBSI descriptors to Lac under the composite similarity score considering wavelet features.
FeaturePearsonSpearmanDcorMICCosineSimilarity_Composite
GLCM_IMC20.9610.9690.9580.8740.9611.157
GLCM_IDN0.9630.9640.9520.8540.9631.137
GLCM_Homogeneity20.9590.9620.9500.8510.9591.129
GLCM_Correlation0.9570.9630.9510.8410.9571.119
GLCM_InverseVariance0.9510.9560.9410.8360.9511.100
GLCM_ID0.9550.9580.9440.8180.9551.093
GLCM_Homogeneity10.9460.9500.9340.7880.9461.050
GLCM_Energy0.9310.9380.9200.7500.9310.990
GLCM_ASM0.9240.9380.9160.7500.9240.981
GLRLM_RunEntropy0.9330.9350.9160.7400.9330.979
Composite score is the z-average of | r | , | ρ | , distance correlation, MIC, and cosine similarity (when available).
Table 8. Compact guidance for using fractal descriptors with IBSI radiomics.
Table 8. Compact guidance for using fractal descriptors with IBSI radiomics.
ScenarioInterpretationRecommended Action
Very high similarity in wavelet setting ( | r | > 0.90 )Fractal measures align with existing contrast, difference, or homogeneity features; limited incremental value.Treat as redundant. Keep a single representative per correlated cluster or exclude Frac and Lac.
Moderate similarity in baseline setting ( 0.50 | r | 0.80 )Partial overlap with potential complementary signal for multiscale irregularity or void patterns.Retain conditionally. Verify added value by nested models or feature importance.
Multiscale heterogeneity is a priori relevantMechanistic rationale favors fractal summaries.Include Frac and Lac with a predefined role. Report sensitivity to discretization and scale choices.
Acquisition or segmentation variability expectedPossible instability under voxel size, reconstruction, noise, and mask changes.Harmonize when appropriate, standardize preprocessing, and run quick checks: resampling, mild blur or noise, and small mask perturbations.
High-dimensional setting (features much greater than samples)Risk of overfitting and unstable selection.Use nested cross-validation, penalized models, and cluster representatives or dimensionality reduction.
Note: Use thresholds as screening heuristics (for example, remove pairs with | r | > 0.90 ) and confirm redundancy with distance correlation when helpful.
Table 9. Stability of composite similarity structure for Frac and Lac under baseline (no wavelet) and wavelet-augmented settings. The composite is the z-average of | r | , | ρ | , distance correlation (dCor), maximal information coefficient (MIC), and cosine similarity. Leave-one-metric-out (LOO) robustness shows the minimum (across five LOO variants) Spearman rank correlation with the original composite and the minimum overlap of the top-30 neighbors.
Table 9. Stability of composite similarity structure for Frac and Lac under baseline (no wavelet) and wavelet-augmented settings. The composite is the z-average of | r | , | ρ | , distance correlation (dCor), maximal information coefficient (MIC), and cosine similarity. Leave-one-metric-out (LOO) robustness shows the minimum (across five LOO variants) Spearman rank correlation with the original composite and the minimum overlap of the top-30 neighbors.
ScenarioSD
(Composite)
IQRMaxTop Feature
(by Composite)
Top
| r |
LOO:
min ρ s
LOO:
Top-30
Overlap
Frac–No Wavelet0.8061.4091.226GLCM difference variance0.9780.2450.700
Frac–Wavelet0.8071.2872.111GLCM difference variance0.9800.0520.867
Lac–No Wavelet0.8201.3451.155GLCM IMC2 (information correlation 2)0.9610.3910.733
Lac–Wavelet0.8231.3731.964GLCM correlation0.9590.2790.933
Notes: (i) SD, IQR, and Max summarize the composite distribution across all IBSI features. (ii) LOO recomputes the composite using four of five metrics; the table reports the most conservative (minimum) Spearman correlation with the original composite ranking and the most conservative top-30 overlap proportion. High overlaps (0.70–0.93) indicate that the neighborhood of Frac and Lac is stable even when the composite recipe is perturbed.
Table 10. Quantitative comparison between 2D and 3D radiomic feature analogs under IBSI definitions. Pilot tests used synthetic volumes ( 64 × 64 × 64 voxels) with isotropic spacing. Correlation values are Spearman rank correlations between 2D and 3D feature vectors across matched texture realizations.
Table 10. Quantitative comparison between 2D and 3D radiomic feature analogs under IBSI definitions. Pilot tests used synthetic volumes ( 64 × 64 × 64 voxels) with isotropic spacing. Correlation values are Spearman rank correlations between 2D and 3D feature vectors across matched texture realizations.
Feature FamilyNeighborhood TopologyDimensional Change2D–3D ρ Relative RuntimeReference(s)
GLCM (gray-level co-occurrence)8 to 26 neighbors2D to 3D0.93–0.95≈4.5×[4,5]
GLRLM (gray-level run length)4 to 13 orientations2D to 3D0.90–0.92≈3.8×[4,6]
GLSZM (gray-level size zone)8 to 26 connectivity2D to 3D0.88–0.91≈5.2×[4,7]
GLDM (gray-level dependence)2D to 3D isotropic kernel2D to 3D0.89–0.93≈4.0×[4,8]
NGTDM (gray-tone difference)Mean over 26-voxel context2D to 3D0.86–0.90≈3.7×[4,9]
Fractal Dimension (box counting)Squares to cubes2D to 3D0.94≈2.8×[11]
Lacunarity (gliding-box)Square to cubic window2D to 3D0.92≈2.9×[12,13]
Notes: Correlations are median Spearman ρ across 100 simulated pairs of 2D and 3D textures. Runtime ratios measured on a 16-core system (Python, version 3.11; available at https://www.python.org/, accessed on 28 October 2025) implementation. All feature definitions follow IBSI 3D guidelines [4].
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zahed, M.; Skafyan, M. Positioning Fractal Dimension and Lacunarity in the IBSI Feature Space: Simulation With and Without Wavelets. Radiation 2025, 5, 32. https://doi.org/10.3390/radiation5040032

AMA Style

Zahed M, Skafyan M. Positioning Fractal Dimension and Lacunarity in the IBSI Feature Space: Simulation With and Without Wavelets. Radiation. 2025; 5(4):32. https://doi.org/10.3390/radiation5040032

Chicago/Turabian Style

Zahed, Mostafa, and Maryam Skafyan. 2025. "Positioning Fractal Dimension and Lacunarity in the IBSI Feature Space: Simulation With and Without Wavelets" Radiation 5, no. 4: 32. https://doi.org/10.3390/radiation5040032

APA Style

Zahed, M., & Skafyan, M. (2025). Positioning Fractal Dimension and Lacunarity in the IBSI Feature Space: Simulation With and Without Wavelets. Radiation, 5(4), 32. https://doi.org/10.3390/radiation5040032

Article Metrics

Back to TopTop