Previous Article in Journal
Colorectal Polyp Segmentation Based on Deep Learning Methods: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Algorithm-Aided Segmentation of Retinal Nerve Fibers Using Fundus Photographs

by
Diego Luján Villarreal
Departamento de Mecatrónica y Biomédica, Escuela de Ingeniería y Ciencias, Instituto Tecnológico y de Estudios Superiores de Monterrey, Monterrey 64700, Mexico
J. Imaging 2025, 11(9), 294; https://doi.org/10.3390/jimaging11090294 (registering DOI)
Submission received: 9 July 2025 / Revised: 11 August 2025 / Accepted: 14 August 2025 / Published: 28 August 2025
(This article belongs to the Special Issue Progress and Challenges in Biomedical Image Analysis—2nd Edition)

Abstract

This work presents an image processing algorithm for the segmentation of the personalized mapping of retinal nerve fiber layer (RNFL) bundle trajectories in the human retina. To segment RNFL bundles, preprocessing steps were used for noise reduction and illumination correction. Blood vessels were removed. The image was fed to a maximum–minimum modulation algorithm to isolate retinal nerve fiber (RNF) segments. A modified Garway-Heath map categorizes RNF orientation, assuming designated sets of orientation angles for aligning RNFs direction. Bezier curves fit RNFs from the center of the optic disk (OD) to their corresponding end. Fundus images from five different databases (n = 300) were tested, with 277 healthy normal subjects and 33 classified as diabetic without any sign of diabetic retinopathy. The algorithm successfully traced fiber trajectories per fundus across all regions identified by the Garway-Heath map. The resulting trace images were compared to the Jansonius map, reaching an average efficiency of 97.44% and working well with those of low resolution. The average mean difference in orientation angles of the included images was 11.01 ± 1.25 and the average RMSE was 13.82 ± 1.55. A 24-2 visual field (VF) grid pattern was overlaid onto the fundus to relate the VF test points to the intersection of RNFL bundles and their entry angles into the OD. The mean standard deviation (95% limit) obtained 13.5° (median 14.01°), ranging from less than 1° to 28.4° for 50 out of 52 VF locations. The influence of optic parameters was explored using multiple linear regression. Average angle trajectories in the papillomacular region were significantly influenced (p < 0.00001) by the latitudinal optic disk position and disk–fovea angle. Given the basic biometric ground truth data (only fovea and OD centers) that is publicly accessible, the algorithm can be customized to individual eyes and distinguish fibers with accuracy by considering unique anatomical features.

1. Introduction

Any State-of-the-Art clinical device measuring structural damage must have a map that connects local regions on the optic disk (OD) (or the optic nerve head) to local trajectories of retinal nerve fiber layer (RNFL) bundles on the retina in order to compare the sensitivity loss of standard automatic perimetry [1]. As for glaucoma, which was estimated to affect 76 million people worldwide in 2020 [2], a topographically accurate prediction of restricted visual field loss due to a limited optic nerve head or peripapillary RNFL injury requires a thorough understanding of RNFL bundle traces, which sequentially is a requirement for developing new diagnostic techniques, such as fundus-oriented perimetry [3] and scotoma-oriented perimetry [4]. Combining perimetry and imaging in the case of glaucoma is also therapeutically useful for integrating the findings from structural and functional tests and increasing the sensitivity and specificity of illness detection and progression assessments [5]. Structure–function research is currently focused on mapping the spatial connection between retinal locations and peripapillary RNFL sectors [6]. These precise topographical localizations of the path of RNFL bundles on the retinal surface and their point of entrance into the OD are necessary for precise analyses of visual field defects brought on by localized optic nerve damage typically observed and detected via optical coherence tomography (OCT) and fundus images [1]. In particular, eye fundus images detailed as color photographs also provide valuable information about the condition of the retina and its key structures: the macula, fovea (FO), OD, vascularization, and the distribution of RNFL bundles. Similarly, imaging data could be helpful in creating customized functional tests that are specific to a patient and tailoring the areas examined during a perimetric technique [5]. Research into the composition and functioning of RNFL bundles is fundamental to understanding the visual system [6]. According to studies by Iwata et al. [7], inspecting and assessing RNFL bundles proves to be a valuable diagnostic tool for several types of optic neuropathy. Consistent with Diekmann et al. [8], axonal damage to the optic disk is the initial trigger for the pathogenesis of glaucoma, which leads to the degradation and apoptosis of retinal ganglion cells, the loss of retinal nerve fibers, and the thinning of the nerve fiber layer. Early diagnosis of glaucoma is critical, as the currently established treatment options can prevent or slow the progression of damage, but are not yet able to repair the loss of RGC axons, which results in lost connections to the central visual system [6].
Numerous attempts have been made to describe RNFL bundle trajectories, and several different techniques applied to patient populations to relate OD and RNFL traces (or visual field locations) have been published. Garway-Heath et al. [9] traced the detectable RNFL bundles of 69 eyes from the OD to visual field points on a 24-2 grid pattern overlaid on fundus photographs. Lamparter et al. [10] traced the RNFL bundle distribution of 100 eyes by setting a minimum of three marker points and connecting them using a nonparametric cubic spline model, and providing the impact of ocular parameters on the mapping of retinal locations to the OD. In a similar manner, Jansonius et al. [11] traced the visible RNFL bundles of 55 eyes in fundus photos and then fitted a mathematical model to the data, enabling extrapolation to additional central retinal regions. Denniss et al. [5] recently developed a simulation model of the association between OD sectors and visual field point locations for a range of clinically plausible anatomical parameters. While time-consuming and subjective, manual tracing by experts on fundus photographs or OCT images serves as a “ground truth” or a baseline for comparison. Yet, a significant distinction between an automatic generated model and those currently published represents significant potential for betterment. The ground truth models necessitate either very costly ophthalmic equipment to obtain some of the required input data [5,12] or excessively long man-hours of labor and time-consuming procedures to manually [9,10] or electronically [11,13] trace RNFL bundles as far as a person can see. Therefore, it is required to generate a publicly available automated model that generates personalized RNFL bundle maps of individual eyes with high precision using publicly available inputs.
Recent advances in artificial intelligence have demonstrated how deep learning algorithms can be used to automatically learn task-specific features from high-dimensional data in a variety of medical domains and model intricate non-linear relationships. The primary benefit of the deep learning algorithm is that it is an end-to-end learning algorithm, meaning that rather than requiring a specific mechanism to tackle complicated issues, it learns these mechanisms during training. Zhiqi et al. [14] developed a deep learning model to estimate pointwise functional outcomes directly from segmentation-free 3D OCT volumes and compared the performance with the model trained with segmentation-dependent 2D OCT thickness maps. Park et al. [15] developed a deep learning architecture according to Inception V3 to predict visual fields using OCT imaging. A convolutional neural network architecture was constructed to predict visual fields using the combination of the acquired two OCT images, the macular ganglion cell–inner plexiform layer and peripapillary retinal nerve fiber layer thicknesses. Globally (the entire visual field area), for all patients, the root mean square error (RMSE) between the actual and predicted visual fields was 4.79 ± 2.56 dB, with 3.27 dB and 5.27 dB for the normal and glaucoma groups, respectively. Christopher et al. [16] developed a deep learning architecture to predict glaucomatous visual fields from OCT images. ResNet was employed as the deep learning architecture for predicting visual field global indices. Several SD-OCT OD images (RNFL thickness maps, RNFL en-face images, and confocal scanning laser ophthalmoscopic images) were inputted and the predictions for each type were compared. The best mean absolute errors between the real and predicted values were 2.5 dB (mean deviation) and 1.5 dB (standard deviation). To better understand how neural systems function and provide measurements for glaucoma progression analysis [17], it is crucial to gather the most relevant anatomical, physiological, and pathological data available about the nervous system’s structure. Similarly, an algorithmic methodology can help to direct efforts towards generating RNFL tracings using computerized models [1,5,9,10,11,12,13,18,19,20,21,22], and also towards diagnosing RNFL defects by analyzing fundus images using image processing techniques [23]. Despite the advances, artificial intelligence’s efforts exclude the segmentation of RNFL bundle trajectories, which is crucial for a topographically accurate prediction of the restricted visual field loss [3], the reach specificity of illness detection and progression assessments [5], the localization of damage and the structure–function correlation [9], and the quantitative assessment and monitoring of disease progression [23]. Deep learning training models have still not yet fully or robustly solved the problem of RNFL bundle segmentation. Several reasons for this include the variability in image acquisition protocols and database characteristics (e.g., resolution, patient demographics). Narrow population data reduces group accuracy, and retinal morphology variations across populations can amplify this effect, leading to disparities in disease results [24]. Distinguishing between histologically distinct sub-layers is challenging for deep learning models because of the reduced resolution of deeper tissues caused by the light intensity decreasing with depth. Similarly, segmentation methods can be complicated by speckle noise, which is a natural feature of OCT images [25].
The purpose of this work is to develop an algorithm for the segmentation of personalized mapping estimates of RNFL bundle tracings using image processing techniques. The algorithm considers personalized anatomical parameters that can be customized to individual eyes using input ground truth data of the FO and OD centers. The algorithm was verified via two different tests. In the first test, (1) a quantitative comparison of the pointwise angles of orientation between the Jansonius map and the algorithm was computed using five distinct fundus image datasets. This local angular difference tests the performance of the curvature-related analysis of the RNFL bundles in the algorithm. The difference in the pointwise angles of orientation was evaluated using the root mean square error, the mean, and the standard deviation. In the second test, (2) a quantitative comparison of the entry angles of RNFL bundles into the OD was carried out between Garway-Heath et al. [9], Lamparter et al. [10], Jansonius et al. [13], and the algorithm. This quantitative comparison tests the performance efficacy of the RNFL bundles’ convergence at the OD by quantitatively assessing the topographical relationship between particular OD regions and the associated retinal RNFL bundles. The RMSE values of the average mean entry angles of [9,10,13] and that of the algorithm were also computed. Possible bias and variation were inspected between individual maps according to personalized anatomy. Visible nerve fiber bundle tracings found using the algorithm were the foundation for comparing bias and variation with previous studies. The computational algorithm provides robust results and holds the advantage that any given combination of anatomical parameters can produce individual maps, involving those that are uncommon and so unlikely to occur in small-scale population research.

2. Materials and Methods

2.1. Subjects and Fundus Photographs

Digitized fundus images of 300 eyes of 300 subjects (277 healthy individuals without any sign of a history of ocular diseases and 33 diabetic subjects without any sign of diabetic retinopathy) were collected from five different databases: the High-Resolution Fundus (HRF) [26], Joint Shantou International Eye Centre (JSIEC) [27], Retina Identification Database (RIDB) [28], DRIVE [29], and MESSIDOR [30]. According to [31], diabetes does not affect the trajectories or their visibility as long as there is no retinopathy. The HRF (Friedrich-Alexander University Erlangen—Nuremberg) dataset consists of 45 8-bit RGB fundus photographs of 2336 × 3504 pixels using a Canon CR-1 fundus camera (Canon Inc., Tokyo, Japan) in a 45° field setting. A total of 15 images were extracted from healthy patients. The JSIEC dataset (collected between September 2009 and December 2018) consists of 8-bit RGB fundus images of 3046 × 2572 pixels using a ZEISS FF450 Plus IR Fundus Camera (Carl Zeiss Meditec AG, Jena, Germany) (2009–2013) and Topcon TRC-50DX Mydriatic Retinal Camera (Topcon Corporation, Tokyo, Japan) (2013–2018) in a 35–50° field setting. A total of 38 images were extracted from healthy patients. The RIDB dataset consists of 100 8-bit RGB healthy retinal fundus images (individuals with no retinal disease or anomalies) of 1504 × 1000 pixels, captured using a TOPCON-TRC-50 EX (Topcon Corporation, Tokyo, Japan) Fundoscope in a 45° field setting. A total of 100 images were extracted from healthy patients. The DRIVE database consists of 400 8-bit RGB fundus images of 768 × 584 pixels, captured using a Canon CR5 non-mydriatic 3CCD camera (Canon Inc., Tokyo, Japan) in a 45° field of view. The screening population consisted of 400 diabetic subjects between 25 and 90 years of age. A total of 33 photographs were selected since they did not show any sign of diabetic retinopathy. The MESSIDOR database consists of 1200 8-bit RGB fundus images of 1440 × 960, 2240 × 1488, or 2304 × 1536 pixels, captured using a color video 3CCD camera mounted on a Topcon TRC NW6 (Topcon Corporation, Tokyo, Japan) non-mydriatic retinograph in a 45° field of view. An excel file is provided, with medical diagnoses for each image. Care was taken to select 114 healthy fundus images. The image database collections are thought to comply with all relevant ethical regulations for studies involving human subjects and research protocols, as approved by the corresponding Ethics Committee.

2.2. Image Preprocessing

Fundus images were preprocessed in order to obtain the binary mask field of view (FOV). Unwanted text images such as patient information and neighboring low pixel noise were removed by determining the binary mask FOV, obtained by—in the following order—filtering the fundus using Prewitt operator kernels, image dilation of size nxn with structuring element neighborhoods of 1, thresholding low-pixel areas, and computing the convex hull with eight connected neighborhoods. The binary mask extracted the FOV region of interest (ROI) in the entire image.

2.3. Segmentation of Personalized Mapping Algorithm

Removing blood vessels is essential for accurately detecting RNFL defects and distinguishing the orientation of RNFLs from that of the vessels. Before eliminating the vessels, two separate procedures were performed for reducing noise and illumination correction using the green channel of the fundus image. The green channel maximizes the contrast between blood vessels, RNFL bundles, and neighboring regions, indicating a bright image with high contrast amongst its blue and red channel competitors [32]. For noise reduction, a median filter was applied to the masked image, replacing the pixel value with the median of its neighbors. For correcting illumination, a sigmoid curve function (or S-shaped curve) was used as a non-linear mapping function to adjust the intensity of pixels and improve low-light images. The S-shaped mid-tone section (or the middle of the curve where the slope is at its steepest) stretches the intermediate tones of the original image, making them brighter and increasing their contrast to reveal textures and features that were previously hidden. Several techniques have been reported for blood vessel extraction from fundus images [33,34]. In the current study, blood vessels were removed by applying contrast-limited adaptive histogram equalization (CAHE) [35] for contrast enhancement and dilating the resultant of enhancing elongated structures using a Hessian-based multiscale Frangi vesselness (HFV) filter [36]. Achieving better results involves “cleansing” the image by applying thresholding to low-pixel areas. CAHE amplifies contrast until the output region’s histogram roughly resembles a distributed curve histogram. Contrast enhancement is limited to prevent enhancing any noise that may be present in the image. An alternative algorithm was employed for optimizing the filter’s response by estimating the thickness of the tubular structures present in the fundus. HFV filtering is a prominent case of a feature enhancement filter, particularly for the segmentation and analysis of tubular structures. The elements of the Hessian matrix, also known as second-order local derivatives, quantify the local curvature of the intensity landscape. The elements capture how the gradient of the image intensity changes, which is crucial for identifying distinct geometric patterns [36]. A complementary algorithm preprocesses the green additive primary color of the RGB fundus for retinal nerve fiber (RNF) segmentation via—in the following order—box–kernel low-pass filtering, a modified CAHE algorithm for contrast boosting, and sigmoid function intensity transformation for high-contrast stretching. The preprocessed monochromatic image is fed into a maximum–minimum modulation (MM) algorithm that isolates RNF segments from the surroundings. The MM algorithm redistributes pixel intensities through intensity normalization, which increases the luminance difference between the target segment and its surroundings. As a result, the segment’s borders are more salient for later edge detection and thresholding. The resulting monochromatic image and blood vessel extraction are binarized using the Otsu method, followed by image subtraction. Multiple algorithms for RNF alignment are executed as follows. Unless otherwise stated, RNFs described by their centroid and the angle of orientation are known as “pointwise RNF extraction”. A modified Garway-Heath map [9] classifies the pointwise RNF orientation, assuming expected sets of angles seen from the OD with high variance for adjusting the extractions across several areas. The sets of angles were based on the direction angles per sample point of the Jansonius map seen from the OD center. The direction angles were relocated to a pixel image, placing the FO in the image center. Image dilation was carried out to avoid pixel spacing. Each area was superimposed, and the set of angles was determined by obtaining the maximum and minimum of the angles within the area. The classification used the criterion based on the pointwise RNF angle; any angle falling outside the defined set was excluded from further analysis. This criterion eliminates false positive cases by highlighting non-RNFL bundle structures. For the complete process, see Validation Procedure. A complementary algorithm incorporates “group” and “individual” averages to ensure the continuity of orientation between neighboring extractions. An interconnected algorithm assembles RNFs by identifying and connecting the optimal nearest pointwise extractions based on the orientation angle. The 4°-order Bezier curve function fitted the assembled RNFs. Key representative points were chosen for the fitting procedure, including the OD center, the final, and the highest point. Bezier curves were selected for a homogeneous solution, including the fitting of RNFL bundles that fail the vertical line test (vertical line intersects the curve more than once), typically seen entering the inferonasal sector of the OD and thus presenting significant challenges for fitting procedures. This singularity was observed in Figure 2a in [11], Figure 3 in [13] and Figure 4a,b in [31]. The curves are traced from the center of the OD to their corresponding end. At least five pointwise RNF extractions are required for fitting a 4°-order Bezier curve. The spatial trajectories from local regions on the retina are connected to the OD by intersecting individual bundle maps with the OD boundary to determine the entry OD angle locations. A standard value of 60 degrees was used, since several databases omit the OCT field of view (visual angle).
A MATLAB (MathWorks, Inc., United States, Version 9.14) script and Image Processing ToolboxTM were used to develop and implement the algorithm for the segmentation of personalized estimate mapping. For accessibility, the PES (personalized estimated segmentation) software application was built using a graphical user interface (GUI) in MATLAB by an app designer that offers interactive, user-friendly, and straightforward tools to generate nerve fiber estimate trajectories. The application is ready-to-download and can be used in MALTAB (see Data Availability Statement).

2.4. Ocular Parameters

The influence of the following ocular parameters on the average RNF orientation angle was investigated. The binary mask field of view obtained before calculated the FOV size, known as the radius at the retina (or one half of the mask in the horizontal line). Two algorithms separately enhanced the FO ROI and two algorithms for the OD ROI in order to obtain different results by specifically targeting contour edges for accurate identification. Consideration was first given to excluding fundus images for cases when the FO and/or the OD could not be located precisely. Concerning the assumption of the OD center, the clearest image representation of the OD ROI contour from the two algorithms was chosen to select five recognizable points around the OD boundary, and the best fit for an elliptical-shaped closed contour was computed using the Gal [37] algorithm using the fit estimation method of the Least Squares method. The OD center and horizontal and vertical OD radii were computed. Despite the Gal algorithm being able to identify rotated elliptical shapes, the OD tilt was not used as an independent variable since Spectralis OCT validation is still required. Concerning the FO center, the clearest image representation of the FO ROI contour was chosen to select its center. The position of the OD in relation to the FO is assumed as the horizontal distance between the OD center and FO ( O D x ), and the vertical distance between the OD center and the horizontal meridian ( O D y ). Distances were measured in degrees. The ellipticity ratio is assumed to be the ratio between the shortest and largest OD radius. The disk–fovea angle is assumed to be the inverse function of the tangent of the O D y to O D x ratio, measured in degrees. The OD area is assumed to be the product of the horizontal and vertical OD radii and π, measured in degrees.

2.5. Validation Procedure

2.5.1. Pointwise Validation (1st Test)

The FO and OD center positions were marked on each fundus image. As recently described by Jansonius et al. [11,13], each traced image was superimposed via translation for centering the FO, followed by zooming and rotation for aligning the OD centers at an eccentricity of 15°, 2° above the horizontal meridian. The resulting traces were compared to that of the Jansonius map [11,13] via the following procedure. The Jansonius map was traced with a 1-degree step, and the orientation angles per sample point were calculated as seen from the OD center, centered at (15°, 2°). An algorithm identifies the sample point within the dataset closest to a specific RNF centroid and assigns the orientation angle to that of the Jansonius map. The model’s performance was optimized through iterative adjustment of the limits for each set of angles using coarse sampling interval for computational efficiency. This provides fine-tuning for each set. The percentage of effectiveness was computed for each extraction as follows:
P e f f = 100 1 a i a ^ i 2 π
where a i is the corresponding angle in the dataset closest to a specific RNF extraction for the ith point and a ^ i is the orientation angle for RNF extraction in units of radians. The set yielding the highest average percentage was recorded. A total of 150 out of 300 fundus images were used to fit the nerve fiber estimate trajectories. The remaining 150 images were reserved as a test sample for the independent validation of the model. This quantitative comparison evaluated the performance of the curvature-related analysis of the RNFL bundles in the algorithm. b and c values used in the Jansonius map corresponded to the 95% limit central range covering 95% of the corresponding fits included (see Figures 3 and 5 in [11], and Figures 5 and 6 in [13]). The root mean square error (RMSE) is computed as
R M S E = 1 n i = 1 n a i a ^ i 2
where n is the number of sampled data points in the RNFL bundles. The mean and standard deviation of the difference, i.e., a i a ^ i , was also computed.

2.5.2. Structure–Function Mapping (2nd Test)

A properly scaled Humphrey Field Analyzer of a 24-2 visual field–test grid pattern (test points in a grid 6° apart) was superimposed onto the fundus image, centered in the FO. Only the MESSIDOR fundus image database (114 healthy, nondiabetic, without any sign of diabetic retinopathy) was used, since all images show the entire OCT field of view area with high resolution. The position of the OD in relation to the FO was recorded, and the entry points of RNFL bundles into the OD were related to all visual field locations. The entry angle of each bundle was determined by intersecting individual bundle maps with the OD boundary. The nasal margin (3 o’clock position, right eye) was designated 0°, and degrees were counted in a clockwise direction. Each RNFL bundle had to be within 0.86° (twice the diameter of the Goldman Size III stimulus) of a visual field point. To account for the inverse relationship between retinal location and visual field sites, the superior visual field places were assigned the OD entry location of their mirror image position in the inferior hemifield, and vice versa for inferior visual field locations. After assigning the entrance angles, the RNFL bundles were relocated to a pixel image, placing the FO to the image center, and image dilatation was carried out to avoid pixel spacing. The scaled visual field–test grid pattern was superimposed and the entry angles per visual field point were computed. This quantitative comparison tests the performance efficacy of the RNFL bundles’ convergence at the OD by quantitatively assessing the topographical relationship between particular OD regions and the associated retinal RNFL bundles. The number of pixels in the circumference of the visual sites, which is determined by the chosen image resolution, is equal to the maximum number of crossing bundles. As the OD sectors formed a circular domain (that is, sectors 1 and 360 were adjacent despite their large numerical variance), ranges of mapped OD sectors for each visual field location were examined for discontinuities around the 1/360 border. In all cases where discontinuity was higher than 180°, it was subtracted 360° from all sectors to make the range continuous. The RMSEs of the entry angles at the OD were computed as in Equation (2). n is the number of visual field points, a i is the average mean of the entry angles of [9,10,13], and a ^ i is the entry angles obtained in the algorithm.

2.6. Statistical Analysis

Multiple linear regression studied the associations between individual deviations by determining the influence of the following factors possibly associated with the trajectories: OD area, OD position, disk–fovea angle, ellipticity ratio, and FOV size (pixels). The outcome measure (or dependent variable) is the average RNF orientation angle (of extractions) in the papillomacular region, where noticeable sinuous deviations occur, as the curve’s apex—where the first derivative is zero, indicating a horizontal tangent line—and in some cases the inflection point can be identified along the RNFL bundle trajectory. As the RNF sectors created a circular domain with discontinuities near the adjacent borders of 1/360 and 180/-180, the algorithm evaluated all situations by subtracting 360 from any sectors less than 0 to maintain continuity for the regression and develop a convention of angles. Thus, for the inferior hemifield, the angles ranged 0° to 180°, indicating a transition from the nasal side to the temporal side. Conversely, for the superior hemifield, the angles ranged from 180° to 360°, representing a movement from the temporal side back to the nasal side. Consistent with Leung et al. [38], the papillomacular region includes the papillomacular and papillofoveal bundle regions. It is important to note that multiple lineal regression does not include normalization/rotation techniques, which decrease the between-subject variability by combining the OD and FO centers in line [10]. Multicollinearity was found via Pearson’s correlation matrix. The components with a correlation coefficient greater than 0.5 were included in the multiple linear regression analysis. Statistical significance in the regression was defined as p < 0.00001. All statistical analyses were carried out in Statistics and Machine Learning ToolboxTM in the MATLAB environment.

3. Results

Of the 300 fundus images, 33 were classified as diabetic without any sign of diabetic retinopathy. The algorithm presented here successfully traced RNFL trajectories per fundus across all regions identified by the Garway-Heath map. The algorithm extracted an average of 7854 (HRF), 10,139 (JSIEC), 4539 (RIDB), 7977 (DRIVE), and 4445 (MESSIDOR) RNF centroids per fundus image and per database, for an average estimated total number of 1,732,657 (which is equal to P e f f ). Fundus images and the status of the eye (healthy, glaucoma subject, or glaucoma) were not included in the algorithm, but were instead reserved for future tests according to eye status. The algorithm assembled an average of 6058 (HRF), 8334 (JSIEC), 4138 (RIDB), 5122 (DRIVE), and 3862 (MESSIDOR) RNFs per fundus image and per database, for an average estimated total number of 1,435,128. On average across all fundus images, the proposed algorithm reached 97.44% effectiveness and worked well with fundus images of low resolution. On average, the number of fiber extractions was 6990, which assembled 5502 retinal nerve fibers per fundus image.
Figure 1 shows the block diagram of the algorithm presented here for the segmentation of the personalized mapping of the optic pathways using image processing techniques. The RGB fundus was preprocessed by determining the binary mask field of view (FOV). Noise reduction and illumination correction were applied using a low-pass filter and a low-light enhancement sigmoid curve function. Blood vessels were removed by applying CAHE and Frangi vesselness filter algorithms, see Figure 1b. A complementary algorithm optimizes the filter’s response by estimating the thickness of the blood vessel structures across the fundus. For RNF segmentation, a complementary algorithm processes the green channel of the RGB fundus image, which is then fed in a maximum–minimum modulation algorithm to isolate RNF segments. The monochromatic image and extracted blood vessels are binarized, followed by image subtraction, see Figure 1c.
A modified Garway-Heath map categorizes RNF orientation, assuming designated sets of angles for aligning RNFs’ direction across various regions. The convention of angles is shown around the OD boundary, see Figure 1d. An additional algorithm utilizes several averages to guarantee the consistency of orientation among adjacent extractions, see Figure 1e. A connected algorithm compiles RNFs by connecting the nearest optimal extraction based on the orientation angle. Bezier curves were used to fit the assembled RNFs from the center of the OD to their corresponding end. The spatial trajectories on the retina are connected to the OD by crossing individual bundle maps with the OD boundary to determine the entry OD angle locations (shown beside colormap, see Figure 1f).
Figure 2 shows the percentage of effectiveness per database as part of the validation process. The fundus images were superimposed per database and to the Jansonius map via translation for centering the FO, followed by zooming and rotation for aligning the OD centers at 15° of eccentricity, 2° over the horizontal meridian. The percentage of effectiveness is computed as in Equation (1).
The algorithm extracted an average of 117,810 (HRF), 395,421 (JSIEC), 453,900 (RIDB), 263,241 (DRIVE), and 502,285 (MESSIDOR) RNF centroids per database. A modified Garway-Heath map [9] used to classify the results in Table 1 is shown. For validation, the frequencies (percentage) higher than 92.5 (percentage of effectiveness) were summed per region and database and shown in Table 1. The average per database is 97.22% (HRF), 94.30% (JSIEC), 97.82% (RIDB), 94.37% (DRIVE), and 97.32% (MESSIDOR). Figure 3 shows the distribution of RNFL bundles and their consistent visual field positions.
At the top are the quantitative comparison of the spreading of mapped OD sectors from this study and the outcomes reported by Garway-Heath et al. [9], Lamparter et al. [10], and Jansonius et al. [13]. Each plot corresponds to the hemifield regions (superior and inferior) of the Garway-Heath et al. map [9], and each row within corresponds to the visual field sites. The blue line represents the mean standard deviation (95% limit) as reported by Garway-Heath et al. [9]. The red line denotes the mean and upper/lower limits of the predicted OD sectors from the Jansonius map. The green line represents the mean standard deviation as published by Lamparter et al. [10]. The black line represents the mean standard deviation (95% limit) as reported in the current study. At the bottom left is the RNF bundle distribution; 34,718 RNFL bundles, based on roughly 48,930 pointwise RNF extractions, were derived from only seven subjects. A total of 52 visual field points were superimposed onto the RNFL bundle image. For reporting, the superior visual field sites were assigned to the OD entry position of their mirror image location in the inferior hemifield, and vice versa for the inferior visual field site. At the bottom right is the RNFL bundle distribution.
Figure 4 shows the quantitative comparison of the root mean square error, the mean, and the standard deviation of the differences in orientation angles between this study and the results reported in the Jansonius map.
Table 2 presents the results of multiple regression that show that vertical OD distance and disk–fovea angle are statistically significant on the average RNF orientation (R2 = 0.64, β = −4.43, p < 0.00001) and (R2 = 0.62, β = −1.56, p < 0.00001), respectively. Figure 5 shows (left side) the influence of O D y on the average RNF orientation angle in the papillomacular region. The average RNF orientation angle along the y-axis is subtracted from 180. On the right side, standardized residuals are shown against the expected value for O D y .

4. Discussion

This work addresses the inherent challenges of accurately segmenting RNFL bundle traces in the human retina, particularly by circumventing the limitations often encountered in deep learning-based methodologies. While deep learning models excel in automatically learning task-specific features from high-dimensional data and modeling intricate non-linear relationships [14], they can hinder their clinical applicability and interpretability in complex anatomical tasks [24,25]. In contrast, the proposed algorithm leverages advanced image processing techniques to intrinsically identify and delineate RNFL bundles, reducing dependency on vast training data and offering greater transparency in its decision-making process. A key innovation lies in the integral use of a structure–function map, which guides patient-specific bundle segmentation by incorporating known anatomical–functional correlations of the OD and its corresponding visual field. This integration not only enhances the accuracy and biological plausibility of the segmented traces, but also provides a clinically relevant framework for evaluating RNFL health, leading to more robust and interpretable results than purely data-driven, deep learning alternatives. Using a single algorithm to track RNFL bundles on the retinal surface can significantly enhance the identification of localized optic nerve damage and visual field defects. Such advances could greatly improve clinical evaluations in the diagnosis and treatment of visual disorders, including visual field loss and glaucoma.

4.1. Literature Review

As mentioned in several works, many other descriptions of nerve fiber tracings have been published before. Wigelius et al. [19] provided a mathematical description in his thesis that included an implicit solution similar to the trajectory path described by the Jansonius map. Nevertheless, the fitting of the trajectory density on a circle surrounding the optic disk and local RNFL thickness was not achieved. An RNFL bundle map based on scotoma borders in bundle defects was created by Weber and Ulrich [18]. This map, while less precise and comprehensive, revealed a pattern that was somewhat identical to the concept described by the Jansonius map. Numerous publications have previously presented descriptions of RNFL bundle trajectories. The publications were derived from postmortem human eyes [39], monkey histology [40], fundus photography [9], associations among optic disk anatomy and perimetric sensitivities by Ferreras et al. [20] and Turpin et al. [21], and optical coherence tomography by Garvin et al. [41]. Jansonius et al. [11,13] developed a mathematical model that explains the typical trajectory and variations in retinal nerve fiber trajectories. Leung et al. [12] presented an RNFL optical texture analysis (ROTA) algorithm, which integrates RNFL reflectance and RNFL thickness data obtained from OCT scans to uncover the optical textural details of RNFL bundles via a series of non-linear transformations. Dennis et al. [5] proposed a similar method to the one presented here, developing an anatomically customizable fiber mapping method by dividing the OD head and the retina into sectors and into a grid of elements, respectively. The amount of RGCs in each grid element was based on empirical data. The pathway from each element to the optic nerve head was determined by the number of axons that could enter the optic nerve head in each sector, as well as the sequence in which axons “grew” from different regions at the retina.

4.2. Concordance and Discordance with Existing RNFL Bundle Paths

The concordance of the algorithm’s output with the mathematical model proposed by Jansonius et al. [13] is in good quantitative and qualitative agreement for most locations, see Figure 2. The average mean difference in the orientation angles of the included images is 11.01 ± 1.25, and the average RMSE is 13.82 ± 1.55, see Figure 4. This concordance holds except with discordance occurring in the horizontal line at the latitude temporal to the fovea, see Figure 2. According to Jansonius et al. [31], “horizontal” is defined as parallel to the straight section temporal to the fovea, known as the location of the raphe. The discordance in the horizontal line temporal to the fovea may be explained by the fact that when tracing RNFL bundles close to the location of the raphe, it is extremely difficult to tell which fibers originate from and which pass over that point of interest. According to previous studies such as Jansonius et al. [11], all visible RNFL bundles from 55 eyes of 55 subjects were electronically traced as far as visible by one author from high-quality red-free fundus images, resulting in 1660 RNFL bundles. Jansonius et al. [31] provided another dataset of 28 fundus pictures of the right eyes of 28 subjects, including only those without diseases affecting the RNFL or its visibility (mostly diabetic patients without diabetic retinopathy). Lamparter et al. [10] manually traced RNFL bundle distribution by setting a minimum of three marker points and connecting them via a nonparametric cubic spline model, providing a continuous analytical function for each nerve bundle and obtaining 6388 RNFL bundles based on 42,914 sampling points from 100 subjects. According to those studies, RNFL bundles close to the horizontal line temporal to the fovea are neither visible nor manually drawn, excluding those crossing the line at nearly 180° minus the fovea–disk angle in the superior hemifield [10,11], and those crossing the line at nearly 180° plus the fovea–disk angle in the inferior hemifield [31]. This is consistent with the observed discordance between the algorithm presented here and the previous studies, see Figure 2. Another reason for such discordance may be attributed to the algorithmic limitations from disregarding the structural data of the optic disk, such as the tilt angle, the axial length, the spherical equivalent, and more unknowns that are not taken into consideration, which may provide hints about how to map correctly in some eyes. The architecture of the RNFL bundle may also be affected by other potential factors, such as the axial elongation that increases the OD–FO distance in myopic eyes [42], the differences in the number and distribution of ganglion cells, the layout and depth of blood vessels, and developmental growth signals [43,44]. A significant obstacle to merging imaging data with the segmentation of RNFL bundles is the significant noise that is typical of fundus images. Finally, the image resolution mostly affects contrast identification, contrast enhancement, and image filtering. Similarly, the visibility of RNFL bundles and layer defects is best recognizable in the peripapillary region and diminishes toward the periphery [45,46]. Fundus photographs provide only a two-dimensional representation of the RNFL, which is actually organized in a three-dimensional structure. The most superficial RNFL bundles are located near the vitreous body and correspond to the central region, while the deeper layers originate from the retinal periphery [47]. Radius et al. [48] and Minckler [49] demonstrated that fibers starting in the peripheral retina are located deep in the RNFL. Consistent with Garway-Heath et al. [9], the retinal nerve fibers do not seem to be highly organized within the deeper retinal layers. Consequently, fibers from the periphery, despite being visible, cannot be traced back to the optic disk [11].

4.3. Quantitative Performance of the Structure–Function Map

The structure–function map presented here is in good qualitative agreement with the previously published results, see Figure 3. Jansonius et al.’s [11] mean standard deviation of the angular location at the OD circumference (calculated as one quarter of the 95% limits) ranged from less than 1° to 18°, with an average value of 8.8° (median 7.3°). Lamparter et al. [10] showed a between-individual variability in the OD entry position, with a mean standard deviation of 12.4°. The variability at the level of the OD circumference found by Garway-Heath et al. [9] exhibited a mean standard deviation of 7.2°. In this work, the mean standard deviation, including a 95% limit (calculated as plus or minus two times the standard deviation), showed a value of 13.5° (median 14.01°) for 50 out of 52 visual field locations (missing only 19h and 27h data). Here it was shown that the standard deviation depends largely on the location of the visual field test point (from less than 1° to 28.4°). A possible explanation can be attributed to the wide range of disk–fovea angles (9.1 ± 3.9, ranged from 0.6° to 19.7°) of the recruited subjects, and similarly to the diverse variety of OD locations, which clarify the most variance in the structure–function maps [5,9]. The comparison of the current model’s output with the mathematical model proposed by Jansonius et al. [13], and the maps produced by Lamparter et al. [10] and Garway-Heath et al. [9], appears to show reasonable qualitative agreement for most locations, see Figure 2 and Figure 3. Even if the means differ numerically, all but one case (fourth visual field point) display overlapping standard deviation ranges, indicating that those groups have a common range of values because of the intrinsic variability. These findings hold potential for producing customized structure–function tests tailored to each patient and increasing the sensitivity and specificity of disease progression and detection metrics. The differences of entry angles between the average mean of [9,10,13] and that obtained here range from less than 1° to 24.9° (9.88 ± 5.95). Similarly, the RMSE had a value of 11.51. For a fair comparison, the value at the 34th visual field point is excluded because of the discontinuities around the 1/360 border.

4.4. Validation up to 30° of Eccentricity

An agreement has been found between the retinal wiring of the nerve fibers presented here and the principle of orderly arrangement available today [11,13,50,51], rigorously up to an eccentricity of 30°. RNFL bundles converge in an ordered fashion to the optic nerve. RNFL bundles that supply the nasal side at the superior and inferior retinas have relatively straight pathways [50,51]. This orderly assembly is visible on fibers with angular positions at the superior, inferior, and nasal edges of the OD (see Figure 1f). The trajectory starting at the OD continues in an orderly fashion by arranging fibers with consecutive angular positions, obeying the natural spatial movement.
Those originating from the peripheral temporal retina arc positioned above or below the macular region enter the optic nerve in an arcuate pattern, into the superior and inferior temporal portions, respectively [50,51]. In this work, fibers with angular positions at the optic disk’s temporal edge are arranged in an arcuate pattern coming from the temporal hemifield in a consecutive manner. Those accommodated in the papillomacular bundle area follow arcuate patterns from the macular region toward to the optic nerve’s temporal edge.

4.5. Close Methodological Comparison of the Algorithm with Prior Literature

A noteworthy difference between the present algorithm and that reported previously is that the most accurate models require excessively long man-hours of work and time-consuming procedures to manually [9,10] or electronically [11,13] trace RNFL bundles as far as visible by a person, or extremely expensive ophthalmic equipment for obtaining the necessary input data [12]. Jansonius et al. [13] used high-quality fundus photograph acquisition, image processing techniques for making the fibers stand out, foveola and OD marking, image superimposition and alignment, and manual tracing with inclusion criteria as inputs. Lamparter et al. [10] used fundus photography, RNFL bundle optimization, manual delineation, and cubic splines to connect the marker points (at least three) as inputs. Dennis et al. [5] used the axial length of the eye, OD position, horizontal and vertical diameter of the OD, horizontal and vertical cup diameter, and fovea diameter. Leung et al. [12] used RNFL reflectance and RNFL thickness data obtained from OCT scans. In this work, RNFL bundles are assembled using image processing algorithms and publicly accessible inputs, see Section 2.4. The inclusion criteria were first given to exclude fundus images for cases when the FO and/or the OD could not be located precisely. Similarly, the criterion requires at least five marker points (or five pointwise RNF extractions) to fit the assembled RNFs using a 4°-order Bezier curve. The general method of precision can be attributed to the sequence of three phases. The first phase involves the CAHE for reducing undesired noise amplification while simultaneously improving local contrast, and the HFV filter for enhancing and detecting tubular patterns, which is crucial for identifying distinct curvilinear RNF structures. The second phase involves the MM algorithm, which inherently involves finding the local maximum and minimum pixel intensities within a region in order to assess edge sharpness and isolate RNFL bundle segments. Pixel intensities are redistributed, thereby improving the luminance difference between the target RNF extractions and their surroundings. The third phase aligns the orientation of RNF extractions by assuming specific sets of angles in a modified Garway-Heath map [9]. A simple criterion states that if the RNF angle falls outside the defined range for a specific area, the RNF extraction is discarded from further evaluation. This criterion eliminates false positive cases by highlighting non-RNFL bundle structures. The average range of angles for adjusting RNFs’ direction provides a broad variability (70° ± 21°). The proposed segmentation algorithm has a runtime complexity of less than 45 s for processing around 8.1 megapixels in the image (2330 × 3500), which involves a fixed number of sequential passes over the image for feature computation (including segmentation, Bezier curve fitting, structure–function mapping, and image plotting).

4.6. OD and FO Localization

This analysis was conducted using a diverse collection of fundus images to develop an algorithm that can be customized to individual eyes and distinguish fibers with great accuracy by considering unique anatomical features. A total of four algorithms were developed to independently boost the OD and FO ROI, focusing on contour edges for precise identification and to address problems that arise while using several databases. First, fundus images in the used collections have a broad variety of color tones surrounding the FO and OD. As a result, several of the images in this collection obscure those regions. The success rate is adversely affected, due to the fact that the OD region’s intensity is similar to that in the remainder of the fundus images. Second, there is variation in color and intensity among the images in the collection. Consequently, the methods presented here emphasize contrast illumination changes via different filtering techniques.

4.7. Intersubject Variability for RNFL Bundle Traces

Probable influences that could cause error bias or RNFL bundle trajectory variability are (1) the fundus photography, (2) patient selection, (3) tracing process, and (4) anatomical differences between subjects. In this work, five different fundus image databases were used, and the results revealed small variabilities from the Jansonius map’s comparison of 97.22% (HRF), 94.29% (JSIEC), 97.82% (RIDB), 94.36% (DRIVE), and 97.32% (MESSIDOR) (average per database of the summation of frequencies higher than 92.5% effectiveness), suggesting that fundus photography is neither a major source of error bias nor variability. As mentioned in the Section 2, the patients selected were healthy, without any sign of a history of ocular diseases, and diabetic subjects without any sign of diabetic retinopathy. According to [31], diabetes does not affect the trajectories or their visibility as long as there is no retinopathy. The tracing method is likewise not a significant source of variability or error bias due to the previously described similar percentage results. Concerning anatomical differences between subjects, multiple linear regression analyses show that the vertical OD distance and disk–fovea angle are statistically significant on the average RNF orientation (R2 = 0.64, β = −4.43, p < 0.00001) and (R2 = 0.62, β = −1.56, p < 0.00001), respectively. Anatomical differences between subjects have been examined before. Tanabe et al. [52] found that refraction and the architecture of the retinal arteries have significant influences on bundle angles using scanning laser ophthalmoscopy. According to Denniss et al. [5], the OD’s position in relation to the fovea, axial length, and horizontal and vertical OD diameters heavily influences the structure–function relationship. In accordance with Lamparter et al. [10], the OD’s latitudinal and longitudinal position and disk area, followed by variation in orientation and the tilt of the OD, influenced intersubject variability on entry angle into the optic nerve head. Garway-Heath et al. [9] found the influence of the OD’s position to be a parameter explaining between-subject variability. The RNFL thickness and the RNFL bundle trajectories are probably associated. As reported by Hwang et al. [53], eyes with counterclockwise rotation and a myopic temporal optic disk tilt exhibit a more temporally positioned superior peak location and a thicker temporal RNFL. According to Tong et al. [54], OD parameters and the RNFL thickness obtained with scanning laser ophthalmoscopy are strongly influenced by the tilting of the OD, but not by refractive error or axial length.
Consistent with the findings of Lamparter et al. [10], Denniss et al. [5], and Garway-Heath et al. [9], the most prominent predictor for the variations presented here is the latitudinal OD location, followed by the disk–fovea angle, which is in good agreement. Similarly, the linear regression analysis revealed robust explanatory power, with an R2 value of 0.64 and 0.62. This indicates that 64% and 62% of the variance accounted for O D y and the disk–fovea angle included in the model, respectively. These R2 values represent a considerable proportion of the explained variance, suggesting that the algorithm may provide a strong understanding of the factors influencing the average RNF orientation in the papillomacular region. Diagnostic analysis of the residuals showed a random distribution around zero, confirming that the assumptions of linear regression, including linearity and homoscedasticity, were adequately met. This validates the appropriateness of the linear model, reinforcing the reliability presented here, see Figure 5 (only one independent variable is shown). While the model explains a large proportion of the variance, 36% and 38% of the variability remains unexplained. This indicates that other factors, not captured in the current model, also contribute to the dependent variable. Future research could explore these unmeasured variables, such as the axial length, to further refine the understanding. Numerous studies support the concept that an increase in axial length, particularly in myopic eyes, leads to the stretching of the RNFL bundles, particularly the papillomacular bundle [55,56]. The variability in the axial length (23.4 ± 1.04 mm, range 20.29–28.68 mm [57]) imposes that any variation in its spatial relationship directly influences the length, curvature, and overall path of the papillomacular bundles [55]. This greater susceptibility results from the papillomacular bundle’s very straight trajectory, which may restrict its capacity to adjust to the growing OD–FO gap. The more peripheral nerve fibers, on the other hand, that are attached to nontemporal rim sectors, have a more arcuate path, and may thus adjust by straightening more effectively [56]. As discussed previously, one limitation imposed by the fundus images is that the axial length is not accessible and thus not used in the current study. Therefore, the clinical relevance of these significant effects awaits careful consideration.
Developing an assessment that includes the variability of the RNFL bundle’s location as a function of orientation angles could be beneficial for evaluating myopic eyes. Consistent evidence of RNFL bundle variability has been previously documented. According to several studies, the location of the peak RNFL thickness and the thickness of supratemporal and infratemporal RNFL bundles are shifted toward the fovea in eyes with longer axial lengths. This is consistent with the reports by Yoo et al. [58] and Hong et al. [59]. As the axial length increases, the RNFL bundles are likewise displaced nearer the fovea in relation to longer eyes. In a two-dimensional fundus picture, these alterations can be detected as a straightening and a shift toward the fovea of the RNFL bundles’ eyes (see Figure 1 in [60]). This RNFL bundle alteration may account for the RNFL defect frequently manifesting at the paracentral region when myopic glaucoma is first developing [60]. The RNFL bundle shift can be quantified by assessing the location of the supratemporal and infratemporal peak RNFL thickness of the peripapillary RNFL thickness profile [58]. Still, for a precise glaucoma diagnosis, it is critical to understand how myopic alterations affect the thickness of the peripapillary RNFL. However, the RNFL bundle’s displacement is not effective for the temporal raphe. Bedggood et al. [61] tested the specific hypotheses of whether myopia alters the normal trajectory of RNFL bundles temporal to the macula at the temporal raphe. The axial length had a minimal effect on the path of nerve fibers in the temporal raphe area, according to this study, but it did have some distinct effects on the thickness of papillary and peripapillary nerve fibers. Although the study conducted by Bedggood et al. [61] only examined the raphe’s shape and not the entire arc of fibers passing through the retina, the results showed that any readjustment of the fiber trajectories would not appear to include the temporal raphe (to the extent that the study was able to do so).

4.8. Limitations

The premise of a personalized model is that clinically measurable biometric information can improve the accuracy of mapping visual fields. This work focuses exclusively on parameters that can be determined based on research expertise such as the ground truth location data of FO and OD centers, which are publicly available. The structural data of the optic disk, such as the tilt angle, the axial length (typically assessed via partial coherence interferometry), the spherical equivalent, and the orientation of the OD (assessed via Spectralis HRA reflectance image (RNFL scan)) [10] was not considered. The fundus images used in this study were obtained from anonymized databases, and biometric data related to the axial length and refraction were not accessible. The impact of both could not be ascertained due to their absence. Accordingly, a number of factors that are not taken into consideration may provide hints about how to map correctly in some eyes. For instance, Lamparter et al. [10] found that variation in the axial length and spherical equivalent influenced the mapping of 9 out of 33 visual field sites. Deviations may also result from other possible factors influencing the architecture of the RNFL bundle, such as variations in developmental growth, the depth of blood vessels, and in the number and distribution of ganglion cells [43,44].
A significant obstacle to merging imaging data with structural maps is the significant noise that is typical of fundus image associations. There are probably several reasons for this noise, including intersubject differences in the relationship between retinal regions and the OD [5]. This topographic relationship between specific locations of the optic disk and their corresponding nerve fiber bundles on the retina is complex, and individual variation is high [9,10]. For reaching ground truth maps, experienced ophthalmologists require excessively long man-hours of labor and time-consuming procedures to manually [9,10] or electronically [11,13] trace RNFL bundles. In this work, the spatial trajectories from local regions on the retina to the OD were associated by intersecting individual bundle maps obtained with the algorithm with the OD boundary. These may provide some deviations that influence the architecture of RNFL bundles and affect the intersubject differences in the relationship between retinal regions and the OD.
The OD’s position explains the most variance in the structure–function maps, and its alteration is a significant factor affecting the RNFL bundle’s structure [5,9,10,11]. Jansonius et al. [11,13] computed the OD size from the macula–disk center distance in fundus photographs and defined the location of the OD by the papillomacular position. Lamparter et al. [10] used magnification-corrected measurements of the OD size and dislocation from the FO (in the horizontal and vertical directions, separately). In this work, two algorithms separately enhanced the FO ROI and two algorithms for the OD ROI in order to obtain different results by specifically targeting contour edges for accurate identification. However, these additional elements might affect the RNFL bundle’s design and could potentially cause deviations. Nonetheless, computational algorithms offer a useful foundation for future experimental hypothesis generation and empirical testing. The simplicity of updating and improving the algorithm in response to new empirical data can benefit the computational modeling technique. Through regular updates, this flexibility enables the algorithm to stay up to date with a rapidly evolving evidence base, allowing for more strict validation in future studies.
One recognized limitation of the HFV filter are RNF crossings and RNF bifurcations [62] as the local intensity profile at these points differs from a simple singular tubular structure and deviates from being an isolated segment with a relatively consistent intensity profile that can be characterized by its orientation (along the segment) and its cross-sectional curvature (across the segment). This limitation is alleviated by (1) aligning the RNF orientation, which eliminates false positive cases of non-RNFL bundle structures, and (2) including “group” and “individual” averages to ensure the continuity of orientation between neighboring extractions. RNFL defects in early or advanced glaucoma or other optic atrophy cannot be detected with the algorithm presented here, since it has not been tested yet. Therefore, future work should be focused on providing high reliability for the detection of optic neuropathies. The location-specific intersubject variability using the structure–function map and the overall validation using more fundus images will be addressed in future studies.

5. Conclusions

This work presents an image processing algorithm that segments RNFL bundle traces in the human retina. Using the FO and OD centers from fundus images as the input, the algorithm generates an accurate representation of the RNFL bundle pathways and associates each RNFL bundle with an estimated location of the OD. The quantitative comparison of the RNFL bundle traces and that of the entry angle of RNFL bundles into the OD shows three findings: (1) the segmentation of RNFL bundle tracings with great accuracy by considering unique anatomical features that can be customized to individual eyes, (2) structure–function maps can be built using image processing techniques tailored for healthy individuals, and (3) RNF orientation angles in the papillomacular region can be described using the latitudinal OD location and the disk–fovea angle. These advances could greatly improve clinical evaluations in the diagnosis and treatment of visual disorders, including glaucoma.

Funding

This work was supported by the Instituto Tecnológico y de Estudios Superiores de Monterrey through the Publications Support Fund (FAP, by its initials in Spanish).

Institutional Review Board Statement

This study utilized publicly available, de-identified fundus image datasets. As the research did not involve human subjects or the collection of new identifiable data, it was determined to be exempt from review by an Institutional Review Board. All methods and procedures were performed in accordance with the ethical guidelines outlined in the Declaration of Helsinki.

Informed Consent Statement

The fundus images were obtained from publicly available databases, and all personal identifying information was removed to protect patient privacy.

Data Availability Statement

PES (personalized estimated segmentation) software application developed for the segmentation of nerve fiber trajectories is a code programmed and owned by the author written in MATLAB, available on GitHub https://github.com/diegolujv/PES-app.git (accessed on 15 August 2025).

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Carreras, F.J.; Medina, J.; Ruiz-Lozano, M.; Carreras, I.; Castro, J.L. Virtual Tissue Engineering and Optic Pathways: Plotting the Course of the Axons in the Retinal Nerve Fiber Layer. Investig. Ophthalmol. Vis. Sci. 2014, 55, 3107–3119. [Google Scholar] [CrossRef] [PubMed]
  2. Shan, S.; Wu, J.; Cao, J.; Feng, Y.; Zhou, J.; Luo, Z.; Song, P.; Rudan, I.; Global Health Epidemiology Research Group (GHERG). Global incidence and risk factors for glaucoma: A systematic review and meta-analysis of prospective studies. J. Glob. Health 2024, 14, 04252. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  3. Schiefer, U.; Flad, M.; Stumpp, F.; Malsam, A.; Paetzold, J.; Vonthein, R.; Oliver Denk, P.; Sample, P.A. Increased detection rate of glaucomatous visual field damage with locally condensed grids: A comparison between fundusoriented perimetry and conventional visual field examination. Arch. Ophthalmol. 2003, 121, 458–465. [Google Scholar] [CrossRef] [PubMed]
  4. Paetzold, J.; Sample, P.A.; Selig, B.; Krapp, E.; Vonthein, R.; Schiefer, U. Differentiation between glaucomatous scotoma progression in depth and in size by using scotoma-oriented perimetry (SCOPE) with locally condensed stimulus arrangements. Investigig. Ophthalmol. Vis. Sci. 2005, 46, 636. [Google Scholar]
  5. Denniss, J.; Allison, M.; McKendrick, A.T. An Anatomically Customizable Computational Model Relating the Visual Field to the Optic Nerve Head in Individual Eyes. Investig. Opthalmology Vis. Sci. 2012, 53, 6981–6990. [Google Scholar] [CrossRef]
  6. Sugita, M.; Pircher, M.; Zotter, S.; Baumann, B.; Roberts, P.; Makihira, T.; Tomatsu, N.; Sato, M.; Vass, C.; Hitzenberger, C.K. Retinal nerve fiber bundle tracing and analysis in human eye by polarization sensitive OCT. Biomed. Opt. Express 2015, 6, 1030–1054. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  7. Iwata, K.; Namba, K.; Abe, H. Early fundus changes caused by repeated small crises in the PosnereSchlossman syndrome: A model for glaucoma simplex. Klin. Monbl. Augenheilkd. 1982, 180, 20–26. [Google Scholar] [CrossRef]
  8. Diekmann, H.; Fischer, D. Glaucoma and optic nerve repair. Cell Tissue Res. 2013, 353, 327–337. [Google Scholar] [CrossRef]
  9. Garway-Heath, D.F.; Poinoosawmy, D.; Fitzke, F.W.; Hitchings, R.A. Mapping the visual field to the optic disc in normal tension glaucoma eyes. Ophthalmology 2000, 107, 1809–1815. [Google Scholar] [CrossRef]
  10. Lamparter, J.; Russell, R.A.; Zhu, H.; Asaoka, R.; Yamashita, T.; Ho, T.; Garway-Heath, D.F. The Influence of Intersubject Variability in Ocular Anatomical Variables on the Mapping of Retinal Locations to the Retinal Nerve Fiber Layer and Optic Nerve Head. Investig. Ophthalmol. Vis. Sci. 2013, 54, 6074–6082. [Google Scholar] [CrossRef]
  11. Jansonius, N.M.; Nevalainen, J.; Selig, B.; Zangwill, L.M.; Sample, P.A.; Budde, W.M.; Jonas, J.B.; Lagrèze, W.A.; Airaksinen, P.J.; Vonthein, R. A mathematical description of nerve fiber bundle trajectories and their variability in the human retina. Vis. Res. 2009, 49, 2157–2163. [Google Scholar] [CrossRef]
  12. Leung, C.K.S.; Lam, A.K.N.; Weinreb, R.N.; Garway-Heath, D.F.; Yu, M.; Guo, P.Y.; Chiu, V.S.M.; Wan, K.H.N.; Wong, M.; Wu, K.Z.; et al. Diagnostic assessment of glaucoma and non-glaucomatous optic neuropathies via optical texture analysis of the retinal nerve fibre layer. Nat. Biomed. Eng. 2022, 6, 593–604. [Google Scholar] [CrossRef]
  13. Jansonius, N.M.; Schiefer, J.; Nevalainen, J.; Paetzold, J.; Schiefer, U. A mathematical model for describing the retinal nerve fiber bundle trajectories in the human eye: Average course, variability, and influence of refraction, optic disc size and optic disc position. Exp. Eye Res. 2012, 105, 70–78. [Google Scholar] [CrossRef] [PubMed]
  14. Chen, Z.; Shemuelian, E.; Wollstein, G.; Wang, Y.; Ishikawa, H.; Schuman, J.S. Segmentation-Free OCT-Volume-Based Deep Learning Model Improves Pointwise Visual Field Sensitivity Estimation. Trans. Vis. Sci. Tech. 2023, 12, 28. [Google Scholar] [CrossRef] [PubMed]
  15. Park, K.; Kim, J.; Lee, J. A deep learning approach to predict visual field using optical coherence tomography. PLoS ONE 2020, 15, e0234902. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  16. Christopher, M.; Bowd, C.; Belghith, A.; Goldbaum, M.H.; Weinreb, R.N.; Fazio, M.A.; Girkin, C.A.; Liebmann, J.M.; Zangwill, L.M. Deep Learning Approaches Predict Glaucomatous Visual Field Damage from OCT Optic Nerve Head En Face Images and Retinal Nerve Fiber Layer Thickness Maps. Ophthalmology 2020, 127, 346–356. [Google Scholar] [CrossRef]
  17. Leung, C.K.S.; Cheung, C.Y.L.; Weinreb, R.N.; Qiu, K.; Liu, S.; Li, H.; Xu, G.; Fan, N.; Pang, C.P.; Tse, K.K.; et al. Evaluation of retinal nerve fiber layer progression in glaucoma with optical coherence tomography guided progression analysis. Investig. Ophthalmol. Vis. Sci. 2009, 11, 5172–5178. [Google Scholar] [CrossRef]
  18. Weber, J.; Ulrich, H. A perimetric nerve fiber bundle map. Int. Ophthalmol. 1991, 15, 193–200. [Google Scholar] [CrossRef]
  19. Wigelius, O. A Model for the Retinal Nerve Fiber Layer. Master’s Thesis, Lund University, Lund, Sweden, 2001. [Google Scholar]
  20. Ferreras, A.; Pablo, L.E.; Garway-Heath, D.F.; Fogagnolo, P.; Garcia-Feijoo, J. Mapping standard automated perimetry to the peripapillary retinal nerve fiber layer in glaucoma. Investig. Ophthalmol. Vis. Sci. 2008, 49, 3018–3025. [Google Scholar] [CrossRef] [PubMed]
  21. Turpin, A.; Sampson, G.P.; McKendrick, A.M. Combining ganglion cell topology and data of patients with glaucoma to determine a structure-function map. Investig. Ophthalmol. Vis. Sci. 2009, 50, 3249–3256. [Google Scholar] [CrossRef]
  22. Denniss, J.; Turpin, A.; Tanabe, F.; Matsumoto, C.; McKendrick, A.M. Structure-function mapping: Variability and conviction in tracing retinal nerve fiber bundles and comparison to a computational model. Investig. Ophthalmol. Vis. Sci. 2014, 55, 728–736. [Google Scholar] [CrossRef]
  23. Oh, J.E.; Yang, H.K.; Kim, K.G.; Hwang, J.-M. Automatic computer-aided diagnosis of retinal nerve fiber layer defects using fundus photographs in optic neuropathy. Investig. Ophthalmol. Vis Sci. 2015, 56, 2872–2879. [Google Scholar] [CrossRef] [PubMed]
  24. Quintana-Quintana, O.J.; Aceves-Fernández, M.A.; Pedraza-Ortega, J.C.; Alfonso-Francia, G.; Tovar-Arriaga, S. Deep Learning Techniques for Retinal Layer Segmentation to Aid Ocular Disease Diagnosis: A Review. Computers 2025, 14, 298. [Google Scholar] [CrossRef]
  25. Zhang, H.; Yang, B.; Li, S.; Zhang, X.; Li, X.; Liu, T.; Higashita, R.; Liu, J. Retinal OCT image segmentation with deep learning: A review of advances, datasets, and evaluation metrics. Comput. Med. Imaging Graph. 2025, 123, 102539. [Google Scholar] [CrossRef]
  26. Budai, A.; Bock, R.; Maier, A.; Hornegger, J.; Michelson, G. Robust Vessel Segmentation in Fundus Images. Int. J. Biomed. Imaging 2013, 2013, 154860. [Google Scholar] [CrossRef] [PubMed]
  27. Cen, L.P.; Ji, J.; Lin, J.W.; Ju, S.T.; Lin, H.J.; Li, T.P.; Wang, Y.; Yang, J.F.; Liu, Y.F.; Tan, S.Y. Automatic detection of 39 fundus diseases and conditions in retinal photographs using deep neural networks. Nat. Commun. 2021, 12, 4828. [Google Scholar] [CrossRef]
  28. Akram, M.U.; Abdul Salam, A.; Khawaja, S.G.; Naqvi, S.G.H.; Khan, S.A. RIDB: A Dataset of fundus images for retina based person identification. Data Brief 2020, 33, 106433. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  29. Hu, Q.; Abràmoff, M.D.; Garvin, M.K. Automated separation of binary overlapping trees in low-contrast color retinal images. Med. Image. Comput. Comput. Assist Interv. 2013, 16 Pt 2, 436–443. [Google Scholar] [PubMed]
  30. Decencière, E.; Zhang, X.; Cazuguel, G.; Lay, B.; Cochener, B.; Trone, C.; Gain, P.; Ordonez, R.; Massin, P.; Erginay, A.; et al. Feedback on a publicly distributed database: The Messidor database. Image Anal. Stereol. 2014, 33, 231–234. [Google Scholar] [CrossRef]
  31. Jansonius, N.M.; Schiefer, U. Anatomical location of the raphe and extended raphe in the human retina: Implications for assessment of the optic nerve with OCT. Trans. Vis. Sci. Tech. 2020, 9, 3. [Google Scholar] [CrossRef]
  32. Khalid, N.E.A.; Noor, N.M.; Mahmud, Z.; Yahya, S.; Ariff, N.M. Bridging Quantitative and Qualitative of glaucoma detection. World Acad. Sci. Eng. Technol. 2012, 6, 584–588. [Google Scholar]
  33. Fraz, M.M.; Basit, A.; Barman, S. Application of morphological bit planes in retinal blood vessel extraction. J. Digit. Imaging 2013, 26, 274–286. [Google Scholar] [CrossRef]
  34. Fraz, M.M.; Remagnino, P.; Hoppe, A.; Uyyanonvara, B.; Rudnicka, A.R.; Owen, C.G.; Barman, S.A. Blood vessel segmentation methodologies in retinal images—A survey. Comput. Methods Programs Biomed. 2012, 108, 407–433. [Google Scholar] [CrossRef]
  35. Zuiderveld, K. Contrast Limited Adaptive Histograph Equalization. In Graphic Gems IV; Academic Press Professional: San Diego, CA, USA, 1994; pp. 474–485. [Google Scholar]
  36. Frangi, A.F.; Niessen, W.J.; Vincken, K.L.; Viergever, M.A. Multiscale vessel enhancement filtering. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention—MICCAI’98, Cambridge, MA, USA, 11–13 October 1998; Springer: Berlin/Heidelberg, Germany, 1998; pp. 130–137. [Google Scholar]
  37. Ohad, G. fit_ellipse. MATLAB Central File Exchange. 2025. Available online: https://www.mathworks.com/matlabcentral/fileexchange/3215-fit_ellipse (accessed on 13 May 2025).
  38. Leung, C.K.S.; Guo, P.Y.; Lam, A.K.N. Retinal Nerve Fiber Layer Optical Texture Analysis: Involvement of the Papillomacular Bundle and Papillofoveal Bundle in Early Glaucoma. Ophthalmology 2022, 129, 1043–1055. [Google Scholar] [CrossRef]
  39. Fitzgibbon, T.; Taylor, S.F. Retinotopy of the human retinal nerve fibre layer and optic nerve head. J. Comp. Neurol. 1996, 375, 238–251. [Google Scholar] [CrossRef]
  40. Vrabec, F. The temporal raphe of the human retina. Am. J. Ophthalmol. 1966, 62, 926–938. [Google Scholar] [CrossRef] [PubMed]
  41. Garvin, M.K.; Abramoff, M.D.; Lee, K.; Niemeijer, M.; Sonka, M.; Kwon, Y.H. 2-D pattern of nerve fiber bundles in glaucoma emerging from spectral-domain optical coherence tomography. Investig. Ophthalmol. Vis. Sci. 2012, 53, 483–489. [Google Scholar] [CrossRef] [PubMed]
  42. Bikbov, M.M.; Iakupova, E.M.; Gilmanshin, T.R.; Bikbova, G.M.; Kazakbaeva, G.M.; Panda-Jonas, S.; Gilemzianova, L.I.; Jonas, J.B. Prevalence and associations of nonglaucomatous optic nerve atrophy in high myopia: The ural eye and medical study. Ophthalmology 2023, 130, 1174–1181. [Google Scholar] [CrossRef]
  43. Hood, D.C.; Fortune, B.; Arthur, S.N.; Xing, D.; Salant, J.A.; Ritch, R.; Liebmann, J.M. Blood vessel contributions to retinal nerve fiber layer thickness profiles measured with optical coherence tomography. J. Glaucoma. 2008, 17, 519–528. [Google Scholar] [CrossRef] [PubMed]
  44. Hood, D.C.; Salant, J.A.; Arthur, S.N.; Ritch, R.; Liebmann, J.M. The location of the inferior and superior temporal blood vessels and inter-individual variability of the retinal nerve fiber layer thickness. J. Glaucoma. 2010, 19, 158–166. [Google Scholar] [CrossRef]
  45. Jonas, J.B.; Nguyen, N.X.; Naumann, G.O.H. The retinal nerve fiber layer in normal eyes. Ophthalmology 1989, 96, 627–632. [Google Scholar] [CrossRef]
  46. Jonas, J.B.; Schiro, D. Visibility of the normal retinal nerve fiber layer correlated with rim width and vessel caliber. Graefe’s Arch. Clin. Exp. Ophthalmol. 1993, 231, 207–211. [Google Scholar] [CrossRef]
  47. Airaksinen, P.J.; Alanko, H.I. Effect of retinal nerve fibre loss on the optic nerve head configuration in early glaucoma. Graefes Arch. Clin. Exp. Ophthalmol. 1983, 220, 193–196. [Google Scholar] [CrossRef] [PubMed]
  48. Radius, R.L.; Anderson, D.R. The course of axons through the retina and optic nerve head. Arch. Ophthalmol. 1979, 97, 1154–1158. [Google Scholar] [CrossRef]
  49. Minckler, D.S. The organization of nerve fiber bundles in the primate optic nerve head. Arch. Ophthalmol. 1980, 98, 1630–1636. [Google Scholar] [CrossRef]
  50. Hood, D.C.; Raza, A.S.; de Moraes, C.G.V.; Liebmann, J.M.; Ritch, R. Glaucomatous damage of the macula. Prog. Retin. Eye Res. 2013, 32, 1–21. [Google Scholar] [CrossRef] [PubMed]
  51. Hogan, M.; Alvardo, J.; Weddell, J. Histology of the Human Eye; WB Saunders: Philadelphia, PA, USA, 1971; pp. 523–606. [Google Scholar]
  52. Tanabe, F.; Matsumoto, C.; Okuyama, S.; Takada, S.; Hashimoto, S.; Arimura, E.; Nomoto, H.; Kayazawa, T.; Eura, M.; Shimomura, Y. Factors in the Retinal Nerve Fiber Layer (RNFL) Bundle Angles at the Optic Disc. IVFIS 2012. Available online: https://www.perimetry.org/__static/1dee4d2e5ebbc3d8091e1c689c10b1f5/20-melbourne.pdf?dl=1 (accessed on 15 August 2025).
  53. Hwang, Y.H.; Yoo, C.; Kim, Y.Y. Characteristics of peripapillary retinal nerve fiber layer thickness in eyes with myopic optic disc tilt and rotation. J. Glaucoma. 2012, 21, 394–400. [Google Scholar] [CrossRef] [PubMed]
  54. Tong, L.; Chan, Y.H.; Gazzard, G.; Loon, S.C.; Fong, A.; Selvaraj, P.; Healey, P.R.; Tan, D.; Wong, T.Y.; Saw, S.M. Heidelberg retinal tomography of optic disc and nerve fiber layer in Singapore children: Variations with disc tilt and refractive error. Investig. Ophthalmol. Vis Sci. 2007, 48, 4939–4944. [Google Scholar] [CrossRef]
  55. Jonas, J.B.; Jonas, R.A.; Bikbov, M.M.; Wang, Y.X.; Panda-Jonas, S. Myopia: Histology, clinical features, and potential implications for the etiology of axial elongation. Prog. Retin. Eye Res. 2023, 96, 101156. [Google Scholar] [CrossRef] [PubMed]
  56. Yii, F.; Gibbon, S.; MacGillivra, T. Sectoral Changes in Neuroretinal Rim Pallor Across Refractive Error. Ophthalmol. Sci. 2025, 5, 100705. [Google Scholar] [CrossRef]
  57. Jonas, J.B.; Xu, L.; Wei, W.B.; Pan, Z.; Yang, H.; Holbach, L.; Panda-Jonas, S.; Wang, Y.X. Retinal Thickness and Axial Length. Investig. Ophthalmol. Vis. Sci. 2016, 57, 1791–1797. [Google Scholar] [CrossRef]
  58. Yoo, Y.C.; Lee, C.M.; Park, J.H. Changes in peripapillary retinal nerve fiber layer distribution by axial length. Optom. Vis. Sci. 2012, 89, 4–11. [Google Scholar] [CrossRef] [PubMed]
  59. Hong, S.W.; Ahn, M.D.; Kang, S.H.; Im, S.K. Analysis of peripapillary retinal nerve fiber distribution in normal young adults. Investig. Ophthalmol. Vis. Sci. 2010, 51, 3515–3523. [Google Scholar] [CrossRef] [PubMed]
  60. Yamashita, T.; Asaoka, R.; Tanaka, M.; Kii, Y.; Yamashita, T.; Nakao, K.; Sakamoto, T. Relationship Between Position of Peak Retinal Nerve Fiber Layer Thickness and Retinal Arteries on Sectoral Retinal Nerve Fiber Layer Thickness. Investig. Ophthalmol. Vis. Sci. 2013, 54, 5481–5488. [Google Scholar] [CrossRef]
  61. Bedggood, P.; Mukherjee, S.; Bao, N.; Nguyen, A.T.; Allison, M. McKendrick; Geometry of the Retinal Nerve Fibers From Emmetropia Through to High Myopia at Both the Temporal Raphe and Optic Nerve. Investig. Ophthalmol. Vis. Sci. 2019, 60, 4896–4903. [Google Scholar] [CrossRef] [PubMed]
  62. Jerman, T.; Pernus, F.; Likar, B.; Spiclin, Z. Enhancement of Vascular Structures in 3D and 2D Angiographic Images. IEEE Trans. Med. Imaging 2016, 35, 2107–2118. [Google Scholar] [CrossRef]
Figure 1. Block diagram for the segmentation of personalized mapping of the optic pathways using image processing techniques. (a) original RGB fundus image. (b) Blood vessel segmentation using CAHE and Frangi vesselness filter. (c) RNF segmentation via maximum–minimum modulation algorithm. (d) A modified Garway-Heath map [13] categorizes RNF orientation, assuming designated set of angles for aligning RNFs’ direction across various areas. The convention of angles is drawn around the OD boundary. (e) Pointwise RNF extractions described by their centroid and the angle of orientation seen from the center of the OD. (f) Assembly of RNFs by identifying and connecting the optimal nearest extraction based on the orientation angle. Bezier curves were employed to progressively fit the assembled RNFs from the center of the OD to their corresponding end. Crossing individual bundle maps with the OD boundary was used to determine the entry OD angle locations. Beside (f), a colormap depicting the entry OD angle sites is displayed.
Figure 1. Block diagram for the segmentation of personalized mapping of the optic pathways using image processing techniques. (a) original RGB fundus image. (b) Blood vessel segmentation using CAHE and Frangi vesselness filter. (c) RNF segmentation via maximum–minimum modulation algorithm. (d) A modified Garway-Heath map [13] categorizes RNF orientation, assuming designated set of angles for aligning RNFs’ direction across various areas. The convention of angles is drawn around the OD boundary. (e) Pointwise RNF extractions described by their centroid and the angle of orientation seen from the center of the OD. (f) Assembly of RNFs by identifying and connecting the optimal nearest extraction based on the orientation angle. Bezier curves were employed to progressively fit the assembled RNFs from the center of the OD to their corresponding end. Crossing individual bundle maps with the OD boundary was used to determine the entry OD angle locations. Beside (f), a colormap depicting the entry OD angle sites is displayed.
Jimaging 11 00294 g001
Figure 2. Percentage of effectiveness as part of the validation process per database: (a) HRF, (b) JSIEC, (c) RIDB, (d) DRIVE, and (e) MESSIDOR. The images were superimposed per database and onto the Jansonius map via translation for centering the FO, followed by zooming and rotation for aligning the OD centers at an eccentricity of 15°, 2° above the horizontal meridian. The percentage of effectiveness is computed as in Equation (1). The algorithm extracted an average of 7854 (HRF), 10139 (JSIEC), 4539 (RIDB), 7977 (DRIVE), and 4445 (MESSIDOR) RNF centroids per fundus image per database, for an average estimated total number of 1,732,657 (which correspond to the same number of P e f f ). (f) General Garway-Heath map used to classify the results is shown in Table 1. The map is a modified version of the structure–function map developed by Garway-Heath et al. [9].
Figure 2. Percentage of effectiveness as part of the validation process per database: (a) HRF, (b) JSIEC, (c) RIDB, (d) DRIVE, and (e) MESSIDOR. The images were superimposed per database and onto the Jansonius map via translation for centering the FO, followed by zooming and rotation for aligning the OD centers at an eccentricity of 15°, 2° above the horizontal meridian. The percentage of effectiveness is computed as in Equation (1). The algorithm extracted an average of 7854 (HRF), 10139 (JSIEC), 4539 (RIDB), 7977 (DRIVE), and 4445 (MESSIDOR) RNF centroids per fundus image per database, for an average estimated total number of 1,732,657 (which correspond to the same number of P e f f ). (f) General Garway-Heath map used to classify the results is shown in Table 1. The map is a modified version of the structure–function map developed by Garway-Heath et al. [9].
Jimaging 11 00294 g002
Figure 3. Distribution of RNFL bundles and their consistent visual field positions. At the top are the quantitative comparison of the distribution of mapped OD sectors from this study and the results published by Garway-Heath et al. [9], Lamparter et al. [10], and Jansonius et al. [13]. Each plot represents one hemifield region of the Garway-Heath et al. map [9], and each row within a plot represents the visual field sites. The blue line represents the mean standard deviation (95% limit) as reported by [9]. The red line represents the mean and upper/lower limits of the predicted OD sectors from [13]. The green line represents the mean standard deviation as reported by [10]. The black line represents the mean standard deviation (95% limit) as reported in the current study. At bottom left is the RNF bundle distribution: 34,718 RNFL bundles, based on roughly 48,930 pointwise RNF extractions, were derived from only seven subjects. A total of 52 visual field points are superimposed onto the RNFL bundle image. For reporting, the superior visual field sites are assigned to the OD entry position of their mirror image location in the inferior hemifield, and vice versa for inferior visual field site. At bottom right is the RNFL bundle distribution.
Figure 3. Distribution of RNFL bundles and their consistent visual field positions. At the top are the quantitative comparison of the distribution of mapped OD sectors from this study and the results published by Garway-Heath et al. [9], Lamparter et al. [10], and Jansonius et al. [13]. Each plot represents one hemifield region of the Garway-Heath et al. map [9], and each row within a plot represents the visual field sites. The blue line represents the mean standard deviation (95% limit) as reported by [9]. The red line represents the mean and upper/lower limits of the predicted OD sectors from [13]. The green line represents the mean standard deviation as reported by [10]. The black line represents the mean standard deviation (95% limit) as reported in the current study. At bottom left is the RNF bundle distribution: 34,718 RNFL bundles, based on roughly 48,930 pointwise RNF extractions, were derived from only seven subjects. A total of 52 visual field points are superimposed onto the RNFL bundle image. For reporting, the superior visual field sites are assigned to the OD entry position of their mirror image location in the inferior hemifield, and vice versa for inferior visual field site. At bottom right is the RNFL bundle distribution.
Jimaging 11 00294 g003
Figure 4. Quantitative comparison of the root mean square error, the mean, and standard deviation of the orientation difference between this study and the results reported by Jansonius et al. [13].
Figure 4. Quantitative comparison of the root mean square error, the mean, and standard deviation of the orientation difference between this study and the results reported by Jansonius et al. [13].
Jimaging 11 00294 g004
Figure 5. Left, the influence of the latitudinal distance of the OD center ( O D y ) on the average RNF orientation angle. The average RNF orientation angle along the y-axis is subtracted from 180. Right, standardized residuals against the predicted value for O D y .
Figure 5. Left, the influence of the latitudinal distance of the OD center ( O D y ) on the average RNF orientation angle. The average RNF orientation angle along the y-axis is subtracted from 180. Right, standardized residuals against the predicted value for O D y .
Jimaging 11 00294 g005
Table 1. Frequencies (percentage) higher than 92.5 (percentage of effectiveness) per region and average per database.
Table 1. Frequencies (percentage) higher than 92.5 (percentage of effectiveness) per region and average per database.
DatabaseNasalSuperotemporalInferotemporalInferonasalSuperonasalAverage
HRF93.3294.6299.0799.2299.8897.22
JSIEC85.4890.1098.5197.6399.7494.30
RIDB95.7795.2998.4199.8699.8097.82
DRIVE87.4189.7597.2797.6699.7494.37
MESSIDOR97.6090.8998.6199.6999.8497.32
Table 2. Multiple linear regression analyses with the average RNF orientation angle as a dependent variable and OD area, OD position, disk–fovea angle, ellipticity ratio, and FOV size as independent variables for the superior–temporal and inferior–temporal regions. All independent variables are shown, including those with a correlation coefficient lower than 0.5.
Table 2. Multiple linear regression analyses with the average RNF orientation angle as a dependent variable and OD area, OD position, disk–fovea angle, ellipticity ratio, and FOV size as independent variables for the superior–temporal and inferior–temporal regions. All independent variables are shown, including those with a correlation coefficient lower than 0.5.
Independent VariableCoefficientR2F-Statisticp-Value
Longitudinal OD position2.67 × 10−18.10 × 10−32.27 × 1001.32 × 10−1
Latitudinal OD position−4.43 × 1006.40 × 10−14.87 × 1023.75 × 10−63
OD area−2.95 × 10−23.90 × 10−31.10 × 1002.96 × 10−1
FOV size−6.20 × 10−37.40 × 10−22.23 × 1013.71 × 10−1
Ellipticity ratio3.16 × 1001.70 × 10−34.81 × 10−14.95 × 10−1
Disk–fovea angle−1.56 × 1006.20 × 10−14.55 × 1021.42 × 10−60
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Luján Villarreal, D. Automatic Algorithm-Aided Segmentation of Retinal Nerve Fibers Using Fundus Photographs. J. Imaging 2025, 11, 294. https://doi.org/10.3390/jimaging11090294

AMA Style

Luján Villarreal D. Automatic Algorithm-Aided Segmentation of Retinal Nerve Fibers Using Fundus Photographs. Journal of Imaging. 2025; 11(9):294. https://doi.org/10.3390/jimaging11090294

Chicago/Turabian Style

Luján Villarreal, Diego. 2025. "Automatic Algorithm-Aided Segmentation of Retinal Nerve Fibers Using Fundus Photographs" Journal of Imaging 11, no. 9: 294. https://doi.org/10.3390/jimaging11090294

APA Style

Luján Villarreal, D. (2025). Automatic Algorithm-Aided Segmentation of Retinal Nerve Fibers Using Fundus Photographs. Journal of Imaging, 11(9), 294. https://doi.org/10.3390/jimaging11090294

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop