1. Introduction
Currently, over 34.2 million people in the United States have Type I or Type II diabetes mellitus, and 88 million US adults have prediabetes [
1]. Diabetic retinopathy (DR), the most common form of diabetic-related eye disease, is the leading cause of visual impairment and blindness in working-age Americans, and is expected nearly to double from 2010 to 2050 (7.7 million to 14.6 million) [
2]. Approximately one-third of patients with DR are estimated to have vision-threatening complications such as proliferative diabetic retinopathy (PDR) or diabetic macular edema (DME) [
3].
Vision-threatening complications in DR arise from prolonged hyperglycemia causing a cascade of biochemical pathways. These pathways lead to oxidative stress in the retinal vasculature and subsequently cause microvascular dysfunction. Ultimately, the development of retinal ischemia in DR triggers the release of the Vascular Endothelial Growth Factor (VEGF), a signaling protein that potentiates this cycle, leading to progressive ischemia and resultant neovascularization, the hallmark of proliferative diabetic retinopathy (PDR). The use of fluorescein angiography (FA) can demonstrate microaneurysms (MAs), vascular leakage, retinal ischemia, and neovascularization (NV) [
3]. Retinal vision threatening complications of PDR can include traction retinal detachment, vitreous hemorrhage, macular edema, and ischemic changes [
4,
5,
6]. If neovascularization occurs in the anterior chamber, neovascular glaucoma may develop with potential complications such as hyphema, optic nerve disease, and ensuing vision loss [
7].
Classically, treatment for PDR has been panretinal photocoagulation (PRP). Its efficacy in halting disease progression was demonstrated in multiple studies over the last 50 years [
8,
9,
10,
11]. Therapies specifically targeting the neovascularization process, in particular the VEGF blockade, recently showed the potential to preserve and possibly reverse underlying retinal damage and vision loss that can occur in the setting of PDR. The Diabetic Retinopathy Clinical Research Network (DRCR.net) Protocol S study demonstrated that intravitreal anti-VEGF therapy was non-inferior to PRP and was potentially associated with fewer complications [
12]. Unlike PRP, which slows disease progression, anti-VEGF treatment can not only slow disease progression but can also improve the Diabetic Retinopathy Severity Scale (DRSS) score, as defined by the Early Treatment Diabetic Retinopathy Study (ETDRS). Several longitudinal studies demonstrated at least a 2-step improvement in DRSS with anti-VEGF therapy compared to sham or PRP treatment over 2, 3, and 5 years [
13,
14,
15,
16], with some showing improved visual acuity, slower progression, and decreased macular edema when compared to PRP [
14,
15].
The ETDRS research group previously defined DRSS using 7-frame fundus photographs to characterize and monitor changes in the retina over time [
15,
17,
18]. More recently, ultra-widefield photography and ultra-widefield fluorescein angiography (UWFA) have allowed for a wider field of view of up to 200°, making it possible to monitor panretinal changes with a single image. In order to standardize and more effectively track these changes, automated programs were employed to measure angiographic parameters such as microaneurysm count, leakage, and ischemia [
19,
20,
21,
22,
23,
24,
25]. Although the effects of anti-VEGF on PDR features are recognized, there is limited information regarding the effect on quantitative angiographic features on UWFA. The purpose of this study is to assess longitudinally the MA count and leakage occurring on UWFA images over a 2-year period in patients with PDR being treated with a fixed-interval intravitreal aflibercept.
2. Materials and Methods
RECOVERY (Intravitreal Aflibercept for Retinal Non-Perfusion in Proliferative Diabetic Retinopathy) is a prospective, randomized, multicenter, and open-label clinical trial (NCT02863354; IND131056;
https://clinicaltrials.gov/ct2/show/NCT02863354, accessed on 8 September 2021) as previously described [
26,
27,
28]. Institutional Review Board (IRB)/Ethics Committee approval was obtained (Sterling IRB); the tenets of the Declaration of Helsinki were followed, and the study is in accord with the Health Insurance Portability and Accountability Act of 1996. All subjects were enrolled at the Retina Consultants of Texas (Houston, Katy, and Woodlands, TX, USA).
2.1. Inclusion and Exclusion Criteria
Inclusion in the study required PDR, an EDTRS best corrected visual acuity of ≥19, and substantial nonperfusion (≥20 disk areas). Subjects with previous anti-VEGF treatment, a history of PRP or vitreoretinal surgery, a clinically relevant DME, or ≥320 μm central retinal thickness in the study eye were excluded.
2.2. Study Design
At enrollment, subjects were consented for participation in the trial and then randomized 1:1. Arm 1 (n = 20) received 2 mg (0.05 mL) intravitreal aflibercept injection (IAI) every month (q4weeks; q28 ± 7 days). Arm 2 (n = 20) received 2 mg IAI quarterly (q12weeks; q3 months). After Year 1, treatment regimens were crossed over, and the Arm 1 began receiving quarterly injections whereas the Arm 2 began receiving monthly injections until the study end at Year 2. Each visit consisted of BCVA testing with the ETDRS chart, a slit lamp examination, an indirect ophthalmoscope examination, and spectral-domain optical coherence tomography (SD-OCT) scanning.
If a patient undergoing quarterly treatment met prespecified criteria at any visit, they were to be treated every 4 weeks with IAI. These criteria were (1) increased neovascularization, (2) BCVA decrease by ≥5 letters due to progressive DME or PDR, (3) worsening of DME causing vision loss, (4) total area of retinal ischemia increasing by 10%. If a patient was then determined to be either stable or improved by these criteria, they were continued on their pre-specified treatment. All subjects could also receive rescue treatment with PRP if PDR progressed despite IAI treatment.
UWFA images were obtained after dilation at Baseline, Year 1, and Year 2 via the Optos 200Tx (Optos plc, Dunfermline, UK). Images were taken during early (~0–60 s after fluorescein dye injection), middle (~60–180 s), and late (~300 s) phases of the angiogram. The images were then transformed to stereographic projection images using proprietary software available from the manufacturer based on ray tracing each pixel with a combined optical model of the imaging device, as previously described [
15].
2.3. Outcome Measures
Early and late images were analyzed by an automated assessment platform to provide quantitative image feature extraction, including the MA count and leakage index [
21,
24,
25]. Two masked readers reviewed the program’s automated segmentation and manually corrected any errors. Three pre-specified macula-centered zones were also created to enable the calculation of zonal changes. Zone 1 defines a posterior zone with a 3-disc-diameter boundary including the fovea. Zone 2 defines a midperiphery zone with a 6-disc-diameter boundary centered at the fovea. Zone 3 defines a far periphery zone with a 9-disc-diameter boundary centered at the fovea.
The MA count analysis was performed in the mid-arteriovenous phase of the fluorescein angiogram. MAs were defined as small dots which were significantly hyperfluorescent compared to the surrounding choroidal fluorescence in early-mid-phase fluorescein angiography images [
21,
28]. For leakage analysis, an early and late phase UWFA image was selected. Leakage was defined as an area of increased hyperfluorescence in the late phase compared to the early phase. The panretinal leakage index was calculated as the area of leakage divided by the total analyzable retinal area. Values were multiplied by 100 to express as a percentage [
27,
28]. DRSS scores were derived in accordance with the Early Treatment Diabetic Retinopathy Study (ETDRS).
2.4. Statistical Analysis
Statistical analysis was performed with R software version 3.4.3 (
www.r-project.org; last accessed on 8 September 2021) and the Microsoft Excel statistical function to compare mean differences in each group between Baseline and Year 1, Baseline and Year 2, and between Year 1 and Year 2. Differences in groups were also determined per the region of interest, including the total analyzable retina, Zone 1, Zone 2, and Zone 3. RECOVERY data were also analyzed by visit: Baseline, Year 1, and Year 2.
3. Results
3.1. Clinical Characteristics
Forty subjects were randomly allocated 1:1 to Arm 1 (n = 20) and Arm 2 (n = 20). All characteristics were similar between Arm 1 and Arm 2 at Baseline. In Arm 1, the mean age was 47.7 ± 12.1 years, with 9 (45%) male and 11 (55%) female subjects. Four (20%) subjects had Type I Diabetes Mellitus (DM) while 16 (80%) were diagnosed with Type II DM, with an average Baseline glycated hemoglobin A1C (HbA1c) of 9.7 ± 2.2%. Baseline BCVA was 20/32 (ETDRS 77.8 ± 6.6), and the central subfield thickness (CST) was 279.7 ± 38.8 μm. In Arm 2, the mean age was 48.3 ± 12.0 years. Furthermore, 12 (60%) were male and 8 (40%) were female; 5 (25%) subjects had Type I DM while 15 (75%) had Type II DM. Mean Baseline HbA1c was 9.2 ± 2.7%. At Baseline, mean BCVA was 20/25 (ETDRS 79.0 ± 8.2), and the mean CST was 276.4 ± 22.7 μm.
All subjects received an average of 14.4 ± 3.8 total injections up to Year 2, with no overall significant difference between the cohorts (
p = 0.16). Arm 1 received a mean 10.95 ± 1.8 injections between Baseline and Year 1, and 4.25 ± 1.5 injections between Year 1 and Year 2. Arm 2 received an average of 3.9 ± 0.5 injections by Year 1, and 9.6 ± 4.4 injections between Year 1 and Year 2. All Baseline characteristics are listed in
Table 1.
3.2. Panretinal Leakage Index
At Baseline, the mean panretinal leakage index in Arm 1 and Arm 2 was 6.4 ± 5.4% and 4.4 ± 2.5% (
p = 0.15), respectively. At Year 1, the total leakage index of Arm 1 improved significantly to 1.5 ± 1.45% (
p < 0.001). Following crossover to q12week IAI dosing, the Arm 1 leakage index increased to 3.7 ± 3.5% at Year 2, which is a 43% decrease from Baseline (
p = 0.03), but a 59% increase following the transition to quarterly dosing (
p = 0.03). In Arm 2, panretinal leakage significantly improved to 2.5 ± 2.7% (
p = 0.04) at Year 1. Following treatment crossover to q12week IAI, further improvement was demonstrated as 1.7 ± 3.4% (31% decrease;
p = 0.12) through Year 2, but this was not a significant change from Year 1. However, the improvement from Baseline to Year 2 remained significant, demonstrating a 61% decrease from Baseline (
p = 0.008). A representative case is shown in
Figure 1, and changes in the mean leakage index in both arms are presented graphically in
Figure 2.
Overall, 26 subjects, 46% (12/26) from Arm 1 and 54% (14/26) from Arm 2, showed an improvement in the panretinal leakage index from Baseline to Year 2, while 6 subjects, 83% (5/6) from Arm 1 and 17% (1/6) from Arm 2, worsened in the leakage index. Those with the most improved leakage index had a significantly higher leakage index at Baseline (6.08 ± 4.93% vs. 3.60 ± 1.92%; p = 0.02) and also demonstrated a greater decrease in CST from Baseline to Year 1 (−29.17 ± 28.2 vs. −24.3 ± 16.0; p = 0.009) and to Year 2 (−31.83 ± 32.6 vs. −28.3 ± 16.1; p = 0.03). Notably, all subjects in Arm 1 (19/19) showed an improvement in the panretinal leakage index from Baseline to Year 1. However, in Arm 2, only 13/18 (72%) subjects showed an improvement in the leakage index before treatment crossover, but these subjects showed an overall improvement in the panretinal leakage index from Baseline to Year 2 (−4.10 ± 1.64%). The other five subjects in Arm 2 who had a worsening panretinal leakage index before crossover (5/18; 28%) demonstrated an overall increase in the panretinal leakage index from Baseline to Year 2 (1.42 ± 3.50%; p = 0.046).
3.3. Zonal Leakage Index
At Baseline, mean leakage index values in Zone 1, Zone 2, and Zone 3 were similar between Arm 1 and Arm 2. By Year 1, both cohorts exhibited a significant leakage index improvement in Zone 1, as Arm 1 decreased from 10.5 ± 6.3% at Baseline to 2.5 ± 2.3% (
p < 0.01) at Year 1, while Arm 2 decreased from 9.2 ± 4.2% at Baseline to 5.1 ± 5.9% (
p = 0.02) at Year 1. Moreover, in Arm 1, there was significant leakage index improvement in Zone 2 (6.9% to 1.5%;
p = 0.03) and Zone 3 (1.3% to 0.4%;
p = 0.05). Arm 2 had a significant improvement in Zone 2 (4.5% to 2.6%;
p = 0.045) from Baseline to Year 1, and Zone 3 had a nonsignificant decrease in the leakage index (0.9% to 0.5%;
p = 0.17). Differences in the leakage index per Zone are illustrated in
Figure 3.
Following treatment crossover, the Zone 1 leakage index increased significantly in Arm 1, but continued to decrease in Arm 2. At Year 1, the leakage index of Arm 1 in Zone 1 was 2.5% and increased to 4.9% (p = 0.01) by Year 2, representing a 53% decrease from Baseline (p = 0.006). In Arm 2, the leakage index of Zone 1 decreased from 5.1% to 2.9% (p = 0.05) by Year 2, which represents a 68% decrease from Baseline (p = 0.001). This pattern was also observed in Zone 2 and Zone 3 in both cohorts from Year 1 to Year 2, although the only significant change was seen in Zone 2 of Arm 2, which decreased to 1.2% at Year 2, a 54% decrease from Year 1 (p = 0.01), and a 73% decrease from Baseline (p < 0.001).
3.4. Microaneurysms
From Baseline to Year 1, the panretinal MA count decreased significantly and within each zone. A representative case is shown in
Figure 4. In Arm 1, the mean panretinal MA count decreased by 48% (870.7 to 455.4;
p < 0.001), 63% in Zone 1 (57.5 to 21.2;
p < 0.001), 62% in Zone 2 (148.8 to 60.6;
p < 0.001), and 52% in Zone 3 (380 to 181;
p < 0.001). In Arm 2, the panretinal MA count decreased by 50% (775.7 to 388.6;
p < 0.001), 38% in Zone 1 (39.6 to 25.1;
p = 0.009), 47% in Zone 2 (105.8 to 56.4;
p < 0.001), and 54% in Zone 3 (326.3 to 151.8;
p < 0.001). However, it is important to note that there was a significant difference in the mean MA count between Arm 1 and Arm 2 at Baseline in Zone 1 (57.5 vs. 39.6 respectively;
p = 0.03) and in Zone 2 (148.8 vs. 105.8 respectively;
p = 0.04). The changes in the mean MA count over time can be seen in
Figure 5.
The mean MA count continued to decrease after treatment crossover in both cohorts. When compared to Baseline, the MA count showed significant panretinal improvement and in all zones by Year 2. However, from Year 1 to Year 2, only Arm 2 had significant decreases in the mean MA count in all regions (panretinal,
p < 0.001; Zone 1,
p = 0.002; Zone 2,
p < 0.001; Zone 3,
p = 0.002). When compared to Baseline, Arm 1 showed a significant panretinal decrease of 53% (870.7 to 409.9;
p < 0.001), 70% in Zone 1 (57.5 to 17.5;
p < 0.001), 68% in Zone 2 (148.8 to 46.6;
p < 0.001), and 57% in Zone 3 (379.6 to 162.9;
p < 0.001) by Year 2. In Arm 2, the mean panretinal MA burden significantly decreased by 73% (775.7 to 388.6;
p < 0.001), by 75% in Zone 1 (39.6 to 9.7;
p < 0.001), by 75% in Zone 2 (105.8 to 26.6;
p < 0.001), and by 75% in Zone 3 (326.3 to 80.5;
p < 0.001) from Baseline to Year 2. These region differences are illustrated in
Figure 6.
Between Year 1 and 2, eight subjects had both an increase in the MA count and panretinal leakage index, 7/8 (87.5%) of which were from Arm 1 while 1/8 (12.5%) were from Arm 2. The subjects who exhibited a concurrent increased MA count and panretinal leakage index after treatment crossover showed a worsening in DRSS by 1.63 ± 1.68 steps in that time period, whereas those that did not had an improvement of 0.48 ± 1.21 steps (p = 0.009). Moreover, between Baseline and Year 2, those with a concurrent worsening MA and panretinal leakage index after crossover had a less significant improvement in DRSS than those that did not (+1.25 ± 1.67 vs. +3.35 ± 1.83; p = 0.01).
3.5. Functional and Anatomic Outcomes
At Baseline, the mean BCVA of Arm 1 was 20/32 (ETDRS 77.8 ± 6.6). Visual acuity significantly improved in Arm 1 from 20/32 to 20/25 (ETDRS 82.8 ± 7.8; p = 0.01) by Year 1, and remained at 20/25 (ETDRS 83.2 ± 9.7) by Year 2. Arm 2 remained at 20/25 between Baseline (ETDRS 79.0 ± 8.2) and Year 1 (ETDRS 80.3 ± 15.6; p = 0.69), but improved to 20/20 (ETDRS 87.6 ± 5.8) by Year 2, which is a significant improvement from Baseline (p < 0.001). Overall, both cohorts exhibited a significant improvement in visual acuity from Baseline to Year 2 (p < 0.001).
The mean central subfield thickness (CST) was similar between Arm 1 and Arm 2 at Baseline—279.7 ± 38.8 μm and 276.4 ± 22.7 μm (
p = 0.75), respectively. Both groups improved significantly from Baseline to Year 1, decreasing by 11.5% (279.7 to 247.5 μm;
p < 0.001) in Arm 1 and by 7.6% (276.4 to 255.3 μm;
p = 0.006) in Arm 2. Arm 1 and Arm 2 also improved significantly from Baseline by Year 2 (246.3 ± 30.9 μm;
p < 0.001, 249.6 ± 29.2 μm;
p = 0.01, respectively). The mean change in CST was moderately positively correlated to the change in the leakage index in Arm 1 in all time periods, but was only moderately positively correlated in Arm 2 after switching to monthly dosing at Year 1 (see
Figure 7).
DRSS improvement (i.e., decrease in DRSS value) was also apparent in both cohorts. From Baseline to Year 1, 100% (17/17) of subjects in Arm 1 had at least a 1-step improvement, and 70.6% (12/17) had at least a 2-step improvement. In Arm 2, 83.3% (14/16) had at least a 1-step improvement, with 75% (12/16) having improved by at least 2 steps. Between Year 1 and Year 2, 23.5% (4/17) of Arm 1 and 62.5% (10/16) of Arm 2 had at least a 1-step improvement. However, 47.1% (8/17) of subjects in Arm 1 exhibited a worsening of DRSS (i.e., increase in score value) by at least one step during that period compared to 6.25% (1/16) of subjects in Arm 2. Overall, between Baseline and Year 2, Arm 1 had 23.5% (4/17) of subjects maintain the same DRSS, 76.5% (13/17) improved by at least one step, and 58.8% (10/17) improved by at least 2 steps. In that period, 6.3% (1/16) of Arm 2 had stable DRSS, 92.7% (15/16) had at least a 1-step improvement, and 81.3% (13/16) had at least a 2-step improvement.
Subjects with at least a 1-step improvement in DRSS showed a greater reduction in the leakage index over time, as subjects that had an improved DRSS between Baseline and Year 1 showed a reduction of 2.70 ± 4.98% in the leakage index, while those without improvement had a mean increase of 1.38 ± 2.34% (p = 0.035) in that period. Similarly, between Year 1 and Year 2, subjects with at least a 1-step DRSS improvement had a mean reduction of the panretinal leakage index of 0.68 ± 1.62%, whereas subjects without improvement in DRSS had a mean increase of 1.86 ± 3.51% (p = 0.013) during that period. Although a greater reduction in the MA count over time was also observed in patients with improved DRSS, the change was not significantly different from the group with stable/worsening DRSS from Baseline to Year 1 (−395 vs. −316, respectively; p = 0.55), Year 1 to Year 2 (−206 vs. −100; p = 0.21), and Baseline to Year 2 (−568 vs. −429; p = 0.52).
Nine subjects had worsening DRSS between Year 1 and 2, 8/9 (88.9%) of whom were from Arm 1. When compared to those that did not have worsening DRSS during that period (n = 24), there was a significant difference in the mean change in the panretinal leakage index ( +3.87 ± 3.75%, −0.46 ± 1.66%, respectively; p = 0.008) and MA count (+4 ± 246, −205 ± 217, respectively; p = 0.04) between Year 1 and 2. When comparing the 8 subjects from Arm 1 with worsening DRSS to the other subjects that had stable/improved DRSS (n = 9) after treatment crossover, there was a significant difference in the Baseline MA count, as those in Arm 1 with worsening DRSS had a Baseline mean MA count of 695 ± 411 whereas those that did not worsen had 1098 ± 382 (p = 0.05).
In Arm 1, among the 9 subjects that had DRSS improvement from Baseline to Year 1, five (55.6%) had stable DRSS from Year 1 to Year 2 whereas 4 (44.4%) continued to improve. The panretinal leakage index at Baseline in the stable group was 4.32 ± 3.12% and was 9.25 ± 1.97% (p = 0.02) in the group that showed improvement, and CST measurements were 265.8 ± 36.57 μm and 310.25 ± 18.25 μm (p = 0.05), respectively. The change in the panretinal leakage index after crossover was also significantly different in these groups, as the stable group’s leakage index decreased by 1.64 ± 3.97% whereas the improved group’s leakage index decreased by 8.17 ± 1.63% (p = 0.02). The change in CST was also greater in the improved group, decreasing by a mean of 65.0 ± 22.73 μm by Year 1 and 71.0 ± 25.14 μm by Year 2, while the stable group decreased by 21.8 ± 16.67 μm (p = 0.02) from Baseline to Year 1, and by 25.8 ± 11.39 μm (p = 0.03) by Year 2 in the stable group.
3.6. Optimal Responders
Optimal responders were defined as subjects who show an improvement in the panretinal leakage index and MA count between Baseline and Year 2. In that group, 13/25 (52%) were male, 12/25 (48%) were current/previous smokers, and 14/25 (56%) had hypertension at Baseline. Overall, both Arm 1 and Arm 2 had a similar number of optimal responders (12/25, 48% vs. 13/25, 52%, respectively), with the majority having a DRSS of a moderate PDR (DRSS = 65). The distributions of DRSS at Baseline were as follows: 6 high-risk PDR (three subjects in Arm 1, three subjects in Arm 2), 16 moderate PDR (eight subjects in Arm 1, eight subjects in Arm 2), and 3 mild PDR (one subject in Arm 1, two subjects in Arm 2).
Optimal responders demonstrated a number of significant differences when compared to other subjects. Optimal responders had a significantly higher leakage index at Baseline compared to those with a worsening leakage index and/or MA count (6.32 ± 4.88% vs. 3.60 ± 1.92%; p = 0.05). Between Baseline and Year 1, the decrease in the leakage index was also greater in the optimal responders (−4.63 ± 5.3% vs. −0.12 ± 3.28%; p = 0.03). From Year 1 to Year 2, the optimal responders had a slight increase in the average panretinal leakage, but it was significantly smaller than the other subjects (0.14% vs. 3.62%, respectively, p = 0.018).
With regards to DRSS, optimal responders improved significantly from Year 1 to Year 2 by a mean 0.42 ± 1.2 steps, while other subjects demonstrated a worsening of 2.17 ± 1.6 steps (p = 0.01). Between Baseline and Year 2, optimal responders also had overall a significantly greater improvement in DRSS, improving by a mean 3.17 ± 2.0 steps compared to other subjects, which only improved by 1.0 ± 0.82 (p < 0.01). All non-optimal responders either demonstrated a worsened DRSS after treatment crossover (4/6; 66.7%) or remained with a stable DRSS (2/6; 33.3%).
3.7. Complications
An analysis of vision-threatening complications in both cohorts revealed that no subjects developed center-involving DME at any point during the study. In addition, no subject required Panretinal Photocoagulation (PRP) over the 2-year time period; however, vitreous hemorrhage (VH) did occur. When excluding the 3 subjects with VH at Baseline (one from Arm 1, two from Arm 2), VH occurred in 7 out of 37 subjects (18.9%), with 5/7 (71.4%) occurring in Arm 1. Subjects who developed VH after Baseline contained the top 20% of the highest values for the Baseline leakage index, with one subject having a Baseline leakage index of 26.08%, the highest of all subjects. These subjects also had a smaller step improvement in DRSS between Baseline and Year 2 compared to those without VH (1 ± 0.89, 3.21 ± 1.93, respectively; p < 0.01), as well as a smaller decrease in CST (−20.3 ± 8.69 μm vs. −48.7 ± 58.4 μm; p = 0.03). The occurrence of VH did not show a significant difference in the Baseline leakage index or microaneurysm count. No significant difference in age, BMI, Baseline HbA1C, or duration of diabetes was noted between subjects with VH and those without VH.
In the optimal responders, two (2/25, 8%) had VH at Baseline while three developed VH after Baseline (3/25, 12%). All three optimal responders that developed VH after Baseline were from Arm 1, while the two subjects with Baseline VH were from Arm 2. Notably, the optimal responders that developed VH after Baseline were reported to have VH on examination in an average of 10 out of 18.7 (53.5%) visits: 7.3 mean visits between Baseline and Year 1 and 2.7 mean visits between Year 1 and Year 2. Two subjects developed VH after Baseline that were not optimal responders (both in Arm 1), and they had VH reported in a mean 3.5 out of 20 (17.5%) visits; all of these were reported between Year 1 and Year 2.
4. Discussion
The semi-automated analysis of ultra-widefield angiography images has enabled an in-depth assessment of angiographic features. Changes in retinal leakage and microaneurysm burden were recently quantitatively analyzed to allow a more detailed understanding of the effects of anti-VEGF treatment on the retina [
5,
29,
30,
31]. Recent studies on patients with DR treated by anti-VEGF showed a significant decrease in retinal leakage and microaneurysm count on UWFA [
32,
33,
34,
35,
36]. The 1-year RECOVERY results also highlighted the effects of dosing frequency on the leakage index and microaneurysm count dynamics over 52 weeks [
27,
28].
In this 2-year trial, subjects undergoing IAI therapy for DME exhibited a significant decrease in the leakage index and microaneurysm count from Baseline up to Year 2. However, each cohort had significant differences in the rate of improvement, which is further highlighted after treatment crossover at Year 1. Between Baseline and Year 1, Arm 1 (monthly IAI from Baseline to Year 1) showed significant improvement in panretinal and zonal leakage, while the Arm 2 (q12 week IAI from Baseline to Year 1) cohort had significant improvement in all regions except in Zone 3 (9-disc-diameter). More importantly, after treatment crossover, Arm 1, which switched to q12 week IAI injections, saw a worsening in the panretinal leakage index, increasing significantly from Year 1 to Year 2. In contrast, Arm 2, which switched to q4 week IAI, continued to show improvements in the leakage index in all zones, although those changes were not statistically significant. Microaneurysm burden differences were also apparent between cohorts when comparing changes before and after treatment crossover. From Baseline to Year 1, both cohorts showed significant decreases in the MA count in all zones. However, after crossover, only Arm 2 showed a significant improvement in the mean panretinal MA count and in all zones.
In addition to leakage and MA burden, changes in DRSS scores have been a parameter of significant interest in subjects undergoing anti-VEGF treatment. Specifically, the clinical importance of DRSS changes have been examined in relation to visual acuity outcomes [
13,
15,
37]. A study in 2017 by Ip and colleagues showed that subjects with stable or improved DRSS levels had a greater improvement in BCVA letter scores, whereas those with worsening DRSS had a smaller improvement. Moreover, significant changes in the BCVA letter score (>/=15 letters) were more common in patients with at least a 2-step improvement in DRSS [
37]. Another study in 2018 by Dhoot and colleagues demonstrated only Baseline DRSS having a strong association with a 2-step DRSS improvement by Year 2, while age, duration of diabetes, HbA1C, BMI, and BCVA had no correlation [
38].
In RECOVERY, most subjects (85%) exhibited an improvement of at least 1 step in DRSS between Baseline and the end of the study at Year 2. Similar to the leakage index and MA count, differences were noted when comparing cohorts before and after treatment crossover. Between Baseline and Year 1, all subjects (100%) in Arm 1 and the majority of Arm 2 (83%) had improved DRSS (decrease in score value). However, after crossover, a substantially higher proportion of Arm 2 subjects continued to improve in DRSS by at least 1 step compared to Arm 1. Moreover, from Baseline to Year 2, a higher ratio of subjects in Arm 2 exhibited a 2-step improvement than subjects in Arm 1. Overall, by the end of the study, a larger proportion of subjects in Arm 2 had an improved DRSS than in Arm 1. Notably, the worsening of DRSS (increase in score value) was only seen after crossover, and the majority of cases were from Arm 1. Overall, the subjects with worsening DRSS after crossover showed a significant worsening in the MA count in that period compared to the significant improvement in the MA count seen in those with stable/improved DRSS. Subjects with worsening DRSS after crossover also showed a significantly larger worsening of the panretinal leakage index during that period. Moreover, there was no significant difference in the age, BMI, Baseline HbA1C, and duration of diabetes noted between subjects in those that improved by 1 step or by 2 steps compared to those who did not improve by Year 2.
Changes in DRSS in relation to the leakage index, MA count, and CST were also investigated. The subjects with an improvement in the DRSS score by at least 1 step showed a greater reduction in the leakage index compared to those without improvement. In Arm 1, subjects with an increased DRSS before treatment crossover and who continued to improve after crossover had a significantly larger leakage index and CST at Baseline compared to those with a stable DRSS after crossover. Moreover, those who continued to improve also had a significantly larger decrease in the leakage index and CST between Baseline and Year 1, and a greater decrease in mean CST was shown from Baseline to Year 2. The change in BCVA was also different between those two groups in Arm 1, as those who continued to improve in DRSS showed a significantly greater increase in ETDRS letters by Year 1. In contrast, subjects in Arm 2 who improved in DRSS before crossover and continued to improve had no significant differences in any parameter compared to subjects who improved before crossover and did not improve after crossover.
The results of this study highlight the potential worsening in the leakage index, MA counts, and DRSS associated with a decreasing IAI frequency from monthly to q12 week dosing. Both the PRIME [
39] and PANORAMA [
40,
41] trials suggested that once an improvement in DRSS is achieved, it can be sustained with q12–16 week intravitreal anti-VEGF, but this was not the case in our study. In contrast, our study showed that a gradual induction of q12 week dosing followed by a more frequent q4 week dosing resulted in better functional and anatomic outcomes overall, especially in terms of leakage and DRSS. An important distinction from both the PRIME study and PANORAMA study is the greater severity of PDR in the current study. This difference in findings supports the need for the frequent monitoring of DR and the potential value of concurrent UWFA imaging. It also suggests that applying a treatment regimen across all patients with PDR may be less efficacious than image-guided personalized therapeutic decision-making.
Additionally, this study and recent publications demonstrate a clear link between the leakage index and DRSS [
27,
39]. Specifically, the progression of DR and hence the worsening of DRSS may be preceded by increases in the leakage index on UWFA, as evidenced by the PRIME Trial (a prospective, randomized, phase 2 trial that enrolled 40 eyes to assess the safety and efficacy of the real-time leakage index vs. DRSS scoring at monthly visits to guide PRN retreatment in both PDR and NPDR patients) [
39]. However, repeated regular UWFAs to assess the leakage index in DR is not without drawbacks, and there is a clear appeal to using DRSS for regular DR monitoring because of its non-invasive nature, efficiency, and ease of image capture.
Traditionally, DRSS has been determined by using a central reading center (CRC). The PANORAMA study demonstrated the necessity for trained readers or CRC for the determination of DRSS, when the disparity between the investigator and CRC reader proved the investigator wrong in 50% of cases [
40]. Further, the need for precise DRSS and the ability to monitor changes in DRSS is becoming of greater importance, not just in terms of DR progression, but for morbidity and mortality in DM. A recent study in Diabetes Care by Fahrmann demonstrated for the first time the strong correlation between DCCT-ETDRS levels ≥3 (Level 35; mild NPDR) and cardiovascular complications in young type 1 diabetics [
42]. ACCORDION, a 9-year observation follow-up study to ACCORD, emphasizes the benefits of the personalization of DR treatment and the potential harm in treating all patients with a one-size-fits-all approach [
43]. The authors concluded that patients should be stratified according to their microvascular complications, as their results revealed that DR patients who have undergone vitrectomy or retinal laser photocoagulation derived benefits from intensive glucose control, while those who had not were at an increased cardiovascular risk. While this study used a binary approach to DR, the opportunity for further research in the relationship between DRSS, cardiovascular disease, and glucose control is an unmet need. Multiple other studies demonstrated the relationship between DR and all-cause mortality, further highlighting the increasing importance of DRSS determination in clinical practice [
44,
45,
46,
47].
Specific limitations of this study include the relatively small number of subjects, presence of artifacts such as eyelashes and lenticular opacities, and errors that may occur analyzing different or low-quality images. Strengths of the study include its prospective nature and the standardization of image assessment for each parameter through the use of an image reading center and previously validated techniques in image analysis and quantification.