Next Article in Journal
Improved YOLOv7-Based Algorithm for Detecting Foreign Objects on the Roof of a Subway Vehicle
Next Article in Special Issue
MAM-E: Mammographic Synthetic Image Generation with Diffusion Models
Previous Article in Journal
A Multiplier-Free Convolution Neural Network Hardware Accelerator for Real-Time Bearing Condition Detection of CNC Machinery
Previous Article in Special Issue
Convolutional Networks and Transformers for Mammography Classification: An Experimental Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Inpainting Saturation Artifact in Anterior Segment Optical Coherence Tomography

Electronics Information Engineering College, Changchun Univesity, Changchun 130022, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(23), 9439; https://doi.org/10.3390/s23239439
Submission received: 24 September 2023 / Revised: 22 November 2023 / Accepted: 23 November 2023 / Published: 27 November 2023
(This article belongs to the Special Issue Image Analysis and Biomedical Sensors)

Abstract

:
The cornea is an important refractive structure in the human eye. The corneal segmentation technique provides valuable information for clinical diagnoses, such as corneal thickness. Non-contact anterior segment optical coherence tomography (AS-OCT) is a prevalent ophthalmic imaging technique that can visualize the anterior and posterior surfaces of the cornea. Nonetheless, during the imaging process, saturation artifacts are commonly generated due to the tangent of the corneal surface at that point, which is normal to the incident light source. This stripe-shaped saturation artifact covers the corneal surface, causing blurring of the corneal edge, reducing the accuracy of corneal segmentation. To settle this matter, an inpainting method that introduces structural similarity and frequency loss is proposed to remove the saturation artifact in AS-OCT images. Specifically, the structural similarity loss reconstructs the corneal structure and restores corneal textural details. The frequency loss combines the spatial domain with the frequency domain to ensure the overall consistency of the image in both domains. Furthermore, the performance of the proposed method in corneal segmentation tasks is evaluated, and the results indicate a significant benefit for subsequent clinical analysis.

1. Introduction

The cornea is an important refractive medium in the human eye, participating in the transmission and processing of visual information. The corneal segmentation technique provides support for the diagnosis and treatment of corneal diseases. Accurate segmentation is crucial, as a few micrometers of corneal segmentation errors can lead to significant changes in the derived clinical parameters [1]. Through corneal segmentation, the precise corneal morphology information [2] can be provided for the diagnosis and evaluation of diseases such as keratoconus [3], while also supporting preoperative preparation for procedures such as refractive surgery [1]. Anterior segment optical coherence tomography (AS-OCT) is a non-invasive imaging technique with a range of potential clinical applications [4,5], which provides high-resolution images of the anterior segment at the micron-scale resolution. AS-OCT achieves the resolution of the anterior and posterior surfaces of the entire cornea, as shown in Figure 1. However, during the imaging process, a strip-like saturation artifact (Figure 1a) is commonly generated, which covers the corneal surface and blurs the corneal edge. This is because the incident light source’s behavior at that point is normal to the tangent of the corneal surface. This high-intensity and high-contrast artifact negatively impacts the accuracy of corneal segmentation, making it challenging to obtain precise corneal morphology information. To improve the reliability and accuracy of corneal segmentation in various clinical applications and research studies, it is crucial to develop an inpainting method to eliminate the effects of the saturation artifact [6].
Recently, deep learning technology [7] has shown superiority in fields such as image processing [8]. AS-OCT image inpainting refers to the process of reconstructing the saturation artifact region and simultaneously maintaining the overall consistency of the image. Deep generative methods [9,10,11,12,13,14] are currently dominant, which effectively extract meaningful semantics from the image to be repaired and recover reasonable content with high visual fidelity, due to their powerful feature learning ability. Although these methods have achieved good performance with natural images, there are some differences between AS-OCT images and natural data, including the structure, noise, resolution, etc. These differences make it difficult to apply natural inpainting methods directly to AS-OCT saturation artifact restoration. AS-OCT inpainting needs to recover lost details more precisely to achieve the texture rationality and biological characteristics of eye tissue, such as the posterior surface of Bowman’s layer (BL).
Nowadays, many methods have been proposed for OCT image restoration. In order to avoid the problem of inaccurate quantification of the choroidal vascular system caused by retinal vascular shadows, Zhang et al. [15] proposed a three-stage image-processing framework to remove shadows in the retina, which may not be suitable for diseases with abnormal retinal pigment epithelial cell (RPE) layers. Cheong et al. [16] recommended the DeshadowGAN method for optical coherence tomography images of the optic nerve head vascular shadow, which performed well on healthy-eye OCT images. However, it is uncertain whether it can achieve the same performance in eyes with pathological conditions such as glaucoma. Liu et al. [17] recently projected a dictionary-based sparse representation method for saturation artifact removal in OCT images. Although this method works very well on narrow-range saturation artifacts, it can be less effective when presented in large areas of shadow. To solve this problem, Tang et al. [18] proposed a multi-scale framework for shadow repair to achieve the repair effect of both wide and narrow retinal vascular shadows, but this repair effect on real shadows cannot be consistent with the composite shadow restoration effect. Nevertheless, these methods are proposed for unique OCT images collected from the various OCT devices. Due to the different data characteristics collected using different OCT devices, it is challenging to directly use existing OCT repair methods for saturation artifact restoration in AS-OCT images. There is an urgent for a dedicated inpainting method to remove saturation artifact in AS-OCT images.
To solve the issues of artifacts and speckle noise patterns and precisely segment the shallowest tissue interface in an AS-OCT image, Ouyang et al. [6] proposed a cascaded neural network framework. However, this frame only removes saturation artifacts and speckle noise patterns just above the shallowest tissue interface via cGAN [19], and does not completely remove the saturation artifact penetrating the center of the cornea. Recently, in order to address the issue of stripe artifacts damaging the visual quality of images and affecting automatic ophthalmic analysis, Bai et al. [20] proposed SC_GAN, which can successfully removed artifacts but cannot significantly improve clinical segmentation accuracy.
In this paper, we treat saturation artifact inpainting as converting the artifact image into an artifact-free image. A new generative model is proposed to inpaint the corneal structure and texture obscured by saturation artifacts, as well as maintaining the overall consistency of the image. The main novelties and contributions are as follows:
  • The dual-domain transformation capability of DualGAN [21] is designed to achieve AS-OCT saturation artifact inpainting by converting the artifact image into an artifact-free image. The structural similarity loss for reconstructing the structure and texture of the cornea is incorporated;
  • A frequency loss that combines the spatial and frequency domains is introduced to ensure the overall consistency of the images in both domains;
  • The repair experiments on both synthetic and real artifacts are devised. The results indicate that the proposed methods can restore artifacts in different situations. To confirm the clinical value of saturation artifact inpainting, segmentation experiments are designed on the three corneal boundaries of real artifact-inpainted images, including the anterior surface of the epithelium (EP), the posterior surface of Bowman’s layer (BL), and the posterior surface of the endothelium (EN). The experimental results demonstrate that the method significantly enhances the precision of corneal segmentation, proving to be more accurate than other repair techniques.
The remainder of the paper is organized as follows: We introduce our approach in Section 2. In Section 3, the experiments are described, specifically for synthesis and real artifact restoration (Section 3.3), segmentation experiments for real restoration verification (Section 3.4), and ablation experiments (Section 3.5), followed by the conclusions in Section 4.

2. Proposed Method

In this work, AS-OCT image inpainting is regarded as a translation task from artifact images to artifact-free data. As shown in Figure 2, the proposed method is built on DualGAN, including two generators ( G A , G B ) and two discriminators ( D A , D B ), and introduces two loss terms: structural similarity and frequency loss. By giving two paired images sampled from domains S and W, the forward task of the network is to learn the generator G A : S → W that maps the images s S to w W , and the backward task is to train the generator G B : W → S. The generated image and ground_truth image are constrained using four different loss functions, including L F , L S S I M , L a d v , and L 1 , making the former more similar to the latter. L F combines the spatial and frequency domain of the image. Moreover, the reconstruction loss, L r e c o n , further restrains the reconstructed image to be consistent with the ground_truth. Two discriminators are used to evaluate the fit between the output of the corresponding generator and the ground_truth. In this section, the generator, discriminator, and loss functions are described in detail.

2.1. Network Architecture

The network consists of two generators ( G A , G B ) and two discriminators ( D A , D B ). As shown in Figure 3, the two generators achieve mutual conversion between S-domain and W-domain images, and the two discriminators are used to distinguish the generated image and real data. Generator G A generates artifact-free images, while G B generates an artifact image with a central stripe pattern. Discriminator D A is trained to distinguish between real samples in the W-domain and data generated from the S-domain, while D B is trained in the opposite manner.
Both generators, G A and G B , employ the U-Net [22] architecture with eight convolutional layers for the encoder and decoder, as shown in Figure 4a. The encoder consists of a convolutional layer, six Relu-Conv-LayerNorm (RCL) blocks, and one Relu-Conv (RC) block. The decoder comprises seven Relu-ConvTranspose-BatchNorm-Dropout (RCBD) blocks and one Relu-ConvTranspose-Tanh (RCT) block. The encoder and decoder are connected through skip connections.
Both discriminators, D A and D B , use the PatchGAN [23] structure, as shown in Figure 4b. The receptive field of PatchGAN is 70 × 70, and it consists of five 4 × 4 convolutional layers. Specifically, it includes a Conv-LeakyRelu (CL) block, three Conv-BatchNorm-LeakyRelu (CBL) blocks, and a Conv-Sigmoid (CS) block.

2.2. Objective Functions

Formally, image s S is input into generator G A to derive the image G A s W , and the image is processed through G B to obtain the reconstructed image, G B ( G A s ) S . Similarly, image w W is passed through generator G B to achieve the image G B ( w ) S , which is input into G A to obtain the reconstructed image, G A ( G B w ) W . Discriminator D A evaluates the degree of fit between G A s and the real w . Discriminator D B estimates the measure of fit between the fake s generated with G B and the real s .
In practice, generator G A generates the image s ^ . In order to avoid network damage to other organizational structures of image s , the images, s and s ^ , and the corresponding mask ( m ) were integrated; thus, the inpainted results can be denoted as follows:
G A ( s ) = s   ʘ ( 1 m ) + s ^ ʘ m
where ʘ is the pixel-wise multiplication. Equation (1) ensures that image s is only replaced by the generated image in the artifact region, while other regions remain unchanged.
SSIM Loss: Although DualGAN achieves image translation between two domains, it distorts the corneal boundary structure and texture information in AS-OCT images. Structural similarity (SSIM) [24] is a powerful tool for image quality assessment. Generally, a higher SSIM means that the image has clearer results. SSIM has been widely used in tasks such as image restoration [25], semantic segmentation [26], and dehazing [27] since it was proposed. Therefore, SSIM is adopted as a loss function to train the network to reconstruct the corneal structure and can improve the corneal segmentation accuracy. Since SSIM is pixel-based, this loss function has constraints on corneal structure reconstruction and helps to restore texture information. SSIM for two images, x and y (x is the ground_truth, y denotes the repaired image), is defined as follows:
S S I M ( x , y ) = ( 2 μ x μ y + C 1 ) ( 2 σ x y + C 2 ) ( μ x 2 + μ y 2 + C 1 ) ( σ x 2 + σ y 2 + C 2 )
where µ and σ represent the standard deviation and covariance of images, respectively. C 1 and C 2 are constants. A higher SSIM indicates that the two images are more similar to each other, and the SSIM equals 1 for identical images. The loss function for the SSIM operates on both generators and can then be written as follows:
L S S I M = E W 1 S S I M w , G A s + E S 1 S S I M s , G B w
Frequency Loss: In AS-OCT images, non-tissue structure saturation artifact is usually positioned at high frequencies, which could overshadow or interfere with other frequency components. There are gaps between the real and generated images, especially in the frequency domain [28,29], as shown in Figure 5. For different images, their frequency domain distribution also varies. Figure 5b,c, respectively, show the frequency domain distribution of synthesized and real saturated artifact images, which have significant differences from the that of artifact-free images. Compared to Figure 5a, these have significant changes in the horizontal direction in the spatial domain, so their frequency domain distribution has a bright line in the horizontal direction, especially in Figure 5b, and the brightness in the central low-frequency region is also different.
To minimize the disparities, the frequency loss is introduced, which is the main idea of obtaining frequency domain characteristics of the real and generated images by performing a fast Fourier transform (FFT); in other words, make Figure 5d infinitely close to Figure 5a to achieve the same. Previous studies [30,31] found it beneficial to replace distance L 2 with distance L 1 , since the former often leads to blurriness. The distance formulas for L 1 and L 2 are Equations (4) and (5), respectively. Hence, the L 1 distance is adopted to measure the distance between the real image and the generated image after FFT. The spatial and frequency domains are combined to further improve the quality of restoration. The frequency loss function operates on both generators and can then be written as Equation (6).
| | L | | 1 = i = 1 n | x i y i |
| | L | | 2 = i = 1 n ( x i y i ) 2
L F = E W | F ( G A ( s ) ) F ( w ) | 1 + E S [ | | F ( G B ( w ) ) F ( s ) | | 1 ]
where F represents the fast Fourier transform, transforming the image from the spatial distribution to the frequency domain via FFT.
L 1  Loss: we use L 1 loss to measure the difference between the generated image and the real image, to avoid some pixels being smoothed by transitions, resulting in the resulting image missing detail and texture information. L 1 loss is defined as
L 1 = E W | G A ( s ) w | 1 + E S [ | | G B ( w ) s | | 1 ]
Adversarial Loss: the adversarial loss acts between two pairs of generators–discriminators ( G A D A ,   G B D B ), and is defined as
L a d v G A , D A , G B , D B = E w ϵ W l o g D A w + E s ϵ S [ l o g ( 1 D A ( G A ( s ) ) ) ] + E s ϵ S [ l o g D B ( s ) ] + E w ϵ W [ l o g ( 1 D B ( G B ( w ) ) ) ]
Discriminator D A is trained with the real w as positive samples and the generated G A ( s ) as negative examples, whereas D B takes the real s as positive and G B ( w ) as negative. Generators G A and G B are optimized to emulate ‘fake’ outputs to blind the corresponding discriminators, D A and D B .
Reconstruction Loss: the L 1 distance between the reconstructed image and the real image is adopted as the reconstruction loss, formulated as
L r e c o m = E s ϵ S | | G B G A s s | | 1 + E w ϵ W [ | | G A G B w w | | 1 ]
Total Losses: the whole objective function of the proposed network can be written as
L t o t a l = λ S S I M L S S I M + λ F L F + λ 1 L 1 + λ a d v L a d v + λ r e c o n L r e c o m
where λ S S I M , λ F , λ 1 , λ a d v , and λ r e c o n are the tradeoff parameters, and set 50, 1, 100, 1, and 1, respectively.

3. Experimental Setup and Results

In this section, the details of data preprocessing are presented in Section 3.1, and some parameter settings during the training process are explained in Section 3.2. Quantitative and qualitative evaluations of the synthetic artifact restoration are conducted in Section 3.3, and a visualization of real restoration is performed. Then, in Section 3.4, the real repair quality is verified through corneal segmentation experiments. Section 3.5 introduces extensive ablation studies to validate the effectiveness of various components of the model.

3.1. Data Preprocessing

The AS-OCT image used in this article is from a CASIA1 [32] ophthalmology device (Tomey Inc., Nagoya, Japan), using a swept-source OCT (SS-OCT) with a scanning speed of 30,000 A-ultrasound/second, a wavelength of 1310 nm, and a frequency of 50 Hz, whilst preserving an original image size of 1689 × 1000 . To more accurately repair saturation artifacts throughout the corneal tissue, the AS-OCT image is cropped to the size of 256 × 256 to obtain the image, I.
Since saturation artifact images do not have paired data, the binary mask ( m ) (with value 0 for known pixels and 1 for the area to be repaired) is manually added to artifact-free image I with varying degrees of inclination in the corneal tissue to simulate saturation artifacts in the spatial domain. Figure 6 shows the process of obtaining the synthetic artifact image s = I × m , where the position and width of the mask are different. The paired data for image s are the artifact-free image w = I . A total of 3774 pairs of synthetic images were obtained; hence, 3374 were employed as a training dataset, while the remaining subsets served as test set. To verify the effectiveness of the network, the mask in the test set was wider than the mask added in the training set, and 80 AS-OCT real artifact images were tested.

3.2. Training Parameters

All training and testing is performed on a single NVIDIA GeForce RTX 3090 GPU (24 GB). Using the RMSprop optimizer, the learning rates of the generator and discriminator are 0.0005, and 0.0001, respectively, and the batch size is 4.

3.3. Evaluation of Inpainting

To validate the proposed method, the model is applied for a comparison with existing deep-learning inpainting methods, including MADF [9], CTSDG [10], AOT-GAN [11], RFR [12], G&L [13], PICNet [14], and DualGAN [21].
Synthetic Saturation Artifacts’ Qualitative Comparison: the comparative experiments on synthetic artifact restoration are conducted in three different situations:
  • The repair effects of different methods on corneal tissue images with different tilt degrees under the same mask conditions are shown in Figure 7;
  • The inpainting results of different methods on corneal tissue images with the same inclination combined with different masks are shown in Figure 8. Briefly, Figure 8 shows the image restoration effects of three groups, adding different masks to the same AS-OCT image. Figure 8(I), (II), and (III) respectively show the repair results of images with downward tilt of corneal tissue, images with no tilt degree of corneal tissue, and images with upward tilt of corneal tissue with different masks added.
  • The results of adding different masks into the corneal tissue images with different degrees of inclination using different methods are shown in Figure 9.
As shown in these renderings, the MADF method can achieve non-tilted corneal tissue image restoration, but there may be some structural defects for wide shadows and tilted corneal tissue images. For CTSDG, AOT-GAN, and G&L, these methods can repair the corneal structure, but they retain blurry or overly smooth textures in shaded areas, especially in wide-shaded images. The RFR approach may effectively rectify backdrop pixels, yet manifests a limited capacity when addressing the corneal structure and texture. The PICNet method has a certain repair effect on the mask area, but its processing ability of the narrow-band artifact’s edge is significantly weak, resulting in incomplete restoration. Significate repair marks occur with the DualGAN method, which can distort the corneal structure and texture. The proposed method can successfully rectify artifacts across varying widths and positions, irrespective of the degree of inclination of the corneal tissue.
Synthetic Saturation Artifacts’ Quantitative Comparison: To quantitatively evaluate the inpainting quality, the means of the peak signal-to-noise ratio (PSNR), SSIM, and learned perceptual image patch similarity (LPIPS) [33] are calculated for the repaired regions from different techniques on the synthetic testing set. Table 1 shows the results, with the best performance value written in bold. For CTSDG, AOT-GAN, and G&L, these methods add noise beyond the original image to the repair results. As the PSNR is more sensitive to noise, these methods achieve higher values in this metric. In addition, the SSIM values indicate that these methods have a positive effect on the reconstruction of corneal structures. The RFR method is unable to repair the structure and texture of the cornea, resulting in low values for various indicators. The MADF, PICNet, and DualGAN methods have varying degrees of inpainting effects, but due to the distortion of the corneal structure in the repair results, low SSIM values are achieved. Although our method cannot achieve optimal results in the PSNR and SSIM metrics for synthetic saturated artifacts, the image quality is also crucial for corneal boundary segmentation [34]. Therefore, it is necessary to pay attention to the corneal tissue structure and texture information simultaneously. LPIPS can better simulate human visual perception and subjective quality judgment. For synthetic artifact restoration, the recommended method achieves the best performance in LPIPS.
Real Saturation Artifact Qualitative Comparison: Figure 10 shows the inpainting performance of different methods for real artifacts, and these comparative methods show consistent repair results with the synthetic saturation artifacts. Specifically, the MADF method can repair narrow-range artifacts, while repairing wide-range artifacts can lead to corneal structural protrusions. The CTSDG, AOT-GAN, and G&L methods bring blurry or smooth texture details, regardless of the wideness of the artifacts. The RFR method can only play a certain role in repairing background pixels. The PICNet method performs well in wide artifact restoration, but it may have blurring for narrow-range artifacts and has weak edge-restoration capabilities. The baseline network, DualGAN, distorts the corneal structure. Our model can repair saturation artifacts with different widths and positions, both synthetic and real, reconstructing the corneal structures and repairing textural details.

3.4. Evaluation on Segmentation

To further evaluate the inpainting performance of different methods on machine analysis, experiments on a corneal segmentation task are conducted. Concretely, the widely used U-Net segmentation model was trained using the labelled AS-OCT image; 31 AS-OCT images are randomly selected from the real repaired saturation artifact to test the segmentation effect. The inpainting and segmentation results of each method are shown in Figure 11. The visual segmentation effect shows that the proposed method maintains overall image consistency while removing the saturation artifact and reconstructing the corneal structure and textural details, due to the constraints of structural similarity and frequency domain loss. In addition, the performance of different inpainting methods is analyzed by calculating the Dice similarity coefficient (DSC), pixel accuracy (PA), F1-score, and Jaccard value of the segmentation results. The calculation formulas for DSC, PA, F1-score, and Jaccard are shown in Equations (11)–(14).
D S C = 2 | X Y | | X | + | Y |  
where X and Y represent the generated image and the ground_truth image, respectively, and X Y is the intersection of the two.
P A = T P + T N T P + T N + F P + F N
where TP is the number of true positives, TN is the number of true negatives, FN is the number of false negatives, and FP is the number of false positives.
F 1 = 2 T P 2 T P + F N + F P
J a c c a r d = | X Y | | X Y |
where X Y signals the union of the generated image and the ground_true.
As shown in Table 2, our method is ahead of other methods in calculating various indicators, especially DSC, with a maximum DSC value of 0.732, which is not presented in other methods. The average values calculated for each indicator are provided in the table, indicating that our inpainting method has significantly improved the corneal segmentation.

3.5. Ablation Studies

In this section, two sets of experiments are conducted to verify the contribution. Here, the impact of structural similarity loss and frequency loss on the repair of saturation artifacts are studied. Specifically, the structural similarity and frequency losses are removed, respectively, and their impact on the experiment is observed. From the results in Figure 12, it can be seen that removing the structural loss will seriously affect the texture and structural information, especially in Bowman’s layered structure. Meanwhile, the segmentation indicators in Table 3 also show low segmentation accuracy due to the incomplete corneal structure. Unlike removing structural loss, although removing frequency loss can maintain almost the same visual effect as the suggested method, Table 3 also indicates that adding frequency loss can improve the accuracy of corneal segmentation. In summary, the results reveal the role of structural similarity loss and frequency loss in the repair of saturation artifacts, and also confirm the superiority of the proposed method.

4. Conclusions

In this study, a novel deep generative method is proposed to remove saturation artifacts generated during AS-OCT imaging. The proposed model is a bilateral network and introduces two loss terms, namely structural similarity loss and frequency loss. The structural similarity loss reconstructs structural and textural details while ensuring the overall image consistency. In the frequency domain, the saturated artifact is located in the high-frequency component, and frequency loss is designed to eliminate image distortion in the high-frequency region. Moreover, frequency loss combines the spatial domain with the frequency domain to reduce the gap between the generated image and the real image. The repair comparison experiment and ablation experiment demonstrate the effectiveness of the innovative points. The results show that this method can successfully remove saturation artifacts and restore the structure and texture information of the cornea. The corneal boundary segmentation experiment further verifies that the proposed method effectively improves the quality of AS-OCT images and significantly improves the accuracy of corneal segmentation.
Although this method has achieved satisfactory results, there are still some limitations to the framework. Specific device data are adopted to obtain these valuable results, while the image restoration performance of other OCT devices is not determined in this study.
In summary, an inpainting method was proposed to remove AS-OCT saturation artifact while preserving the surroundings. The proposed network is of great significance for the improvement of corneal segmentation accuracy.

Author Contributions

Conceptualization, J.L., H.Z. and G.B.; methodology, H.Z.; software, H.Z.; validation, H.Z.; formal analysis, J.L.; investigation, J.L.; resources, J.L.; data curation, H.Z.; writing—original draft preparation, H.Z.; writing—review and editing, J.L., X.W. and H.Z.; visualization, H.Z.; supervision, J.L., H.W. and J.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are not publicly available due to the privacy of clinical data.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, Y.; Shekhar, R.; Huang, D. Segmentation of 830-and 1310-nm LASIK corneal optical coherence tomography images. In Proceedings of the Medical Imaging 2002, Image Processing, SPIE, San Diego, CA, USA, 23–28 February 2002; Volume 4684, pp. 167–178. [Google Scholar] [CrossRef]
  2. Elsawy, A.; Gregori, G.; Eleiwa, T.; Abdel-Mottaleb, M.; Shousha, M.A. Pathological-Corneas Layer Segmentation and Thickness Measurement in OCT Images. Transl. Vis. Sci. Technol. 2020, 9, 24. [Google Scholar] [CrossRef] [PubMed]
  3. dos Santos, V.A.; Schmetterer, L.; Stegmann, H.; Pfister, M.; Messner, A.; Schmidinger, G.; Garhofer, G.; Werkmeister, R.M. CorneaNet: Fast segmentation of cornea OCT scans of healthy and keratoconic eyes using deep learning. Biomed. Opt. Express 2019, 10, 622–641. [Google Scholar] [CrossRef] [PubMed]
  4. Vincent, S.J.; Alonso-Caneiro, D.; Collins, M.J. Optical coherence tomography and scleral contact lenses: Clinical and research applications. Clin. Exp. Optom. 2018, 102, 224–241. [Google Scholar] [CrossRef] [PubMed]
  5. Ang, M.; Baskaran, M.; Werkmeister, R.M.; Chua, J.; Schmidl, D.; dos Santos, V.A.; Garhöfer, G.; Mehta, J.S.; Schmetterer, L. Anterior segment optical coherence tomography. Prog. Retin. Eye Res. 2018, 66, 132–156. [Google Scholar] [CrossRef] [PubMed]
  6. Ouyang, J.; Mathai, T.S.; Lathrop, K.; Galeotti, J. Accurate tissue interface segmentation via adversarial pre-segmentation of anterior segment OCT images. Biomed. Opt. Express 2019, 10, 5291–5324. [Google Scholar] [CrossRef]
  7. Yang, Y.; Gao, D.; Xie, X.; Qin, J.; Li, J.; Lin, H.; Yan, D.; Deng, K. DeepIDC: A Prediction Framework of Injectable Drug Combination Based on Heterogeneous Information and Deep Learning. Clin. Pharmacokinet. 2022, 61, 1749–1759. [Google Scholar] [CrossRef] [PubMed]
  8. Khan, Y.D.; Mahmood, M.K.; Ahmad, D.; Al-Zidi, N.M. Content-Based Image Retrieval Using Gamma Distribution and Mixture Model. J. Funct. Spaces 2022, 2022, 1–10. [Google Scholar] [CrossRef]
  9. Zhu, M.; He, D.; Li, X.; Li, C.; Li, F.; Liu, X.; Ding, E.; Zhang, Z. Image Inpainting by End-to-End Cascaded Refinement with Mask Awareness. IEEE Trans. Image Process. 2021, 30, 4855–4866. [Google Scholar] [CrossRef] [PubMed]
  10. Guo, X.; Yang, H.; Huang, D. Image inpainting via conditional texture and structure dual generation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 14134–14143. [Google Scholar] [CrossRef]
  11. Zeng, Y.; Fu, J.; Chao, H.; Guo, B. Aggregated Contextual Transformations for High-Resolution Image Inpainting. IEEE Trans. Vis. Comput. Graph. 2022, 29, 3266–3280. [Google Scholar] [CrossRef]
  12. Li, J.; Wang, N.; Zhang, L.; Du, B.; Tao, D. Recurrent feature reasoning for image inpainting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 7760–7768. [Google Scholar] [CrossRef]
  13. Uddin, S.M.N.; Jung, Y.J. Global and Local Attention-Based Free-Form Image Inpainting. Sensors 2020, 20, 3204. [Google Scholar] [CrossRef] [PubMed]
  14. Zheng, C.; Cham, T.J.; Cai, J. Pluralistic image completion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1438–1447. [Google Scholar] [CrossRef]
  15. Zhang, H.; Yang, J.; Zhou, K.; Li, F.; Hu, Y.; Gao, S.; Zhang, X.; Liu, J. Eliminating Shadow Artifacts via Generative Inpainting Networks to Quantify Vascular Changes of the Choroid. arXiv 2019, arXiv:1907.01271. [Google Scholar]
  16. Cheong, H.; Devalla, S.K.; Pham, T.H.; Zhang, L.; Tun, T.A.; Wang, X.; Perera, S.; Schmetterer, L.; Aung, T.; Boote, C.; et al. DeshadowGAN: A Deep Learning Approach to Remove Shadows from Optical Coherence Tomography Images. Transl. Vis. Sci. Technol. 2020, 9, 23. [Google Scholar] [CrossRef]
  17. Liu, H.; Cao, S.; Ling, Y.; Gan, Y. Inpainting for Saturation Artifacts in Optical Coherence Tomography Using Dictionary-Based Sparse Representation. IEEE Photon J. 2021, 13, 1–10. [Google Scholar] [CrossRef] [PubMed]
  18. Tang, Y.; Li, Y.; Liu, H.; Li, J.; Jin, P.; Gan, Y.; Ling, Y.; Su, Y. Multi-scale sparse representation-based shadow inpainting for retinal OCT images. In Proceedings of the Medical Imaging 2022, Image Processing, SPIE, San Diego, CA, USA, 23–28 February 2002; Volume 12032, pp. 9–17. [Google Scholar] [CrossRef]
  19. Isola, P.; Zhu, J.-Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar] [CrossRef]
  20. Bai, G.; Li, S.; Zhang, H.; Higashita, R.; Liu, J.; Zhang, M. A Structure-Consistency GAN for Unpaired AS-OCT Image Inpainting. In International Workshop on Ophthalmic Medical Image Analysis; Springer: Berlin/Heidelberg, Germany, 2023; pp. 142–151. [Google Scholar] [CrossRef]
  21. Yi, Z.; Zhang, H.; Tan, P.; Gong, M. Dualgan: Unsupervised dual learning for image-to-image translation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2849–2857. [Google Scholar] [CrossRef]
  22. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention, Proceedings of International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar] [CrossRef]
  23. Li, C.; Wand, M. Precomputed real-time texture synthesis with markovian generative adversarial networks. In Computer Vision, Proceedings of the ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 702–716. [Google Scholar] [CrossRef]
  24. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  25. Zhao, H.; Gallo, O.; Frosio, I.; Kautz, J. Loss Functions for Image Restoration With Neural Networks. IEEE Trans. Comput. Imaging 2017, 3, 47–57. [Google Scholar] [CrossRef]
  26. He, H.; Yang, D.; Wang, S.; Wang, S.; Li, Y. Road Extraction by Using Atrous Spatial Pyramid Pooling Integrated Encoder-Decoder Network and Structural Similarity Loss. Remote. Sens. 2019, 11, 1015. [Google Scholar] [CrossRef]
  27. Yang, H.H.; Yang, C.H.H.; Tsai, Y.C.J. Y-net: Multi-scale feature aggregation network with wavelet structure similarity loss function for single image dehazing. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 2628–2632. [Google Scholar] [CrossRef]
  28. Jiang, L.; Dai, B.; Wu, W.; Loy, C.C. Focal frequency loss for image reconstruction and synthesis. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 13919–13929. [Google Scholar] [CrossRef]
  29. Fuoli, D.; Van Gool, L.; Timofte, R. Fourier space losses for efficient perceptual image super-resolution. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 2360–2369. [Google Scholar] [CrossRef]
  30. Larsen, A.B.L.; Sønderby, S.K.; Larochelle, H.; Winther, O. Autoencoding beyond pixels using a learned similarity metric. In Proceedings of the 33rd International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016; Available online: http://proceedings.mlr.press/v48/larsen16 (accessed on 22 November 2023).
  31. Han, S.W.; Suh, D.Y. PIINET: A 360-degree panoramic image inpainting network using a cube map. Comput. Mater. Contin. 2021, 66, 213–228. [Google Scholar] [CrossRef]
  32. Feldman, R.M.; Kim, G.; Chuang, A.Z.; Shiraishi, A.; Okamoto, K.; Tsukamoto, M. Comparison between the CASIA SS-1000 and Pentacam in measuring corneal curvatures and corneal thickness maps. BMC Ophthalmol. 2023, 23, 10. [Google Scholar] [CrossRef] [PubMed]
  33. Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 586–595. [Google Scholar] [CrossRef]
  34. Garcia-Marin, Y.F.; Alonso-Caneiro, D.; Fisher, D.; Vincent, S.J.; Collins, M.J. Patch-based CNN for corneal segmentation of AS-OCT images: Effect of the number of classes and image quality upon performance. Comput. Biol. Med. 2023, 152, 106342. [Google Scholar] [CrossRef] [PubMed]
Figure 1. AS-OCT can visualize the entire cornea.
Figure 1. AS-OCT can visualize the entire cornea.
Sensors 23 09439 g001
Figure 2. The overall architecture of the proposed model includes two generators ( G A , G B ) and two discriminators ( D A , D B ). Image s S is input to generator G A and image w W is input to generator G B . The proposed structural similarity loss and frequency loss are represented as L S S I M and L F , respectively. L F combines spatial and frequency domains.
Figure 2. The overall architecture of the proposed model includes two generators ( G A , G B ) and two discriminators ( D A , D B ). Image s S is input to generator G A and image w W is input to generator G B . The proposed structural similarity loss and frequency loss are represented as L S S I M and L F , respectively. L F combines spatial and frequency domains.
Sensors 23 09439 g002
Figure 3. The mutual conversion between S-domain and W-domain images through two generators. Two discriminators are used to discern between generated images and real data.
Figure 3. The mutual conversion between S-domain and W-domain images through two generators. Two discriminators are used to discern between generated images and real data.
Sensors 23 09439 g003
Figure 4. The proposed method structure of generators and discriminators. The generators adopt a U-Net architecture, and the discriminators employ PatchGAN with five convolutional layers.
Figure 4. The proposed method structure of generators and discriminators. The generators adopt a U-Net architecture, and the discriminators employ PatchGAN with five convolutional layers.
Sensors 23 09439 g004
Figure 5. Feature distribution of different images in space and frequency domains: (a) artifact-free image; (b) synthetic saturation artifact image; (c) real saturation artifact image; (d) after inpainting data.
Figure 5. Feature distribution of different images in space and frequency domains: (a) artifact-free image; (b) synthetic saturation artifact image; (c) real saturation artifact image; (d) after inpainting data.
Sensors 23 09439 g005
Figure 6. The process of creating the synthesized saturated artifact image, s = I × m .
Figure 6. The process of creating the synthesized saturated artifact image, s = I × m .
Sensors 23 09439 g006
Figure 7. Experimental results of synthetic saturation artifact repair using corneal tissues with different slopes combined with the same mask: (a) synthetic saturation artifact input; (b) MADF; (c) CTSDG; (d) AOT-GAN; (e) RFR; (f) G&L; (g) PICNet; (h) DualGAN; (i) ours; and (j) ground_truth.
Figure 7. Experimental results of synthetic saturation artifact repair using corneal tissues with different slopes combined with the same mask: (a) synthetic saturation artifact input; (b) MADF; (c) CTSDG; (d) AOT-GAN; (e) RFR; (f) G&L; (g) PICNet; (h) DualGAN; (i) ours; and (j) ground_truth.
Sensors 23 09439 g007
Figure 8. Three groups of experimental results of synthetic saturation artifact inpainting, using corneal tissue with the same inclination degree combined with different masks: (a) synthetic saturation artifact input; (b) MADF; (c) CTSDG; (d) AOT-GAN; (e) RFR; (f) G&L; (g) PICNet; (h) DualGAN; (i) ours; and (j) ground_truth.
Figure 8. Three groups of experimental results of synthetic saturation artifact inpainting, using corneal tissue with the same inclination degree combined with different masks: (a) synthetic saturation artifact input; (b) MADF; (c) CTSDG; (d) AOT-GAN; (e) RFR; (f) G&L; (g) PICNet; (h) DualGAN; (i) ours; and (j) ground_truth.
Sensors 23 09439 g008
Figure 9. Experimental results of synthetic saturation artifact restoration, using corneal tissue with different tilt angles combined with different masks: (a) synthetic saturation artifact input; (b) MADF; (c) CTSDG; (d) AOT-GAN; (e) RFR; (f) G&L; (g) PICNet; (h) DualGAN; (i) ours; and (j) ground_truth.
Figure 9. Experimental results of synthetic saturation artifact restoration, using corneal tissue with different tilt angles combined with different masks: (a) synthetic saturation artifact input; (b) MADF; (c) CTSDG; (d) AOT-GAN; (e) RFR; (f) G&L; (g) PICNet; (h) DualGAN; (i) ours; and (j) ground_truth.
Sensors 23 09439 g009
Figure 10. Experimental results for real saturation artifacts: (a) saturation artifact input; (b) MADF; (c) CTSDG; (d) AOT-GAN; (e) RFR; (f) G&L; (g) PICNet; (h) DualGAN; and (i) ours.
Figure 10. Experimental results for real saturation artifacts: (a) saturation artifact input; (b) MADF; (c) CTSDG; (d) AOT-GAN; (e) RFR; (f) G&L; (g) PICNet; (h) DualGAN; and (i) ours.
Sensors 23 09439 g010
Figure 11. The inpainting and segmentation results: (a) saturation artifact image before inpainting; (b) MADF; (c) CTSDG; (d) AOT-GAN; (e) RFR; (f) G&L; (g) PICNet; (h) DualGAN; and (i) ours. EP, BL, and EN are represented by red, green, and yellow lines, respectively.
Figure 11. The inpainting and segmentation results: (a) saturation artifact image before inpainting; (b) MADF; (c) CTSDG; (d) AOT-GAN; (e) RFR; (f) G&L; (g) PICNet; (h) DualGAN; and (i) ours. EP, BL, and EN are represented by red, green, and yellow lines, respectively.
Sensors 23 09439 g011
Figure 12. The ablation study results: (a) saturation artifact image; (b) w/o L F ; (c) w/o L S S I M ; (d) ours.
Figure 12. The ablation study results: (a) saturation artifact image; (b) w/o L F ; (c) w/o L S S I M ; (d) ours.
Sensors 23 09439 g012
Table 1. Quantitative comparison of synthetic artifact inpainting results using different methods. ↑ Higher is better. ↓ Lower is better. The best results are presented in bold.
Table 1. Quantitative comparison of synthetic artifact inpainting results using different methods. ↑ Higher is better. ↓ Lower is better. The best results are presented in bold.
MethodPSNR↑SSIM↑LPIPS↓
Before Inpainting4.8990.3140.687
MADF20.5330.3920.091
CTSDG22.8860.5540.080
AOT-GAN22.1680.5180.125
RFR16.1680.4310.213
G&L22.8790.5610.225
PICNet19.9770.4610.125
DualGAN17.91540.2500.106
Ours21.3330.4960.072
Table 2. Quantitative comparison of corneal segmentation results in AS-OCT images using different methods. ↑ Higher is better. The best results are presented in bold.
Table 2. Quantitative comparison of corneal segmentation results in AS-OCT images using different methods. ↑ Higher is better. The best results are presented in bold.
MethodDSC↑PA↑F1-Score↑Jaccard↑
Before Inpainting0.4310.9860.9840.299
MADF0.4940.9870.9850.357
CTSDG0.5360.9880.9860.390
AOT-GAN0.5200.9870.9850.381
RFR0.4090.9860.9850.280
G&L0.5310.9880.9860.387
PICNet0.4820.9870.9850.340
DualGAN0.4700.9870.9850.335
Ours0.5850.9890.9870.444
Table 3. Quantitative comparison of corneal segmentation results in ablation study. ↑ Higher is better. The best results are presented in bold.
Table 3. Quantitative comparison of corneal segmentation results in ablation study. ↑ Higher is better. The best results are presented in bold.
MethodDSC↑PA↑F1-Score↑Jaccard↑
w/o L F 0.5700.9890.9870.425
w/o L S S I M 0.4790.9870.9850.338
Ours0.5850.9890.9870.444
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, J.; Zhang, H.; Wang, X.; Wang, H.; Hao, J.; Bai, G. Inpainting Saturation Artifact in Anterior Segment Optical Coherence Tomography. Sensors 2023, 23, 9439. https://doi.org/10.3390/s23239439

AMA Style

Li J, Zhang H, Wang X, Wang H, Hao J, Bai G. Inpainting Saturation Artifact in Anterior Segment Optical Coherence Tomography. Sensors. 2023; 23(23):9439. https://doi.org/10.3390/s23239439

Chicago/Turabian Style

Li, Jie, He Zhang, Xiaoli Wang, Haoming Wang, Jingzi Hao, and Guanhua Bai. 2023. "Inpainting Saturation Artifact in Anterior Segment Optical Coherence Tomography" Sensors 23, no. 23: 9439. https://doi.org/10.3390/s23239439

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop