Next Article in Journal
Characterization of COVID-19-Related Lung Involvement in Patients Undergoing Magnetic Resonance T1 and T2 Mapping Imaging: A Pilot Study
Previous Article in Journal
Seismic Waveform Inversion Capability on Resource-Constrained Edge Devices
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient and Robust Method to Detect the Location of Macular Center Based on Optimal Temporal Determination

by
Helmie Arif Wibawa
1,2,
Agus Harjoko
2,*,
Raden Sumiharto
2 and
Muhammad Bayu Sasongko
3
1
Department of Informatics, Faculty of Science and Mathematics, Diponegoro University, Semarang 50275, Indonesia
2
Department of Computer Science and Electronics, Faculty of Mathematics and Natural Sciences, Gadjah Mada University, Yogyakarta 55281, Indonesia
3
Department of Ophthalmology, Faculty of Public Health Medicine and Nursing, Gadjah Mada University, Yogyakarta 55281, Indonesia
*
Author to whom correspondence should be addressed.
J. Imaging 2022, 8(12), 313; https://doi.org/10.3390/jimaging8120313
Submission received: 20 September 2022 / Revised: 4 November 2022 / Accepted: 11 November 2022 / Published: 22 November 2022

Abstract

:
The location of the macular central is very important for the examination of macular edema when using an automated screening system. The erratic character of the macular light intensity and the absence of a clear border make this anatomical structure difficult to detect. This paper presents a new method for detecting the macular center based on its geometrical location in the temporal direction of the optic disc. Also, a new method of determining the temporal direction using the vascular features visible on the optic disc is proposed. After detecting the optic disc, the temporal direction is determined by considering blood vessel positions. The macular center is detected using thresholding and simple morphology operations with optimum macular region of interest (ROI) direction. The results show that the proposed method has a low computation time of 0.34 s/image with 100% accuracy for the DRIVE dataset, while that of DiaretDB1 was 0.57 s/image with 98.87% accuracy.

1. Introduction

Diabetic macular edema (DME) often threatens the vision of people with diabetes when not treated properly, and fundus photos have been recommended for routine eye screening [1]. Analysis of fundus images with computer technology can make such examinations more effective and efficient.
DME severity is measured with polar coordinates that are determined by the appearance of a hard exudate on the retinal fundus image and its distance to the macular area [2,3,4]. The center of the macular is used as the focus in this polar coordinate system [5,6], which helps in measuring the distance of the hard exudate when it appears. Therefore, macular center detection is considered the initial step for their determination [7]. Figure 1 shows a polar coordinate system for the appearance and distribution of hard exudate in the macular area.
In the retinal fundus image, the macula is identified as the small circular portion with the darkest intensity [5] without the clear border of the optic disc area. Furthermore, as this area does not contain blood vessels [8], it is more difficult to detect the macular than anatomical structures such as the optic disc and blood vessels.
Therefore, the location of the macular center must be detected as the first step towards establishing the polar coordinates. In the past, the detection of the macular center has been performed using deep learning and traditional approaches. In one study, a deep-learning approach using Convolutional Neural Network (CNN) to simultaneously detect the optic disc and fovea [9] In another study, multi-stage segmentation of the fovea was performed with fully convolutional neural networks and explained that the fovea was segmented and localized using an end-to-end encoder–decoder network [10]. It has been found that deep-learning usage involves very complex parameters that influence each other, even though it provides high accuracy results [9]. Furthermore, the inspection and repair process in this approach is more complicated compared to the traditional method because of its black-box nature. This is consistent with [11,12,13,14], which stated that the deep-learning method requires large amounts of data to achieve optimal results.
The traditional approach has been used to directly detect the macula based on the darkest intensity [15], while other studies have employed template matching to identify the fovea [16,17,18,19]. For example, a template based on the Gauss function was used in [16,19], while a histogram of the mean intensity was employed in [17]. A similar process was also utilized by [8] to perform matching for the extracted local features. Support Vector Machine (SVM) has also been used to determine macula candidates [3]. Furthermore, [20,21,22] employed geometric techniques in center detection and compared the location of the macular center to the optic disc or blood vessels. Zheng et al. [23] detected the macular center based on the location of the optic disc (OD) by using the temporal direction. However, the method failed to detect the macular center when the macular area was not clear.
It is important to note that the traditional methods mentioned above focus on determining and formulating specific characters for describing the fovea [24], such as the area with the darkest intensity or the fewest blood vessels. This formulation is difficult when conditions that do not fulfill the criteria are encountered, such as the appearance of hard exudate or a large black area in the fovea. The method proposed in this paper attempts to overcome these shortcomings and focuses on parameter formulation as well as the features for determining the macular center in the temporal area. Moreover, this method does not require large amounts of data to formulate an optimal model.
This paper also proposes an approach for obtaining the location of the macular center via determination of the optimum macular region of interest (ROI) based on the temporal direction. After the optic disc is detected, a feature for determining the temporal direction is extracted, which helps with accurate identification of the macular center location. The contributions of this study can be summarized as follows:
  • The macular center can be detected based only on its geometrical location in relation to the optic disc. This often leads to robust variations when detecting the intensity.
  • The method uses the inherent features in the optic disc to determine the temporal direction in which the macula is located, thereby making the process run faster.
  • Macular ROI with the right direction, location, and size reduces the detection area, facilitating a simpler detection process.
The study is presented in the following order: Section 2 describes the materials and methods used, including the sequence of processes for macular center detection. Section 3 and Section 4 respectively present the results and discussion, while the conclusions are contained in Section 5.

2. Materials and Methods

2.1. Materials

2.1.1. Dataset

A total of four datasets were used in this study: 3 public datasets, namely DRIVE, DiaretDB1, and Messidor, and 1 local dataset. DRIVE [25] consists of 40 color images with 8-bit depth. The images have a 768 × 584 size with a 45° field of view (FOV).
DiaretDB1 contains 89 retinal images taken from Kuopio University Hospital. The images have a size of 1500 × 1552 pixels and were taken with a 50° FOV [26]. The Messidor dataset contains 1200 retinal images; of these, 212, 400, 588 images with a size of 2304 × 1536, 2240 × 1488, 1440 × 960 pixels, respectively, were included [27].
The images from the local dataset were taken from the Jogjakarta Eye Diabetic Study in the Community (JOGED.com) [28]. This local dataset has two sizes, namely 2124 × 2056 and 3696 × 2448, with 73 and 26 images, respectively.

2.1.2. Environment

The experimental results were obtained using MATLAB R2018b on a computer with a 2.50 GHz Intel (R) Core (TM) i5– 4200 CPU, 4GB RAM, and Intel (R) HD Graphics 4600 graphics card.

2.2. Methods

In this study, the macular center was determined based on its geometrical location relative to the optic disc. The macula was considered as the area with the darkest intensity in the retinal image, and its center was located at 2.5 optic disc diameter (DD) temporally from the optic disc [23]. The search for the optic disc location and the temporal direction are prerequisites for determining the macula position. The method used for identifying the macular center consists of several main steps, which include optic disc localization, determining temporal area direction, identifying the macular region of interest (ROI), and macular center point coordinates extraction. Figure 2 shows the flow of macular center point detection.

2.2.1. Optic Disc Localization

The OD center served as a reference point when determining the location of the macular coordinates. Furthermore, the blood vessels visible on the OD were useful for determining the direction of the temporal area on the retinal image [23]. It is important to note that the OD is an anatomical structure in the retinal image that has a higher light intensity compared to others, and this is the reason for it being localized through its intensity character. In this study, localization was conducted by combining the thresholding and morphological operation methods [29]. This thresholding technique was employed to determine the intensity value limit that distinguishes the OD from other areas. Meanwhile, the morphological operation was conducted to improve the thresholding results in order to provide an optimal OD area.
Before the OD localization process, image preprocessing was conducted, consisting of image resizing and intensity normalization. This image resizing was performed to equalize the image height between datasets for more uniformity and to reduce the processing time. The height of the image was resized to 565 pixels, while the width of the image was adjusted to the proportion of the input image. The value of 565 was taken from the height of the image in the DRIVE dataset, which has the smallest size compared to other datasets. This resizing process is formulated in Equations (1) and (2).
h = 565
w = h 565 × w
where h is the normalized height, w denotes the normalized width, h represents the height, and w represent the width of the input image.
The next step was intensity normalization to overcome uneven lighting in the retinal image, which causes some areas other than the OD to appear brighter, thereby leading to localization errors. Intensity normalization is performed to minimize the effect of non-uniform lighting. Normalization was conducted by combining several morphological operations, as formulated in Equations (3)–(5). These operations were applied to the green layer of the image.
I p r e 1 ( I ) = I + I b t θ ( I ) + I b g ( I )
I p r e 2 ( I ) = I o p σ ( I ) + I b g ( I )
I c ( I ) = I p r e 1 ( I ) + I p r e 2 ( I ) 2
I p r e 1 and I p r e 2 represented precondition images used for uniform lighting while maintaining the brightness level of the OD area (Figure 3c,d). I represents the green layer of the resized input image. I b t θ ( I ) = ε θ ( δ θ ( I ) ) I was the morphological top-hat of I, performed with a disc structuring element, θ . Here, ε is morphological erosion and δ represents morphological dilation. Ibg denotes the background image generated through filtering operations using an average filter. According to [30], the filter utilized was 89 × 89, in which I p r e 2 was generated from the sum of I o p and the background image, while I o p σ = δ σ ( ε σ ( I ) ) was the morphological opening of I with structuring element of σ. The average of I p r e 1 and I p r e 2 produced an image called I c with more even lighting, and the optic disc area was maintained as shown in Figure 3e.
In order to obtain the OD ROI, Ic was first binarized through the thresholding operation on the I c image as conducted in [29], while the threshold value used was 0.89 of the maximum I c intensity. The threshold value was obtained through a series of empirically conducted experiments against a range of possible values. The resulting image is shown in Figure 3f. Then, the coordinates of this center point candidate were employed as the retinal image cropping center, while the crop size was ( w × l ) 4 . The retinal image cropping was performed on the red layer of the input image (Figure 3g). In the red layer, the OD was still clearly visible, and the presence of blood vessels did not have much effect [23].
The OD ROI was further processed to obtain the optic disc. The process began with contrast enhancement using Contrast-Limited Adaptive Equalization (CLAHE) and continued with opening morphology to remove blood vessels. The next step was binarization of the resultant image using the Otsu’s threshold. In order to obtain the candidate blob of the OD, morphological closing morphology followed by morphological opening were performed. Disc-structuring elements of radius 10 and 15 were used in these morphological operations (Figure 3h). Then, the resulting image was cropped according to the bounding box of the blob to obtain the OD (Figure 3i). The cropped optic disc image was called I O D and was used in the process of determining the temporal direction. Furthermore, the center point of the blob was determined to be the center point of the OD, which was then plotted on the retinal image to show the results of OD localization (Figure 3j). Figure 3 shows the series of results for the OD localization.

2.2.2. Temporal Area Determination

The macula is a small area on the circular retina that has low intensity but does not contain blood vessels. According to [31], the center of the macula is located at 2.5 DD from the optic disc center. In [32], the retinal was vertically divided into temporal and nasal areas but horizontally divided into inferior and superior. One study found that the macular area is temporally located on the retina from the optic disc center [23]. Therefore, the temporal direction information was used in the proposed method to determine the macular ROI.
The main blood vessels of the retina converged at the optic disc. It was also observed that the blood vessels’ appearance has a unique character. For example, the blood vessels on the optic disc tended to gather at one side within the optic disc, indicating temporal and nasal directions on retinal images. This means that a relationship was formed between the nasal and temporal directions on the retinal image with the appearance of blood vessels. Moreover, when the optic disc was vertically divided by 2, the temporal direction was indicated by the area containing fewer blood vessels. The relationship between the temporal directions and the appearance of blood vessels on the retinal image viewed from the optic disc center is shown in Figure 4.
In this proposed method, the number of blood vessel pixels on the optic disc was utilized as a feature for determining the retinal image’s temporal direction. Furthermore, the optic disc area containing fewer blood vessels showed the temporal direction, and the pixel number was calculated on the binary image generated from the green layer of the OD ROI. This green layer was selected because it shows the blood vessels more clearly [23]. After the contrast enhancement process using CLAHE, a bottom-hat operation was conducted on the CLAHE image to further emphasize the blood vessels. A disc-structuring element of suitable pixel radius was used. This type of structuring element was appropriate for maintaining the shape of blood vessels. The size of the structuring element was selected to be slightly larger than the width of the blood vessel; 5 were selected.
It is important to note that blood vessel pixels were computed from black and white pixels generated through an adaptive thresholding process using Otsu. Unlike the technique employed by Zheng [23], this proposed method only involved vertically oriented blood vessels because there is a possibility that an optic disc is filled with blood vessels, thereby causing an error in detecting the temporal direction. On closer inspection, the blood vessels that tend to be vertical occupied the area opposite the temporal direction. Therefore, this method aims to eliminate horizontally oriented blood vessels before calculating the number of pixels in the optic disc, indicating that only the pixels showing the vertical blood vessels remain. The horizontal blood vessels were eliminated through a combination of opening and closing morphological operations. Therefore, a rectangular structuring element with a vertical orientation was used to maintain vertical vessels. Rectangular-shaped structuring elements of size 15 × 1 and 50 × 15, respectively, were used in this process. Figure 5 shows an example of a blood vessel extraction on the OD.
After the vertically oriented blood vessel images were obtained, the next step was determining the temporal area direction in the retinal image. The eliminated image was divided vertically into left and right segments, and then the number of white pixels on each segment was calculated. The temporal area was located in the segment with fewer blood vessels, indicating that when the left segment has more white pixels, the temporal direction was to the right, but when the right segment has more white pixels, the temporal direction was to the left.

2.2.3. Macular ROI Determination

Macular ROI was determined to reduce the search area of the macular center and search time. In this method, the macular ROI was obtained through the following limits:
  • The determination of macular ROI was based on the temporal direction. Furthermore, the macula located in the temporal area was obtained geometrically with reference to the OD center point [23].
  • The macular center was 2.5 times of OD diameter from the OD center [5,20,33] and located slightly below the OD [34].
Using these limits, the central location of the macular ROI was determined in this study via Algorithm 1. Figure 6 shows an illustration of macular ROI determination. The inputs for Algorithm 1 were the temporal direction, center point of OD, and diameter of OD. The OD center had been obtained when performing OD localization, while the diameter of the OD was obtained by measuring the ratio between the width of the OD and the width of the retinal image. The experimental results showed that the appropriate OD diameter was v 12 where v′ was the width of the retinal in the image segmented by Otsu thresholding on the grayscale image. The use of these three parameters, as shown in Algorithm 1, could provide an optimal macular ROI.
Algorithm 1: Macular ROI determination
Input:  d i r e c t i o n t e m p o r a l , OD center coordinates ( x O D , y O D ) , OD diameter (DD)
Parameter: abscissa factor (p), ordinate factor (q), ROI box factor (r)
1: if   a r a h t e m p o r a l = L E F T  then
2: x M x O D p × D D
3: else
4: x M x O D + p × D D
5 :   y M   y O D + q × D D
6: d i m R O I   r × D D
7: determine the macular ROI with the center ( x M ,   y M )   a n d   t h e   s i z e   o f   d i m R O I
As shown in Figure 6, the ROI of the macula was determined in the form of a square shape with a size of r × DD, where r was the ROI box factor. In this study, the best values of the parameters p, q, and r were selected through parameter tuning, with possible values of p = {3.6, 3.8, 4.0}, q = {0,25, 0,5}, and r = {2.0, 2.25, 2.5}.

2.2.4. Macular Center Extraction

After determining the macular ROI, its center was obtained through thresholding and morphological operations. The first step was to increase the contrast of the macular ROI image using CLAHE with a clip limit value of 1, while the value of parameter number of tile was [8 8]. In order to obtain the candidate blob of macula, binarization was performed on resultant. The threshold value was selected based on the maximum intensity, as formulated in Equation (6).
τ = 0.98 × max ( I m )
In order to minimize detection errors when the macula was in a dark area, a dilation morphological operation was performed. This operation was followed by an opening process, which enlarged the candidate of macular area. The largest candidate blob was selected as the macular area, and then the center of the blob was designated as the center of the macula. The extraction process results of the macular center point is shown in Figure 7.

3. Results

This method has been tested on the DRIVE, DiaretDB1, and Messidor public datasets, as well as the local dataset from JOGED.com. All images in the DiaretDB1, Messidor, and JOGED.com datasets were utilized in this test, but five images were not used for macular center detection in the DRIVE dataset. This is in line with [23], which stated that five images in the macular center area were invisible. The five images include image#4, image#15, image#23, image#31, and image#34. The macula center detection validation was conducted based on the Euclidean distance measure of the detected centers of macula from the ground truth (GT) centers. According to Medhi [35], if the automatically located center of macula was located less than of 50 pixels from ground truth, it was considered as a correct detection. The Euclidian distance formula is shown by Equation (7) [5,36]:
d i s t a n c e = ( x s y s t e m x G T ) 2 + ( y s y s t e m y G T ) 2
These results were plotted on a Cartesian diagram, as shown in Figure 8. The ground truth value validated by experts was employed as the center point of the coordinates, while the points on the diagram showed the distribution of the points generated by the proposed method. In addition, a red circle with a radius of 50 indicated the acceptable distance limit for well-detected results. Therefore, the point distribution in this diagram showed the distribution of the accuracy of the proposed method in detecting the center of the macula in a dataset. There were several points that lie outside the red circle, as shown in Figure 8b–d. These points showed the coordinates of the macular center points that were not precisely detected. Examples of detection results are presented in Figure 9.
Based on experiments on the four datasets, this method works well and is stable. Out of the 35 images in the DRIVE dataset, the system was able to detect the macular center with 100% accuracy. Meanwhile, on the DiaretDB1 dataset, the proposed method achieved an accuracy of 98.87%, and on the Messidor dataset, 94.67%. Furthermore, the accuracy of this proposed method on the local JOGED.com dataset was 93%. The optimum value for the four datasets was successfully obtained by setting the abscissa distance to 3.8DD and the ordinate to 0.25DD. Consequently, the optimal macular ROI box size was obtained at 2DD × 2DD. The distance between the points recorded by the system and the ground truth produced the shortest average distance in the DRIVE with a value of 7.1, while those of DiaretDb1, Messidor, and JOGED.com were 15.8, 8.7, and 13.5, respectively.
The computation time test results showed that the method requires very little time. Specifically, it involved an average time of 0.34 s/image to detect the macular center in the DRIVE, while the DiaretDB, the Messidor, and the local JOGED.com were 0.57, 0.64, and 0.78 s/image, respectively. Table 1 shows that the time achieved by this proposed method exceeded others. This comparison was performed in studies that utilized the DRIVE, DiaretDB1, and Messidor datasets.

4. Discussion

The results showed that macular center detection by temporal area selection provided stable outcomes. It was observed that the selection of the right temporal direction provided a simple process for determining macular ROI. Furthermore, determining the optimal geometric macular ROI was able to provide stable detection results, which were not affected by image conditions that sometimes have uneven lighting.
In addition to the high-accuracy results, Table 1 also shows that the computational time of macular center detection in the proposed method surpassed others. For example, it significantly outperformed other methods on the DiaretDB1 dataset with large image size. Although the accuracy obtained for the DiaretDB1 and Messidor datasets was slightly below that of the other methods as reported in [24], the proposed method was much faster. Resizing and determining the macular ROI with an appropriate temporal reference are the keys to obtaining these results. This makes computing lighter and provides faster execution time, even with large images, namely the DiaretDB1 dataset.
The average computation time obtained by this method was slightly slower than the average time reported by Sedai, which had an average time of 0.4 s/image. However, the method in [23], which was based on deep learning, used high computational facilities supported by a 12 GB GPU. This can reduce computational time significantly, as stated by Chalakkal in [17].
The proposed method failed to detect the macular center of image059.png in the DiaretDb1 dataset and several images in the JOGED.com dataset. The detection failure of the macular center was caused by the inability to recognize the optic disc location due to irregular lighting in the image. It was observed that the non-optic disc area has a higher intensity compared to the optic disc. This condition causes OD candidate selection failures as well as OD localization and further leads to inaccuracy when determining the macular center. The geometric properties used in this method, which make the optic disc center a reference point, require precision in the localization of the optic disc. Therefore, failure in optic disc center localization resulted in failure when detecting the macular center. An example of a macular center detection failure that occured on image059.png in the DiaretDB1 dataset is shown in Figure 10.

5. Conclusions

This study described a new method for detecting the macular center location based on the temporal direction in retinal images. The proposed method was based on the geometric relationship between the macular area and the optic disc. The temporal direction determination provided optimal macular ROI, thereby leading to a detection process with low computation time and high accuracy. When compared to other methods, the results showed that the proposed method was faster while maintaining high accuracy.

Author Contributions

Conceptualization, H.A.W. and A.H.; methodology, H.A.W.; software, H.A.W.; validation, H.A.W. and M.B.S.; formal analysis, H.A.W. and A.H.; investigation, H.A.W. and M.B.S.; data curation, H.A.W. and M.B.S.; writing—original draft preparation, H.A.W.; writing—review and editing, A.H., R.S. and M.B.S.; visualization, H.A.W.; supervision, A.H., R.S. and M.B.S.; project administration, A.H.; funding acquisition, A.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Directorate of Higher Education, Research, and Technology, the Ministry of Education, Culture, Research, and Technology Indonesia under Grant No. 1898/UN1/DITLIT/Dit-Lit/PT.01.03/2022.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

This work utilized publically available datasets and local dataset obtained directly from the database manager so informed consent statement is not applicable.

Data Availability Statement

Images from the public datasets, DRIVE, DiaretDB1 and Messidor were used in this article. The DRIVE dataset and documentation can be found at https://drive.grand-challenge.org/ (accessed on 15 April 2021), while DiaretDb1 can be found at https://www.it.lut.fi/project/imageret/diaretdb1/ (accessed on 21 June 2021). The Messidor dataset is kindly provided by the Messidor program partners at https://www.adcis.net/en/third-party/messidor/ (accessed on 29 October 2021). The ground truth of the macula center of both datasets is available at https://github.com/helmieaw/Ground-Truth-of-Macular-center (accessed on 22 September 2022). The JOGED.com local dataset was used in this article, but was not fully accessible.

Acknowledgments

The authors would like to thank the Jogja Eye Diabetic Study in The Community (JOGED).com providers for access to their databases.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ciulla, T.A.; Amador, A.G.; Zinman, B. Diabetic retinopathy and diabetic macular edema: Pathophysiology, screening, and novel therapies. Diabetes Care 2003, 26, 2653–2664. [Google Scholar] [CrossRef] [Green Version]
  2. Deepak, K.S.; Sivaswamy, J. Automatic assessment of macular edema from color retinal images. IEEE Trans. Med. Imaging 2012, 31, 6–76. [Google Scholar] [CrossRef] [Green Version]
  3. Syed, A.M.; Akram, M.U.; Akram, T.; Muzammal, M. Fundus images-based detection and grading of macular edema using robust macula localization. IEEE Access 2018, 6, 58784–58793. [Google Scholar] [CrossRef]
  4. Abràmoff, M.D.; Garvin, M.K.; Sonka, M. Retinal imaging and image analysis. IEEE Rev. Biomed. Eng. 2010, 3, 169–208. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Welfer, D.; Scharcanski, J.; Marinho, D.R. Fovea center detection based on the retina anatomy and mathematical morphology. Comput. Methods Programs Biomed. 2011, 104, 397–409. [Google Scholar] [CrossRef]
  6. Early Treatment Diabetic Retinopathy Study Research Group. Grading diabetic retinopathy from stereoscopic color fundus photographs—An Extension of the Modified Airlie House Classification: ETDRS report number 10. Ophthalmology 2020, 127, S99–S119. [Google Scholar] [CrossRef] [PubMed]
  7. Niemeijer, M.; Abràmoff, M.D.; van Ginneken, B. Fast detection of the optic disc and fovea in color fundus photographs. Med. Image Anal. 2009, 13, 859–870. [Google Scholar] [CrossRef] [Green Version]
  8. Guo, X.; Li, Q.; Sun, C.; Lu, Y. Automatic localization of macular area based on structure label transfer. Int. J. Ophthalmol. 2018, 11, 422–428. [Google Scholar]
  9. Al-Bander, B.; Al-Nuaimy, W.; Williams, B.M.; Zheng, Y. Multiscale sequential convolutional neural networks for simultaneous detection of fovea and optic disc. Biomed. Signal Process. Control 2018, 40, 91–101. [Google Scholar] [CrossRef]
  10. Sedai, S.; Tennakoon, R.; Roy, P.; Cao, K.; Garnavi, R. Multi-stage segmentation of the fovea in retinal fundus images using fully Convolutional Neural Networks. In Proceedings of the 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017), Melbourne, VIC, Australia, 18–21 April 2017. [Google Scholar]
  11. Mamoshina, P.; Vieira, A.; Putin, E.; Zhavoronkov, A. Applications of deep learning in biomedicine. Mol. Pharm. 2016, 13, 1445–1454. [Google Scholar] [CrossRef]
  12. Camara, J.; Neto, A.; Pires, I.M.; Villasana, M.V.; Zdravevski, E.; Cunha, A. Literature review on artificial intelligence methods for glaucoma screening, segmentation, and classification. J. Imaging 2022, 8, 19. [Google Scholar] [CrossRef] [PubMed]
  13. Royer, C.; Sublime, J.; Rossant, F.; Paques, M. Unsupervised approaches for the segmentation of dry armd lesions in eye fundus cslo images. J. Imaging 2021, 7, 143. [Google Scholar] [CrossRef]
  14. Lakshminarayanan, V.; Kheradfallah, H.; Sarkar, A.; Balaji, J.J. Automated detection and diagnosis of diabetic retinopathy: A comprehensive survey. J. Imaging 2021, 7, 165. [Google Scholar] [CrossRef]
  15. Medhi, J.P.; Nath, M.K.; Dandapat, S. Automatic Grading of Macular Degeneration from Color Fundus Images. In Proceedings of the 2012 World Congress on Information and Communication Technologies, Trivandrum, India, 30 October–2 November 2012. [Google Scholar]
  16. Sinthanayothin, C.; Boyce, J.F.; Cook, H.L.; Williamson, T.H. Automated localisation of the optic disc, fovea and retinal blood vessels from digital color fundus images. Br. J. Ophthalmol. 1999, 4, 902–910. [Google Scholar] [CrossRef]
  17. Chalakkal, R.J.; Abdulla, W.H.; Thulaseedharan, S.S. Automatic detection and segmentation of optic disc and fovea in retinal images. IET Image Process. 2018, 12, 2100–2110. [Google Scholar] [CrossRef]
  18. Fleming, A.D.; Goatman, K.A.; Philip, S.; Olson, J.A.; Sharp, P.F. Automatic detection of retinal anatomy to assist diabetic retinopathy screening. Phys. Med. Biol. 2007, 52, 331–345. [Google Scholar] [CrossRef]
  19. Kao, E.; Lin, P.; Chou, M.; Jaw, T.S.; Liu, G.C. Automated detection of fovea in fundus images based on vessel-free zone and adaptive Gaussian. Comput. Methods Programs Biomed. 2014, 117, 92–103. [Google Scholar] [CrossRef]
  20. Aquino, A. Establishing the macular grading grid by means of fovea centre detection using anatomical-based and visual-based features. Comput. Biol. Med. 2014, 55, 61–73. [Google Scholar] [CrossRef]
  21. Qureshi, R.J.; Kovacs, L.; Harangi, B.; Nagy, B.; Peto, T.; Hajdu, A. Combining algorithms for automatic detection of optic disc and macula in fundus images. Comput. Vis. Image Underst. 2012, 116, 138–145. [Google Scholar] [CrossRef]
  22. Nugroho, H.A.; Listyalina, L.; Wibirama, S.; Oktoeberza, W.K. Automated determination of macula centre point based on geometrical and pixel value approaches to support detection of foveal avascular zone. Int. J. Innov. Comput. Inf. Control 2018, 14, 1453–1463. [Google Scholar]
  23. Zheng, S.; Pan, L.; Chen, J.; Yu, L. Automatic and Efficient Detection of The Fovea Center in Retinal Images. In Proceedings of the 2014 7th International Conference on BioMedical Engineering and Informatics, Dalian, China, 14–16 October 2014. [Google Scholar]
  24. Romero-Oraá, R.; García, M.; Oraá-Pérez, J.; López, M.I.; Hornero, R. A robust method for the automatic location of the optic disc and the fovea in fundus images. Comput. Methods Programs Biomed. 2020, 196, 1–12. [Google Scholar] [CrossRef] [PubMed]
  25. Staal, J.; Abràmoff, M.D.; Niemeijer, M.; Viergever, M.A.; Ginneken, B.V. Ridge-based vessel segmentation in color images of the retina. IEEE Trans. Med. Imaging 2004, 23, 501–509. [Google Scholar] [CrossRef] [PubMed]
  26. Kauppi, T.; Kalesnykiene, V.; Kamarainen, J.K.; Lensu, L.; Sorri, I.; Raninen, A.; Voutilainen, R.; Pietilä, J.; Kälviäinen, H.; Uusitalo, H. The DIARETDB1 Diabetic Retinopathy Database and Evaluation Protocol. In Proceedings of the British Machine Vision Conference 2007, University of Warwick, Warwick, UK, 10–13 September 2007. [Google Scholar]
  27. Decenciere, E.; Zhang, X.; Cazuguel, G.; La¨y, B.; Cochener, B.; Trone, C.; Charton, B. Feedback on a publicly distributed image database: The Messidor database. Image Anal. Stereol. 2014, 33, 231–234. [Google Scholar] [CrossRef] [Green Version]
  28. Sasongko, M.B.; Agni, A.N.; Wardhana, F.S.; Kotha, S.P.; Gupta, P.; Widayanti, T.W.; Supanji; Widyaputri, F.; Widyaningrum, R.; Wong, T.Y.; et al. Rationale and methodology for a community-based study of diabetic retinopathy in an indonesian population with type 2 diabetes mellitus: The Jogjakarta eye diabetic study in the community. Ophthalmic Epidemiol. 2017, 24, 48–56. [Google Scholar] [CrossRef]
  29. Septiarini, A.; Harjoko, A.; Pulungan, R.; Ekantini, R. Optic disc and cup segmentation by automatic thresholding with morphological operation for glaucoma evaluation. Signal. Image Video Process. 2017, 11, 945–952. [Google Scholar] [CrossRef]
  30. Abdullah, M.; Fraz, M.M.; Barman, S.A. Localization and segmentation of optic disc in retinal images using circular Hough transform and grow-cut algorithm. PeerJ 2016, 4, 1–22. [Google Scholar] [CrossRef]
  31. Mookiah, M.R.K.; Acharya, U.R.; Chua, C.K.; Lim, C.M.; Ng, E.Y.K.; Laude, A. Computer-Aided Diagnosis of Diabetic Retinopathy: A Review. Comput. Biol. Med. 2013, 43, 2136–2155. [Google Scholar] [CrossRef]
  32. Calvo-Maroto, A.M.; Esteve-Taboada, J.J.; Pérez-Cambrodí, R.J.; Madrid-Costa, D.; Cerviño, A. Pilot study on visual function and fundus autofluorescence assessment in diabetic patients. J. Ophthalmol. 2016, 2016, 1–10. [Google Scholar] [CrossRef] [Green Version]
  33. Siddalingaswamy, P.C.; Prabhu, K.G. Automatic Grading of Diabetic Maculopathy Severity Levels. In Proceedings of the International Conference on Systems in Medicine and Biology, ICSMB 2010, Kharagpur, India, 16–18 December 2010. [Google Scholar]
  34. Chin, K.S.; Trucco, E.; Tan, L.; Wilson, P.J. Automatic fovea location in retinal images using anatomical priors and vessel density. Pattern Recognit. Lett. 2013, 34, 1152–1158. [Google Scholar] [CrossRef]
  35. Medhi, J.P.; Dandapat, S. An effective fovea detection and automatic assessment of diabetic maculopathy in color fundus images. Comput. Biol. Med. 2016, 74, 30–44. [Google Scholar] [CrossRef] [PubMed]
  36. Gegundez-Arias, M.E.; Marin, D.; Bravo, J.M.; Suero, A. Locating the fovea center position in digital fundus images using thresholding and feature extraction techniques. Comput. Med. Imaging Graph. 2013, 37, 386–393. [Google Scholar] [CrossRef] [PubMed]
Figure 1. An example of polar coordinates in retinal image (DD: disc diameter).
Figure 1. An example of polar coordinates in retinal image (DD: disc diameter).
Jimaging 08 00313 g001
Figure 2. The flow of macular center point detection (OD: optic disc, ROI: region of interest).
Figure 2. The flow of macular center point detection (OD: optic disc, ROI: region of interest).
Jimaging 08 00313 g002
Figure 3. OD localization process, (a) input image, (b) grayscale image of I (c) Ipre1 image, (d) Ipre2 image, (e) I c , (f) OD center area candidate (g) OD ROI (h) closing+opening result of binary OD ROI, (i) OD, (j) localization result of OD.
Figure 3. OD localization process, (a) input image, (b) grayscale image of I (c) Ipre1 image, (d) Ipre2 image, (e) I c , (f) OD center area candidate (g) OD ROI (h) closing+opening result of binary OD ROI, (i) OD, (j) localization result of OD.
Jimaging 08 00313 g003
Figure 4. Temporal direction viewed from the appearance of blood vessels in the optic disc.
Figure 4. Temporal direction viewed from the appearance of blood vessels in the optic disc.
Jimaging 08 00313 g004
Figure 5. The extraction process of blood vessels on the optic disc; (a) optic disc on the green layer; (b) the extraction result of blood vessels on the optic disc; (c) the result of elimination of horizontal blood vessels.
Figure 5. The extraction process of blood vessels on the optic disc; (a) optic disc on the green layer; (b) the extraction result of blood vessels on the optic disc; (c) the result of elimination of horizontal blood vessels.
Jimaging 08 00313 g005
Figure 6. Illustration of macular ROI determination. The red box shows the macular ROI.
Figure 6. Illustration of macular ROI determination. The red box shows the macular ROI.
Jimaging 08 00313 g006
Figure 7. The extraction process results of the macular center point, (a) retinal image, (b) macular ROI localization result (red box), (c) macular ROI gray image, (d) macular ROI binarization result, (e) dilation + opening operation result, (f) macular center point detection result.
Figure 7. The extraction process results of the macular center point, (a) retinal image, (b) macular ROI localization result (red box), (c) macular ROI gray image, (d) macular ROI binarization result, (e) dilation + opening operation result, (f) macular center point detection result.
Jimaging 08 00313 g007
Figure 8. Euclidian distance distribution between the proposed method with GT. A red circle with a radius of 50 shows the acceptable distance limit for well-detected results, (a) DRIVE, (b) DiaretDB1, (c) Messidor, and (d) JOGED.com (accessed on 22 July 2022).
Figure 8. Euclidian distance distribution between the proposed method with GT. A red circle with a radius of 50 shows the acceptable distance limit for well-detected results, (a) DRIVE, (b) DiaretDB1, (c) Messidor, and (d) JOGED.com (accessed on 22 July 2022).
Jimaging 08 00313 g008
Figure 9. Examples of the detection results and location of the ground truth. The red ‘+’ symbol represents the detection result while the blue ‘+’ indicates the location of the ground truth.
Figure 9. Examples of the detection results and location of the ground truth. The red ‘+’ symbol represents the detection result while the blue ‘+’ indicates the location of the ground truth.
Jimaging 08 00313 g009
Figure 10. Macular center detection failure on image059.png of DiaretDB1.
Figure 10. Macular center detection failure on image059.png of DiaretDB1.
Jimaging 08 00313 g010
Table 1. Comparison of macular center detection results with other studies.
Table 1. Comparison of macular center detection results with other studies.
MethodDRIVEDiaretDB1MessidorJOGED.com
AccuracyTime (s)AccuracyTime (s)AccuracyTime (s)AccuracyTime (s)
Zheng [23]100%1293.3%12----
Medhi [35]100%-95.51%-97.98%---
Chalakkal [17]100%2595,5%2598.5%25--
Sedai [10]100%0.4------
Romero-oraá [24]100%0.54100%14.5599.67%27.04--
Proposed method100%0.3498.78%0.5794.67%0.6493%0.78
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wibawa, H.A.; Harjoko, A.; Sumiharto, R.; Sasongko, M.B. Efficient and Robust Method to Detect the Location of Macular Center Based on Optimal Temporal Determination. J. Imaging 2022, 8, 313. https://doi.org/10.3390/jimaging8120313

AMA Style

Wibawa HA, Harjoko A, Sumiharto R, Sasongko MB. Efficient and Robust Method to Detect the Location of Macular Center Based on Optimal Temporal Determination. Journal of Imaging. 2022; 8(12):313. https://doi.org/10.3390/jimaging8120313

Chicago/Turabian Style

Wibawa, Helmie Arif, Agus Harjoko, Raden Sumiharto, and Muhammad Bayu Sasongko. 2022. "Efficient and Robust Method to Detect the Location of Macular Center Based on Optimal Temporal Determination" Journal of Imaging 8, no. 12: 313. https://doi.org/10.3390/jimaging8120313

APA Style

Wibawa, H. A., Harjoko, A., Sumiharto, R., & Sasongko, M. B. (2022). Efficient and Robust Method to Detect the Location of Macular Center Based on Optimal Temporal Determination. Journal of Imaging, 8(12), 313. https://doi.org/10.3390/jimaging8120313

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop