Next Article in Journal
Analysis of Physiological Signals for Stress Recognition with Different Car Handling Setups
Next Article in Special Issue
A Fusion Decision-Making Architecture for COVID-19 Crisis Analysis and Management
Previous Article in Journal
Digital Implementation of Harmonic and Unbalanced Load Compensation for Voltage Source Inverter to Operate in Grid Forming Microgrid
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Use of Convolutional Neural Networks and Digital Camera Images in Cataract Detection

Department of Information Management, National Chi Nan University, Nantou 54561, Taiwan
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(6), 887; https://doi.org/10.3390/electronics11060887
Submission received: 23 February 2022 / Revised: 8 March 2022 / Accepted: 10 March 2022 / Published: 11 March 2022

Abstract

:
Cataract is one of the major causes of blindness in the world. Its early detection and treatment could greatly reduce the risk of deterioration and blindness. Instruments commonly used to detect cataracts are slit lamps and fundus cameras, which are highly expensive and require domain knowledge. Thus, the problem is that the lack of professional ophthalmologists could result in the delay of cataract detection, where medical treatment is inevitable. Therefore, this study aimed to design a convolutional neural network (CNN) with digital camera images (CNNDCI) system to detect cataracts efficiently and effectively. The designed CNNDCI system can perform the cataract identification process accurately in a user-friendly manner using smartphones to collect digital images. In addition, the existing numerical results provided by the literature were used to demonstrate the performance of the proposed CNNDCI system for cataract detection. Numerical results revealed that the designed CNNDCI system could identify cataracts effectively with satisfying accuracy. Thus, this study concluded that the presented CNNDCI architecture is a feasible and promising alternative for cataract detection.

1. Introduction

According to a World Health Organization report [1], the estimated number of visually impaired people worldwide in 2021 is at least 2.2 billion. More than 1 billion can be prevented or have not been resolved. The leading cause of vision impairment or blindness is cataracts, which account for roughly 94 million. Its early detection and cataract surgery can decrease the possibility of blindness, avoid its deterioration, and improve patients’ visions.
A cataract is a crystalline opacity clouding the clear lens in human eyes. The blurring lenses block the light focus on the retina, resulting in poor vision. In addition, lens capsule or protein denaturation causes protein clumps, and pigments deposited on crystals are the main reasons for cataracts. Besides, many factors also cause cataracts, including genetic, aging, and smoking [2]. In the early stage, cataracts are generally painless, nonitching, and have almost no significant influence on sight. Thus, most patients are not aware of cataracts at the onset. Meanwhile, delays in detection and treatment could greatly increase the possibility of blindness. There are several ways to classify cataracts. The most common standard is based on the area of protein deposition, through which cataracts can be classified into three types, namely nuclear cataract, cortical cataract, and posterior subcapsular cataract. A nuclear cataract develops in the nucleus of the lens, becoming yellow and brown. On the other hand, a cortical cataract, mostly triggered by diabetes, occurs in the cortex of the lens. Meanwhile, a posterior subcapsular cataract, often happening to those who have taken a high dose of steroid medicine, exists at the back of the lens [3]. Another classification is based on the severity of the cataract, including normal, mild, medium, and severe cataracts.
To this day, regular cataract screening proves to be an effective way of preventing blindness and determining patients who need surgery [4]. Cataract surgery is the most successful and safest way to restore vision [1]. When cataracts have deteriorated enough to impact vision, removing the blurring lens by cataract surgery is necessary. In the medical field, ophthalmologists directly use slit lamps to screen a patient’s ocular tissues by a high-intensity light source [5] and manually grade by certain diagnostic criteria, such as the Lens Opacities Classification System III [6]. However, these manual judgments and instruments are time consuming, costly, and not easy to carry. Li et al. [7] pointed out that manual diagnosis may be influenced by personal experience accumulation, owning to subjective and error-prone judgment. The scarcity of well-experienced ophthalmologists, poor eye care resources, and economic considerations resulted in people having no access to effective treatment in time, leading to the blindness of most patients [8]. Nevertheless, many studies have made a great effort in developing highly portable and automatic cataract detection systems in recent years to treat cataracts early and enhance the accuracy of diagnoses.
When employing machine learning methods for cataract detection, feature extraction requires abundant engineering techniques and ophthalmic domain knowledge. Thus, the feature selection procedure is laborious and highly dependent on experience [9]. Simultaneously, predefined features might oversimplify the problem and omit some important hidden patterns. This investigation exploits the strength of convolutional neural networks in feature extraction and uses digital images provided by smartphones to conduct cataract detection. Meanwhile, the motivations, contributions, and innovations of this study are to develop user-friendly, portable, and automatic cataract detection systems for areas short of medical resources for instruments and ophthalmologists. In this way, users can pre-examine cataracts via mobile devices with digital images. The rest of this study is organized as follows: Section 2 examines related literature in recent years; Section 3 introduces convolutional neural networks; Section 4 presents the CNNDCI architecture for cataract detection and numerical results; Section 5 addresses the conclusions and directions for future study.

2. Related Work

Ophthalmic medical images have been widely employed in analyzing cataract severity and are expected to provide a better degree of accuracy in identifying and classifying cataract disease [10]. Generally, six types of ophthalmic images (i.e., slit-lamp images [11,12,13], retro illumination images [14], ultrasonic images [15,16], anterior segment optical coherence tomography images [17], fundus images [9,18,19,20,21,22], and digital camera images [5,8,23,24,25,26,27,28]) have been used for cataract detection. Fundus images and slit-lamp images are the two most frequently used in cataract detection. However, instruments for fundus images and slit-lamp images are not easy to access for people living in rural areas. Comparatively, digital camera images are more available as an alternative for ophthalmic images when cataract detection is performed [29]. Therefore, a cataract detection system based on digital camera images is highly required for early cataract detection that is more popular and user friendly.
Nayak [23] used the image preprocessing technique, such as big ring area and edge pixel count, to select features from the pupil area of images and conducted classification tasks by support vector machines. Fuadah et al. [24] gathered the features in digital camera images and converted them into grayscale manually. The gray level co-occurrence matrix was used to extract features divided into contrast, dissimilarity, and uniformity. The k-nearest neighbor was used to classify the images into normal or cataract. Pathak and Kumar [5] proposed a texture features-based algorithm for cataract detection by the occurrences of cataracts from the true color images. Khan et al. [25] presented a computer-aided diagnostic system that aids cataract detection in resource-lacking areas. In preprocessing, this study utilized Daugman’s operator to isolate the iris and pupil from optical images for the region of interest, where six features were gathered. The support vector machine was employed in the presented system to classify images into cataract and noncataract. Tawfik et al. [26] used two classifiers—support vector machines and artificial neural networks—to perform cataract detection. The discrete wavelet transform and Log Gabor transform were employed to select features from pupil areas. This study reported that support vector machines outperformed artificial neural networks in terms of classification accuracy. Agarwal et al. [27] designed an android application system to detect the presence of a cataract in an individual’s eye using a smartphone as a medium. Three classifiers were conducted: support vector machines, Naïve Bayes, and the K-nearest neighbors algorithm. Numerical results showed that the K-nearest neighbors algorithm could generate more accurate results than the other two classifiers. Sigit et al. [8] developed a cataract detection system using a single-layer perceptron with smartphones to reduce ophthalmologists’ workload. The proposed system was able to perform classification with satisfying accuracy. Yusuf et al. [28] presented a web-based cataract detection system employing a convolutional neural network with digital camera images. They pointed out that the classification accuracy was influenced by transfer learning trained on ImageNet.

3. Convolutional Neural Networks

Convolutional neural networks with both feature extraction and classification capabilities [29] can cope with classification problems effectively and efficiently. In addition, weight sharing and feature extraction in the convolutional neural network have highly improved the computation efficiency. Thus, convolutional neural networks can handle problems with large-scale data [18]. A convolutional neural network includes a convolutional layer, a max-pooling layer, a flatten layer, and a dense layer. The convolutional layer can effectively extract essential features. The feature map F has a convolution operation with kernel maps, an operation represented by Equation (1).
F j l = f i = 1 N l 1 F i l 1 K i j l + b j l
where Nl − 1 is the number of feature maps of l − 1th layer, Flj is the jth feature map of l layer, Fl − 1i is the ith feature map of l − 1 layer, ⨂ represents the convolutional operation, K i j l represents the kernel map connecting the ith feature map of l − 1 layer and jth feature map of l layer, and b j l is the bias. The activation function f ( ) , common Sigmoid, and ReLU are shown as Equations (2) and (3), which are used to learn complex patterns that can solve problems linear models could not deal with.
S i g m o i d :           f x = 1 1 + e x
R e L U :           f x = m a x 0 ,   x
The Max-pooling layer is used to reduce the size of the feature map, not only to avoid overfitting but also to decrease the computation time. Its operation can be expressed by Equation (4).
F j l = M a x p o o l i n g F i l 1
where Max-pooling () represents the max-pooling operation. In the flatten layer, the 2-dimensional feature maps F were converted into a 1-dimensional array, as illustrated in Equation (5).
n l = F H l 1 × F W l 1 × F C l 1
where n represents the number of neurons in the flatten layer, and F H ,   F W , F C represent the height and width of the feature map and channel size, respectively. The flattened neurons were used as the input of the dense layer, and its operation is determined by Equations (6) and (7).
x j l = i = 1 N l 1 F i l 1 w i ,   j l + b j l
y j l = f x j l
where x j l represents the output, N l 1 is the number of neurons of l − 1 layer, w i , j l represents the weight between the ith neuron of l − 1 layer and jth neuron of lth layer. The output layer is the last layer of the network and produces the final result.

4. The Proposed CNNDCI System for Cataracts Detection and Numerical Results

4.1. The Proposed CNNDCI System

This study proposed a cataract detection system and employed CNN with digital camera images to identify cataracts. Figure 1 illustrates the proposed CNNDCI architecture. This study collected two datasets from GitHub.com, which were published by krishnabojha [30] and piygot5 [31]. In this study, datasets of krishnabojha [30] and piygot5 [31] are denominated as dataset I and dataset II, respectively. All images were photographed by digital camera. Both datasets contain two classes, namely cataract and noncataract. Dataset I contains 9668 images, including 4514 cataract images and 5154 noncataract images. Dataset II contains 89 images, including 43 cataract images and 46 noncataract images. The ImageDataGenerator [32], one of the Keras functions, was employed to preprocess image data. This study used the dataset I to train CNN with fivefold cross-validation procedures and then provided classification accuracies. Dataset II was employed to examine the classification performances of the trained CNN models by different data. Table 1 indicates the numbers of data in each partition for dataset I with fivefold cross-validation. Table 2 illustrates the training data and testing data for dataset I with fivefold cross-validation. Figure 2 illustrates the convolutional neural networks coping with digital camera eye images used in this investigation.
The proposed model comprised seven layers, including two convolutional layers, two Max-pooling layers, a flatten layer, and two dense layers. Meanwhile, the input images contained three basic colors: red, green, and blue, with 64 × 64 sizes. The proposed CNN included two convolutional layers, 32 filters with kernel sizes of 3 × 3, a stride of 1. Convolutional layers were activated by rectified linear unit (ReLU). The ReLU only conveyed positive values; the outputs were zero when values were negative. A pool size of 2 × 2 was employed in two max-pooling layers. The ReLU activation function and 128 neurons were used in the first dense layer. As a binary classifier, one neuron and the Sigmoid activation function were employed in the last dense layer. This study could obtain a highly satisfying performance when the dropout rate approached zero. Thus, the dropout rate was set to zero.
The learning algorithm using binary cross-entropy as a loss function aimed to minimize the training error between predicted and actual values. In this study, an epoch was set to 1000 for the training procedure, and the Adam optimizer with a learning rate of 0.001 was utilized. The components of the proposed CNN model are presented in Table 3. After training, the CNN model was built and performed on a server. Users were required to use mobile devices to photograph eyeball appearances. Then, digital camera images were uploaded to the server through a website. Finally, the detected results were immediately generated on mobile devices for users.
To implement the proposed CNNDCI system, the hardware environment used for the convolutional neural network comprised NVIDIA GeForce GTX 1080 GPU, Intel(R) Core (TM) i7-7700 CPU @ 3.6oGHz, 32GB RAM, and Windows 10 as the operating system. The python deep learning library Keras version employed was 2.4.3, used to run TensorFlow.

4.2. Numerical Results

In this study, a confusion matrix, represented in Table 4, was used to measure the performance of models. Three indices—accuracy, precision, and recall, expressed as Equations (8)–(10), respectively—were employed to measure the performance of forecasting models [33].
A c c u r a c y = T p + T n T p + F p + T n + F n  
P r e c i s i o n = T p T p + F p
R e c a l l = T p T p + F n
where T p , F p , T n , and F n are the number of true positive, false positive, true negative, and false negative.
Table 5 shows three measurements of accuracy generated by fivefold cross-validation. Figure 3 indicates the convergence curves of the five-fold cross-validation of training and testing accuracy. In terms of classifying accuracy, the average testing accuracy of the dataset I and dataset II are 98.5% and 92%, respectively. Comparing numerical results of the present study with the previous study of dataset I, this study obtained a higher average testing accuracy with a fivefold cross-validation procedure than the previous study without cross-validation procedures. In addition, the recall values are 97.9% and 91%, respectively, for Dataset I and Dataset II. The high recall values indicate that the portion of the false negative was smaller and imply that the probability that a cataract could not be detected was very low. Table 6 lists the numerical results of related studies using digital camera images. It can be observed that the proposed CNNDCI with Dataset I outperformed the other related studies in terms of average classification testing accuracy with a fivefold cross-validation procedure. For Dataset II, the proposed CNNDCI model can obtain a satisfying average classification accuracy compared with the current related studies.
Figure 4 illustrates the graphical user interface, including four steps: the start page, uploading images of eyes, cropping images of eyes, and detecting results. First, users pressed the Start button and then uploaded images of eyes for detection. After uploading the eye images, users needed to crop them according to the proper position. After cropping the images, the uploaded images were detected by the CNNDCI built up in a server. Finally, the CNNDCI delivered the detected results immediately to users through mobile devices. The progress against the most recent state-of-the-art similar studies is that this study developed a very user-friendly, convenient, and accurate cataract detection system.

5. Conclusions

Regular screening and early treatment can hugely decrease the probability of deterioration and blindness. Meanwhile, an inexpensive, robust, and convenient tool for screening cataracts is essential in rural or underdeveloped areas. In addition, with the rapid development of the deep learning technique, the convolutional neural network has been one of the most powerful classifiers. Thus, this study used digital camera images to train a convolutional neural network classifier to detect cataracts. The numerical results revealed that the proposed model is robust with two datasets. Furthermore, a user-friendly graphical user interface was provided to increase the easy-to-use feature and popularity of the proposed CNNDCI system. For future studies, more detailed data, such as classifications of severe levels or the locations of protein deposition, could be included by improving the current CNNDCI system into a multiclassification system. The level of cataracts includes normal, mild, medium, and severe cataract, while the location of protein deposition includes nuclear cataracts, cortical cataracts, and posterior subcapsular cataracts [34]. In particular, the convolutional neural network classifier can provide more detailed results with more detailed data and the multiclassification function. Another possible direction for future work is improving the quality of training and image enhancement [35,36] by continually collecting noise-free images [37]. It must be noted that this study was limited by the environmental conditions of users. Furthermore, it was found that the reflection of light and the positions of images influenced the quality of the performance of CNNDCI. Therefore, a notice function to remind users of light reflections and positions of images could also be a direction for future work.

Author Contributions

Conceptualization, P.-F.P.; data curation; C.-J.L., M.M., H.-H.H., S.-H.W. and D.-N.C.; formal analysis, P.-F.P. and C.-J.L.; funding acquisition, P.-F.P.; methodology, P.-F.P.; software, C.-J.L. and M.M.; visualization, C.-J.L. and M.M.; writing—original draft, C.-J.L. and P.-F.P., review and editing, P.-F.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Technology, Taiwan, under the Contract Number MOST 109-2410-H-260-023.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. World Health Organization. Blindness and Vision Impairment. Available online: https://www.who.int/en/news-room/fact-sheets/detail/blindness-and-visual-impairment (accessed on 14 October 2021).
  2. Luo, X.; Li, J.; Chen, M.; Yang, X.; Li, X. Ophthalmic Disease Detection via Deep Learning With A Novel Mixture Loss Function. IEEE J. Biomed. Health Inform. 2021, 25, 3332–3339. [Google Scholar] [CrossRef] [PubMed]
  3. Patil, D.; Nair, A.; Bhat, N.; Chavan, R.; Jadhav, D. Analysis and study of cataract detection techniques. In Proceedings of the 2016 International Conference on Global Trends in Signal Processing, Information Computing and Communication (ICGTSPICC), Jalgaon, India, 22–24 December 2016. [Google Scholar]
  4. Hu, S.; Wang, X.; Wu, H.; Luan, X.; Qi, P.; Lin, Y.; He, X.; He, W. Unified Diagnosis Framework for Automated Nuclear Cataract Grading Based on Smartphone Slit-Lamp Images. IEEE Access 2020, 8, 174169–174178. [Google Scholar] [CrossRef]
  5. Pathak, S.; Kumar, B. A robust automated cataract detection algorithm using diagnostic opinion based parameter thresholding for telemedicine application. Electronics 2016, 5, 57. [Google Scholar] [CrossRef] [Green Version]
  6. Chylack, L.T.; Wolfe, J.K.; Singer, D.M.; Leske, M.C.; Bullimore, M.A.; Bailey, I.L.; Friend, J.; McCarthy, D.; Wu, S.Y. The lens opacities classification system III. Arch. Ophthalmol. 1993, 111, 831–836. [Google Scholar] [CrossRef] [PubMed]
  7. Li, H.; Lim, J.H.; Liu, J.; Wing, D.; Wong, K.; Wong, T.Y. Feature analysis in slit-lamp image for nuclear cataract diagnosis. In Proceedings of the 2010 3rd International Conference on Biomedical Engineering and Informatics, Yantai, China, 16–18 October 2010. [Google Scholar]
  8. Sigit, R.; Triyana, E.; Rochmad, M. Cataract Detection Using Single Layer Perceptron Based on Smartphone. In Proceedings of the 2019 3rd International Conference on Informatics and Computational Sciences (ICICoS), Semarang, Indonesia, 29–30 October 2019. [Google Scholar]
  9. Xu, X.; Zhang, L.; Li, J.; Guan, Y.; Zhang, L. A hybrid global-local representation CNN model for automatic cataract grading. IEEE J. Biomed. Health Inform. 2019, 24, 556–567. [Google Scholar] [CrossRef] [PubMed]
  10. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Qian, X.; Patton, E.W.; Swaney, J.; Xing, Q.; Zeng, T. Machine learning on cataracts classification using SqueezeNet. In Proceedings of the 2018 4th International Conference on Universal Village (UV), Boston, MA, USA, 21–24 October 2018. [Google Scholar]
  12. Liu, X.; Jiang, J.; Zhang, K.; Long, E.; Cui, J.; Zhu, M.; An, Y.; Zhang, J.; Liu, Z.; Lin, Z.; et al. Localization and diagnosis framework for pediatric cataracts based on slit-lamp images using deep features of a convolutional neural network. PLoS ONE 2017, 12, e0168606. [Google Scholar] [CrossRef] [PubMed]
  13. Xu, Y.; Gao, X.; Lin, S.; Wong, D.W.K.; Liu, J.; Xu, D.; Cheng, C.Y.; Cheung, C.Y.; Wong, T.Y. Automatic grading of nuclear cataracts from slit-lamp lens images using group sparsity regression. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Nagoya, Japan, 22–26 September 2013. [Google Scholar]
  14. Zhang, W.; Li, H. Lens opacity detection for serious posterior subcapsular cataract. Med. Biol. Eng. Comput. 2017, 55, 769–779. [Google Scholar] [CrossRef] [PubMed]
  15. Caixinha, M.; Amaro, J.; Santos, M.; Perdigão, F.; Gomes, M.; Santos, J. In-vivo automatic nuclear cataract detection and classification in an animal model by ultrasounds. IEEE Trans. Biomed. Eng. 2016, 63, 2326–2335. [Google Scholar] [CrossRef] [PubMed]
  16. Caxinha, M.; Velte, E.; Santos, M.; Perdigão, F.; Amaro, J.; Gomes, M.; Santos, J. Automatic cataract classification based on ultrasound technique using machine learning: A comparative study. Phys. Procedia 2015, 70, 1221–1224. [Google Scholar] [CrossRef] [Green Version]
  17. Zhang, X.; Xiao, Z.; Higashita, R.; Chen, W.; Yuan, J.; Fang, J.; Hu, Y.; Liu, J. A Novel Deep Learning Method for Nuclear Cataract Classification Based on Anterior Segment Optical Coherence Tomography Images. In Proceedings of the 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Toronto, ON, Canada, 11–14 October 2020. [Google Scholar]
  18. Khan, M.S.M.; Ahmed, M.; Rasel, R.Z.; Khan, M.M. Cataract Detection Using Convolutional Neural Network with VGG-19 Model. In Proceedings of the 2021 IEEE World AI IoT Congress (AIIoT), Seattle, WA, USA, 10–13 May 2021. [Google Scholar]
  19. Yang, J.J.; Li, J.; Shen, R.; Zeng, Y.; He, J.; Bi, J.; Li, Y.; Zhang, Q.; Peng, L.; Wang, Q. Exploiting ensemble learning for automatic cataract detection and grading. Comput. Methods Programs Biomed. 2016, 124, 45–57. [Google Scholar] [CrossRef] [PubMed]
  20. Guo, L.; Yang, J.J.; Peng, L.; Li, J.; Liang, Q. A computer-aided healthcare system for cataract classification and grading based on fundus image analysis. Comput. Ind. 2015, 69, 72–80. [Google Scholar] [CrossRef]
  21. Zheng, J.; Guo, L.; Peng, L.; Li, J.; Yang, J.; Liang, Q. Fundus image based cataract classification. In Proceedings of the 2014 IEEE International Conference on Imaging Systems and Techniques (IST) Proceedings, Santorini, Greece, 14–17 October 2014. [Google Scholar]
  22. Yang, M.; Yang, J.J.; Zhang, Q.; Niu, Y.; Li, J. Classification of retinal image for automatic cataract detection. In Proceedings of the 2013 IEEE 15th International Conference on e-Health Networking, Applications and Services (Healthcom 2013), Lisbon, Portugal, 9–12 October 2013. [Google Scholar]
  23. Nayak, J. Automated classification of normal, cataract and post cataract optical eye images using SVM classifier. In Proceedings of the World Congress on Engineering and Computer Science, San Francisco, CA, USA, 23–25 October 2013. [Google Scholar]
  24. Fuadah, Y.N.; Setiawan, A.W.; Mengko, T.L.R. Performing high accuracy of the system for cataract detection using statistical texture analysis and K-Nearest Neighbor. In Proceedings of the 2015 International Seminar on Intelligent Technology and Its Applications (ISITIA), Surabaya, Indonesia, 20–21 May 2015. [Google Scholar]
  25. Khan, A.A.; Akram, M.U.; Tariq, A.; Tahir, F.; Wazir, K. Automated Computer Aided Detection of Cataract. In Proceedings of the International Afro-European Conference for Industrial Advancement, Marrakesh, Morocco, 21–23 November 2016. [Google Scholar]
  26. Tawfik, H.R.; Birry, R.A.; Saad, A.A. Early Recognition and Grading of Cataract Using a Combined Log Gabor/Discrete Wavelet Transform with ANN and SVM. Int. J. Comput. Inf. Eng. 2018, 12, 1038–1043. [Google Scholar]
  27. Agarwal, V.; Gupta, V.; Vashisht, V.M.; Sharma, K.; Sharma, N. Mobile application based cataract detection system. In Proceedings of the 2019 3rd International Conference on Trends in Electronics and Informatics (ICOEI), Tirunelveli, India, 23–25 April 2019. [Google Scholar]
  28. Yusuf, M.; Theophilous, S.; Adejoke, J.; Hassan, A.B. Web-Based Cataract Detection System Using Deep Convolutional Neural Network. In Proceedings of the 2019 2nd International Conference of the IEEE Nigeria Computer Chapter (NigeriaComputConf), Zaria, Nigeria, 14–17 October 2019. [Google Scholar]
  29. Zhang, X.; Fang, J.; Hu, Y.; Xu, Y.; Higashita, R.; Liu, J. Machine Learning for Cataract Classification and Grading on Ophthalmic Imaging Modalities: A Survey. arXiv 2020, arXiv:2012.04830. [Google Scholar]
  30. Krishnabojha. Cataract_Detection-Using-CNN. Available online: https://github.com/krishnabojha/Cataract_Detection-using-CNN (accessed on 27 March 2021).
  31. Piygot5. Cataract-Detection-and-Classification. Available online: https://github.com/piygot5/Cataract-Detection-and-Classification (accessed on 27 March 2021).
  32. Keras. Image Data Generator. Available online: https://keras.io/zh/preprocessing/image/ (accessed on 27 March 2021).
  33. Jahanbakhshi, A.; Abbaspour-Gilandeh, Y.; Heidarbeigi, K.; Momeny, M. A novel method based on machine vision system and deep learning to detect fraud in turmeric powder. Comput. Biol. Med. 2021, 136, 104728. [Google Scholar] [CrossRef] [PubMed]
  34. Patwari, M.A.U.; Arif, M.D.; Chowdhury, M.N.; Arefin, A.; Imam, M.I. Detection, categorization, and assessment of eye cataracts using digital image processing. In Proceedings of the First International Conference on Interdisciplinary Research and Development, Thailand, China, 31 May–1 June 2011. [Google Scholar]
  35. Khaldi, Y.; Benzaoui, A.; Ouahabi, A.; Jacques, S.; Taleb-Ahmed, A. Ear recognition based on deep unsupervised active learning. IEEE Sens. J. 2021, 21, 20704–20713. [Google Scholar] [CrossRef]
  36. Khaldi, Y.; Benzaoui, A. A new framework for grayscale ear images recognition using generative adversarial networks under unconstrained conditions. Evol. Syst. 2021, 12, 923–934. [Google Scholar] [CrossRef]
  37. Khan, A.; Jin, W.; Haider, A.; Rahman, M.; Wang, D. Adversarial Gaussian Denoiser for Multiple-Level Image Denoising. Sensors 2021, 21, 2998. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The architecture of the proposed CNNDCI system.
Figure 1. The architecture of the proposed CNNDCI system.
Electronics 11 00887 g001
Figure 2. The architecture of CNN used in this study.
Figure 2. The architecture of CNN used in this study.
Electronics 11 00887 g002
Figure 3. Convergence curves of the fivefold cross-validation of training accuracy and testing accuracy.
Figure 3. Convergence curves of the fivefold cross-validation of training accuracy and testing accuracy.
Electronics 11 00887 g003aElectronics 11 00887 g003b
Figure 4. Four steps of the graphical user interface with CNNDCI.
Figure 4. Four steps of the graphical user interface with CNNDCI.
Electronics 11 00887 g004
Table 1. The numbers of data in each partition for dataset I with fivefold cross-validation.
Table 1. The numbers of data in each partition for dataset I with fivefold cross-validation.
PartitionsCataract CasesNormal Cases
Partition 19021030
Partition 29031031
Partition 39031031
Partition 49031031
Partition 59031031
Table 2. The training data and testing data for dataset I with fivefold cross-validation.
Table 2. The training data and testing data for dataset I with fivefold cross-validation.
CVTraining DataTesting Data
CV 1Partition 1, Partition 2, Partition 3, Partition 4Partition 5
CV 2Partition 1, Partition 2, Partition 3, Partition 5Partition 4
CV 3Partition 1, Partition 2, Partition 4, Partition 5Partition 3
CV 4Partition 1, Partition 3, Partition 4, Partition 5Partition 2
CV 5Partition 2, Partition 3, Partition 4, Partition 5Partition 1
Table 3. The illustration of components in the proposed CNN architecture.
Table 3. The illustration of components in the proposed CNN architecture.
LayersComponentsOutputsKernel SizeStride
0Image input (width, height, channels)(64, 64, 3)
1Convolutional (width, height, channels)(62, 62, 32)3 × 31
2Max-pooling (width, height, channels)(31, 31, 32)2 × 2
3Convolutional (width, height, channels)(29, 29, 32)3 × 31
4Max-pooling (width, height, channels)(14, 14, 32)2 × 2
5Flatten(nodes)(6272)
6Dense(nodes)(128)
7Dense(nodes)(1)
Table 4. The confusion matrix.
Table 4. The confusion matrix.
Confusion
Matrix
Actual
TrueFalse
PredictedTrueTrue positiveFalse positive
FalseFalse negativeTrue negative
Table 5. Experiment results of fivefold the cross-validation.
Table 5. Experiment results of fivefold the cross-validation.
MetricsCV1CV2CV3CV4CV5Average
Dataset ITraining accuracy99.1%98.2%98.6%99.5%99.3%98.9%
Testing accuracy99.1%98.2%97.7%99.1%98.8%98.5%
Recall98.8%97.4%97.1%98.5%98.0%97.9%
Precision99.2%98.7%97.9%99.4%99.4%98.9%
Dataset IITesting accuracy94.3%89.8%92.1%92.1%92.1%92%
Recall95.3%90.6%88.3%90.6%90.6%91%
Precision93.1%88.6%95.0%92.8%92.8%92.4%
Table 6. Numerical results of related studies using digital camera images.
Table 6. Numerical results of related studies using digital camera images.
ReferencesMethodAccuracyTraining DataTesting DataCross-Validation
[5]K-means Clustering98%Not available200No
[8]Single Layer Perceptron85%3020No
[23]Support Vector Machine88.39%12549No
[24]K-nearest Neighbor94.5%8080No
[25]Support Vector Machine69.4%5836No
[26]Support Vector Machine96.8%7842No
Artificial Neural Network92.3%
[27]K-nearest Neighbor83%Not available1152No
Support Vector Machine75.2%
Naïve Bayes76.6%
[28]Convolutional Neural Network78%10030No
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lai, C.-J.; Pai, P.-F.; Marvin, M.; Hung, H.-H.; Wang, S.-H.; Chen, D.-N. The Use of Convolutional Neural Networks and Digital Camera Images in Cataract Detection. Electronics 2022, 11, 887. https://doi.org/10.3390/electronics11060887

AMA Style

Lai C-J, Pai P-F, Marvin M, Hung H-H, Wang S-H, Chen D-N. The Use of Convolutional Neural Networks and Digital Camera Images in Cataract Detection. Electronics. 2022; 11(6):887. https://doi.org/10.3390/electronics11060887

Chicago/Turabian Style

Lai, Chi-Ju, Ping-Feng Pai, Marvin Marvin, Hsiao-Han Hung, Si-Han Wang, and Din-Nan Chen. 2022. "The Use of Convolutional Neural Networks and Digital Camera Images in Cataract Detection" Electronics 11, no. 6: 887. https://doi.org/10.3390/electronics11060887

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop