Next Article in Journal
Is Methionyl-tRNA Synthetase Applicable as a Diagnostic Marker for Lung Cancer in Bronchial Ultrasound-Guided Brushing Cells?
Previous Article in Journal
Multiparametric MRI and Radiomics in Prostate Cancer: A Review of the Current Literature
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep-Learning-Based Hair Damage Diagnosis Method Applying Scanning Electron Microscopy Images

AI/SC Lab, Computer Engineering of Gachon University, Seongnam-si 461-701, Gyeonggi-do, Korea
*
Author to whom correspondence should be addressed.
Diagnostics 2021, 11(10), 1831; https://doi.org/10.3390/diagnostics11101831
Submission received: 25 August 2021 / Revised: 28 September 2021 / Accepted: 29 September 2021 / Published: 3 October 2021
(This article belongs to the Topic Artificial Intelligence in Healthcare)

Abstract

:
In recent years, with the gradual development of medicine and deep learning, many technologies have been developed. In the field of beauty services or medicine, it is particularly important to judge the degree of hair damage. Because people in modern society pay more attention to their own dressing and makeup, changes in the shape of their hair have become more frequent, e.g., owing to a perm or dyeing. Thus, the hair is severely damaged through this process. Because hair is relatively thin, a quick determination of the degree of damage has also become a major problem. Currently, there are three specific methods for this purpose. In the first method, professionals engaged in the beauty service industry make a direct judgement with the naked eye. The second way is to observe the damaged cuticle layers of the hair using a microscope, and then make a judgment. The third approach is to conduct experimental tests using physical and chemical methods. However, all of these methods entail certain limitations, inconveniences, and a high complexity and time consumption. Therefore, our proposed method is to use scanning electron microscope to collect various hair sample images, combined with deep learning to identify and judge the degree of hair damage. This method will be used for hair quality diagnosis. Experiment on the data set we made, compared with the experimental results of other lightweight networks, our method showed the highest accuracy rate of 94.8%.

1. Introduction

Hair quality inspection and damage determination in beauty salons rely only on the judgment of professionals such as beauticians, which is mostly based on the tactile experience of beauticians and observations with the naked eye. However, in the absence of excessive professional experience, errors can occur when judging the quality of the hair. The current methods of judging hair damage are to determine the hair damage by detecting the moisture content, cystine content, coagulation, and/or relaxation; apply a dye absorption method, the alkali solubility, copper absorption method, or lithium bromide method [1]; or consider the absorbance and tensile strength. Ref. [2] Using these methods, we need to test the composition of many aspects in a chemical or physical experiment. Compared with our proposed method, this type of chemical and physical method for detecting hair is complicated and time-consuming. Therefore, we need a faster and simpler approach to determine the degree of hair damage. This method will be applied to portable devices in the future, and everyone can easily judge the degree of damage to one’s hair anytime and anywhere, and then decide whether to receive a perm, dye, or other beauty services. To better observe cuticle layers in the hair, we used a scanning electron microscope (SEM) and observed it at ×800. Thus, according to the form of cuticle layers (Figure 1), damaged cuticle layers (the white part) accounts for the proportion of hair. We define a lifting up of the cuticle edge and an irregular overlay of cuticles without cracks or holes as weak damage; cracks or holes due to severe lifting up of cuticle layers as damage; exposure of cortical cells and complete disappearance of cuticles as high damage. Regarding the analysis of the SEM image, the segmentation method was developed using the watershed segmentation algorithm, the global–local threshold method, Laplacian of Gaussian filter, and non-maximum suppression in [3].
We, therefore, judged that the degree of hair damage directly seen in the hair images was too simple to understand. Thus, we need to recognize and judge the degree of hair damage in the hair images obtained through observations under a microscope. In this study, we combine deep learning image classification technology for recognition and classification.
The main contributions of this study through the use of a lightweight Convolutional Neural Network (CNN) [4] are as follows:
  • Decreases the number of free parameters.
  • Achieves a high classification accuracy using small datasets.
  • Reduces the training time required for convergence.
  • Decreases the complexity of the network, thus, enabling mobile applications.
Establishes a future research direction that will extend the applicability of the method to digital microscopy devices (for example:SVBONY SM401 Microscope Wireless Digital Microscope 50×–1000×).
Weak damage
Damage
High damage
Damaged cuticle layers

2. Materials and Methods

2.1. Datasets of SEM Images (DHI Dataset)

For deep learning, it is important to choose a dataset. To identify and judge the degree of hair damage, we need to identify and classify cuticle layers in the hair images. However, according to our investigation, there are very few datasets on microscopic images of hair that clearly show the cuticle layers. For this reason, we collected and produced a dataset directly, called the damage hair image (DHI) dataset, as shown (Figure 2). The DHI dataset was collected through our own observation using SEM. During the process of data collection, we collected various hair samples for observation and image collection. The dataset contains hair samples from young to old people, from 19 to 55 years old, as well as perm and non-perm processed hair samples. Using distance from the scalp less than 1 cm, from 3 to 5 cm, and over 5 cm, we sectioned the upper, middle, and lower parts of the hair to be hair image data sample (from the root to the tip), and then contrasted and observed them under SEM. After the observation, the cuticle layers in the sample were arranged in an orderly manner. A small amount of the middle part of the cuticle layers is generally missing, whereas the cuticle layers in the lower part is largely missing, and in some areas is completely gone. However, the perm hair samples were irregularly distributed, and cuticle layers were shown to be shedding from the center. After a long period of time, we collected images through SEM, and then analyzed and organized them. The final sample data images of minor, moderate, and severe injuries numbered 286 in total. However, this number of datasets is still too low for deep learning. Therefore, we applied data expansion technology [5] including rotation, blur, and a noise increase to expand the dataset to 2900 images. The original image resolution was 640 × 480, and to make the data training of the neural network faster and more convenient, the resolution was adjusted to 224 × 224.

2.2. Methodology

2.2.1. Convolutional Neural Networks

A convolutional neural network (CNN) is a type of feedforward neural network that includes convolution calculations and has a deep structure and is a representative deep learning algorithm [6]. CNNs have representation learning capabilities and can classify input information according to their structure, and in recent years, recognition and classification networks in deep learning have become increasingly mature, including MobileNet, Googlenet, and VggNet. These networks have demonstrated a high accuracy in the recognition and classification of animals, plants, faces, and other fields. However, we need a small network for experimentation, which can be applied to mobile phones or portable devices in the future, and because our dataset is small, it is prone to a poor training and easy overfitting on large networks. We, therefore, need a smaller model than a VGG or ResNet. For greater accuracy in the classification and recognition results of hair damage, we propose a network suitable for hair classification, i.e., Hair-Diagnosis-Mobilenet (HDM-NET). In addition, we use HDM-NET to extract and select the features, and finally use an SVM [7] to classify hair damage images (Figure 3). In this study, we used an SVM for a single experiment and HDM-NET combined with multilayer perceptron (MLP) [8], random forest (RF) [9], k-nearest neighbor (KNN) [10], and the original MobileNet model combined with the classification methods to conduct multi-group control experiments. Finally, we discuss the advantages and disadvantages of the proposed method.

2.2.2. Select Features

When the feature vector is generated from the image information, we choose the attributes. The significance of this choice is to eliminate unnecessary features when analyzing the image, and then reduce the computational complexity, thereby optimizing the prediction model, obtaining better results.According to [11], attribute selection technology is mainly used to identify basic information. Herein, we use the gain ratio [12] algorithm for selection. When calculating the information gain, entropy is used to measure the complexity and decision trees [13] to observe the vector attributes. In this way, the information gain is improved. It, therefore, provides a higher accuracy when spanning the tree. The dataset (D) consists of data samples with different categories (T). Equation (1) presents the classification calculation information for a given sample.
I n c a ( D ) = i = 1 t p i log 2 ( p i )
where 𝑝𝑖 is the probability of that the sample belongs to class 𝐶𝑖. The entropy calculation of a given attribute 𝐴t having 𝑣 values is shown in Equation (2),
E n t r o p y ( A t ) = i = 1 t I n c a ( D ) S 1 i + S 2 i + + S t i s  
where D i j represents the number of samples belonging to class C i of subset D i . The attribute gain 𝐴t is represented through Equation (3):
              G a i n ( A t ) = I n c a ( D ) E n t r o p y ( A t )
Equation (4) presents the information value generated by dividing the dataset of D into 𝑣 partitions:
    S I n c a ( D ) = i = 1 v ( | S i | | S | ) log 2 ( | S i | | S | )
Finally, the gain ratio is defined as the result of dividing the solution into Equation (3) by solving Equation (4). The attributes are ranked according to the values of their gain ratios, and 𝑛 attributes with highest values are selected.

2.2.3. MobileNet

We selected and improved MobileNet [14], which can be used in mobile devices and small microscopes or mobile cameras to obtain and diagnose hair images. MobileNet is based on depthwise separable convolutions, which consist of two core layers: depthwise [15] and pointwise [16] convolutions. A depthwise convolution is the step of filtering the input without creating new features. Thus, the process of generating new features, called a pointwise convolution, was also applied. Finally, the combination of the two layers is called a depthwise separable convolution. This model uses depthwise convolutions to apply a single filter for each channel of input, and then uses a 1 × 1 convolution (pointwise) to create a linear combination of output from the depthwise layer. Batch normalization (BN) [17,18] and a rectified linear unit (ReLU) [19,20] were used after each convolution. (Figure 4) shows the architecture of MobileNet, which consists of a convolution layer, a depthwise convolution layer followed by a BN layer and an ReLU layer, a pointwise convolution layer, a BN and ReLU layer, a global average pooling layer, a reshape layer, a dropout layer, a convolutional layer, a softmax layer, and a reshape layer. This model contains approximately four million parameters, which is a much smaller number than those of the other models. The structure of MobileNet (depthwise + pointwise convolutions).

2.2.4. Hair-Diagnosis Mobilenet (HDM-NET)

Because our dataset is relatively small, when testing the original MobileNet model, the classification results were unsatisfactory. Therefore, we made some improvements based on the MobileNet model. We removed three layers from the five layers of depthwise convolution in MobileNet. The purpose here was to reduce the redundant parameters of the model with almost no effect on the results. We then changed the final average pooling to global pooling [21] and added batch normalization. Relu will cause some neurons to output a 0, resulting in a sparsity of the network, reducing the interdependence between parameters, and alleviating the occurrence of an overfitting. We, therefore, used Relu. Finally, the output [22,23] applies an SVM (Figure 5). After such changes, this model achieved good results in our DHI dataset. Because there is no new structure, the number of parameter calculations of the improved model [24] is slightly lower than that of the original MobileNet model. A standard convolutional layer inputs the feature map F of D F × D F × M, and obtains an output feature map G of D G × D G × N, where DF represents the width and height of the input feature map, M is the number of input channels (input depth), DG is the width and height of the output feature map, N is the number of output channels (output depth), and K is the size of the depth convolution kernel. The table of parameters distributed in HDM-NET with classifier is (Table 1). The expression is as follows:
Depthwise convolutional parameters,
D K × D K × M × D F × D F .  
Pointwise convolutional parameters,
M × N × D F × D F .  
Thus, all parameters are expressed as follows:
P T = D K × D K × M × D F × D F + M × N × D F × D F .  

3. Experiment Results

In the classification process, the most commonly used classifiers are SVM, multilayer perceptron (MLP), random forest (RF), K-nearest neighbors (KNN), radial basis function network (RBFN) [25], and Naive Bayes (NB) [26]. During the experiment, we conducted multiple sets of controlled trials. The four groups of experiments were compared and verified using the HDM-NET and SVM, HDM-NET and MLP, HDM-NET and RF, HDM-NET and KNN, MobileNet and SVM, MobileNet and MLP, MobileNet and RF, and MobileNet and KNN classification process. Compared to performance and accuracy of the HDM-NET+SVM architecture.
For training, we used the Adam optimizer with a batch size of eight images, and a learning rate of 0.001 [27]. Because the image sizes collected in the original dataset are all 640 × 480, the optimizer will be slow during training and often lack video memory. Thus, the image is uniformly reduced to a pixel resolution of 224 × 224 and filled with zeros to maintain the original aspect ratio. A minimal data enhancement process was applied in the form of sample image rotation and rescaling. Apart from the standard pixel normalization of between zero and 1, no further preprocessing steps were applied to the input image. For experimental implementation, we used the Keras API in the Python3.6 TensorFlow2.1 framework. The training was conducted on an NVIDIA GeForce GTX 1660ti GPU equipped with 1536 CUDA cores, 6 GB of RAM, and a base clock speed of 1455 MHz. The comparative classification performance results shown in (Figure 6) show that, after testing on the same dataset DHI, the accuracy of HDM-NET is increased by 2.1% compared with the traditional CNN network. In particular, using our HDM-NET for feature extraction, combined with SVM, RF, MLP, and other methods for hair feature extraction and classification, the accuracy of the combined classification methods, such as Mobilenetv1, is generally higher. In many of the experiments, with HDM-NET combined with SVM, the final classification accuracy rate was the highest, reaching up to 94.8%.
Herein, we only used MobileNet and HDM-NET to complete the experiment. In addition, we did not use other CNN models for the experiments because such models (Table 2), such as VGG16 and GoogleNet, despite obtaining a high accuracy, have excessively large parameters for operation on mobile devices. Moreover, a small microscope was applied. The proposed model was improved and optimized on the basis of MobilenetV1, and, thus, we only conducted experiments for MobileNetV1 and HDM-NET. The proposed model has the least number of parameters, and its accuracy is slightly higher than that of the original MobileNet model. Because the MobileNet model has an ultra-small network framework, the classification accuracy is lower than that of the general CNN model. Our improved model based on MobileNet (HDM-NET) is far lower in accuracy than large-parameter models such as VGG16 (Table 3); however, compared with small networks such as MobileNet, our model has better classification results. Moreover, the parameters of the proposed model are small, and, thus, it is easier to directly complete the classification of hair damage images on a mobile microscope. We also used the HDM-NET and other models to complete the experiment (Table 4). We also tested and drew a conclusion of how long it took when one network was trained and when it had finished detecting certain images (Table 5).
In this study, the structures of HDM-NET and SVM were used to analyze and classify the hair images collected under SEM. The HDM-NET architecture is proposed and a new SEM image dataset on hair was constructed separately to solve the limitations of other models in diagnosing hair images, including the following:
1. General CNN model training requires a large number of datasets, and the datasets we can use now are extremely small. 2. Using a larger model with too many parameters leads to poor training effects. 3. Although the pre-trained model obtained on ImageNet [33,34] can be used for migration learning to alleviate the results of a poor training effect owing to the lack of a dataset, in our survey, the ImageNet contains few SEM images of hair, and, thus, the effect is poor. Therefore, we can only choose a CNN model that has smaller parameters for improvement.

4. Conclusions

In this study, we proposed a fast and easy method for diagnosing the health of hair and proposed a lightweight deep learning framework. We completed the task of identifying and classifying our dataset extremely well. Although the final classification result is lower than that of the general CNN model, the accuracy is better than that of other lightweight CNN models when compared with models similar to our proposed lightweight CNN model (HDM-NET), e.g., SqueezeNet. In addition, the parameters are less than the original model. Owing to the compactness of the model, in the future, it can be easily transplanted to mobile phones or portable mobile devices [35] for conducting a hair diagnosis at any time. It will also promote other services such as hair care in the beauty industry. We have also proposed a new type of dataset on hair health, which we are still compiling. Deep learning provides convenience for fast hair health image diagnosis and classification tasks.
In general, this method is faster and more convenient than traditional physical and chemical diagnostic methods for hair health. At the same time, the high accuracy of the deep learning method shows the significant potential of deep learning in the diagnosis of hair health, and it can also be applied to micro or digital microscopes [36] (for example: SVBONY SM401 Microscope Wireless Digital Microscope 50×–1000×) connected to mobile phones in the future. Our experiments also proved that the attribute selection step in the feature selection process is beneficial for improving the accuracy of recognition and classification of microscope images.

Author Contributions

Conceptualization, Y.I.C. and L.Z.; methodology, L.Z. and Q.M.; software, L.Z.; validation, L.Z., Q.M., Y.I.C.; resources, Y.I.C.; data curation, L.Z. and Q.M.; writing—original draft preparation, L.Z.; writing—review and editing, Y.I.C., Q.M.; visualization, L.Z. and Q.M.; supervision, L.Z. and Y.I.C.; formal analysis, L.Z. and Q.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Ministry of Science and ICT (MSIT), Korea, under the Information Technology Research Center (ITRC) support program (IITP-2021-2017-0-01630) supervised by the Institute for Information & Communications Technology Promotion (IITP).This research was supported by the MSIT, Korea, under the ITRC support program (IITP-2021-2017-0-01630) supervised by the IITP, 2018R1D1A1A09084151 by NRF, and supported by TIPA (G21S312068101).All individuals included in this section have consented to the acknowledgement.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data are not publicly available due to privacy. The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhao, T.; Chen, X.; Chen, Q. Heat current method-based modeling and optimization of the single effect lithium bromide absorption chiller. Appl. Therm. Eng. 2020, 175, 115345. [Google Scholar] [CrossRef]
  2. Zhang, H.; Berg, A.C.; Maire, M.; Malik, J. SVM-KNN: Discriminative nearest neighbor classification for visual category recognition. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; Volume 2, pp. 2126–2136. [Google Scholar]
  3. Yang, R.; Buenfeld, N.R. Binary segmentation of aggregate in SEM image analysis of concrete. Cem. Concr. Res. 2001, 31, 437–441. [Google Scholar] [CrossRef]
  4. Kalchbrenner, N.; Grefenstette, E.; Blunsom, P. A convolutional neural network for modelling sentences. arXiv 2014, arXiv:1404.2188. [Google Scholar]
  5. Li, X.; Li, J.; Li, B.; Yang, B. High-fidelity reversible data hiding scheme based on pixel-value-ordering and prediction-error expansion. Signal Process. 2013, 93, 198–205. [Google Scholar] [CrossRef]
  6. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  7. Noble, W.S. What is a support vector machine? Nat. Biotechnol. 2006, 24, 1565–1567. [Google Scholar] [CrossRef]
  8. Gardner, M.W.; Dorling, S.R. Artificial neural networks (the multilayer perceptron)—A review of applications in the atmospheric sciences. Atmos. Environ. 1998, 32, 2627–2636. [Google Scholar] [CrossRef]
  9. Pal, M. Random forest classifier for remote sensing classification. Int. J. Remote Sens. 2005, 26, 217–222. [Google Scholar] [CrossRef]
  10. Keller, J.M.; Gray, M.R.; Givens, J.A. A fuzzy k-nearest neighbor algorithm. IEEE Trans. Syst. Man Cybern. 1985, SMC-15, 580–585. [Google Scholar] [CrossRef]
  11. Guyon, I.; Elisseeff, A. An Introduction to Feature Extraction; Feature extraction; Springer: Berlin/Heidelberg, Germany, 2006; pp. 1–25. [Google Scholar]
  12. Karegowda, A.G.; Manjunath, A.S.; Jayaram, M.A. Comparative study of attribute selection using gain ratio and correlation based feature selection. Int. J. Inf. Technol. Knowl. Manag. 2010, 2, 271–277. [Google Scholar]
  13. Quinlan, J.R. Induction of decision trees. Mach. Learn. 1986, 1, 81–106. [Google Scholar] [CrossRef] [Green Version]
  14. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  15. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  16. Hua, B.S.; Tran, M.K.; Yeung, S.K. Pointwise convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 984–993. [Google Scholar]
  17. Santurkar, S.; Tsipras, D.; Ilyas, A.; Mądry, A. How does batch normalization help optimization. arXiv 2018, arXiv:1805.11604. [Google Scholar]
  18. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, Lille, France, 7–9 July 2015; pp. 448–456. [Google Scholar]
  19. Hara, K.; Saito, D.; Shouno, H. Analysis of function of rectified linear unit used in deep learning. In Proceedings of the 2015 International Joint Conference on Neural Networks (IJCNN), Killarney, Ireland, 12–17 July 2015; pp. 1–8. [Google Scholar]
  20. Xu, B.; Wang, N.; Chen, T.; Li, M. Empirical evaluation of rectified activations in convolutional network. arXiv 2015, arXiv:1505.00853. [Google Scholar]
  21. Zhang, S.; Zhang, S.; Zhang, C.; Wang, X.; Shi, Y. Cucumber leaf disease identification with global pooling dilated convolutional neural network. Comput. Electron. Agric. 2019, 162, 422–430. [Google Scholar] [CrossRef]
  22. Lin, Y.; Lv, F.; Zhu, S.; Yang, M.; Cour, T.; Yu, K.; Cao, L.; Huang, T. Large-scale image classification: Fast feature extraction and svm training. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 1689–1696. [Google Scholar]
  23. Li, X.; Wang, L.; Sung, E. Multilabel SVM active learning for image classification. In Proceedings of the 2004 International Conference on Image Processing, 2004 ICIP’04, Singapore, 24–27 October 2004; Volume 4, pp. 2207–2210. [Google Scholar]
  24. Sam, S.M.; Kamardin, K.; Sjarif, N.N.A.; Mohamed, N. Offline signature verification using deep learning convolutional neural network (CNN) architectures GoogLeNet Inception-v1 and Inception-v3. Procedia Comput. Sci. 2019, 161, 475–483. [Google Scholar]
  25. Billings, S.A.; Zheng, G.L. Radial basis function network configuration using genetic algorithms. Neural Netw. 1995, 8, 877–890. [Google Scholar] [CrossRef]
  26. Murphy, K.P. Naive bayes classifiers. Univ. Br. Columbia 2006, 18, 1–8. [Google Scholar]
  27. Zeiler, M.D. Adadelta: An adaptive learning rate method. arXiv 2012, arXiv:1212.5701. [Google Scholar]
  28. Rezende, E.; Ruppert, G.; Carvalho, T.; Theophilo, A.; Ramos, F.; de Geus, P. Malicious software classification using VGG16 deep neural network’s bottleneck features. In Information Technology-New Generations; Springer: Cham, Switzerland, 2018; pp. 51–59. [Google Scholar]
  29. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  30. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arxiv arXiv:1602.07360, 2016.
  31. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6848–6856. [Google Scholar]
  32. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  33. He, K.; Girshick, R.; Dollár, P. Rethinking imagenet pre-training. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 4918–4927. [Google Scholar]
  34. Venkatesh; Nagaraju, Y.; Siddhanth, U.H.; Stalin, S. Fine-tuned MobileNet Classifier for Classification of Strawberry and Cherry Fruit Types. In Proceedings of the 2021 International Conference on Computer Communication and Informatics (ICCCI), Coimbatore, India, 27–29 January 2021; pp. 1–18. [Google Scholar] [CrossRef]
  35. Gragnaniello, D.; Sansone, C.; Verdoliva, L. Iris liveness detection for mobile devices based on local descriptors. Pattern Recognit. Lett. 2015, 57, 81–87. [Google Scholar] [CrossRef]
  36. Dickerson, J.; Kubasko, D. Digital microscopes: Enhancing collaboration and engagement in science classrooms with information technologies. Contemp. Issues Technol. Teach. Educ. 2007, 7, 279–292. [Google Scholar]
Figure 1. Image of degree of hair damage under SEM microscope ×800; from left to right: damage, high damage, and weak damage. The bottom image shows damaged cuticle layers.
Figure 1. Image of degree of hair damage under SEM microscope ×800; from left to right: damage, high damage, and weak damage. The bottom image shows damaged cuticle layers.
Diagnostics 11 01831 g001
Figure 2. The processing of hair damage data collection, (a) is the many style of hair, (b) is the samples divided into three parts, (c) is the data collected on scanning electron microscope, (d) is the SEM image of hair, (e) is the many SEM hair images made into a DHI dataset.
Figure 2. The processing of hair damage data collection, (a) is the many style of hair, (b) is the samples divided into three parts, (c) is the data collected on scanning electron microscope, (d) is the SEM image of hair, (e) is the many SEM hair images made into a DHI dataset.
Diagnostics 11 01831 g002
Figure 3. Flowchart of hair damage diagnosis.
Figure 3. Flowchart of hair damage diagnosis.
Diagnostics 11 01831 g003
Figure 4. Basic structure convolution layers of MobileNet.
Figure 4. Basic structure convolution layers of MobileNet.
Diagnostics 11 01831 g004
Figure 5. Proposed model architecture: Hair-Diagnosis Mobilenet (HDM-NET + SVM) and process of training the CNN model (HDM-NET), convolution (Conv), depthwise (DW), and batch normalization (BN). The classifier can also be changed for RF, MLP, and KNN.
Figure 5. Proposed model architecture: Hair-Diagnosis Mobilenet (HDM-NET + SVM) and process of training the CNN model (HDM-NET), convolution (Conv), depthwise (DW), and batch normalization (BN). The classifier can also be changed for RF, MLP, and KNN.
Diagnostics 11 01831 g005
Figure 6. Accuracy of MobileNet (a) and HDM-NET (b) and the choice of attributes used for MobileNet (c) and HDM-NET (d).
Figure 6. Accuracy of MobileNet (a) and HDM-NET (b) and the choice of attributes used for MobileNet (c) and HDM-NET (d).
Diagnostics 11 01831 g006
Table 1. Parameters distributed in HDM-NET with classifiers (SVM), dw means depthwise.
Table 1. Parameters distributed in HDM-NET with classifiers (SVM), dw means depthwise.
Layer TypeStrideKernelFeature Map
Conv. Layer23 × 33 × 32
Depthwise Conv.13 × 332 dw
Conv. Layer11 × 132 × 64
Depthwise Conv.23 × 364 dw
Conv. Layer11 × 164 × 128
Depthwise Conv.13 × 3128 dw
Conv. Layer11 × 1128 × 128
Depthwise Conv.23 × 3128 dw
Conv. Layer11 × 1128 × 256
Depthwise Conv.13 × 3256 dw
Conv. Layer11 × 1256 × 256
Depthwise Conv.23 × 3256 dw
Conv. Layer11 × 1256 × 512
2×Depthwise13 × 3512 dw
2×Conv. Layer11 × 1512 × 512
Depthwise Conv.23 × 3512 dw
Conv. Layer11 × 1512 × 1024
Depthwise Conv.13 × 31024 dw
Conv. Layer11 × 11024 × 1024
Global AvgPool17×71024 × 1024
BN------1024 × 1024
ReLU------1024 × 1024
Classifier (SVM)------3
Table 2. Results of MobileNetV1 and HDM-NET on DHI datasets.
Table 2. Results of MobileNetV1 and HDM-NET on DHI datasets.
ModelClassifiersAccuracy (%)
MobileNetV1(Softmax)85.3
MobileNetV1SVM92.6
MobileNetV1RF91.9
MobileNetV1MLP90.5
MobileNetV1KNN89.8
HDM-NET(ours)(Softmax)87.4
HDM-NET(ours)SVM94.8
HDM-NET(ours)RF91.4
HDM-NET(ours)MLP92.3
HDM-NET(ours)KNN90.3
Table 3. Parameters and results of general CNN models on DHI datasets.
Table 3. Parameters and results of general CNN models on DHI datasets.
ModelDHI Dataset Accuracy (%)Million Parameters
VGG1699.1138 M
GoogleNet97.26.8 M
ResNet5098.625.5 M
SqueezeNetV183.41.2 M
ShufflenetV184.82.3 M
MobileNetV185.34.3 M
HDM-NET(ours)87.41.4 M
Table 4. Results of HDM-net and other lightweight networks on DHI datasets.
Table 4. Results of HDM-net and other lightweight networks on DHI datasets.
ModelClassifiersAccuracy (%)
SqueezeNetV1(Softmax)88.4
SqueezeNetV1SVM93.7
SqueezeNetV1RF87.3
SqueezeNetV1MLP90.2
SqueezeNetV1KNN89.8
ShufflenetV1(Softmax)86.5
ShufflenetV1SVM92.3
ShufflenetV1RF89.4
ShufflenetV1MLP90.6
ShufflenetV1KNN91.4
HDM-NET(ours)SVM94.8
Table 5. The training time and detection time for HDM-net and other networks (VGG16 [28], GoogleNet [29], ResNet50 [30]) and lightweight CNN models (SqueezeNetV1 [31], ShufflenetV1 [32], MobilenetV1).
Table 5. The training time and detection time for HDM-net and other networks (VGG16 [28], GoogleNet [29], ResNet50 [30]) and lightweight CNN models (SqueezeNetV1 [31], ShufflenetV1 [32], MobilenetV1).
ModelTraining Time (hours)Detection Time (s)
SqueezeNetV12.60.33
ShufflenetV12.60.52
MobileNetV12.70.55
VGG165.60.68
GoogleNet3.10.46
ResNet503.40.51
HDM-NET(ours)2.40.22
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, L.; Man, Q.; Cho, Y.I. Deep-Learning-Based Hair Damage Diagnosis Method Applying Scanning Electron Microscopy Images. Diagnostics 2021, 11, 1831. https://doi.org/10.3390/diagnostics11101831

AMA Style

Zhang L, Man Q, Cho YI. Deep-Learning-Based Hair Damage Diagnosis Method Applying Scanning Electron Microscopy Images. Diagnostics. 2021; 11(10):1831. https://doi.org/10.3390/diagnostics11101831

Chicago/Turabian Style

Zhang, Lintong, Qiaoyue Man, and Young Im Cho. 2021. "Deep-Learning-Based Hair Damage Diagnosis Method Applying Scanning Electron Microscopy Images" Diagnostics 11, no. 10: 1831. https://doi.org/10.3390/diagnostics11101831

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop