Next Article in Journal
Deep Learning Based Automatic Ankle Tenosynovitis Quantification from MRI in Patients with Psoriatic Arthritis: A Feasibility Study
Previous Article in Journal
Stacked Ensemble Learning for Classification of Parkinson’s Disease Using Telemonitoring Vocal Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Landmark Detection and Lip Thickness Classification Using a Convolutional Neural Network in Lateral Cephalometric Radiographs

1
Department of Orthodontics, Nanjing Stomatological Hospital, Affiliated Hospital of Medical School, Research Institute of Stomatology, Nanjing University, Nanjing 210008, China
2
School of Internet of Things, Nanjing University of Posts and Telecommunications, Nanjing 210003, China
3
Discipline of Orthodontics, Department of Oral Science, Faculty of Dentistry, University of Otago, Dunedin 9016, New Zealand
*
Authors to whom correspondence should be addressed.
Diagnostics 2025, 15(12), 1468; https://doi.org/10.3390/diagnostics15121468
Submission received: 24 April 2025 / Revised: 23 May 2025 / Accepted: 5 June 2025 / Published: 9 June 2025
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

Objective: The objective of this study is to develop a convolutional neural network (CNN) for the automatic detection of soft and hard tissue landmarks and the classification of lip thickness on lateral cephalometric radiographs. Methods: A dataset of 1019 pre-orthodontic lateral cephalograms from patients with diverse malocclusions was utilized. A CNN-based model was trained to automatically detect 22 cephalometric landmarks. Upper and lower lip thicknesses were measured using some of these landmarks, and a pre-trained decision tree model was employed to classify lip thickness into the thin, normal, and thick categories. Results: The mean radial error (MRE) for detecting 22 landmarks was 0.97 ± 0.52 mm. Successful detection rates (SDRs) at threshold distances of 1.00, 1.50, 2.00, 2.50, 3.00, and 4.00 mm were 72.26%, 89.59%, 95.41%, 97.66%, 98.98%, and 99.47%, respectively. For nine soft tissue landmarks, the MRE was 1.08 ± 0.87 mm. Lip thickness classification accuracy was 0.91 ± 0.04 (upper lip) and 0.90 ± 0.04 (lower lip) in females and 0.92 ± 0.03 (upper lip) and 0.88 ± 0.05 (lower lip) in males. The area under the curve (AUC) values for lip thickness were ≥0.97 for all gender–lip combinations. Conclusions: The CNN-based landmark detection model demonstrated high precision, enabling reliable automatic classification of lip thickness using cephalometric radiographs.

1. Introduction

As a critical anatomical structure in the lower third of the face, the morphological characteristics of the lips play a decisive role in overall facial harmony. In facial esthetic evaluation, the proportional relationship between lip thickness and other facial structures is an essential quantitative indicator for assessing facial esthetics [1]. Changes in the upper lip following tooth movement have been widely studied. However, the extent of this change varies significantly depending on factors such as gender, age, ethnicity, soft tissue thickness, and lip strain [2]. Orthodontic treatment alters tooth position, which can in turn affect lip shape, thickness, length, and strain. The upper lip length and thickness have been found to increase after incisor retraction, with smaller increases observed in patients with thicker lips than those with thinner lips [3]. Thus, measuring lip thickness is essential for developing a customized orthodontic treatment plan for patient-centered healthcare.
In the literature, the thickness of the upper lip base is conventionally defined as the distance from the subnasal point to the subspinale point [4]. However, notable inconsistencies exist in the measurement methods for upper lip vermilion thickness [5,6]. Consequently, no standardized measurement protocol has been established for assessing the thickness of the upper lip, lower lip, or chin [7].
In orthodontics, cephalometric analysis plays a crucial role in clinical diagnosis, treatment planning, and outcome evaluation [8]. Lip thickness can be measured through soft and hard tissue landmark identification on lateral cephalometric radiographs [4,5]. However, landmark identification, measurement, and lip thickness calculation are time-consuming and labor-intensive for clinicians.
Artificial intelligence (AI), a core field in computer science, aims to develop systems and algorithms simulating human intelligence [8]. Deep learning technologies based on artificial neural networks exhibit powerful feature extraction and learning capabilities, with wide applications in image recognition, natural language processing, and decision optimization. As a core deep learning model, convolutional neural networks (CNNs) are designed to process grid-structured data (e.g., images), efficiently extracting spatial features via local perception, weight sharing, and pooling operations [9]. Currently, AI has been successfully applied in various dental domains, including caries detection, periapical lesion diagnosis, alveolar bone resorption assessment, cyst/tumor classification, cephalometric analysis, and osteoporosis screening [9,10,11,12]. In modern orthodontics, semi-automatic computer-assisted cephalometric measurement is widely used, where clinicians manually identify landmarks before software-based measurements [13]. However, this method remains highly dependent on clinicians’ expertise; thus, recent research has focused on developing fully automated AI-based cephalometric systems to enhance clinical efficiency and reduce subjective bias. Existing studies primarily target automated hard tissue landmark localization, but soft tissue landmark identification accuracy remains suboptimal due to challenges like low imaging clarity/contrast and anatomical variability [12].
Therefore, the first aim of this study was to automatically identify landmarks in lateral cephalometric radiographs, with a particular focus on the precise localization of soft tissue landmarks. The second aim was to measure and summarize the distribution characteristics of lip thickness in the East China population and establish a classification framework for lip thickness. The third objective was to achieve automatic lip thickness classification using a decision tree model based on the automatically identified landmarks, thus providing an efficient and objective assessment method for clinical applications.

2. Materials and Methods

This study was approved by the ethics committee of Nanjing Stomatological Hospital, Medical School of Nanjing University (approval number: NJSH-2024NL-063). The experimental design flowchart is shown in Figure 1.

2.1. Data Collection and Preprocessing

Data were collected from the radiology department of Nanjing Stomatological Hospital, Nanjing University, China. A total of 1019 subjects (318 males and 701 females) who sought orthodontic treatment at Nanjing University Stomatological Hospital between 2021 and 2024 were included in this study. All included subjects had lateral cephalograms as part of their medical records. Informed consent was obtained from all subjects. Lateral cephalometric radiographs were taken using a BDS221 Digital Panoramic Cephalometric System (Tuusula, Finland) from subjects with various malocclusions.
Inclusion criteria: (1) Patients with permanent dentition and Class I, II, or III malocclusions (regardless of severity) and (2) aged 18–45 years old.
Exclusion criteria: (1) History of craniofacial deformities, orthodontic treatment, orthognathic surgery, or facial plastic surgery (e.g., hyaluronic acid lip augmentation). (2) Cephalometric images obtained in non-intercuspal occlusion.
Imaging parameters included a tube voltage of 77 kV, a tube current of 16 mA, an exposure time of 8 s, and a distance of 140 cm from the center of the X-ray tube to the sensor. Subjects were instructed to (1) maintain a natural head position, (2) keep their head stable, (3) relax and breathe calmly, and (4) avoid chewing or swallowing during image acquisition. All images were saved in .jpg format.
Annotation personnel: A single orthodontist with 15 years of clinical experience annotated 200 lateral cephalometric films to establish a gold standard and standardized annotation criteria for two additional orthodontists. The two orthodontists subsequently identified landmarks on the remaining 819 lateral cephalometric images. The two orthodontists recorded the X and Y coordinates of each landmark on all radiographs. Intraclass correlation coefficients (ICCs) were calculated separately for each landmark’s X and Y coordinates.
The formula is as follows:
ICC = I n t e r d o c t o r   v a r i a n c e I n t e r d o c t o r   v a r i a n c e + e r r o r   v a r i a n c e
Annotation: A total of 24 landmarks on lateral cephalometric radiographs were identified using the Labelme (version 5.0.1) software. These include nine soft tissue landmarks: Sn, UT, UL, LT, LL, Bs, Pos, Gns, and Mes; thirteen bone tissue landmarks: S, N, A, Spr, UJ, UI, LI, LJ, Id, B, Po, Gn, and Me (Figure 2a, Table 1); and two ruler points. Upper lip thickness was measured as the average length of three line segments: A-Sn, Spr-UT, and UT-UJ [4,5,6]. Due to the lack of consensus in the literature on lower lip thickness measurement methods, the current study defined lower lip thickness as the average length of three line segments: B-Bs, Id-LT, and LT-LJ (Figure 2b).
Image Processing: The collected lateral cephalometric radiographs were cropped to a uniform size of 800 pixels by 640 pixels, and personal patient information (e.g., ID number and name) was removed. Each pixel corresponded to 0.2 mm. During the training phase of the automatic cephalometric landmarking model, data augmentation operations, such as random contrast enhancement, random histogram equalization, and random half-body flipping, were employed.

2.2. The Automatic Cephalometric Landmark Detection Model

The network architecture of the automatic cephalometric landmark detection model follows the backbone–neck–head structure, as shown in Figure 3. HRNet [14] is selected as the backbone network to extract multi-scale features. After inputting the lateral cephalogram image, the backbone network extracts multi-scale feature maps, which will be fused in the neck later. The neck is designed based on SRPose [15], and the feature fusion module uses separable convolutional layers to fuse features. The neck module adopts a structure similar to the feature pyramid network (FPN) [16] to fuse multi-scale features, including two feature fusion paths from small-scale to large-scale feature maps (M5 to M2 and G5 to G2) and one feature fusion path from large-scale to small-scale feature maps (W2 to W5). The head utilizes SRPose and super-resolution technology to encode and upsample the fused features into high-resolution heatmaps. After obtaining the high-resolution heatmaps, the coordinates of the landmark points are determined by finding the peak positions in the heatmaps.
The Adam optimizer was used during model training with an initial learning rate of 0.0001. The learning rate was reduced by a factor of ten every 50 epochs. The batch size was set to 2, and training was conducted for 300 epochs. Various data augmentation techniques were applied to mitigate overfitting due to limited training data, including random contrast enhancement, histogram equalization, and half-body flipping. Each training sample underwent all augmentation steps in sequence, with randomized parameters for each operation (e.g., rotation angle and contrast intensity). These augmentations enhanced the diversity of the training data, improving model generalization.

2.3. Model Evaluation and Accuracy Assessment

After the automatic cephalometric landmark detection model is trained, the test datasets are input into the model, and the predicted heatmaps for each key point are obtained. The coordinates of landmarks are recovered by performing non-maximum suppression (NMS) over the predicted heatmaps [17]. The evaluation metrics used include the mean radial error (MRE, in millimeters; the smaller, the better) and the successful detection rates (SDR; the larger, the better) within radii of 1.00, 1.50, 2.00, 2.50, 3.00, and 4.00 mm. MRE is the average distance between predicted and ground-truth landmarks, and SDR is the percentage of predicted landmarks within a pre-defined range of ground-truth landmarks.
The formulas are as follows:
R = x i 2 + y i 2
MRE = i = 1 n R i
SD = i = 1 n ( R i M R E ) 2 N 1
SDR = N a c c N a l l
where x i and y i represent the absolute differences between the predicted and ground-truth coordinates of the landmark on the X-axis and Y-axis, respectively, and n is the total number of landmarks. N a c c and N a l l represent the number of correctly predicted landmarks and the total number of landmarks within the radius ranges (1.00, 1.50, 2.00, 2.50, 3.00, and 4.00 mm), respectively.

2.4. The Decision Tree Model

The classification method for lip thickness was based on previous studies [18]. The upper and lower lip thickness values of 1019 patients were measured and then sorted in ascending order for each gender. The first 25% were classified as the thin-lip group, the middle 50% (25–75%) as the normal-lip group, and the last 25% as the thick-lip group. The overall performance of the decision tree model was evaluated using receiver operating characteristic (ROC) curves, the area under the curve (AUC), accuracy, sensitivity, specificity, precision, and the F1 score. The metrics above were calculated using the true positive (TP), false positive (FP), true negative (TN), and false negative (FN).
The calculation formula is as follows:
Accuracy = T P + T N T P + T N + F P + F N
Recall = Sensitivity = T P T P + F N
Specificity = T N T N + F P
Precision = T P T P + F P
F 1   score = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l

3. Results

3.1. Accuracy of Automatic Landmark Localization

The two orthodontists’ intraclass correlation coefficient (ICC) for landmark locations exceeded 90%. The automatic cephalometric landmark detection model demonstrated a rapid automatic landmarking speed, taking an average of 0.13 s to recognize and locate 22 landmarks in each lateral cephalometric X-ray image. Excluding ruler point 1 and ruler point 2 (which were not included in the training), the remaining 22 landmarks were successfully identified in each image, resulting in a recognition success rate of 100%. The mean radial error (MRE) for the 22 cephalometric landmarks was 0.97 ± 0.52 mm. The successful detection rates (SDRs) within various error margins were as follows: 72.26% within 1.00 mm, 89.59% within 1.50 mm, 95.41% within 2.00 mm, 97.66% within 2.50 mm, 98.98% within 3.00 mm, and 99.47% within 4.00 mm (Table 2). Using the manually annotated landmark coordinates as the reference, we obtained the coordinate distribution of AI-predicted landmarks and plotted an ellipse representing the 95% confidence interval, as shown in Figure 4. Additionally, we calculated the discrepancies between the machine-annotated and manually annotated landmarks and analyzed the relative frequency distribution of these discrepancies. The Gaussian-fitted curve provided a more precise visualization of the error concentration and distribution characteristics (Appendix A Figure A1). Visualized landmark heatmaps output by the automated landmark detection model is shown in Appendix A Figure A2.
The loss curves of the landmark detection model are shown in Appendix A Figure A3. During the first 10 epochs, both the training and test losses decreased rapidly. By the 50th epoch, the losses approached zero and stabilized, indicating that the performance of the landmark detection model progressively improved, approaching human-level precision automatically.

3.2. Lip Thickness Classification Performance

The classification standards for the lip thickness of the upper and lower lips based on different genders were established (Table 3). The measured lip thickness values were categorized according to the preliminary lip thickness classification criteria. Confusion matrices for the three categories, which comprise both the AI-predicted lip thickness classification and the manually annotated lip thickness classification, are shown in Figure 5.
The performance of the upper and lower lip thickness classification for males and females is as follows: excellent accuracy (0.88–0.93), sensitivity and recall (0.81–0.87), specificity (0.90–0.94), precision (0.86–0.90), and F1 score (0.82–0.88) (Table 4). The AUC values were calculated for each class, and the ROC curves were generated, as shown in Table 4 and Figure 6. The AUC values were as follows: For upper lip thickness in females and males, they were (0.97 ± 0.02) and (0.98 ± 0.01), respectively. For lower lip thickness in females and males, they were (0.98 ± 0.01) and (0.98 ± 0.02), respectively.

4. Discussion

This study focuses on automatic landmark detection in pre-treatment patients’ lateral cephalograms and the classification of lip thickness. It provides a data-driven foundation for personalized orthodontic treatment planning. By assessing lip thickness before treatment, the model can predict how much a patient’s lateral profile may change following extraction orthodontic treatment [18]. For example, in patients with thick lips, the lateral profile is less impacted by retraction. As a result, greater anterior teeth retraction can be performed to enhance facial harmony. In contrast, patients with thin lips are more likely to experience significant changes in their lateral profile, so minimal retraction is recommended to avoid lip area collapse and premature facial aging [19]. This predictive capability helps clinicians select treatment strategies that balance functional correction with esthetic outcomes.
Multiple studies have explored the application of deep learning using cephalometric analysis. For example, Kang et al. reported a mean radial error (MRE) of 1.45 ± 0.92 mm for 18 anatomical landmarks [20], while Yao et al. noted the highest localization error at the Pos landmark (2.03 ± 5.95 mm) [21]. Hong et al. used a deep learning network based on reinforcement learning and found that the successful detection rate of landmarks within 2 mm was 67.33% [22]. Using an automatic program, Bao et al.’s study [23] showed that the mean radial error (MRE) for 19 landmarks was 2.07 ± 1.35 mm. The average success detection rates (SDRs) within 1 mm, 2 mm, 2.5 mm, 3 mm, and 4 mm were 18.82%, 58.58%, 71.70%, 82.04%, and 91.39%, respectively. By comparison, our model achieved an MRE of 1.08 ± 0.87 mm specifically for nine soft tissue landmarks, indicating improved precision, particularly in regions typically associated with higher localization error. This comparison highlights the performance advantage of our model and supports its effectiveness in enhancing soft tissue landmark detection in lateral cephalometric radiographs. Furthermore, based on the lip thickness data of 1019 patients with various types of malocclusion, this study preliminarily summarized the characteristics of lip thickness in the East China population and proposed a lip thickness classification framework. Additionally, the thicknesses of the upper and lower lips were measured, and automated lip thickness classification was achieved using a decision tree model based on the automatically detected landmarks.
Various factors can influence the changes in the soft tissue of the lips. For example, tooth extraction and non-extraction approaches can affect lip thickness at the lip base and vermilion; extraction treatment has been shown to cause more significant increases [24]. Different ethnicities may also respond differently to treatment, influenced by lip tension and dental crowding [24]. Previous studies have mainly focused on the ratio of incisor retraction to soft tissue changes [25,26,27,28,29], neglecting individual differences in lip thickness, a crucial factor influencing facial protrusion and esthetic outcomes [30]. Facial soft tissue thickness has been found to differ between sexes, with males generally having thicker soft tissue than females [29]. When predicting soft tissue changes, factors such as lip thickness, lip strain, mandibular divergence, incisor position, gender, and ethnicity should be considered.
Over the past few decades, a large number of computer vision and machine learning techniques have been applied to automated medical image landmark detection. In the early stage, researchers defined landmarks as corner points, endpoints, local extreme points, etc., and designed image filters, operators, and template-matching algorithms to address this problem [31]. However, these models generally exhibited poor generalization performance. With the development of deep-learning technology, CNN-based methods have demonstrated superior performance in landmark detection tasks. The mainstream approaches can be classified into direct landmark location regression and landmark heatmap estimation [32].
Direct landmark location regression uses CNN to extract image features and then directly maps these image features to landmark coordinates via the fully connected network (FCN). However, soft tissue on lateral cephalometric radiographs is usually complex, making it challenging to map CNN-extracted features directly to landmark coordinates. In contrast, landmark heatmap estimation is more effective for soft tissue landmark detection [33]. This approach also employs CNN for feature extraction. The key difference lies in that it maps features to heatmaps and finally determines landmark coordinates by identifying the highest heat values within these heatmaps.
Among soft tissue landmarks, the Pos point had the most significant positioning error. It represents the most anterior point of the soft tissue in the chin. This point is easily located in a normally developed chin due to its prominence. However, the soft tissue in the chin slopes without apparent protrusion in patients with skeletal Class II malocclusion and a vertical growth pattern, making Pos point localization difficult. The highest error was observed at the soft tissue of the menton, possibly because of the thin chin soft tissue and indistinct edge delineation. The detection challenges for these soft tissue landmarks include low contrast in soft tissue radiographs, limited imaging clarity, and high anatomical variability. To address these issues, it is essential to develop a deep learning model that possesses enhanced resistance to interference and adaptability across different scenarios. This model should be capable of effectively handling anomalies and challenging samples, ultimately improving overall performance.
In this study, for hard tissue landmark localization, the mean radial error (MRE) of point B was higher than that of other hard tissue landmarks. Anatomically, the B point represents the deepest concavity of the mandible; it is easily identifiable in individuals with normal chin development due to the distinct curvature, but it is uncertain in cases of retrognathia or insufficient chin growth. The error at point A is attributed to overlapping anatomical structures, obscuring the deepest point of the maxilla in imaging. The nasopalatine suture (N point) may not always appear in radiographic images. Overlapping bilateral structures likely cause the error in Po point localization. This finding aligns with previous studies [34] indicating that landmarks such as A, ANS, B, Po, and Pos are relatively difficult to identify.
There are also limitations in our study. Although this study included 1019 patients with various malocclusions, the relatively limited geographic distribution of the sample may restrict the generalizability of the findings to the overall lip thickness characteristics of the Chinese population. Future research should explore multi-center validation to enhance the robustness of the findings. For example, the proposed model could automatically validate cephalometric point localization on lateral cephalograms from diverse populations across multiple hospital settings. Conducting such multi-center studies would substantially improve the reliability, generalizability, and scientific impact of the results.
As the first systematic investigation to summarize and explore lip thickness distribution in the East China population, this study represents a significant contribution with innovative value. In the future, the proposed system can be integrated either as a standalone tool or within widely used clinical cephalometric analysis software, e.g., Dolphin. Automatic landmark detection can be performed on radiographs from diverse patient populations across multiple clinical sites, followed by manual verification and correction by trained clinicians. To ensure accuracy and improve model generalizability, the system is designed to log manual adjustments, which can be used to retrain and refine the model iteratively. We also outline preliminary user training protocols and propose routine quality control procedures, including inter- and intra-operator consistency checks, to support safe and effective clinical deployment.

5. Conclusions

The automatic cephalometric landmark detection model with a backbone–neck–head network structure demonstrated high accuracy in detecting cephalometric landmarks, especially for soft tissue landmarks. By using the automatically detected landmarks, the thicknesses of the upper and lower lips can be measured. When these measurements are input into the decision tree classification model, automatic lip classification can be accomplished. This novel automated landmark detection and lip thickness classification method holds great potential for application in orthodontic clinical practice.

Author Contributions

Data curation, J.R.; funding acquisition, H.L. and J.L.; methodology, H.Z.; software, Z.H.; writing—original draft, M.H.; writing—review and editing, L.M. All authors have read and agreed to the published version of the manuscript.

Funding

Cultivation Program for Junior Talents of Nanjing Stomatological School, Medical School of Nanjing University, China (grant number: 0222E305), and High-Level Hospital Construction Project of Nanjing Stomatological Hospital, Affiliated Hospital of Medical School, Institute of Stomatology, Nanjing University (grant number: 0224C044).

Institutional Review Board Statement

All experiments were performed in accordance with the Declaration of Helsinki. Ethical approval for this retrospective study was granted by the ethics committee at the Nanjing Stomatological Hospital, Affiliated Hospital of Medical School, Research Institute of Stomatology, Nanjing University, Nanjing, China [approval No. NJSH-2024NL-063, 05/08/24].

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Informed consent for the anonymous use of patient data was obtained from all individual participants included in the study.

Data Availability Statement

The datasets supporting the conclusions of this article are presented within the article. The raw data can be obtained from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Figure A1. (AF) sequentially depict the cumulative frequency distribution and Gaussian fitting curves for the soft tissue landmarks: Pos, Gns, Mes, UT, LT, and Sn.
Figure A1. (AF) sequentially depict the cumulative frequency distribution and Gaussian fitting curves for the soft tissue landmarks: Pos, Gns, Mes, UT, LT, and Sn.
Diagnostics 15 01468 g0a1
Figure A2. Visualized landmark heatmaps output by the automated landmark detection model.
Figure A2. Visualized landmark heatmaps output by the automated landmark detection model.
Diagnostics 15 01468 g0a2
Figure A3. The loss curves of landmark detection models in test datasets and training datasets.
Figure A3. The loss curves of landmark detection models in test datasets and training datasets.
Diagnostics 15 01468 g0a3

References

  1. Peck, H.; Peck, S. A concept of facial esthetics. Angle Orthod. 1970, 40, 284–318. [Google Scholar] [PubMed]
  2. Sodagar, A.; Borujeni, D.G.; Amini, G. Prediction of soft tissue profile changes following orthodontic retraction of incisors in Iranian girls. World J. Orthod. 2010, 11, 262–268. [Google Scholar]
  3. Ramos, A.L.; Sakima, M.T.; Pinto Ados, S.; Bowman, S.J. Upper lip changes correlated to maxillary incisor retraction—A metallic implant study. Angle Orthod. 2005, 75, 499–505. [Google Scholar] [PubMed]
  4. Kiełczykowski, M.; Perkowski, K.; Kamiński, K.; Zadurska, M.; Czochrowska, E. Evaluation of the soft tissue facial profile in different skeletal malocclusions in relation to age. BMC Oral. Health 2024, 24, 711. [Google Scholar] [CrossRef]
  5. Mariscal, J.F.; Ravelli, D.B.; Pinto, A.d.S.; Dominguez, G.C.; Aidar, L.A.d.A.; Jacob, H.B. Comparisons of soft tissue changes after Herbst appliance treatment in three different growth phases. J. World Fed. Orthod. 2024, 13, 78–85. [Google Scholar] [CrossRef]
  6. Khurshid, H.M.; Rizwan, S.; Ahmed, I.; Mehmood, H.T.; Nagani, N.I.; Tanvir, F. Comparison of upper and lower lip position, length and thickness in sagittal and vertical malocclusion. J. Pak. Med. Assoc. 2023, 73, 2423–2426. [Google Scholar] [CrossRef] [PubMed]
  7. Hamid, S.; Abuaffan, A.H. Facial soft tissue thickness in a sample of Sudanese adults with different occlusions. Forensic Sci. Int. 2016, 266, 209–214. [Google Scholar] [CrossRef]
  8. Schwendicke, F.; Singh, T.; Lee, J.-H.; Gaudin, R.; Chaurasia, A.; Wiegand, T.; Uribe, S.; Krois, J.; on behalf of the IADR e-Oral Health Network; The ITU WHO Focus Group AI for Health. Artificial intelligence in dental research: Checklist for authors, reviewers, readers. J. Dent. 2021, 107, 103610. [Google Scholar] [CrossRef]
  9. McBee, M.P.; Awan, O.A.; Colucci, A.T.; Ghobadi, C.W.; Kadom, N.; Kansagra, A.P.; Auffermann, W.F. Deep Learning in Radiology. Acad. Radiol. 2018, 25, 1472–1480. [Google Scholar] [CrossRef]
  10. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  11. Hosny, A.; Parmar, C.; Quackenbush, J.; Schwartz, L.H.; Aerts, H. Artificial intelligence in radiology. Nat. Rev. Cancer 2018, 18, 500–510. [Google Scholar] [CrossRef] [PubMed]
  12. Khanagar, S.B.; Al-Ehaideb, A.; Maganur, P.C.; Vishwanathaiah, S.; Patil, S.; Baeshen, H.A.; Sarode, S.C.; Bhandi, S. Developments, application, and performance of artificial intelligence in dentistry-A systematic review. J. Dent. Sci. 2021, 16, 508–522. [Google Scholar] [CrossRef] [PubMed]
  13. Wu, Z.; Hu, M.; Gong, X.; Wei, X. Advances in the Application of Machine Learning in Orthodontic Diagnosis and Treatment. Chin. J. Stomatol. 2021, 56, 1277–1281. [Google Scholar]
  14. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep high-resolution representation learning for human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5693–5703. [Google Scholar]
  15. Wang, H.; Liu, J.; Tang, J.; Wu, G. Lightweight Super-Resolution Head for Human Pose Estimation. arXiv 2023, arXiv:2307.16765. [Google Scholar]
  16. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  17. Bodla, N.; Singh, B.; Chellappa, R. Soft-NMS: Improving object detection with one line of code. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5561–5569. [Google Scholar]
  18. Guan, Y.-X.; Xin, L.; Tian, P.-F.; Kyung, H.-M.; Kwon, T.-G.; Bing, L.; Wu, X.-P. Effect of Soft Tissue Thickness on the Morphology of Lip in Orthodontic Treatment. Int. J. Morphol. 2019, 37, 1245–1251. [Google Scholar] [CrossRef]
  19. Albertini, P.; Barbara, L.; Albertini, E.; Willeit, P.; Lombardo, L. Soft-tissue profile changes in adult patients treated with premolar extractions. Am. J. Orthod. Dentofacial Orthop. 2024, 166, 171–178. [Google Scholar] [CrossRef]
  20. Kang, S.; Kim, I.; Kim, Y.; Kim, N.; Baek, S.; Sung, S. Accuracy and clinical validity of automated cephalometric analysis using convolutional neural networks. Orthod. Craniofac Res. 2024, 27, 64–77. [Google Scholar] [CrossRef]
  21. Yao, J.; Zeng, W.; He, T.; Zhou, S.; Zhang, Y.; Guo, J.; Tang, W. Automatic localization of cephalometric landmarks based on convolutional neural network. Am. J. Orthod. Dentofac. Orthop. 2022, 161, e250–e259. [Google Scholar] [CrossRef]
  22. Hong, W.B.; Kim, S.-M.B.; Choi, J.B.; Ahn, J.; Paeng, J.-Y.; Kim, H. Automated Cephalometric Landmark Detection Using Deep Reinforcement Learning. J. Craniofac. Surg. 2023, 34, 2336–2342. [Google Scholar] [CrossRef]
  23. Bao, H.; Zhang, K.; Yu, C.; Li, H.; Cao, D.; Shu, H.; Liu, L.; Yan, B. Evaluating the accuracy of automated cephalometric analysis based on artificial intelligence. BMC Oral. Health 2023, 23, 191. [Google Scholar] [CrossRef]
  24. Sadry, S.; Koru, B.E.; Kayalar, E. Analyzing the effects of tooth extraction on the lip in orthodontic treatment. J. Stomatol. Oral. Maxillofac. Surg. 2022, 123, e126–e132. [Google Scholar] [CrossRef] [PubMed]
  25. Snow, J.I.; Chung, D.D. Hispanic adolescent lip response to extraction and nonextraction orthodontic treatment. Am. J. Orthod. Dentofac. Orthop. 2023, 163, 68–78. [Google Scholar] [CrossRef] [PubMed]
  26. Perlaza Ruiz, N.A. Facial soft tissue thickness of Colombian adults. Forensic Sci. Int. 2013, 229, 160.e1–160.e9. [Google Scholar] [CrossRef]
  27. Kamak, H.; Celikoglu, M. Facial soft tissue thickness among skeletal malocclusions: Is there a difference? Korean J. Orthod. 2012, 42, 23–31. [Google Scholar] [CrossRef] [PubMed]
  28. Noman, M.; Hashmi, G.; Ali, M.M.; Yousaf, U.; Hussain, M.; Mujeeb, R. Comparison of Soft Tissue Chin Thickness in Adult Patients With Various Mandibular Divergence Patterns. Cureus 2024, 16, e59150. [Google Scholar] [CrossRef]
  29. Veni, S.K.; Elsayed, M.; Singh, I.S.; Nayan, K.; Varma, P.K.; Naik, M.K. Changes in Soft Tissue Variable of Lips Following Retraction of Anterioir Teeth- A Cephalometric Study. J. Pharm. Bioallied Sci. 2023, 15 (Suppl. S1), S248–S251. [Google Scholar] [CrossRef]
  30. Negrutiu, B.M.; Vaida, L.L.; Todor, B.I.; Judea, A.S.; Lile, I.E.; Moca, A.E.; Judea-Pusta, C.T. An important morphological feature of the face: Upper lip length. Rom. J. Morphol. Embryol. 2019, 60, 537–541. [Google Scholar]
  31. Yin, X.X.; Ng, B.W.H.; Yang, Q.; Pitman, A.; Ramamohanarao, K.; Abbott, D. Anatomical landmark localization in breast dynamic contrast-enhanced MR imaging. Med. Biol. Eng. Comput. 2012, 50, 91–101. [Google Scholar] [CrossRef]
  32. Geng, Z.; Sun, K.; Xiao, B.; Zhang, Z.; Wang, J. Bottom-up human pose estimation via disentangled keypoint regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 14676–14686. [Google Scholar]
  33. Bulat, A.; Tzimiropoulos, G. Human pose estimation via convolutional part heatmap regression. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; pp. 717–732. [Google Scholar]
  34. Lindner, C.; Wang, C.-W.; Huang, C.-T.; Li, C.-H.; Chang, S.-W.; Cootes, T.F. Fully Automatic System for Accurate Localisation and Analysis of Cephalometric Landmarks in Lateral Cephalograms. Sci. Rep. 2016, 6, 33581. [Google Scholar] [CrossRef]
Figure 1. The experimental design of the present study. The red dots represent the auto-located output markers, and the green line segments indicate the lip thickness measurement lines.
Figure 1. The experimental design of the present study. The red dots represent the auto-located output markers, and the green line segments indicate the lip thickness measurement lines.
Diagnostics 15 01468 g001
Figure 2. Location instructions and definition of the 24 cephalometric landmarks on the cephalogram (a) and schematic (b). The measurement segments for the upper lip are A-Sn, Spr-UT, and UT-UJ, as shown in red in (b). For the lower lip, they are B-Bs, Id-LT, and LT-LJ, as shown in blue in (b).
Figure 2. Location instructions and definition of the 24 cephalometric landmarks on the cephalogram (a) and schematic (b). The measurement segments for the upper lip are A-Sn, Spr-UT, and UT-UJ, as shown in red in (b). For the lower lip, they are B-Bs, Id-LT, and LT-LJ, as shown in blue in (b).
Diagnostics 15 01468 g002
Figure 3. The network architecture of the automatic cephalometric landmark detection model. The framework follows the backbone–neck–head structure. The red dots indicate the auto-located output markers.
Figure 3. The network architecture of the automatic cephalometric landmark detection model. The framework follows the backbone–neck–head structure. The red dots indicate the auto-located output markers.
Diagnostics 15 01468 g003
Figure 4. Error distribution of each AI-predicted landmark along the X-axis and Y-axis (using manually annotated landmarks as the coordinate origin). Red indicates manually located markers, and blue indicates AI-located markers. The coordinate distribution of AI-predicted landmarks is shown, along with an ellipse representing the 95% confidence interval.
Figure 4. Error distribution of each AI-predicted landmark along the X-axis and Y-axis (using manually annotated landmarks as the coordinate origin). Red indicates manually located markers, and blue indicates AI-located markers. The coordinate distribution of AI-predicted landmarks is shown, along with an ellipse representing the 95% confidence interval.
Diagnostics 15 01468 g004
Figure 5. The confusion matrices for upper and lower lip thickness classification in females and males: (A) upper lip thickness in females, (B) upper lip thickness in males, (C) lower lip thickness in females, and (D) lower lip thickness in males. The x-axis represents the AI-predicted lip thickness classification, while the y-axis represents the manually annotated lip thickness classification. The green color indicates agreement between manual and AI-based lip thickness classifications; other colors indicate disagreement.
Figure 5. The confusion matrices for upper and lower lip thickness classification in females and males: (A) upper lip thickness in females, (B) upper lip thickness in males, (C) lower lip thickness in females, and (D) lower lip thickness in males. The x-axis represents the AI-predicted lip thickness classification, while the y-axis represents the manually annotated lip thickness classification. The green color indicates agreement between manual and AI-based lip thickness classifications; other colors indicate disagreement.
Diagnostics 15 01468 g005
Figure 6. The ROC curves for upper and lower lip thickness classification in females and males: (A) upper lip thickness in females, (B) upper lip thickness in males, (C) lower lip thickness in females, and (D) lower lip thickness in males.
Figure 6. The ROC curves for upper and lower lip thickness classification in females and males: (A) upper lip thickness in females, (B) upper lip thickness in males, (C) lower lip thickness in females, and (D) lower lip thickness in males.
Diagnostics 15 01468 g006
Table 1. Cephalometric landmarks used in this study.
Table 1. Cephalometric landmarks used in this study.
No.Landmarks
1Sella (S)
2Nasion (N)
3Subspinale (A)
4Superior prosthion (Spr)
5The most labial surface of the upper incisor (UJ)
6Upper incisor (UI)
7Lower incisor (LI)
8The most labial surface of the lower incisor (LJ)
9Infradentale (Id)
10Supramental (B)
11Pogonion (Po)
12Gnathion (Gn)
13Menton (Me)
14Subnasale (Sn)
15Labrale superius (UT)
16Stomion superius (UL)
17Stomion superius (LL)
18Labrale inferius (LT)
19Inferior labial sulcus (Bs)
20Pogonion of soft tissue (Pos)
21Gnathion of soft tissue (Gns)
22Menton of soft tissue (Mes)
23Ruler point 1
24Ruler point 2
Table 2. The success detection rates and MRE for 22 landmarks ( x ¯ ± s).
Table 2. The success detection rates and MRE for 22 landmarks ( x ¯ ± s).
LandmarkSuccess Detection Rates (%)Mean ± SD (mm)
1 mm1.5 mm2 mm2.5 mm3 mm4 mm
Soft tissue
Sn75.9895.5999.02100.00100.00100.000.71 ± 0.41
UL75.4992.6598.0498.5399.51100.000.76 ± 0.46
UT76.9695.1096.5797.5599.0299.510.85 ± 0.47
LL79.4193.6398.0499.51100.00100.000.70 ± 0.45
LT83.8295.1099.5199.51100.00100.000.73 ± 0.37
Bs78.9292.6596.57100.00100.00100.000.79 ± 0.49
Pos34.3161.2778.9285.7891.1896.571.54 ± 1.09
Gns38.7366.1883.8291.6795.5998.531.34 ± 0.83
Mes50.0075.0086.7689.7192.1695.592.50 ± 1.27
Hard tissue
S84.3197.5599.51100.00100.00100.000.66 ± 0.37
N72.5590.6996.0897.5599.0299.021.95 ± 1.56
A59.3184.3195.5998.0498.53100.000.96 ± 0.56
Spr87.7598.0499.51100.00100.00100.000.64 ± 0.32
UJ86.7699.02100.00100.00100.00100.000.65 ± 0.30
UI86.2796.5799.0299.51100.00100.000.65 ± 0.38
LI79.4194.1296.5799.0299.0299.510.78 ± 0.52
LJ84.3195.5998.5399.0299.5199.510.72 ± 0.43
Id85.7896.0898.53100.00100.00100.000.69 ± 0.35
B61.2783.8290.6996.0898.04100.001.02 ± 0.78
Po60.2981.8693.1499.51100.00100.000.98 ± 0.55
Gn75.4995.5998.0499.51100.00100.000.77 ± 0.41
Me72.5590.6996.5798.0499.51100.000.83 ± 0.52
Average72.2689.5995.4197.6698.6999.470.97 ± 0.52
Table 3. Upper and lower lip thicknesses in male and female subjects in different lip thickness classifications (mm).
Table 3. Upper and lower lip thicknesses in male and female subjects in different lip thickness classifications (mm).
ClassificationUpper LipLower Lip
FemaleMaleFemaleMale
Thin lip6.69–10.146.84–10.817.72–11.518.88–11.47
Normal lip10.15–12.6410.83–13.9311.52–13.7411.49–14.61
Thick lip12.67–15.9813.96–20.8813.75–17.0714.67–20.35
Table 4. The performance of the classification for upper and lower lip thicknesses.
Table 4. The performance of the classification for upper and lower lip thicknesses.
AccuracySensitivitySpecificityPrecisionF1-ScoreAUC
Upper lip thickness
Female0.91 ± 0.04 0.86 ± 0.04 0.92 ± 0.070.88 ± 0.040.87 ± 0.010.97 ± 0.02
Class 10.95 0.82 0.98 0.92 0.87 0.98
Class 20.87 0.89 0.84 0.84 0.87 0.94
Class 30.92 0.86 0.94 0.88 0.87 0.97
Male0.92 ± 0.030.87 ± 0.760.93 ± 0.070.90 ± 0.050.88 ± 0.040.98 ± 0.02
Class 10.93 0.78 0.97 0.91 0.84 0.98
Class 20.88 0.93 0.84 0.84 0.88 0.97
Class 30.95 0.89 0.97 0.94 0.91 0.99
Lower lip thickness
Female0.90 ± 0.040.85 ± 0.02 0.91 ± 0.070.86 ± 0.030.85 ± 0.010.98 ± 0.01
Class 10.94 0.85 0.96 0.85 0.85 0.99
Class 20.85 0.88 0.82 0.84 0.86 0.97
Class 30.91 0.80 0.96 0.89 0.85 0.98
Male0.88 ± 0.050.80 ± 0.110.90 ± 0.140.86 ± 0.130.82 ± 0.040.98 ± 0.02
Class 10.91 0.80 0.95 0.85 0.82 0.99
Class 20.82 0.92 0.74 0.73 0.81 0.95
Class 30.90 0.70 1.00 1.00 0.82 0.97
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Han, M.; Huo, Z.; Ren, J.; Zhu, H.; Li, H.; Li, J.; Mei, L. Automated Landmark Detection and Lip Thickness Classification Using a Convolutional Neural Network in Lateral Cephalometric Radiographs. Diagnostics 2025, 15, 1468. https://doi.org/10.3390/diagnostics15121468

AMA Style

Han M, Huo Z, Ren J, Zhu H, Li H, Li J, Mei L. Automated Landmark Detection and Lip Thickness Classification Using a Convolutional Neural Network in Lateral Cephalometric Radiographs. Diagnostics. 2025; 15(12):1468. https://doi.org/10.3390/diagnostics15121468

Chicago/Turabian Style

Han, Miaomiao, Zhengqun Huo, Jiangyan Ren, Haiting Zhu, Huang Li, Jialing Li, and Li Mei. 2025. "Automated Landmark Detection and Lip Thickness Classification Using a Convolutional Neural Network in Lateral Cephalometric Radiographs" Diagnostics 15, no. 12: 1468. https://doi.org/10.3390/diagnostics15121468

APA Style

Han, M., Huo, Z., Ren, J., Zhu, H., Li, H., Li, J., & Mei, L. (2025). Automated Landmark Detection and Lip Thickness Classification Using a Convolutional Neural Network in Lateral Cephalometric Radiographs. Diagnostics, 15(12), 1468. https://doi.org/10.3390/diagnostics15121468

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop