Next Article in Journal
Synthesis, In Vitro Anti-Inflammatory Activity, Molecular Docking, Molecular Dynamics and DFT Calculations of Thiazoline-2-Thione Derivatives
Previous Article in Journal
Impact of HF-Free Synthesis Modification on Purity and Adsorption Performances of MOF MIL-100(Fe) for Gas Capture and Energy Storage Applications
Previous Article in Special Issue
A Comparative Analysis of Hyper-Parameter Optimization Methods for Predicting Heart Failure Outcomes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Computer-Aided Facial Soft Tissue Reconstruction with Computer Vision: A Modern Approach to Identifying Unknown Individuals

1
Faculty of Applied Computer Sciences and Biosciences, University of Applied Sciences Mittweida, 09649 Mittweida, Germany
2
Faculty of Medicine, Institute of Legal Medicine, 80539 Munich, Germany
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(11), 6086; https://doi.org/10.3390/app15116086
Submission received: 1 April 2025 / Revised: 16 May 2025 / Accepted: 22 May 2025 / Published: 28 May 2025

Abstract

Facial soft tissue reconstruction is an important tool in forensic investigations, especially when conventional identification methods are unsuccessful. This paper presents a digital workflow for facial reconstruction and identity verification using computer vision techniques applied to two forensic cases. The first case involves a cold case from 1993, in which a manual reconstruction by Prof. Helmer was conducted in 1994. We digitally reconstructed the same individual using CAD software (Blender), enabling a direct comparison between manual and digital techniques. To date, the deceased remains unidentified. The second case, from 2021, involved a digitally reconstructed face that was later matched to a missing person through DNA analysis. Here, comparison material was available, including an official photograph. A police officer involved in the case noted a “striking resemblance” between the reconstruction and the photograph. To evaluate this subjective impression, we performed quantitative analyses using three face recognition models (Dlib-based method, VGG-Face, and GhostFaceNet). The models did not indicate significant similarity, highlighting a gap between human perception and algorithmic assessment. These findings suggest that current face recognition algorithms may not yet be fully suited to evaluating reconstructions, which tend to deviate in subtle but critical facial features. To achieve better facial recognition results, further research is required to generate more anatomically accurate and detailed reconstructions that align more closely with the sensitivity of AI-based identification systems.

1. Introduction: The Significance and Aims of Forensic Facial Soft Tissue Reconstruction

Facial soft tissue reconstruction (FSTR) is a method of reconstructing the face of an unknown deceased person directly on the skull or on a replica of the skull [1,2,3]. Reconstructions are carried out for both forensic and archaeological purposes. However, in the forensic context, the aim of an FSTR is to reconstruct the image of the unknown individual as accurately as possible and to use the resulting evidence to identify the person, whereas, in the archaeological context, the focus is more on the symbolic representation of the person during their lifetime. Forensic FSTR can be used in cases where typical identification methods such as DNA, dental, and X-ray data are not available to identify the deceased [4]. This is the case when there is no antemortem comparative data of a potentially suitable person, and thus, the attribution of human remains to a specific person is severely limited [5]. The lack of comparison material often results in the failure of investigations into the identification of unknown deceased individuals. Given the significant methodological effort and associated costs incurred years ago, the concept of high-throughput identification through facial soft tissue reconstruction was unthinkable. A digital approach combined with software-specific developments and medical insights enables a time-effective and flexible implementation of facial soft tissue reconstruction. Once the image of an unknown deceased individual has been reconstructed, the next step is identification. By presenting a reconstructed face to the public, investigators hope to obtain valuable information about the individual’s identity from witnesses or photographs of potentially relevant persons. Once such comparison material is available, a quantitative comparison can be made using basic facial landmarks and measurements or employing computer vision techniques.
The creation of a biological (bio-anthropological) profile of an unknown individual, as well as well-established reconstruction methods, form the basis of any forensic facial soft tissue reconstruction. Useful information about the biological profile can be obtained by applying classical and modern methods of forensic anthropology when examining the remains of unknown individuals. In detail, the biological profile includes information on sex, age, and ethnic origin [4,6,7]. Standard procedures include morphognostic (visual assessment of form-based anatomical features) and osteometric (based on bone measurement techniques) examinations of the bones, particularly the skull, pelvis, and individual long bones, which allow sex to be determined by assessing and measuring sex-specific characteristics such as the shape of the eye sockets. In addition, age-specific features such as the racemization of aspartic acid in the root dentin of teeth (a biochemical technique for age estimation based on amino acid conversion in teeth), adhesions of bone joints, and analyses of the histological structure of bones can provide an estimate of age at death [8]. Morphognostic sex determination on the skeleton is carried out by analyzing sexually dimorphic skeletal characters, e.g., on the skull or pelvis, which are particular carriers of these sexual dimorphisms [4]. For the morphognostic evaluation, the characteristics are categorized according to their degree of expression on a scale from hyperfeminine to feminine, indifferent, masculine, and hypermasculine. The overall view of all assessed traits leads to a diagnosis of feminine, masculine, or indifferent [4]. Osteometric methods have been developed to counter the accusation of the subjectivity of morphognostic methods [9]. Discriminant analysis, in particular, has become established in sex determination [10]. This involves taking sex-specific length and distance measurements on bone material and developing so-called discriminant functions. These, in turn, make it possible to assign a possible sex to the unknown individual. In addition to the classical methods, modern techniques such as DNA analysis or the radiocarbon method are used to determine the age at death or the lifespan of the individual.
In addition to the creation of the biological profile mentioned at the beginning, the reconstruction methods that have been established over the years form an essential basis for any facial soft tissue reconstruction. The three most prominent methods that initially relied on manual 3D modeling of soft tissues directly onto the skull are the Russian, American, and Manchester methods. The first is the Russian method, also known as the Gerasimov method, which is based on a purely muscular reconstruction of the face, taking into account the individual facial muscles and muscle attachments to the bone structures [11,12]. The second method is the American method of Betty Pat Gatliff and Clyde Snow, which is based on the knowledge and application of craniometric reference markers, so-called anatomical soft tissue markers, which are characteristic of the age, sex, and ethnic group of the person whose face is to be reconstructed [13]. Anatomical soft tissue markings are applied to specific areas of the skull, and the spaces between them are filled in to accentuate the facial features [12,14,15]. The final method is the British or Manchester method of Richard Neave, which combines the Russian and American techniques to create a successful method of facial reconstruction [2,16,17].
In recent years, digital methods have become increasingly important compared to manual FSTR [18,19,20,21]. This is not only due to time-saving and high flexibility but also to the possibility of optimizing individual steps of the reconstruction process through automation.
Digital facial reconstruction methods can fundamentally be divided into two categories: 2D and 3D techniques. In 2D reconstruction, one common approach is digital drawing, in which a forensic artist, often in collaboration with an anthropologist, sketches the facial features directly onto an image of the skull [22]. This technique follows similar principles to traditional manual methods, such as the application of average soft tissue thicknesses and anatomical guidelines for feature placement [22]. Another 2D technique is superimposition, where an algorithm reconstructs the face by overlaying skull images with portrait photo databases to find the best visual match [23]. In contrast, 3D digital reconstructions include both the virtual implementation of traditional methods, such as the American, Russian, or Manchester approaches, using computer-aided design (CAD) software [24] and more advanced techniques involving the approximation of facial soft tissues through 3D morphing based on large anatomical databases [25,26]. A detailed comparison of both classical and digital methods of forensic facial soft tissue reconstruction, including their methodological foundations, advantages, and limitations, is provided in Table 1.
However, in order to be able to carry out a computer-aided reconstruction in a virtual environment, and for ethical reasons also for manual reconstructions, the remains, in particular the skull, must first be digitized or replicated. Various methods, such as photogrammetry, laser scanning, imaging techniques, and 3D printing techniques, can be used for this purpose [26,27,28,29,30,31]. Photogrammetric methods, in particular, are becoming increasingly important for the production of digitized data alongside imaging techniques such as CT scans [32,33,34,35]. A large number of commercial and open-source software applications are available for processing the captured data in the form of image and video recordings or sectional images for CT scans [34,36,37,38]. Another digitization option is laser scanners, which are also widely used depending on the objects to be digitized [39,40,41,42].
Depending on the evidence and the underlying technology, different methods can be used to create high-quality digital copies of remains, which form the basis for computer-aided reconstructions. These are carried out using open-source software such as Blender, a 3D suite that covers the entire 3D pipeline from modeling, rigging, animation, simulation, rendering, compositing, and motion tracking [27,43,44]. A particular advantage of such applications is that already integrated functions can be constantly extended by means of specially developed scripts and modules, so-called add-ons (an add-on is a small piece of software or a feature that enhances or extends the functionality of a main program or application like Blender), such as those used in this publication. As with the manual approach, the virtual face is created using one of the reconstruction methods presented. Established methods for positioning the eyeballs and generating facial features such as the nose, mouth, and ear [2,45,46,47,48,49,50] help.
However, in order to improve the accuracy and plausibility of facial soft tissue reconstructions, it is important to analyze the technical possibilities for producing digitized data, the extent to which manual approaches can be further digitized, and how existing digital procedures can be improved.
Therefore, using two real cases, this paper presents an optimized procedure for forensic facial soft tissue reconstruction based on [51]. Furthermore, a comparison is made between a manual reconstruction and the results of a digital reconstruction based on the optimized reconstruction process. Additionally, using computer vision, a facial comparison was performed between photographs of the missing and identified person and the digital reconstruction of another case.

2. Background Information on the Two Real Cases

2.1. Case 1: Göttingen (Germany) 1993

In 1993, an unknown corpse was found in Holzerode near Göttingen (Germany). A post-mortem examination carried out at the Institute of Forensic Medicine Göttingen (Germany) revealed massive violence to the neck with bony injuries to the larynx and hyoid bone as the cause of death. On the basis of the available remains, the death was considered a homicide. The victim was not yet identified. In order to clarify the identity of the unidentified man, extensive examinations were carried out at the time, including the alleged victim’s tattoos, teeth, and dental status, in particular the crowns. Furthermore, records were used to create a biological profile, which formed the basis of a manual facial soft tissue reconstruction in 1993, and the digital FSTR was also part of this publication. The following information regarding the biological profile and external appearance of the individual was known before the reconstruction:
  • Sex: male; age: about 30–40 years; height: about 170 cm.
  • Dark hair, about 10 to 15 cm long.
  • European ancestry and a robust skeletal structure.
  • Broad, oval face with a slightly protruding midface.
  • Slightly arched eyebrows and a relatively low forehead.
  • Rather voluminous eyelids and pupil distance of approximately 60 mm.
  • Large nose with a saddle-shaped, straight bridge and slightly protruding nostrils, with the width of the nostrils being approximately 35 mm.
  • Mouth with narrow lips, moderately curved and weak philtrum, with a broad and firm chin.
  • Neutral gray–brown eye color, taking into account typical correlations between eye and hair color, as well as general population traits associated with ethnic background.
In addition to the above information, Prof. Richard P. Helmer (Richard Helmer was Germany’s leading expert in forensic facial reconstruction at the Institute of Forensic Medicine at the University of Bonn), MD, described particular features of possible facial injuries at the time. Specifically, Prof. Helmer described an oval bone defect on the left maxilla (upper jawbone) (6 mm below the inferior orbital rim (lower edge of the eye socket) and 20 mm from an imaginary midline, counting from the left side) with evidence of complete bone healing. This defect includes areas where sensory nerves run from large parts of the left side of the face. From all indications, it could be a facial injury or a consequence of inflammatory reactions following the opening of the left maxillary sinus. The bony injury could have resulted in scarring just below the left eye, which would have been visible on a living person and might have been conspicuous. A nerve injury to the infraorbital (area located below the eye socket) nerve could also have resulted in partial trophic disturbances on the left side of the face. During the medical treatment of a possible injury and the removal of the injured tissue, the surgical wound could have been closed with surgical sutures. This might leave a scar that looks like a line. Overall, according to Prof. Helmer, MD, the bone injury provides too little information about the shape of a possible scar.
Based on all this conclusion, Prof. Helmer carried out a manual soft tissue reconstruction of the face of the unknown deceased in 1994 (Figure 1). This reconstruction will later be used to compare traditional methods with a digital reconstruction based on an optimized method of computer-aided FSTR. This study received the necessary permissions (approval dated 17 December 2024) from the Göttingen Public Prosecutor’s Office.

2.2. Case 2: Hildesheim (Germany) 2021

The second case presented in this article regards an unidentified body found in early 2021. The body was found in a wooded area near Hildesheim (Germany). The following data on the biological profile of the deceased were obtained by examining the remains as part of a forensic post-mortem examination at the Institute of Forensic Medicine Hannover (Germany). All of the personal details are taken from the autopsy report:
  • Sex: male; age: about 40–50 years old; height: about 170 cm.
  • Brown hair.
  • Caucasian of unknown nutritional status/size.
  • Very elaborate and high quality teeth and fillings.
  • Well-developed bony structures.
This study received the necessary permissions (approval dated 11 December 2024) from the Hildesheim Public Prosecutor’s Office.
The information from the two cases served as the basis for facial soft tissue reconstruction with the goal of clarifying the identity of the decedent. The following Section 3 first describes the process of forensic facial soft tissue reconstruction in general, followed by the individual steps of the process in the selected cases.

3. Procedure for Forensic Facial Soft Tissue Reconstruction in Detail

The process of computer-aided forensic facial soft tissue reconstruction is divided into the areas of information acquisition, digitization of the underlying skeletal material, modeling process, and plausibility checks. The database used in FSTR typically includes photographs and video recordings, as well as data from imaging techniques such as CT or MRI scans. These imaging modalities can be used, for example, to generate 3D models of the skull or to extract anatomical landmarks (e.g., the anterior nasal spine) that guide the reconstruction of specific facial features. Anthropological and molecular biological methods provide useful information on the biological profile of the unknown individual. Evidence, such as clothing, can also provide additional information. Once the information has been obtained and the remains have been digitized, the actual reconstruction steps are carried out within the appropriate software applications.
We chose the Manchester method for facial reconstruction due to its recognized accuracy in forensic identification. Developed by Neave in 1977, it combines the objectivity of depth markers with individual muscle structure, making it a reliable and widely accepted method [52]. This approach is frequently used in forensic cases and provides a solid foundation for detailed facial reconstruction [52], making it a suitable choice for our study.
The positioning of anatomical soft tissue markers, the modeling of individual facial features, and the modeling of age-specific features are the key process steps. To optimize the reconstruction process, these steps have been partially automated based on the preliminary work of [51]. Once the facial and age characteristics have been reconstructed, each model is individualized by integrating further details such as accessories and hairstyles; then, the model is subjected to a final plausibility check. A detailed overview of the individual process steps, including optimizations, can be found in the following Figure 2. The shown process is computer-aided.
The reconstructions were originally carried out using Blender version 2.7 (2021). For the generation of the final visualizations included in this publication, Blender version 3.4 was used.
In the following sections, we will provide a detailed procedure for digital forensic facial soft tissue reconstruction. This will include the methodologies used, the step-by-step process of reconstructing facial features, and the tools and technologies employed in the reconstruction.

3.1. Preceding Preparation and Information Gathering

Prior to the actual reconstruction, extensive research and fact-finding are required to obtain as much information as possible about the subject. In most cases, the investigating authorities and forensic experts provide information in the form of photographs of the crime scene and post-mortem reports. The latter primarily contain information on the biological profile of the individual as well as possible clues as to the circumstances of death or, if applicable, the use of force [54].
In addition to the information provided, databases of photographs of people of different genders and ages are searched to provide reconstructors an overview of, for example, signs of aging and age-related changes [55]. It is also important to take the spirit of the times into account. This means, for example, a typical hairstyle at a certain time (e.g., the 1990s). This is particularly important for cold cases, such as Case 1. The most important prerequisite for facial soft tissue reconstruction is the presence of a largely intact skull and mandible. Although facial reconstruction without the mandible is possible, the results are not satisfactory. However, procedures are now available to reconstruct the lower jaw, which in turn allows the soft parts of the face to be reconstructed. Examples include [8,56,57]. Injuries or fractures of the skull must also be reconstructed beforehand [58]. In addition to these methods of obtaining information, the methods mentioned above for predicting facial features, such as the shape of the nose and mouth, are indispensable.
Once a biological profile has been established and the anatomical soft tissue markers and corresponding facial reconstruction methods have been researched, the actual practical process steps of computer-aided forensic soft tissue facial reconstruction of the unknown individuals in both cases are followed.

3.2. Digitization of the Mortal Remains

As a basis for further steps in the facial soft tissue reconstruction process, available remains, more specifically the skull, are digitized using photogrammetric and imaging techniques. The latter was particularly important in Case 1. This was because the facial soft tissue reconstruction carried out by Prof. Helmer in 1994 was based on the original skull, not a replica. As this reconstruction had to be preserved, the only option was to use imaging techniques, in this case, CT scans, to visualize the bone structure of the skull under the plaster and then to calculate a model of the skull. The same procedure was followed for the skull of the victim in Case 2, with the addition of photogrammetric images. A detailed description of the procedures can be found in the following sections.

3.2.1. Modeling on the Basis of CT Data for the Victim in Case 1

In Case 1, due to the situation described above, it was only possible to collect data using imaging techniques, more specifically, CT scans. The CT dataset comprised a total of 2082 individual images in the DICOM format. The open-source software application InVesalius version 3.1.1 on Windows 7 and 10 (both 64-bit) [37] was used for data processing. With the help of the application, a volume model was calculated from the axial, sagittal, and coronal sectional images. As both the modeling material and the ocular prostheses were available in the CT data from the 3D model previously acquired by computed tomography, the bony structure of the skull in the CT dataset had to be determined first using software-specific parameters. InVesalius provides the ability to select different tissue planes for 3D visualization based on a minimum and maximum threshold value. A minimum threshold of 850 and a maximum threshold of 3071 were used to visualize the bony structures.
Artifacts that occurred during the volume calculation of the 3D skull model (e.g., small surface irregularities or unnatural protrusions caused by the interpolation and meshing of CT data) were manually removed in Blender using the edit mode. This post-processing step ensures that facial reconstruction is based on anatomically plausible bony structures and not distorted by technical artifacts introduced during model generation. Figure 3 and Figure 4 show the volume calculation in InVesalius and the final result after artifact removal.

3.2.2. Photogrammetric Images and Modeling of the Victim in Case 2

Photogrammetric images and photogrammetry itself are based on non-contact measurement methods and evaluation procedures that indirectly determine the position and shape of an object from photographs by measuring the image. This procedure is known in the literature as the “photogrammetric method” [59]. The use of specially developed algorithms, such as the SIFT algorithm (SIFT stands for scale-invariant feature transform), enables the calculation of 3D models of the depicted objects from overlapping photogrammetric images. In Case 2, the skull, or more precisely, the cranium, skullcap, and lower jaw, were positioned on a turntable (company: stageonair, Penzing, Germany; [60]) and photographed 90 times, i.e., every 4°, with reference marks from three different perspectives using three Nikon D7100 and D7500 SLR cameras with simultaneous shutter release (Nikon Corporation, Tokyo, Japan; a total of 270 images per run, i.e., 90 images per SLR camera). Three NG-220A continuous light studio lights from NANGUANG (Guangdong Nanguang Photo&Video Systems Co., Ltd., Dongguan, China) were used as a light source, some of which were fitted with softboxes to ensure homogeneous illumination. The principle of photogrammetry and the test setup are shown in Figure 5. The photogrammetric images of the skull, more precisely the cranium, skullcap, and mandible, are shown in Figure 6.
The photogrammetric images of the bony structures of the skull were processed into 3D models using the open-source software Meshroom (Version 2021.1.0) [38]. The basis for the 3D modeling was 270 photographs (90 images per SLR camera) for the cranium, skullcap, and mandible, respectively. The primary processing steps involved the extraction of image features, the alignment of the overlapping images, the calculation of the point cloud, the meshing of the point clouds, and the subsequent texturing of the calculated 3D models. Following data processing, a 3D model was generated for each of the images of the cranium, the skull, and the mandible. These were then combined into an overall model in the Blender software application. An overview of the calculated individual models and the combined model is presented in Figure 7.

3.3. Placement of Anatomical Soft Tissue Markers

Anatomical soft tissue markers, known as landmarks, represent average soft tissue thicknesses at defined anatomical points on the face or skull [26]. The mean millimeter values have been determined using scientific measurement methods and differ between men and women and between leptosome and obese individuals. Of particular importance to the reconstruction process is the extent to which the actual values required for optimal reconstruction of the face of an unknown deceased person deviate from the average soft tissue thicknesses at the respective anatomical points. Deviations result from, among other things, the aging process, nutritional status, and lifestyle of the person to be identified. Data on these soft tissue thicknesses used to be obtained from cadavers [61]. Today, more modern methods such as ultrasound, MRI, and CT are available. Another advantage of imaging techniques is that they allow measurements to be made on living people, eliminating post-mortem changes and artifacts [62]. In addition, differences in soft tissue thickness between sitting and lying positions can be studied [63]. There are many publications that provide information on soft tissue thickness in different ethnicities, sexes, and age groups [13,64].
The data for soft tissue markers are based on the 2006 study by S. De Greef et al. [13]. In this study, the soft tissue thicknesses of 967 adult Caucasians were measured, taking into account the age and BMI of the subjects. The datasets were then categorized into age groups, e.g., 18 to 29 years and 30 to 39 years. The positions of the anatomical soft tissue markers on the human skull were also taken from [13] for subsequent facial reconstruction. For the reconstruction, the anatomical soft tissue markers were applied in the Blender software application to the 3D models of the skulls with the corresponding dimensions from Table 3b in [13] for the victim in Case 1 and Table 3c in [13] for the victim in Case 2. In both cases, a normal BMI was assumed, which was taken into account in the anatomical soft tissue thickness data collected by [13] for the unknown individuals. In the cold case, the body was described as a “medium body type”; in the case of 2021, no information was available. An overview of the positions of the anatomical soft tissue markers on the skulls of the present cases is shown in the following Figure 8 and Figure 9.
The add-on, called the enchanted soft tissue marker (ESTM), optimizes the creation and placement of anatomical soft tissue markers. In previous facial soft tissue reconstructions, it was necessary to search the literature outside of Blender to find the optimal soft tissue thicknesses for each individual. Depending on age and gender, as well as other factors such as leptosomal or obese physique, data sheets with average soft tissue thicknesses have been collected by numerous research groups. In addition to these preparations, process steps such as the manual creation of each individual soft tissue marker in terms of its length and its application to the skull itself were very time-consuming. Subsequent changes also required several additional software-specific steps. This ESTM add-on provides the ability to import preferred marker lists into the software. This allows the user to select the most appropriate anatomical soft tissue markers based on the information available about the biological profile of the individual. The selection includes data from De Greef et al. and Rhine and Moore. The data are available in .csv format, which is read by the add-on, and are divided into depth, name, and description. The soft tissue markers can be selected and imported from a drop-down menu. If necessary, the name and length can then be adjusted using numeric input or other functions. This includes the percentage adjustment of each soft tissue marker depending on the changes that may result from an increased or decreased BMI [13,65,66]. The adjustments ±5% and ±10% have been chosen here.

3.4. Reconstruction of Selected Facial Features and Modeling of the Soft Tissue and Skin

3.4.1. Reconstruction and Modeling of the Victim in Case 1

The musculature of the face was built up step by step in relation to the anatomical soft tissue markers up to the skin tissue. The muscle markers on the bones provided indications of the strength of the musculature attached to these areas.
The eyeballs were positioned in the 3D model according to C. Wilkinson and K. Taylor [2,67]. For this purpose, an eye add-on can be utilized. The add-on can be used to create a model of the right eye at the position of the 3D cursor in the Blender software application, whose position, size, and rotation can then be changed using appropriate sliders. To create the second eye, the user simply places the 3D cursor in the appropriate position and then creates the eye using the function regarding it. This eye model is an exact copy of the other eyeball, which can be duplicated and mirrored. The orbicularis muscle (the circular muscle around the eyes) can then be added to both eyes, and the distance between the eyes can be adjusted.
The shape of the mouth was reconstructed according to the mathematical basis of C.N. Stephan [68]. To support this process, a custom-developed Blender add-on (“Mouth Editor”) was used, which facilitates the application of Stephan’s method within the 3D modeling environment. It allows the user to import a mouth model and customize it to their needs. In addition, the predefined phenotypes Caucasian, Mongoloid, and Negroid can be applied to the mouth model. This is divided into sub-areas, such as the lower and upper lip and the mandible, which can be transformed independently.
The reconstruction of the nose was based on the mathematical principles developed by C. Rynn (European dataset) and Gerasimov [45]. A nose construction add-on allows the user to make changes to an existing nose model or import and edit a completely new model. By selecting predefined areas, specifically columella (nasal base), side, tip, dorsum (bridge), rhinion (bony–cartilaginous junction on the nasal bridge), radix (nasal root between the eyes), and alae (nostril wings), the model can be adapted to the needs or underlying reconstruction method of the nose. Once the model is complete, it can be integrated into the existing reconstruction process.
All the measurements required for this, taken on the 3D model of the skull, and the values calculated from them for the reconstruction of the nose and mouth are shown in Table 2, Table 3 and Table 4. Figure 10 shows the reconstructed soft tissue and the facial features based on the measured values.

3.4.2. Reconstruction and Modeling for the Victim in Case 2

As in the previous case, the eyeballs were positioned in the 3D model according to C. Wilkinson and K. Taylor [2,67], the shape of the mouth was reconstructed according to the mathematical basis of C.N. Stephan [68], and the shape of the nose according to the mathematical basis of C. Rynn (European dataset) and Gerasimov [45]. Reconstruction of the eyes, nose, and mouth also utilized the previously described add-ons. All measured values for the reconstruction of the nose and mouth are shown in Table 5, Table 6 and Table 7. The following Figure 11 shows the reconstructed soft tissue and facial features based on the measurements taken.

3.5. Modeling of Hair and Age Characteristics, as Well as Individualization of Faces by Incorporating Special Features and Accessories

3.5.1. Modeling of of Individual Characteristics in Case 1

The following procedures were used for the victim in Case 1. By including the described face databases [55], corresponding wrinkles, which can occur in the age range of 30–40 years, were transferred to the 3D model. For the realization of hairstyles, extensive Internet research was carried out on typical hairstyles in Europe from 1993. Hair combed backwards, slicked back, and worn loose was widespread. In the Blender software application, it is possible to transfer hair of different lengths, colors, and textures to models and create different hairstyles for them. In the present case, the most frequently represented hairstyles were modeled on the basis of the hair length interval of 10 to 15 cm specified in the files in accordance with the research carried out. A dark hair color was chosen for the file entries. Based on Prof. Helmer’s statements, one model was modeled with a scar. As no information about eye color was available, the final results were presented with different eye colors. The results of the reconstruction are shown in Figure 12.

3.5.2. Modeling of of Individual Characteristics in Case 2

The following procedures were used for the victim in Case 2. By including the described face databases [55], corresponding wrinkles, which can occur in the age range of 40–50 years, were transferred to the 3D model. Based on the information about the hair attachments to the skull, different facial features were modeled. The results of the reconstruction are shown in Figure 13.

3.6. Integration of the Model into Everyday Life (Rotation, Presentation, and Visualization)

The primary goal of forensic facial soft tissue reconstruction is to identify the unknown individual. Once the face is finally modeled and individualized, the identification process can start using a variety of techniques. Firstly, searches and comparisons with missing person databases are carried out, followed by media coverage of the discovery of an unknown deceased person in the hope of obtaining information relevant to the investigation from the civilian population. Nowadays, social networks are also available to disseminate such information to the general public.
A common prerequisite for all procedures is the availability of high-resolution photographs of the facial soft tissue reconstruction in order to provide the best possible conditions for the identification process. A few years ago, it was common practice to take photographs of the final facial models in photo studios, but digital reconstructions and virtual space offer many advantages in generating and providing such information.
Thanks to the ability to parameterize virtual cameras according to all the technical parameters of their analogue counterparts, high-resolution images of computer-aided facial soft tissue reconstructions from Full HD, QHD, to UHD and beyond can be produced in a very short time. Perspective is irrelevant, i.e., models can be presented and new images calculated as required by simple rotation operations. Thanks to modern and simple software and web applications (such as Paint3D or Sketchfab) that allow the visualization of 3D models by dragging and dropping, the models themselves can be used for presentations without the need to photograph them.
The presentation of three-dimensional data is now widely used in other fields, such as archaeology, demonstrating the clear added value of such presentation [69,70]. The ongoing development of visualization software, which optimizes the effects of light, shadow, and texture, means that the models and scenes created are becoming ever more realistic, which can only benefit third parties. Finally, it should be noted that while traditional photography is perfectly adequate for the purpose, the use of computer-aided visualization techniques provides valuable additional assistance in the identification process.
An overview of the methodological workflow, including the software applications, applied techniques, and corresponding literature, is provided in Table 8.

3.7. Discussion of Methodological Foundations

While the process of facial reconstruction inherently involves some degree of estimation, the core components of our method are grounded in established anatomical and statistical principles. The placement of the eyes, the reconstruction of the nose, the width of the mouth, and the positioning and average thickness of soft tissue markers are all derived from well-documented anthropological studies and follow reproducible mathematical procedures. The anatomical landmarks used for soft tissue marker placement are clearly defined and leave minimal room for subjective interpretation. It should be noted, however, that the objectivity of these elements also depends on the quality, sample size, and demographic representativeness of the underlying anthropological datasets.
Nevertheless, certain facial features remain less constrained by skeletal morphology. For instance, while mouth width and vertical positioning can be estimated based on dentition, the shape and fullness of the lips cannot be reliably inferred from bone structures due to the lack of direct anatomical correlates. Similarly, features such as the hairline and eyebrow shape exhibit significant individual variation and are only partially informed by underlying craniofacial structures. As such, these elements are reconstructed with greater artistic freedom, although we aim to minimize arbitrariness by adhering to population-level norms and anatomical plausibility.

4. Comparison of Classical and Computer-Aided Reconstruction Results in Case 1

When considering the reconstruction results of a forensic facial soft tissue reconstruction, it should not be assumed that the individual appearance of an unknown deceased person can be reconstructed in every detail. However, according to experience and scientific publications on facial soft tissue reconstructions carried out to date, a considerable approximation to the overall appearance of an unknown individual is achieved. Even if there are strong influences on the external appearance of a reconstructed face, such as the reproduction of the correct hairstyle, the overall impression of the face and the facial proportions remain very decisive for a reconstruction. This means that there is at least the certainty that the basic character of a face will be reproduced exactly.
It is not uncommon for the reconstructed face to achieve a very high degree of similarity with the actual appearance of the deceased. In other cases, however, the unavoidable margin of judgment in the interpretation of the empirical reconstruction rules leads to a deviation in the reconstruction from the individual appearance. It is irrelevant whether the reconstruction is carried out classically or computer-aided. A comparison of the reconstruction results is usually made on a quantitative basis. However, as the identity of the unknown man in Case 1 remains unconfirmed, no reference images or identity-validating data are currently available. Consequently, no formal validation can be conducted at this stage. Therefore, any further discussion regarding this case will be based solely on qualitative criteria.
Above all, the choice of the basic method for reconstructing the face and the individual facial features, as well as the underlying data on the anatomical soft tissue markings, can influence the result. In the present case, computer-aided reconstruction was carried out using the Manchester method. It was not clear from the documents provided which method was used for the classic reconstruction. However, according to the descriptions of the procedure in the protocols from 1994 and in conjunction with the statements in [12], it can be assumed that the American method was used for the classic facial soft tissue reconstruction.
Another difference could lie in the data used for the anatomical soft tissue markings. In computer-aided reconstruction, data on anatomical soft tissue thicknesses from de Greef et al. [13] were available. In comparison, the documents do not contain any information on the underlying tissue thicknesses but only indicate that the data were collected from living individuals. For this reason, only a limited statement can be made at this point about possible influences on the reconstruction results based on the anatomical soft tissue markers. The following statements can be made regarding the reconstruction methods used for the individual facial features. In the computer-aided reconstruction, the nose was modeled according to the established method of Rynn et al. [45], and in the classic method, according to A. Macho [71].

5. Comparison of Computer-Aided Reconstruction Results with Photographs of the Deceased Person in Case 2

In Case 2, the identity of the body found could be clarified in the course of the investigation. This means that photographs of the deceased were available for comparison, which allowed for a quantitative evaluation of how similar the reconstructed face appears to the original. This approach should help to assess the accuracy of the reconstruction in a more objective manner. It is important to note that the focus of this work was less on evaluating face recognition models themselves and more on quantitatively assessing the similarity between the reconstructed face and the actual appearance of the individual.
A quantitative comparison can be made using basic facial landmarks and measurement distances. However, these methods are generally limited to images captured under strictly controlled conditions. These conditions include, for example, a homogeneous background and a neutral expression but primarily focus on well-defined frontal upright captures of faces [72,73].
A quantitative comparison using computer vision, especially through machine learning, is often not constrained by specific conditions and can yield robust results, even with faces of varying perspectives [74]. Facial recognition can be used in many applications, including surveillance, security, crime prevention, and access control.
One possibility for performing face recognition is the use of the Dlib software library (Dlib version 19.3.0 was used in this work) [75]. Dlib is an open-source toolkit written in C++ that provides machine learning and computer vision algorithms. It enables automatic face detection, comparison, and identification using artificial intelligence techniques. The facial features are not explicitly defined but are learned from large datasets such that the resulting numerical representations (embeddings) are nearly identical for different images of the same individual.
In this work, Dlib serves as one of the core systems for face recognition. To compare two faces, the framework requires an input image of the target person and an image of the 3D model’s face. The comparison process involves multiple preprocessing steps, including face detection and localization, facial landmark detection for alignment, and feature vector extraction using Dlib’s CNN-based pipeline [75]. The implementation is based on the open-source project by [76].
In addition to Dlib, two modern deep learning-based face recognition models, VGG-Face [77] and GhostFaceNet [78], were also employed. These models offer high recognition accuracy and robust feature extraction capabilities. They are integrated via the DeepFace framework and utilize YOLOv8 (Version 8.0.0) [79] for initial face detection. Unlike Dlib, they do not rely on explicit facial landmark detection for alignment but instead process the cropped and scaled face images directly. The individual preprocessing steps and model differences are explained in more detail in the following sections.

5.1. Procedure of the Comparison Using Dlib-Based Face Recognition

5.1.1. Step 1: Face Detection and Localization

To detect the presence of faces in complex scenes and accurately delineate their positions and areas, face detection must be performed as a preliminary step. Once a face is identified, the region containing the face is globally analyzed, and the positions of all faces are marked based on the input image data. The coordinates of the bounding boxes for the faces are then output [75].
There are multiple selectable alternatives for face detection, each with its own disadvantages and advantages. One of the methods used in this study is a convolutional neural network (CNN)-based face detector, specifically the CNN Face Detection Model V1 provided by the Dlib library [80], along with the corresponding pre-trained model file (mmod_human_face_detector.dat). The advantage of this method is its ability to work well, even from odd angles [81,82]. Convolutional neural networks (CNNs) are a prime example of classical neural networks and are increasingly being utilized across various fields. Their primary application lies in pattern recognition within images [83], making them well-suited for detecting objects in images.
In addition to the Dlib pipeline, an alternative approach was implemented using YOLOv8 [79] as the detection module. YOLOv8 is a state-of-the-art object detection model known for its high speed and accuracy in identifying objects, including human faces [79]. From experience, YOLOv8 proves to be very reliable in detecting faces across various lighting and background conditions. However, in contrast to Dlib, YOLOv8 does not output facial landmarks, it only provides bounding boxes. This limitation directly affects the alignment step described below.

5.1.2. Step 2: Feature Point Location for Face Alignment

To acquire comprehensive information about the detected face image, it is crucial to pinpoint facial feature points. This process enables the identification of facial contour boundaries, including the eyebrows, eyes, nose tip, and mouth corners, among other significant features. By accurately locating these points, we can precisely determine the face’s position and structure [75].
For feature point localization in the Dlib-based pipeline, the CNN-based five-point shape predictor (shape_predictor_5_face_landmarks.dat) provided by Dlib was employed, which detects five key facial landmarks: the inner and outer corners of both eyes and the tip of the nose [84]. Based on these landmarks, facial alignment is performed using Dlib’s get_face_chip function [84], which rotates, scales, and crops the face region to produce a standardized, centered, and frontal view, thereby reducing pose-related variance. Research indicates that face alignment enhances face recognition accuracy [85]. To normalize the faces, basic image transformations such as rotation and scale, which preserve parallel lines (referred to as affine transformations), are applied [75] until the key points are centered as accurately as possible.
In contrast, the YOLOv8-based setup [79] does not support explicit landmark-based alignment since YOLOv8 does not return facial landmarks. As a result, step 2 is omitted in this pipeline. Nevertheless, the preprocessing ensures that the faces are cropped based on the bounding box and rescaled to a fixed input size, which introduces a basic level of normalization.
The test images used in this study are controlled to a high degree. Image 1 is a passport-style frontal photograph, which does not require alignment. Image 2, however, exhibits a slightly lateral view, which may introduce minor pose-related differences. While this could be mitigated by applying facial alignment, it was not considered substantial enough to necessitate an additional alignment procedure in this context. Prior studies suggest that such alignment becomes especially important when dealing with larger pose variations or uncontrolled input conditions [75,85].

5.1.3. Step 3: Face Feature Vector Extraction

To facilitate the mathematical comparison of faces, a numerical representation known as face encoding is necessary. These encodings are matrix representations of faces, where similar faces yield similar matrix representations, as previously mentioned [75]. Conversely, dissimilar faces produce distinct matrix representations.
In the Dlib-based approach, the aligned face images are converted into 128D feature vectors representing each face, using the pretrained ResNet model for face recognition (dlib_face_recognition_resnet_model_v1.dat) [84]. This model achieved a verification accuracy of 0.9938 [84] on the Labeled Faces in the Wild (LFW) dataset [86], making it a very effective and lightweight solution for practical use cases.
In addition, two modern deep learning models were also used for feature extraction in the YOLOv8-based pipeline:
  • VGG-Face [77]: A deep convolutional neural network trained on a large-scale dataset of celebrity faces. It is known for its robust performance and has achieved a verification accuracy of 0.9895 [77] on the LFW dataset [86], classifying it as a highly accurate model for face recognition.
  • GhostFaceNet [78]: A lightweight yet powerful model that incorporates ArcFace loss to generate highly discriminative facial embeddings. GhostFaceNet achieved a verification accuracy of 0.9971 [78] on the LFW dataset [86], thereby outperforming both Dlib and VGG-Face.
The models were used via the DeepFace framework, which internally handles preprocessing and supports multiple backends [87].
The use of multiple recognition models enables a broader evaluation and allows the comparison of classical and modern face representation techniques under the same preprocessing conditions.

5.1.4. Step 4: Face Comparison

As a final step, a mathematical comparison between two faces is conducted using the extracted feature vectors. Typically, the Euclidean distance or cosine similarity is employed to calculate the distance between these vectors. The result is a distance measure, where higher values indicate greater differences and thus lower similarity, while lower values suggest higher similarity.
It is important to note that each face recognition model defines its own optimal threshold for deciding whether two faces belong to the same person. These thresholds are typically determined empirically, based on the model’s performance during training and validation on a specific dataset, aiming to balance the rates of false positives (incorrectly matching different individuals) and false negatives (failing to match the same individual). For instance, using the Euclidean metric, Dlib uses a recommended threshold of 0.60 [84], and VGG-Face typically applies a threshold of around 1.17 [87]. To ensure comparability of results across different models, all distance scores were normalized by dividing by the respective model-specific threshold, resulting in a uniform decision boundary of 1.0 for all models. In this normalized setting a value below 1.0 indicates a match (same identity), a value above 1.0 indicates a non-match (different identities).
It is important to remember, however, that this normalization is a heuristic for comparison purposes only. The resulting values are not linearly related to the semantic similarity between faces and must not be interpreted as probabilities. They merely indicate the relative position of a score with respect to the model’s acceptance threshold.
This approach allows consistent evaluation across models with different embedding characteristics and scales while preserving their individual performance behavior.

5.2. Results of the Comparison Using Dlib-Based Face Recognition

The comparison of the results from facial soft tissue reconstruction of the victim in Case 2 with two seized photographs of the deceased person (not shown to protect the individual’s right to privacy) was carried out using computer vision, employing three different face recognition models: Dlib [75], VGG-Face [77], and GhostFaceNet [78]. The precise results of the computer-vision-aided comparison between the face of the person in question and the reconstructed 3D face in Case 2, normalized and rounded to the fourth decimal place, are shown in Table 9.
The comparison of the reconstructed face in Case 2 with two reference photographs (Image 1: frontal passport-style; Image 2: slightly lateral view) yielded normalized similarity scores above the model-specific thresholds, indicating no similarity between the faces. All scores were normalized such that a value of 1.0 corresponds to the respective model’s verification threshold.
For Image 1, the models produced normalized distances above 1.0, suggesting that the face in this frontal image does not match the reconstructed face closely. The closest match was achieved by VGG-Face (Euclidean: 1.1396) and GhostFaceNet (Euclidean: 1.1664), with Dlib showing a slightly higher value (Euclidean: 1.2827).
For Image 2, GhostFaceNet displayed a much higher distance (Cosine: 1.4823), indicating a clear mismatch. VGG-Face and Dlib both returned similar values (Euclidean: 1.1382 for VGG-Face and 1.2950 for Dlib), still above the threshold and suggesting no significant similarity.
In conclusion, all models returned distances that exceeded the normalized threshold of 1.0, implying that the reconstructed face does not exhibit a strong resemblance to the reference images of the deceased person, according to the evaluated face recognition models.
To provide context and improve the interpretability of the similarity scores, we additionally compared the passport photograph of the identified individual with a photo of a subject who is subjectively perceived as clearly dissimilar (e.g., female, more angular facial features, different hair color, mouth, and nose shape). Interestingly, the resulting Euclidean distances (VGG-Face: 1.1883; GhostFaceNet: 1.1732) were only slightly higher than those observed in the comparisons with the reconstructed face. Cosine distances were more distinct (VGG-Face: 1.4213; GhostFaceNet: 1.556), indicating greater dissimilarity. However, this comparison shows that absolute distance values do not differ as strongly as one might expect based on human perception. It also highlights that these similarity metrics, especially when close to model thresholds, must be interpreted cautiously, as they are not linearly scaled or normally distributed and do not directly correspond to perceptual similarity.

5.3. Discussion of the Comparison Using Computer Vision

As mentioned above, when considering the reconstruction results of a forensic facial soft tissue reconstruction, it should not be assumed that the individual appearance of an unknown deceased person can be reconstructed in every detail. It is more of an approximation to the individual’s appearance.
The minimal differences in the normalized similarity scores between Image 1 (frontal passport photo) and Image 2 (slightly lateral view) suggest that the pose angle had little impact on the face recognition results. Despite the slight lateral angle in Image 2, the face remains largely frontal, meaning the models did not encounter significant pose variation. Both VGG-Face and GhostFaceNet are known for their robustness to small pose changes, which is reflected in the consistent results across both images.
It is important to note that all the face recognition models used here were trained on large benchmark datasets that include a wide variety of individuals in diverse environments. These datasets typically feature significant variations in lighting, exposure, angles, facial expressions, and even occlusions [86], which makes the models highly robust to such factors. Consequently, these models perform well even when the images deviate slightly from the ideal frontal view, as seen in the comparison of Image 1 and Image 2.
These findings indicate that the slight pose change did not significantly affect the results, though larger pose variations might yield different outcomes.
Human perception seems capable of recognizing similarities even when deviations in facial features are noticeable upon closer examination. This ability was reflected in Case 2, where the police officers involved noted that the similarity to the reconstructed face was “astonishing” (translated from German), with one remarking, “As a layperson, I see a significant match” (translated from German). Additionally, the press reported, “Indeed, based on comparison photos of the missing person […], a resemblance to the reconstructed face was noticeable, the investigators reported” (translated from German) [88]. However, facial recognition software is particularly sensitive to minor facial features and will reject matches if there are discrepancies. The ultimate goal is to enable the automatic and rapid search of large databases using facial recognition software based on soft tissue reconstructions. To achieve this, further research is needed to generate more accurate reconstructions.
An additional approach for facial recognition could involve training models to compare a 3D-generated or reconstructed face with a photograph of the person. However, this approach is quite challenging due to the need for a large dataset containing both 3D models of individuals and corresponding photographs. A possible method would resemble the approach taken in the work of Moreno et al., which tackles the recognition of composite sketches and photographs using deep learning techniques [89].

6. Summary and Prospects

6.1. Summary

This section summarizes the advantages of the computerized approach. Essentially, using modern digitization methods such as photogrammetry, digital images of a skull can be produced in the form of three-dimensional models in a matter of minutes. These are then used as the basis for further modeling steps. For ethical reasons, the costly reproduction of the original is not necessary, as the modeling is always carried out on the digital image and not on the original skull.
Particularly in the case of people who have been lying down for a long time or victims of severe violence in the facial area, skull finds often show fragmentation and destruction, which must be reconstructed before the actual work can begin. Traditionally, this is carried out by hand, using plaster, clay, or other materials. Computer-aided approaches work in a similar way but are much more flexible. Where it is not possible to reconstruct missing parts of the face using traditional methods, methods such as computer-aided morphing open up new possibilities for reconstructing missing areas of the skull.
One of the most important factors in facial soft tissue reconstruction is the use of anatomical soft tissue markers. An exception is the Russian method, developed by Gerasimov in 1971, in which only the anatomy of the skull forms the basis of the reconstruction. The reference to these markers is essential for the individualization of a face, as some of them are deliberately over- or under-represented at any given time, depending on the information available about the individual in question. Thanks to the many visualization features of facial soft tissue reconstruction software, such problems can be solved literally at the touch of a button. In contrast, reference to anatomical soft tissue markers is possible with conventional methods, but only to a limited extent.
A problem that tends to arise with historical findings is the subsequent adaptation of the models, especially when the information base is sparse. Classical methods are clearly limited in their flexibility, as individual facial features of hardened plaster models cannot be easily exchanged and optimized or hairstyles quickly colored. The use of digital model libraries makes it possible to adapt computer-generated models quickly and easily, for example, to color hair or reduce the size of a nose. In addition, the computer-aided reconstruction process can be partially automated using developed scripts and modules in the form of so-called add-ons, in particular for the selection and positioning of anatomical soft tissue markers and the generation of individual facial features.
In summary, classical methods are not obsolete or ineffective. However, computer-aided approaches offer advantages that make it possible to deal with the known problems in this field more flexibly and, in some cases, more effectively and efficiently.
The advantages of digital reconstruction are particularly evident in so-called cold cases, where conventional investigations often fail due to a lack of comparative material. In the past, the high methodological requirements and costs made widespread use of such techniques almost impossible. Digital approaches, combined with specialized software development and medical expertise, now allow for time-efficient and flexible implementation. A practical example of the application of this method was illustrated using a case from 1993 (Case 1) and from 2021 (Case 2). Digital reconstruction techniques were used to illustrate the benefits of this modern technique. Digital facial soft tissue reconstruction not only offers greater accuracy but also significantly speeds up the entire identification process. This is particularly valuable in cases where traditional methods fail, and rapid identification can be critical. In summary, by integrating advanced software and medical expertise, digital facial soft tissue reconstruction offers an effective solution for the identification of unknown persons, especially in complex cold cases.

6.2. Prospects

However, these processes can be further improved and optimized. This is where approaches from the field of machine vision or AI in general offer opportunities. The gradual integration of these approaches in various disciplines has enabled significant progress to be made in dealing with complex problems. In this context, the application of AI technologies to forensic anthropology opens up new dimensions in facial soft tissue reconstruction. Not only does it promise to increase the accuracy of reconstruction and speed up the entire identification process, but it also offers the possibility of increased objectivity and standardization.
In particular, the objectivity of AI compared to studies such as those by Vanezis et al. [1] or Stephan et al. [15] is crucial. Both research groups investigated and critically examined the dependence of reconstruction accuracy on the skill, experience, and subjective bias of the reconstructor. The results of these studies showed significant differences in the resulting reconstructions due to the personal influence of the reconstructor.
AI approaches offer advantages in this context due to their objectivity, as they operate based on predefined algorithms and are not influenced by human bias [90]. A key strength lies in their internal consistency: the same model will produce identical results when given the same input, regardless of the user or the time of execution [90]. However, it is important to note that consistency cannot be expected across different models, especially if they are trained on varying datasets or follow different architectures. Nevertheless, the reproducibility of individual AI models remains a valuable property, particularly in applications that demand standardized and transparent procedures. The integration of more advanced machine learning algorithms and 3D modeling techniques is expected to build upon this foundation and further improve reliability and performance in this field.
Yet, the effective application of such AI-driven methods relies on a solid foundation of extensive and high-quality data. To generate reconstructions that reflect individual-specific features with high accuracy, a large and diverse dataset of cranial and facial information is essential. This need highlights the importance of developing a dedicated dataset that includes both 3D cranial models and corresponding facial data. Such a dataset would enable systematic evaluation and benchmarking of facial reconstruction techniques, particularly in the context of digital and automated workflows. However, assembling such data poses significant ethical, legal, and logistical challenges, especially regarding data availability and consent.

Author Contributions

Conceptualization, S.P. and S.B.; methodology, S.P., S.B. and J.R.; software, S.P. and J.R.; validation, S.P.; formal analysis, S.P.; investigation, S.P. and S.B.; resources, J.R. and D.L.; data curation, S.P. and S.B.; writing—original draft preparation, S.P. and S.B.; writing—review and editing, S.P.; visualization, S.P. and S.B.; supervision, D.L.; project administration, S.P.; funding acquisition, D.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Ethical review and approval were waived for this study due to the fact that the data used in the forensic investigations were lawfully provided by the Hildesheim and Göttingen Public Prosecutor’s Offices pursuant to § 476 of the German Code of Criminal Procedure (StPO). The use of the anonymized case data, including 3D models and images, for research and publication was explicitly authorized by the respective authorities.

Informed Consent Statement

Patient consent was waived due to the fact that the individuals were deceased and the legal authorization for data use was granted by the responsible Public Prosecutor’s Offices (Hildesheim and Göttingen) in accordance with § 476 of the German Code of Criminal Procedure (StPO).

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Acknowledgments

We would like to express our sincere gratitude to Jan Dreßler and Steffen Peldschus for their dedicated supervision and continuous support. Dreßler, from the Institute of Legal Medicine, Leipzig, Germany, and Peldschus, from the Institute of Legal Medicine, Munich, Germany, provided important mentorship throughout our research.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Vanezis, P.; Blowes, R.; Linney, A.; Tan, A.; Richards, R.; Neave, R. Application of 3-D computer graphics for facial reconstruction and comparison with sculpting techniques. Forensic Sci. Int. 1989, 42, 69–84. [Google Scholar] [CrossRef] [PubMed]
  2. Wilkinson, C. Facial reconstruction–anatomical art or artistic anatomy? J. Anat. 2010, 216, 235–250. [Google Scholar] [CrossRef] [PubMed]
  3. Guleria, A.; Krishan, K.; Sharma, V.; Kanchan, T. Methods of forensic facial reconstruction and human identification: Historical background, significance, and limitations. Sci. Nat. 2023, 110, 8. [Google Scholar] [CrossRef]
  4. Verhoff, M.A.; Kreutz, K.; Ramsthaler, F.; Schiwy-Bochat, K.H. Forensische Anthropologie und Osteologie–Übersicht und Definitionen. Dtsch Arztebl 2006, 103, 782–788. [Google Scholar]
  5. Verhoff, M.; Kreutz, K.; Jopp, E.; Kettner, M. Forensische Anthropologie im 21. Jahrhundert. Rechtsmedizin 2013, 23, 79–84. [Google Scholar] [CrossRef]
  6. Gottschaldt, U. Anthropologische Untersuchungen zur Geschlechtsbestimmung Adulter Individuen Sowie zur Altersschätzung Subadulter Individuen. Ph.D. Thesis, Friedrich-Schiller-Universität, Jena, Germany, 2003. [Google Scholar]
  7. Uerlings, H. Zur Geschlechtsbestimmung von Menschlichen Skelettteilen. Ph.D. Thesis, Friedrich-Schiller-Universität, Jena, Germany, 1991. [Google Scholar]
  8. Babacan, S.; Isiklar, S.; Kafa, I.M.; Gokalp, G. Redesign of missing mandible by determining age group and gender from morphometric features of skull for facial reconstruction (approximation). Archaeol. Anthropol. Sci. 2021, 13, 1–19. [Google Scholar] [CrossRef]
  9. Stewart, T. Sex determination of the skeleton by guess and by measurement. Am. J. Phys. Anthropol. 1954, 12, 385–392. [Google Scholar] [CrossRef]
  10. Giles, E.; Elliot, O. Sex determination by discriminant function analysis of crania. Am. J. Phys. Anthropol. 1963, 21, 53–68. [Google Scholar] [CrossRef]
  11. Ullrich, H.; Stephan, C.N. On Gerasimov’s Plastic Facial Reconstruction Technique: New Insights to Facilitate Repeatability. J. Forensic Sci. 2011, 56, 470–474. [Google Scholar] [CrossRef]
  12. Verzé, L. History of facial reconstruction. Acta Biomed. Atenei Parm. 2009, 80, 5–12. [Google Scholar]
  13. De Greef, S.; Claes, P.; Vandermeulen, D.; Mollemans, W.; Suetens, P.; Willems, G. Large-scale in-vivo Caucasian facial soft tissue thickness database for craniofacial reconstruction. Forensic Sci. Int. 2006, 159, S126–S146. [Google Scholar] [CrossRef]
  14. Neave, R.; Prag, J. Making faces. In Using Forensic and Archaeological Evidence; British Museum Press: London, UK, 1997. [Google Scholar]
  15. Stephan, C.N. Facial approximation—From facial reconstruction synonym to face prediction paradigm. J. Forensic Sci. 2015, 60, 566–571. [Google Scholar] [CrossRef]
  16. Wilkinson, C. Forensic Facial Reconstruction; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  17. Wilkinson, C.; Neave, R. The reconstruction of a face showing a healed wound. J. Archaeol. Sci. 2003, 30, 1343–1348. [Google Scholar] [CrossRef]
  18. La Cava, S.M.; Orrù, G.; Drahansky, M.; Marcialis, G.L.; Roli, F. 3D Face Reconstruction: The Road to Forensics. ACM Comput. Surv. 2023, 56, 1–38. [Google Scholar] [CrossRef]
  19. Navic, P.; Inthasan, C.; Chaimongkhol, T.; Mahakkanukrauh, P. Facial reconstruction using 3-D computerized method: A scoping review of Methods, current Status, and future developments. Leg. Med. 2023, 62, 102239. [Google Scholar] [CrossRef]
  20. Kalafutová, S.; Chmelík, M.; Olexa, L.; Boroňová, I.; Gal’ová, J.; Petrejčíková, E. 3D facial reconstruction and 3D printing of women skull from Bronze age, Slovakia. Digit. Appl. Archaeol. Cult. Herit. 2024, 33, e00344. [Google Scholar] [CrossRef]
  21. La Cava, S.M.; Orrù, G.; Goldmann, T.; Drahansky, M.; Marcialis, G.L. 3D face reconstruction for forensic recognition-a survey. In Proceedings of the 2022 26th International Conference on Pattern Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022; IEEE: New York, NY, USA, 2022; pp. 930–937. [Google Scholar]
  22. Yadav, N.; Panat, R.; Aggarwal, A. CT scans-a compelling tool in forensic facial reconstruction. J. Dent. Sci. Oral Rehabil. 2010, 1, 42. [Google Scholar]
  23. Damas, S.; Cordón, O.; Ibáñez, O. Handbook on Craniofacial Superimposition: The Meprocs Project; Springer Nature: Cham, Switzerland, 2020. [Google Scholar]
  24. Moraes, C.A.d.C.; Dias, P.E.M.; Melani, R.F.H. Demonstration of protocol for computer-aided forensic facial reconstruction with free software and photogrammetry. J. Res. Dent. 2014, 2, 77–90. [Google Scholar] [CrossRef]
  25. Shui, W.; Wu, X.; Zhou, M. A computerized facial approximation method for Homo sapiens based on facial soft tissue thickness depths and geometric morphometrics. J. Anat. 2023, 243, 796–812. [Google Scholar] [CrossRef]
  26. Guyomarc’h, P.; Dutailly, B.; Charton, J.; Santos, F.; Desbarats, P.; Coqueugniot, H. Anthropological Facial Approximation in Three Dimensions (AFA 3D): Computer-Assisted Estimation of the Facial Morphology Using Geometric Morphometrics. J. Forensic Sci. 2014, 59, 1502–1516. [Google Scholar] [CrossRef]
  27. Baldasso, R.P.; Moraes, C.; Gallardo, E.; Stumvoll, M.B.; Crespo, K.C.; Strapasson, R.A.P.; de Oliveira, R.N. 3D forensic facial approximation: Implementation protocol in a forensic activity. J. Forensic Sci. 2021, 66, 383–388. [Google Scholar] [CrossRef]
  28. Jayakrishnan, J.M.; Kumar, R.B.V. Forensic facial reconstruction using CBCT: A systematic review. Stomatologija 2022, 24, 49–55. [Google Scholar]
  29. Marić, J.; Bašić, Ž.; Jerković, I.; Mihanović, F.; Anđelinović, Š.; Kružić, I. Facial reconstruction of mummified remains of Christian Saint-Nicolosa Bursa. J. Cult. Herit. 2020, 42, 249–254. [Google Scholar] [CrossRef]
  30. Miranda, G.E.; Wilkinson, C.; Roughley, M.; Beaini, T.L.; Melani, R.F.H. Assessment of accuracy and recognition of three-dimensional computerized forensic craniofacial reconstruction. PLoS ONE 2018, 13, e0196770. [Google Scholar] [CrossRef]
  31. Stanciu, N.V.; Rosculet, R.T.; Fetecau, C.; Tapu, C. Forensic Facial Reconstruction Using 3D Printing. Mater. Plast. 2020, 57, 248–257. [Google Scholar] [CrossRef]
  32. Donato, L.; Cecchi, R.; Goldoni, M.; Ubelaker, D.H. Photogrammetry vs CT Scan: Evaluation of Accuracy of a Low-Cost Three-Dimensional Acquisition Method for Forensic Facial Approximation. J. Forensic Sci. 2020, 65, 1260–1265. [Google Scholar] [CrossRef]
  33. Lussu, P.; Marini, E. Ultra close-range digital photogrammetry in skeletal anthropology: A systematic review. PLoS ONE 2020, 15, e0230948. [Google Scholar] [CrossRef]
  34. Omari, R.; Hunt, C.; Coumbaros, J.; Chapman, B. Virtual anthropology? Reliability of three-dimensional photogrammetry as a forensic anthropology measurement and documentation technique. Int. J. Leg. Med. 2021, 135, 939–950. [Google Scholar] [CrossRef]
  35. Santoro, V.; Lubelli, S.; De Donno, A.; Inchingolo, A.; Lavecchia, F.; Introna, F. Photogrammetric 3D skull/photo superimposition: A pilot study. Forensic Sci. Int. 2017, 273, 168–174. [Google Scholar] [CrossRef]
  36. Agisoft Metashape. Version 1.8.1. Available online: https://www.agisoft.com/ (accessed on 10 August 2021).
  37. Amorim, P.; Moraes, T.; Silva, J.; Pedrini, H. InVesalius: An Interactive Rendering Framework for Health Care Support. In Advances in Visual Computing; Version 3.1; Bebis, G., Boyle, R., Parvin, B., Koracin, D., Pavlidis, I., Feris, R., McGraw, T., Elendt, M., Kopper, R., Ragan, E., et al., Eds.; Springer: Cham, Switzerland, 2015; pp. 45–54. [Google Scholar]
  38. Meshroom. Version 2021.1.0. Available online: https://alicevision.org/ (accessed on 10 August 2021).
  39. Kogan, I.; Rucki, M.; Jähne, M.; Passos, D.E.; Cvjetkovic, T.; Schmidt, S. One Head, many Approaches–Comparing 3D Models of a Fossil Skull. In Photogrammetrie–Laserscanning–Optische 3D-Messtechnik: Beiträge der Oldenburger 3D-Tage; Luhmann, T., Schumacher, C., Eds.; Wichmann: Berlin, Germany, 2020; pp. 22–31. [Google Scholar]
  40. Petleshkova, T.; Manev, H.; Sivkov, S.; Timonov, P.; Baltadjiev, A.; Raycheva, R. Study of the facial morphology of young bulgarians with 3D laser scan. Trakia J. Sci. 2019, 17, 103–108. [Google Scholar] [CrossRef]
  41. Pfeuffer, C. 4. 3D-Modellierung von Skulpturen mit Laserscan und die quantitative Erfassung der 3D-Flächenmaße von Verwitterungsphänomenen. In Risk Number. Determining and Objectively Assessing Environmental Damage on Marble and Sandstone Sculptures; Drewello, R., Ed.; University of Bamberg Press: Bamberg, Germany, 2018; pp. 43–64. [Google Scholar]
  42. Toneva, D.; Nikolova, S.; Georgiev, I.; Tchorbadjieff, A. Accuracy of linear craniometric measurements obtained from laser scanning created 3D models of dry skulls. In Advanced Computing in Industrial Mathematics; Springer: Cham, Switzerland, 2017; pp. 215–229. [Google Scholar]
  43. Blender. Used for Reconstruction (v2.7, 2021) and Rendering (v3.4, 2024). Available online: https://www.blender.org/ (accessed on 10 August 2021).
  44. Zanatta, A.; Bezzi, L.; Carrara, N.; Moraes, C.; Thiene, G.; Zampieri, F. New technique in facial reconstruction: The case of Giovanni Battista Morgagni. Anthropol. Anz. 2018, 75, 131–140. [Google Scholar] [CrossRef] [PubMed]
  45. Rynn, C.; Wilkinson, C.M.; Peters, H.L. Prediction of nasal morphology from the skull. Forensic Sci. Med. Pathol. 2010, 6, 20–34. [Google Scholar] [CrossRef] [PubMed]
  46. Stephan, C.N.; Henneberg, M.; Sampson, W. Predicting nose projection and pronasale position in facial approximation: A test of published methods and proposal of new guidelines. Am. J. Phys. Anthropol. Off. Publ. Am. Assoc. Phys. Anthropol. 2003, 122, 240–250. [Google Scholar] [CrossRef]
  47. Stephan, C.; Murphy, S. Mouth width prediction in craniofacial identification: Cadaver tests of four recent methods, including two techniques for edentulous skulls. J. Forensic-Odonto-Stomatol. 2008, 27, 2–7. [Google Scholar]
  48. Kim, Y.S.; Lee, W.J.; Yun, J.S.; Kim, D.H.; Lozanoff, S.; Lee, U.Y. Predicting the eyebrow from the orbit using three-dimensional CT imaging in the application of forensic facial reconstruction and identification. Sci. Rep. 2023, 13, 4014. [Google Scholar] [CrossRef]
  49. Fugiwara, F.V.G.; Vassallo, F.N.S.; Moritsugui, D.S.; Mazzilli, L.E.N.; Melani, R.F.H. Forensic facial reconstruction: An anthropometric study of the ear. Aust. J. Forensic Sci. 2024, 57, 142–153. [Google Scholar] [CrossRef]
  50. Kim, D.H.; Yang, S.C.; Kim, H.; Lee, S.S.; Kim, Y.S.; Lozanoff, S.; Kwak, D.S.; Lee, U.Y. Regression analysis of nasal shape from juvenile to adult ages for forensic facial reconstruction. Leg. Med. 2024, 66, 102363. [Google Scholar] [CrossRef]
  51. Becker, S.; Dreßler, J.; Thiele, K.H.; Labudde, D. Gesichtsweichteilrekonstruktion mithilfe einer Open-Source-Software. Rechtsmedizin 2016, 26, 83–89. [Google Scholar] [CrossRef]
  52. Gupta, S.; Gupta, V.; Vij, H.; Vij, R.; Tyagi, N. Forensic Facial Reconstruction: The Final Frontier. J. Clin. Diagn. Res. 2015, 9, 26–28. [Google Scholar] [CrossRef]
  53. Becker, S.; Rosenfelder, J.; Labudde, D. Forensische Gesichtsweichteilrekonstruktion einer historischen Person: So sahen unsere Vorfahren aus! NextGen Sci. Rev. 2023, 36–46. [Google Scholar] [CrossRef]
  54. Burrath, S. Visuelle Personenidentifizierung und Polizeiliche Personenbeschreibung: Praxishandbuch; Verlag für Polizeiwissenschaft: Frankfurt am Main, Germany, 2009. [Google Scholar]
  55. FaceDB. Available online: https://www.face-rec.org/databases/ (accessed on 17 August 2021).
  56. Krogman, W.; Iscan, N. Radiographic analysis. In The Human Skeleton in Forensic Medicine; Charles C Thomas: Springfield, IL, USA, 1986; pp. 82–126. [Google Scholar]
  57. Sassouni, V. Palatoprint, Physioprint, and Roentgenographic Cephalometry, as New Methods in Human Identification: Preliminary Report; Callaghan: Wildwood, MI, USA, 1957. [Google Scholar]
  58. Verhoff, M.A. Forensische Osteologie: Problematische Fragestellungen; Lehmanns Media: Cologne, Germany, 2008. [Google Scholar]
  59. Becker, S.; Spranger, M.; Heinke, F.; Grunert, S.; Labudde, D. A Comprehensive Framework for High Resolution Image-based 3D Modeling and Documentation of Crime Scenes and Disaster Sites. Int. J. Adv. Syst. Meas 2018, 11, 1–12. [Google Scholar]
  60. Stageonair. Available online: https://www.stageonair.com/ (accessed on 10 August 2021).
  61. Kollmann, J.; Büchly, W. Die Persistenz der Rassen und die Reconstruction der Physiognomie Prähistorischer Schädel; Archiv f. Anthrop.: Braunschweig, Germany, 1898. [Google Scholar]
  62. Simpson, E.; Henneberg, M. Variation in soft-tissue thicknesses on the human face and their relation to craniometric dimensions. Am. J. Phys. Anthropol. Off. Publ. Am. Assoc. Phys. Anthropol. 2002, 118, 121–133. [Google Scholar] [CrossRef]
  63. Werner, G. Weichteilverschiebung im Gesicht bei Personen in Stehender und Liegender Position. Ph.D. Thesis, Universität Leipzig, Leipzig, Germany, 2009. [Google Scholar]
  64. Rhine, J.S.; Moore, C.E.; Weston, J. Facial Reproduction: Tables of Facial Tissue Thicknesses of American Caucasoids in Forensic Anthropology; Maxwell Museum of Anthroplogy, The University of New Mexico: Albuquerque, NM, USA, 1982. [Google Scholar]
  65. Baillie, L.J.; Mirijali, S.A.; Niven, B.E.; Blyth, P.; Dias, G.J. Ancestry and BMI influences on facial soft tissue depths for a cohort of Chinese and Caucasoid women in Dunedin, New Zealand. J. Forensic Sci. 2015, 60, 1146–1154. [Google Scholar] [CrossRef]
  66. Baillie, L.J.; Muirhead, J.C.; Blyth, P.; Niven, B.E.; Dias, G.J. Position effect on facial soft tissue depths: A sonographic investigation. J. Forensic Sci. 2016, 61, S60–S70. [Google Scholar] [CrossRef]
  67. Taylor, K.T. Forensic Art and Illustration; CRC Press: Boca Raton, FL, USA, 2000. [Google Scholar]
  68. Stephan, C.; Henneberg, M. Predicting mouth width from inter-canine width-a 75% rule. J. Forensic Sci. 2003, 48, 725–727. [Google Scholar] [CrossRef]
  69. Kersten, T.P.; Tschirschwitz, F.; Deggim, S.; Lindstaedt, M. Virtual reality for cultural heritage monuments–from 3d data recording to immersive visualisation. In Proceedings of the Euro-Mediterranean Conference, Nicosia, Cyprus, 29 October–3 November 2018; Springer: Cham, Switzerland, 2018; pp. 74–83. [Google Scholar]
  70. Rahaman, H.; Champion, E.; Bekele, M. From photo to 3D to mixed reality: A complete workflow for cultural heritage visualisation and experience. Digit. Appl. Archaeol. Cult. Herit. 2019, 13, e00102. [Google Scholar] [CrossRef]
  71. Macho, G.A. An appraisal of plastic reconstruction of the external nose. J. Forensic Sci. 1986, 31, 1391–1403. [Google Scholar] [CrossRef]
  72. Shi, J.; Samal, A.; Marx, D. How effective are landmarks and their geometry for face recognition? Comput. Vis. Image Underst. 2006, 102, 117–133. [Google Scholar] [CrossRef]
  73. Nagamalla, S.; Dhara, B.C. A novel face recognition method using facial landmarks. In Proceedings of the 2009 Seventh International Conference on Advances in Pattern Recognition, Kolkata, India, 4–6 February 2009; IEEE: New York, NY, USA, 2009; pp. 445–448. [Google Scholar]
  74. Kim, I.; Han, S.; Park, S.J.; Baek, J.W.; Shin, J.; Han, J.J.; Choi, C. Discface: Minimum discrepancy learning for deep face recognition. In Proceedings of the Asian Conference on Computer Vision, Kyoto, Japan, 30 November–4 December 2020. [Google Scholar]
  75. Zhang, D.; Li, J.; Shan, Z. Implementation of Dlib deep learning face recognition technology. In Proceedings of the 2020 International Conference on Robots & Intelligent System (ICRIS), Sanya, China, 7–8 November 2020; IEEE: New York, NY, USA, 2020; pp. 88–91. [Google Scholar]
  76. Geitgey, A. Face recognition documentation. Release 2019, 1, 3–37. [Google Scholar]
  77. Parkhi, O.M.; Vedaldi, A.; Zisserman, A. Deep Face Recognition. In Proceedings of the British Machine Vision Conference, Swansea, UK, 7–10 September 2015. [Google Scholar]
  78. Alansari, M.; Hay, O.A.; Javed, S.; Shoufan, A.; Zweiri, Y.; Werghi, N. GhostFaceNets: Lightweight Face Recognition Model From Cheap Operations. IEEE Access 2023, 11, 35429–35446. [Google Scholar] [CrossRef]
  79. Jocher, G.; Qiu, J.; Chaurasia, A. Ultralytics YOLO, 2023. Version 8.0.0. Available online: https://github.com/ultralytics/ultralytics (accessed on 7 May 2025).
  80. King, D.E. CNN Face Detection Using Dlib (Version 19.3.0). 2017. Dlib Version 19.3.0Was Used in This Work. Available online: https://dlib.net/cnn_face_detector.py.html (accessed on 7 May 2025).
  81. Khan, S.S.; Sengupta, D.; Ghosh, A.; Chaudhuri, A. MTCNN++: A CNN-based face detection algorithm inspired by MTCNN. Vis. Comput. 2024, 40, 899–917. [Google Scholar] [CrossRef]
  82. Preuß, S.; Labudde, D. Einsatz von Algorithmen zur Gesichtserkennung in Kipo Verfahren—Vorstellung Einer Pipeline zur Automatisierten Verarbeitung; Polizei-Informatik 2023; Honekamp, W., Fähndrich, J., Eds.; Rediroma Verlag: Remscheid, Germany, 2023. [Google Scholar]
  83. O’shea, K.; Nash, R. An introduction to convolutional neural networks. arXiv 2015, arXiv:1511.08458. [Google Scholar]
  84. King, D.E. Face Recognition Using Dlib (Version 19.3.0). 2017. Dlib Version 19.3.0 Was Used. Available online: https://dlib.net/face_recognition.py.html (accessed on 7 May 2025).
  85. Koestinger, M.; Wohlhart, P.; Roth, P.M.; Bischof, H. Annotated facial landmarks in the wild: A large-scale, real-world database for facial landmark localization. In Proceedings of the 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops), Barcelona, Spain, 6–13 November 2011; IEEE: New York, NY, USA, 2011; pp. 2144–2151. [Google Scholar]
  86. Huang, G.B.; Mattar, M.; Berg, T.; Learned-Miller, E. Labeled Faces in the Wild: A Database forStudying Face Recognition in Unconstrained Environments. In Proceedings of the Workshop on Faces in ‘Real-Life’ Images: Detection, Alignment, and Recognition, Erik Learned-Miller and Andras Ferencz and Frédéric Jurie, Marseille, France, 17 October 2008. [Google Scholar]
  87. Serengil, S.; Ozpinar, A. A Benchmark of Facial Recognition Pipelines and Co-Usability Performances of Modules. J. Inf. Technol. 2024, 17, 95–107. [Google Scholar] [CrossRef]
  88. Presse-Agentur, D. Freden: Erneuter DNA-Abgleich: Unbekannter Toter Ist Identifiziert. 2022. Available online: https://www.zeit.de/news/2022-03/23/erneuter-dna-abgleich-unbekannter-toter-ist-identifiziert (accessed on 6 May 2025).
  89. Galea, C.; Farrugia, R.A. Matching Software-Generated Sketches to Face Photographs with a Very Deep CNN, Morphed Faces, and Transfer Learning. IEEE Trans. Inf. Forensics Secur. 2018, 13, 1421–1431. [Google Scholar] [CrossRef]
  90. Preuß, S.; Labudde, D. Digitale Fortschritte in der Gesichtsweichteilrekonstruktion: Von manuellen Methoden zu Künstlicher Intelligenz. In 24. Nachwuchswissenschaftler: Innenkonferenz; Hochschule Mittweida: Mittweida, Germany, 2024; pp. 266–271. [Google Scholar]
Figure 1. Procedure and results of the manual facial soft tissue reconstruction of the unknown decedent from Case 1. The following illustrations show the procedure and results of manual facial soft tissue reconstruction performed by Prof. Helmer in 1994 (the illustrations are own figures). (A) Illustration of the skull, reconstructed from CT images. (B) Reconstructed soft tissue of the unknown deceased. The illustration was also based on CT images. (C) The final reconstruction was further individualized with hairstyle and clothing. The illustration shows a 3D model that has been created with the help of photogrammetric methods.
Figure 1. Procedure and results of the manual facial soft tissue reconstruction of the unknown decedent from Case 1. The following illustrations show the procedure and results of manual facial soft tissue reconstruction performed by Prof. Helmer in 1994 (the illustrations are own figures). (A) Illustration of the skull, reconstructed from CT images. (B) Reconstructed soft tissue of the unknown deceased. The illustration was also based on CT images. (C) The final reconstruction was further individualized with hairstyle and clothing. The illustration shows a 3D model that has been created with the help of photogrammetric methods.
Applsci 15 06086 g001
Figure 2. Process of the computer-aided forensic facial soft tissue reconstruction by means of an exemplary case. The above figure shows the entire process of computer-aided forensic facial soft tissue reconstruction. It is divided into the digitization of the underlying skeletal material, information acquisition, the modeling process, and plausibility checks. Source: [53] modified.
Figure 2. Process of the computer-aided forensic facial soft tissue reconstruction by means of an exemplary case. The above figure shows the entire process of computer-aided forensic facial soft tissue reconstruction. It is divided into the digitization of the underlying skeletal material, information acquisition, the modeling process, and plausibility checks. Source: [53] modified.
Applsci 15 06086 g002
Figure 3. Calculation of the volume model from the CT data of Case 1 in the software application InVesalius [37]. Display of the three slices: axial (A), sagittal (B), and coronal (C), as well as the calculated volume model (D) of the skull in the software application InVesalius using version 3.1.1. Possible artifacts were removed in post-processing steps.
Figure 3. Calculation of the volume model from the CT data of Case 1 in the software application InVesalius [37]. Display of the three slices: axial (A), sagittal (B), and coronal (C), as well as the calculated volume model (D) of the skull in the software application InVesalius using version 3.1.1. Possible artifacts were removed in post-processing steps.
Applsci 15 06086 g003
Figure 4. Volume model from the CT data of Case 1 in the Blender software application. The display of the volume model, calculated from the CT data in the Blender software application, is shown with artifacts (A) and after the post-processing steps (B). The latter model formed the basis for the actual soft tissue reconstruction of the face.
Figure 4. Volume model from the CT data of Case 1 in the Blender software application. The display of the volume model, calculated from the CT data in the Blender software application, is shown with artifacts (A) and after the post-processing steps (B). The latter model formed the basis for the actual soft tissue reconstruction of the face.
Applsci 15 06086 g004
Figure 5. Principle of photogrammetry and experimental setup for photogrammetric photographs of the skull of the victim in Case 2. (A1,A2) The three-dimensional structure of an object can be calculated using appropriate software applications based on overlapping photographs from different perspectives. The photographs themselves were taken on test setups, as shown in sub-figures (B,C). (B,C) show the positioning of the skull on a turntable. Three Nikon D7100 and D7500 single-lens reflex cameras with a simultaneous shutter release and three NG-220A continuous light studio lamps from NANGUANG were employed for the photographic sessions. The skull was captured a total of 90 times, with one image being taken at a 4° angle. Consequently, 270 images were generated per session (90 images per camera).
Figure 5. Principle of photogrammetry and experimental setup for photogrammetric photographs of the skull of the victim in Case 2. (A1,A2) The three-dimensional structure of an object can be calculated using appropriate software applications based on overlapping photographs from different perspectives. The photographs themselves were taken on test setups, as shown in sub-figures (B,C). (B,C) show the positioning of the skull on a turntable. Three Nikon D7100 and D7500 single-lens reflex cameras with a simultaneous shutter release and three NG-220A continuous light studio lamps from NANGUANG were employed for the photographic sessions. The skull was captured a total of 90 times, with one image being taken at a 4° angle. Consequently, 270 images were generated per session (90 images per camera).
Applsci 15 06086 g005
Figure 6. Exemplary photogrammetric images of the skull of the victim in Case 2, more precisely, the cranium, skullcap, and lower jaw. Presentation of selected photogrammetric photographs of the skull of the victim in Case 2, specifically the cranium, skullcap, and lower jaw from different perspectives, using continuous light studio lights to create a 3D model of the bony structures described. (A1A3) Slightly oblique image from above with almost a frontal image, as well as a slightly oblique image from below the cranium. (B1B3) Slightly oblique image from above with almost a frontal image, as well as a slightly oblique image from below the skullcap. (C1C3) Slightly oblique image from above with almost a frontal image, as well as a slightly oblique image from below the mandible.
Figure 6. Exemplary photogrammetric images of the skull of the victim in Case 2, more precisely, the cranium, skullcap, and lower jaw. Presentation of selected photogrammetric photographs of the skull of the victim in Case 2, specifically the cranium, skullcap, and lower jaw from different perspectives, using continuous light studio lights to create a 3D model of the bony structures described. (A1A3) Slightly oblique image from above with almost a frontal image, as well as a slightly oblique image from below the cranium. (B1B3) Slightly oblique image from above with almost a frontal image, as well as a slightly oblique image from below the skullcap. (C1C3) Slightly oblique image from above with almost a frontal image, as well as a slightly oblique image from below the mandible.
Applsci 15 06086 g006
Figure 7. Calculated individual models of the cranium, skull, and mandible and their combination to form an overall model of the victim in Case 2. Illustration of the calculated 3D models for the cranium, skull, and mandible (A1A3). The models were calculated in the software application Meshroom (Version 2021.1.0). The individual models were combined into an overall model (B) in the Blender software application.
Figure 7. Calculated individual models of the cranium, skull, and mandible and their combination to form an overall model of the victim in Case 2. Illustration of the calculated 3D models for the cranium, skull, and mandible (A1A3). The models were calculated in the software application Meshroom (Version 2021.1.0). The individual models were combined into an overall model (B) in the Blender software application.
Applsci 15 06086 g007
Figure 8. Three-dimensional model of the skull of the victim in Case 1 with applied anatomical soft tissue markers. Display of the anatomical soft tissue markers applied to the skull in Case 1 according to [13] based on male adults aged 30 to 39 years. The soft tissue markers with the specific soft tissue thicknesses were applied to the skull using the Blender software application. (A) Frontal view of the skull. (B) Slightly lateral view of the left side of the skull.
Figure 8. Three-dimensional model of the skull of the victim in Case 1 with applied anatomical soft tissue markers. Display of the anatomical soft tissue markers applied to the skull in Case 1 according to [13] based on male adults aged 30 to 39 years. The soft tissue markers with the specific soft tissue thicknesses were applied to the skull using the Blender software application. (A) Frontal view of the skull. (B) Slightly lateral view of the left side of the skull.
Applsci 15 06086 g008
Figure 9. Three-dimensional model of the skull of the victim in Case 2 with applied anatomical soft tissue markers. Representation of the anatomical soft tissue markers applied to the skull of the victim in Case 2 according to [13] based on male adults aged 40 to 49 years. The soft tissue markers with the specific soft tissue thicknesses were applied to the skull using the Blender software application. (A) Frontal view of the skull. (B) Slightly lateral view of the left side of the skull.
Figure 9. Three-dimensional model of the skull of the victim in Case 2 with applied anatomical soft tissue markers. Representation of the anatomical soft tissue markers applied to the skull of the victim in Case 2 according to [13] based on male adults aged 40 to 49 years. The soft tissue markers with the specific soft tissue thicknesses were applied to the skull using the Blender software application. (A) Frontal view of the skull. (B) Slightly lateral view of the left side of the skull.
Applsci 15 06086 g009
Figure 10. Reconstruction of soft tissue and facial features of the victim in Case 1. Illustration of the soft tissue reconstruction based on anatomical soft tissue markers applied to the skull of the victim in Case 1 (A) and fitting of the eyeballs. The measurements taken on the skull and methods for reconstructing selected facial features were used to model the nose and mouth of the individual in question from 1993. Auxiliary lines were drawn for the reconstruction of the facial features. (B) White outline shows auxiliary lines for the reconstruction of the nose based on the corresponding method selected. This ensures that the relationship between the skull and the modeled soft tissue is maintained at all times during the reconstruction process. Once the soft tissue and facial features (C) have been modeled, further process steps follow, such as the modeling of age-specific features.
Figure 10. Reconstruction of soft tissue and facial features of the victim in Case 1. Illustration of the soft tissue reconstruction based on anatomical soft tissue markers applied to the skull of the victim in Case 1 (A) and fitting of the eyeballs. The measurements taken on the skull and methods for reconstructing selected facial features were used to model the nose and mouth of the individual in question from 1993. Auxiliary lines were drawn for the reconstruction of the facial features. (B) White outline shows auxiliary lines for the reconstruction of the nose based on the corresponding method selected. This ensures that the relationship between the skull and the modeled soft tissue is maintained at all times during the reconstruction process. Once the soft tissue and facial features (C) have been modeled, further process steps follow, such as the modeling of age-specific features.
Applsci 15 06086 g010
Figure 11. Reconstruction of the soft tissue and facial features of the victim in Case 2. Illustration of the soft tissue reconstruction based on anatomical soft tissue markers applied to the skull of the victim in Case 2 (A) and fitting of the eyeballs. The measurements taken on the skull and methods for reconstructing selected facial features were used to model the nose and mouth of the 2021 victim in Case 2 (B). Once the soft tissue and facial features had been modeled, further process steps followed, such as the modeling of age-specific features.
Figure 11. Reconstruction of the soft tissue and facial features of the victim in Case 2. Illustration of the soft tissue reconstruction based on anatomical soft tissue markers applied to the skull of the victim in Case 2 (A) and fitting of the eyeballs. The measurements taken on the skull and methods for reconstructing selected facial features were used to model the nose and mouth of the 2021 victim in Case 2 (B). Once the soft tissue and facial features had been modeled, further process steps followed, such as the modeling of age-specific features.
Applsci 15 06086 g011
Figure 12. Results of facial soft tissue reconstruction of the victim in Case 1. Presentation of the results of the facial soft tissue reconstruction based on the underlying information on the biological profile and external characteristics of the unknown individual in Case 1. The age characteristics were based on research in corresponding facial databases. The so-called zeitgeist was taken into account in the choice of hairstyles. In the available files, Prof. Helmer mentioned a statement about possible scarring below the left eye, which was modeled as an example. As no information about the eye color was available, models with a light and dark eye color were created. The partial figures (A1A3,D1D3) show a model in frontal, slightly lateral, and lateral perspectives with longer hair and dark and light eye color. The partial figures (B1B3,E1E3) show a model in frontal, slightly lateral, and lateral perspectives with shorter hair combed backwards and dark and light eye color. The partial figures (C1C3,F1F3) show a model in frontal, slightly lateral, and lateral perspectives with shorter hair combed backwards and dark and light eye color with a scar under the left eye (according to Prof. Helmer’s statements). For a better overview, partial figure (G) shows one of the models (B1) in a large format.
Figure 12. Results of facial soft tissue reconstruction of the victim in Case 1. Presentation of the results of the facial soft tissue reconstruction based on the underlying information on the biological profile and external characteristics of the unknown individual in Case 1. The age characteristics were based on research in corresponding facial databases. The so-called zeitgeist was taken into account in the choice of hairstyles. In the available files, Prof. Helmer mentioned a statement about possible scarring below the left eye, which was modeled as an example. As no information about the eye color was available, models with a light and dark eye color were created. The partial figures (A1A3,D1D3) show a model in frontal, slightly lateral, and lateral perspectives with longer hair and dark and light eye color. The partial figures (B1B3,E1E3) show a model in frontal, slightly lateral, and lateral perspectives with shorter hair combed backwards and dark and light eye color. The partial figures (C1C3,F1F3) show a model in frontal, slightly lateral, and lateral perspectives with shorter hair combed backwards and dark and light eye color with a scar under the left eye (according to Prof. Helmer’s statements). For a better overview, partial figure (G) shows one of the models (B1) in a large format.
Applsci 15 06086 g012
Figure 13. Results of the facial soft tissue reconstruction of the victim in Case 2. Presentation of the results of the facial soft tissue reconstruction based on the underlying information on the biological profile and external characteristics of the unknown individual in Case 2. The age characteristics were determined based on research in corresponding facial databases. The hairstyles, as well as hair and eye color, were modeled based on the information provided. (AC) show a model in frontal, slightly lateral, and lateral perspectives with short dark hair and dark eye color.
Figure 13. Results of the facial soft tissue reconstruction of the victim in Case 2. Presentation of the results of the facial soft tissue reconstruction based on the underlying information on the biological profile and external characteristics of the unknown individual in Case 2. The age characteristics were determined based on research in corresponding facial databases. The hairstyles, as well as hair and eye color, were modeled based on the information provided. (AC) show a model in frontal, slightly lateral, and lateral perspectives with short dark hair and dark eye color.
Applsci 15 06086 g013
Table 1. Comprehensive comparison of classical manual and modern computer-assisted methods used in forensic facial soft tissue reconstruction, highlighting their procedural types, methodological advantages, and inherent limitations. The table includes traditional sculptural techniques (American, Russian, and Manchester) as well as digital 2D and 3D approaches such as drawing, overlay, CAD modeling, and morphing algorithms.
Table 1. Comprehensive comparison of classical manual and modern computer-assisted methods used in forensic facial soft tissue reconstruction, highlighting their procedural types, methodological advantages, and inherent limitations. The table includes traditional sculptural techniques (American, Russian, and Manchester) as well as digital 2D and 3D approaches such as drawing, overlay, CAD modeling, and morphing algorithms.
MethodTypeAdvantagesDisadvantages
American MethodManual, analog (sculptural)Objective and reproducibleRelies heavily on the accuracy of reference depth tables; loss of skull; 3D replica required for modeling; skill-dependant
Russian MethodManual, analog (sculptural)Anatomically detailed and expressiveSubjective and non-standardized; requires expert anatomical and anthropological knowledge; loss of skull; 3D replica required for modeling; skill-dependant
Manchester MethodManual, analog (sculptural)Combines objectivity of depth markers with individuality of muscular structureInherits limitations of both muscle-based and depth-based methods; loss of skull; 3D replica required for modeling; skill-dependant
Digital DrawingManual, digital (2D illustration)Non-invasive (skull remains intact); fast and low-cost; easy to distributeNo depth info; skill-dependent; limited anatomical accuracy
Superimposition TechniqueManual, digital (2D overlay)Non-invasive (skull remains intact); requires minimal resourcesNeeds antemortem photos; subjective comparison; not a full reconstruction
3D Sculpting using CAD Software (e.g., Blender)Manual/semi-automated, digital (sculptural)Non-invasive (skull remains intact); faster than manual methods; partially automatableSuffers from the same conceptual limitations as its underlying manual approach
3D Morphing from DatabasesSemi-/fully automated (data-driven morphing)Non-invasive (skull remains intact); fast and data-driven; less subjective than manual methodsMay lack individual-specific traits; dependent on dataset quality and diversity; output may appear generalized
Table 2. Measurement of the bony structures on the skull of the victim in Case 1 for the reconstruction of facial features.
Table 2. Measurement of the bony structures on the skull of the victim in Case 1 for the reconstruction of facial features.
DistanceLength in [mm]
Nasal cavity width25.81
Nasion-Subspinale63.67
Nasion-Acanthion56.68
Rhinion-Subspinale45.46
Premolare distance32.40
Molare distance42.11
Table 3. Calculations for the reconstruction of the nose according to [45] for the victim in Case 1.
Table 3. Calculations for the reconstruction of the nose according to [45] for the victim in Case 1.
DistanceLength in [mm]
Pronasale ant34.23
Pronasalae ver49.01
Pronasale pFHP36.28
Nasal length50.61
Nasal height59.16
Nasal depth23.18
Nose width (according to Gerasimov)43.02
Table 4. Calculations for the reconstruction of the mouth according to [68] for the victim in Case 1.
Table 4. Calculations for the reconstruction of the mouth according to [68] for the victim in Case 1.
DistanceLength in [mm]
Mouth width42.11
Table 5. Measurement of the bony structures on the skull of the victim in Case 2 for the reconstruction of facial features.
Table 5. Measurement of the bony structures on the skull of the victim in Case 2 for the reconstruction of facial features.
DistanceLength in [mm]
Nasal cavity width22.31
Nasion-Subspinale62.99
Nasion-Acanthion57.91
Rhinion-Subspinale42.08
Inter-Canine distance37.34
Table 6. Calculations for the reconstruction of the nose according to [45] for the victim in Case 2.
Table 6. Calculations for the reconstruction of the nose according to [45] for the victim in Case 2.
DistanceLength in [mm]
Pronasale ant31.43
Pronasalae ver50.12
Pronasale pFHP33.13
Nasal length50.11
Nasal height58.63
Nasal depth21.83
Nose width (according to Gerasimov)37.18
Table 7. Calculations for the reconstruction of the mouth according to [68] for the victim in Case 2.
Table 7. Calculations for the reconstruction of the mouth according to [68] for the victim in Case 2.
DistanceLength in [mm]
Mouth width49.79
Table 8. Overview of the methodological steps and associated software tools used for the computer-aided forensic facial soft tissue reconstruction. The table includes specific methods, relevant literature sources, and custom-developed Blender add-ons applied throughout the reconstruction process.
Table 8. Overview of the methodological steps and associated software tools used for the computer-aided forensic facial soft tissue reconstruction. The table includes specific methods, relevant literature sources, and custom-developed Blender add-ons applied throughout the reconstruction process.
StepSoftwareMethod and Literature
Digitalization of the mortal remainsCase 1: InVesalius v3.1 [37]Bone Segmentation: Image Threshold Segmentation [37]
Mesh Generation: Marching Cubes [37]
Case 2: Meshroom v2021.1.0 [38]Photogrammetry
Artifact removalBlender v2.7 [43]Manual Removal (Blender Edit Mode)
Placement of soft tissue markersBlender v2.7 [43] (Custom Add-on: Enchanted Soft Tissue Marker (ESTM))Study by S. De Greef et al. 2006 [13]
Reconstruction of musculatureBlender v2.7 [43]Manual Sculpting (Blender Sculpt Mode)
Positioning of eyeballsBlender v2.7 [43] (Custom Add-on: Eye Add-on)According to C. Wilkinson and K. Taylor [2,67]
Mouth ReconstructionBlender v2.7 [43] (Custom Add-on: Mouth Editor)According to C.N. Stephan [68]
Nose ReconstructionBlender v2.7 [43] (Custom Add-on: Nose Add-on)According to C. Rynn and Gerasimov [45]
Hair ModelingBlender v2.7 [43]Blender Hair Particle System
RenderingBlender v3.4 [43]Blender Renderer
Table 9. Normalized similarity scores for two test images (Image 1: frontal passport photo; Image 2: slightly lateral view) across three face recognition models. Both cosine similarity and Euclidean distance were evaluated. All scores are normalized such that a value of 1.0 corresponds to the respective model-specific decision threshold. Note that Dlib does not support cosine similarity; therefore, entries are marked as not applicable (n.a.). Smaller values correspond to higher degrees of facial similarity.
Table 9. Normalized similarity scores for two test images (Image 1: frontal passport photo; Image 2: slightly lateral view) across three face recognition models. Both cosine similarity and Euclidean distance were evaluated. All scores are normalized such that a value of 1.0 corresponds to the respective model-specific decision threshold. Note that Dlib does not support cosine similarity; therefore, entries are marked as not applicable (n.a.). Smaller values correspond to higher degrees of facial similarity.
Image 1 dpImage 2 dp
ModelCosineEuclideanCosineEuclidean
VGG-Face1.30271.13961.27591.1382
GhostFaceNet1.29231.16641.48231.2457
Dlibn.a.1.2827n.a.1.2950
Note: n.a. = not applicable.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Preuß, S.; Becker, S.; Rosenfelder, J.; Labudde, D. Computer-Aided Facial Soft Tissue Reconstruction with Computer Vision: A Modern Approach to Identifying Unknown Individuals. Appl. Sci. 2025, 15, 6086. https://doi.org/10.3390/app15116086

AMA Style

Preuß S, Becker S, Rosenfelder J, Labudde D. Computer-Aided Facial Soft Tissue Reconstruction with Computer Vision: A Modern Approach to Identifying Unknown Individuals. Applied Sciences. 2025; 15(11):6086. https://doi.org/10.3390/app15116086

Chicago/Turabian Style

Preuß, Svenja, Sven Becker, Jasmin Rosenfelder, and Dirk Labudde. 2025. "Computer-Aided Facial Soft Tissue Reconstruction with Computer Vision: A Modern Approach to Identifying Unknown Individuals" Applied Sciences 15, no. 11: 6086. https://doi.org/10.3390/app15116086

APA Style

Preuß, S., Becker, S., Rosenfelder, J., & Labudde, D. (2025). Computer-Aided Facial Soft Tissue Reconstruction with Computer Vision: A Modern Approach to Identifying Unknown Individuals. Applied Sciences, 15(11), 6086. https://doi.org/10.3390/app15116086

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop