Next Article in Journal
Vision-Based People Counting and Tracking for Urban Environments
Previous Article in Journal
A Slicer-Independent Framework for Measuring G-Code Accuracy in Medical 3D Printing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hierarchical Multi-Resolution Self-Supervised Framework for High-Fidelity 3D Face Reconstruction Using Learnable Gabor-Aware Texture Modeling

by
Pichet Mareo
1 and
Rerkchai Fooprateepsiri
2,*
1
Business Administration and Information Technology Faculty, Rajamangala University of Technology Tawan-ok, Bangkok 10400, Thailand
2
Innovative Education and Lifelong Learning Institute, Rajamangala University of Technology Tawan-ok, Chonburi 20110, Thailand
*
Author to whom correspondence should be addressed.
J. Imaging 2026, 12(1), 26; https://doi.org/10.3390/jimaging12010026
Submission received: 24 November 2025 / Revised: 26 December 2025 / Accepted: 28 December 2025 / Published: 5 January 2026

Abstract

High-fidelity 3D face reconstruction from a single image is challenging, owing to the inherently ambiguous depth cues and the strong entanglement of multi-scale facial textures. In this regard, we propose a hierarchical multi-resolution self-supervised framework (HMR-Framework), which reconstructs coarse-, medium-, and fine-scale facial geometry progressively through a unified pipeline. A coarse geometric prior is first estimated via 3D morphable model regression, followed by medium-scale refinement using a vertex deformation map constrained by a global–local Markov random field loss to preserve structural coherence. In order to improve fine-scale fidelity, a learnable Gabor-aware texture enhancement module has been proposed to decouple spatial–frequency information and thus improve sensitivity for high-frequency facial attributes. Additionally, we employ a wavelet-based detail perception loss to preserve the edge-aware texture features while mitigating noise commonly observed in in-the-wild images. Extensive qualitative and quantitative evaluation of benchmark datasets indicate that the proposed framework provides better fine-detail reconstruction than existing state-of-the-art methods, while maintaining robustness over pose variations. Notably, the hierarchical design increases semantic consistency across multiple geometric scales, providing a functional solution for high-fidelity 3D face reconstruction from monocular images.

1. Introduction

The use of high-fidelity 3D facial representations has gained prominence in widely publicized human–machine interaction applications such as facial recognition and analysis [1,2,3,4], intelligent medical systems [5,6], and facial animation [7,8]. The most well-known parametric 3D facial model, which is the 3D morphable model (3DMM), was proposed in [9], and its advanced geometric representation property allows the efficient image-based reconstruction of 3D facial geometry [10,11]. However, the recoverability of fine-grained 3D facial structures from a 2D image is still hindered by limited depth cues and spatial information, primarily inherent in 2D images. So far, progress has been made by using a deep learning methodology of regressing 3DMM parameters to improve reconstruction quality. The authors of [12] generated a weakly supervised method where image-level and perceptual losses were computed to recover facial geometry between the input and reconstructed images. Similarly, the authors of [13] proposed a self-supervised framework, which ensures shape consistency across viewpoint, illumination, and occlusion, allowing a more exact recognition of geometric attributes. While these methods can reconstruct coarse facial geometry, they fail to recover high-fidelity, fine-scale features like crow’s feet and subtle skin microstructures. To overcome this limitation, much research has focused on extracting detailed facial information via a prediction of detail maps, which has achieved significant performance gains [14,15,16].
As illustrated in Figure 1, existing approaches exhibit distinct differences in their strategies for modeling and recovering facial geometry across multiple spatial scales, which directly influence their reconstruction capabilities and limitations. Each method has its limitations, however. The authors of [15] required high-quality scans during training, which limited generalization to in-the-wild imagery. The authors of [16] developed UV displacement maps to detect expression-related wrinkles and obtained much better recovery of fine-scale detail; however, they did not restore medium-scale geometric structures like dimples and deep nasolabial folds. The authors of [14], for example, made a prediction for displacement depth maps to reconstruct medium-scale features better, although fine-scale geometry was still suboptimal. Taken together, these shortcomings indicate the challenges of recovering high-dimensional multi-scale geometric features on a single detail map. In this study, a hierarchical multi-resolution framework based-on self-supervised learning (HMR-Framework) is proposed to address these problems. The framework is structured in a three-stage architecture focused on reconstructing large-, medium-, and fine-scale geometric facial information and trained at each stage using a self-supervised strategy. At the first stage, a 3DMM is used as the geometric prior, in which UV-space detail maps are developed slowly toward the fine detail of medium size- and fine size-scaled surfaces.
Since UV texture maps comprise simultaneously multiple multi-scale and multi-frequency levels of information, directly inputting these results into a neural network would hinder the learning of ultra-high frequency structures [17]. This implies that a robust decoupling mechanism is needed to decouple high-frequency textures and model them separately. Inspired by recent evidence of the high spatial-frequency texture extraction power of Gabor filters on fine-grained recognition tasks [18], a learnable Gabor-aware texture enhancement module is proposed. By coupling the learnable Gabor with a couple of constraints, this module promotes co-decoupling between medium- and fine-scale textures (spatial and frequency). These high-frequency features are fused with the UV texture map and computed by a CNN, facilitating the accurate reconstruction of fine-scale facial structures. Due to the serious impacts of skin imperfections and noise in these in-the-wild facial images, we developed detail perception loss (DPLoss) based on wavelet transform to mitigate its negative effects. Wavelet transform [19] allows facial textures to be encoded as sparse coefficients in the wavelet domain, providing noise robustness with the stability provided by preserving important structural information for performing fine-detail reconstructions. Additionally, fine-grained facial features are mainly centralized near the eye and mouth areas; because of that, they are constrained in isolation [20]. A Markov random field loss [21] is developed to enhance the spatial coherence of global and local representations of features. The main contribution of the work presented in this paper is as follows:
  • We propose to model multi-level geometric facial features in the hierarchical mode using the hierarchical multi-resolution framework based on self-supervised learning (HMR-Framework).
  • A learnable Gabor-aware texture enhancement module can be proposed to enhance fine-scale detail reconstruction by a joint spatial-frequency decoupling. This module constitutes the first incorporation of a learnable Gabor-based convolutional layer into the 3D face reconstruction pipeline to enable adaptive learning of high-frequency detail parameters.
  • Global and local Markov random field loss (GL-MRFLoss) and detail perception loss (DPLoss) are proposed to deliver the global–local perceptual guidance and to ensure the structural properties of fine-scale facial features.
Despite recent advances in single-image 3D face reconstruction, existing approaches remain limited in their ability to jointly recover medium- and fine-scale geometric details under a unified self-supervised framework. Methods relying on a single displacement or detail map often suffer from scale entanglement, leading to either over-smoothed reconstructions or unstable high-frequency artifacts. To address these limitations, a hierarchical multi-resolution self-supervised framework (HMR-Framework) is proposed, in which facial geometry is progressively reconstructed from coarse to fine scales. By explicitly decoupling medium- and high-frequency information and introducing spatial–frequency-aware modeling, the proposed approach enables structurally consistent and perceptually faithful 3D facial reconstruction from a single monocular image.

2. Literature Review

2.1. Morphable Model-Based 3D Face Reconstruction (3DMM)

3DMM is considered a reliable geometric prior for 3D face reconstruction, allowing an estimation of 3D face geometry from a single 2D image in a short time. A variety of deep learning methods [22,23] applied deep neural network architectures to directly regress the 3DMM coefficients and obtained impressive precision. As pseudo-ground-truth labels, 3D facial models gained by optimization-guided procedures were implemented in [24,25] to be able to supervise the learning of the convolutional neural networks. The authors of [26] optimized reconstruction accuracy by using synthetic facial images with their 3D scans as training data. While these learning-based techniques made remarkable progress, their performance was limited by the unreliable nature of the training labels, which created data discrepancy between the reconstruction findings and the images involved. Because there are few paired 2D–3D facial datasets out there, self-supervised learning has emerged as an invaluable research avenue. In [27], a modulation-based deformation model was proposed for the self-supervised learning of shape correspondences between images and deformable meshes, leading to higher geometric fidelity. Ref. [12] suggested a weakly supervised hybrid loss to direct reconstruction by comparing the input and reconstructed images. The authors of [28] applied 2D facial landmark heatmaps and constructed four self-supervised learning methods to achieve relatively higher accuracy in 3DMM reconstruction. The study in [29] utilized the cyclic consistency of dynamic 3D facial features as a signal for in-the-wild reconstruction in a self-supervised manner. The authors of [30] decomposed facial features into identity and expression and modeled them separately; this way, the robustness of 3DMM can be improved with expression differences. However, even with the aforementioned progress in facial representation, the expressive ability of the original 3DMM is only restricted to the problem of producing coarse 3D face shapes without fine-grained information. To address this limitation, some extensions have been suggested. In [31], a nonlinear 3DMM was proposed and has a wider resolution to improve the visualization of fine-scale structure. An animatable displacement model using UV displacement maps was performed in [16] to generate expression-related wrinkles. The representational realism of reconstructions was enhanced by separating reflectance and geometric detail attributes and modeling them independently in [32]. On the basis of texture, some static–dynamic decoupling mechanisms were proposed in [15], to estimate static facial details, while dynamic details were detected and predicted using a pix2pixHD network assisted by facial deformation maps. The coarse-to-fine strategy UDL [14] used UV displacement depth maps on a single channel to extend medium-scale geometry representation capability. Single-image reconstruction has evolved a lot, but realism and detail accuracy are still challenging, especially with fine-scale features. Therefore, a hierarchical model structure is introduced to fill this gap that was analyzed in this research. It allows us to refine the geometric details on different levels, as such a model attempts to reconstruct realistic 3D faces. In comparison to the UDL baseline, we extended the two-stage framework to a three-stage framework in which it is actively recommended to decouple mid- and high-frequency details in order to enhance fine-detail perception. Furthermore, two novel loss functions are added and subsequently used to achieve new global–local contour consistency and high-frequency detail recovery in the second and third parts, respectively.

2.2. Gabor Filter

The Gabor filter, originally introduced in [33], has been demonstrated as a reliable estimate of the receptive field of the mammalian visual cortex. Its performance on both spatial and frequency scales is strong, and it also enables joint spatial–frequency analysis, allowing direct control of frequency and orientation. Thus, due to these merits, Gabor filters have been applied in computer vision applications like fingerprint recognition [34], face recognition [35], and age or gender assessment [36]. Moreover, empirical studies have demonstrated that adding Gabor responses as network inputs has a positive effect on performance of convolutional neural networks (CNNs) [37]. Then, the Gabor jet method was devised to capture multiscale and multi-orientation Gabor responses [36]. Most of these methods apply Gabor filtering as a preprocessing step, which complicates the seamless integration by Gabor filtering into the deep architectures. Recently, attempts to embed Gabor filters directly within deep neural networks have been made [38,39,40]. However, such methods generally suffer from training instability and a high computational cost, which undermines their practical usefulness. A relatively stable approach was introduced in [41], where manipulable Gabor layers were inserted in a cascaded network to enhance scale- and orientation-based image decomposition, thereby enhancing the generalization and computational efficiency. More recently proposed here, a learning-based approach for automatic Gabor parameter optimization of fine-grained recognition tasks was introduced in [18]. By setting constraints on the trained parameters, that method adapts to find the parameter combinations most appropriate to the target task. Since the Gabor filters are shown to effectively identify detailed visual information, we believe that they can be developed to support the 3D facial geometry reconstruction. Their excellent spatial–frequency discrimination is promising for improving the prediction of high-frequency detail models as a challenge in single-image 3D face reconstruction.

3. Proposed Frameworks

3.1. Overview

A hierarchical multi-resolution framework based-on self-supervised learning (HMR-Framework) is then presented to reconstruct 3D facial geometry from a single input image with multi-scale detail. The complete pipeline is depicted in Figure 2, which is divided into three successive stages. For the initial step, a trainable VGG-Encoder [42] is used to regress the parameter values of 3DMM and to recover a coarse geometric prior. In the subsequent step, a global–local Markov random field loss (GL-MRFLoss) is computed for the prediction of a three-channel vertex deformation map, which will be the model. Finally, a Gabor-aware texture enhancement module is implemented to achieve a learnable Gabor-aware texture refinement module and joint decoupling of fine-scale textures in spatial and spectral domains, in order to enhance the network better in high-range detection of delicate, high-frequency facial features. In addition, a detail perception loss (DPLoss) is used to forecast UV displacement maps enriched with finer structural features. The vertex deformation map and UV displacement map predicted from the second and third stages are subsequently combined with the coarse geometry prior to yield a fully detailed 3D facial reconstruction. Every intermediate output from the three stages is generated as 2D images, and the whole process is applied to obtain an end-to-end self-supervised learning goal where discrepancies between the 2D images and the input facial data are minimized.

3.2. Large-Resolution Geometry Prior Reconstruction

Since a single-image 3D facial reconstruction is an ill-posed problem, 3DMM is used as an application method for generating a rough geometric representation at the initial reconstruction stage. The initial 3D face, denoted in Figure 2, is thus regressed using a trainable VGG-Encoder [42] to obtain the 3DMM parameters.
S = S α , β = S ¯ + B i d α + B e x p β
T = T δ = T ¯ + B t δ ,
S ¯ and T ¯ are the mean shape and texture, the B i d and B t are identity and texture bases, B e x p are expression bases, and α , β , and δ the necessary parameter vectors. The latent code from just one facial image is fed through the VGG-Encoder to the parameter outputs α R 60 , β R 30 , and δ R 60 . Like in [14], training is self-supervised, in which the reconstructed 3D face is rendered onto an input image plane, and the gap between the obtained and original image is kept as minimal as possible. The total loss is calculated as follows:
L p r o i r = ω 1 · L p i x e l + ω 2 · L l m + ω 3 · L i d + ω 4 · R p a r a m ,
where L p i x e l is the photometric loss, L l m is the landmark consistency loss, L i d is the identity-aware loss, and R p a r a m is a regularization term for the 3DMM parameters. Weight coefficients ω 1 ,   ω 2 ,   ω 3 , ω 4 = {1.3, 1.0, 1.5, 20.0} follow the settings found in [14].
The pipeline is organized into three progressive stages. In the first stage, a trainable encoder regresses 3D morphable model parameters to establish a coarse geometric prior. In the second stage, a three-channel vertex deformation map refines medium-scale geometric structures under global–local Markov random field supervision, ensuring structural coherence across the facial surface. In the third stage, a Gabor-aware texture enhancement module decouples spatial–frequency information to recover fine-scale details, which are further supervised by a wavelet-based detail perception loss. The outputs of all stages are fused with the geometric prior to generate a high-fidelity 3D facial reconstruction.

3.3. Medium-Resolution Geometry Detail Reconstruction

Although the coarse geometric prior in the first stage affords a stable global facial structure, it cannot represent medium-scale geometric features, including nasolabial folds, dimples, and facial surface variations associated with certain expressions, at the middle level of morphology. To overcome the limitation, a medium-resolution reconstruction stage is introduced through the prediction of a three-channel vertex deformation map that sharpens the coarse geometry while preserving structural consistency. It is based on a combination of photometric loss, regularization, and global–local Markov random field loss (GL-MRFLoss) and supervised. The GL-MRFLoss is deliberately designed to create structural uniformity over the face spatially by incorporating compatible global as well as local constraints. The global aspect induces a consistent shape all over the face (to avoid macro-based geometric distortions), while the local aspects emphasize high-detail parts, like the eyes and mouth, where fine structural differences show more of a visual impact. Unlike the GL-MRFLoss, it is not constructed as a monolithic constraint, and we intentionally trade-off the effects of global and regional interactions with weighted combinations, as described in Equations (6)–(8). It guarantees global facial integrity while also allowing the network to selectively enhance medium-scale details in anatomically critical locations (e.g., edges in a cross-section of the facial area). The weighting scheme is empirically established to stabilize the training sequence and to avoid over-focusing on small local areas, which would introduce geometric irregularities. This intermediate refinement phase provides the critical link between coarse geometry estimation and fine-scale texture improvement. The methodology explicitly models medium-scale deformations under strict global–local supervision and provides a well-founded geometric platform on which high-frequency details can be reconstructed.
Since the coarse geometry established in the first stage lacks mid-level structural information, a three-channel vertex deformation map is added to preserve medium-scale geometric aspects. The green highlighted module in Figure 2 depicts the image-to-image translation network [43] for reconstruction. The medium-resolution model is trained with pixel-wise photometric loss, GL-MRFLoss, and a regularization term, arriving at the following:
L m e d i u m = ω p · L p i x e l + ω m r f · L g l _ m r f + ω r · L r e g ,
with weights ω p ,   ω m r f ,   ω g l _ m r f = {1, 1, 0.01}.
Pixel-wise Photometric Loss: The photometric loss calculates the L2 difference in visible pixels between the input image ( I ) and the rendered image ( I R ):
L p i x e l = 1 M V i M V V · I i I i R 2 ,
where ( M V ) represents the visible facial region.
Global and Local Markov Random Field Loss (GL-MRFLoss): Markov random field regularization has been shown to improve detail restoration for image synthesis tasks [21,44]. As a result, GL-MRFLoss aims to ensure global structural coherence and local consistency in high-detail facial regions. The global MRF loss is computed as follows from conv3_2 and conv4_2 on VGG19 following [44], increasing conv3_2 weights as suggested in [45]:
L g _ m r f = 2 · L M ( c o n v 3 _ 2 g l o b a l ) + L M ( c o n v 4 _ 2 g l o b a l ) .
Local MRF losses are derived from cropped UV patches around the mouth and eyes:
L m _ m r f = 2 · L M ( c o n v 3 _ 2 m o u t h ) + L M ( c o n v 4 _ 2 m o u t h ) .
and similarly for the eye region. So, the combined GL-MRFLoss is given by the following:
L g l _ m r f = ω g · L g _ m r f + ω m · L m o u t h _ m r f + ω e · L e y e _ m r f ,
with weights ω g ,   ω m ,   ω e = {0.3, 0.5, 0.2}.
Regularization Loss: Regularization is also used for smooth appearance of UV normal and position maps to guarantee smoothness:
L r e g = i M V ω n · N i N i D 2 + ω d · P i P i D 2 ,
where ω n ,   ω d = {0.05, 0.01}.

3.4. Fine-Resolution Geometry Detail Reconstruction

3.4.1. Gabor-Aware Texture Enhancement

Fine-scale facial geometry from a single image is particularly challenging because medium- and high-frequency textures are strongly entangled in UV texture representations. When trained without an explicit decoupling mechanism, convolutional networks often adopt a dominance of dominant medium-scale variations, resulting in over-smoothed reconstructions and inadequate sensitivity to subtle high-frequency details, such as wrinkles and skin microstructures. To alleviate this limitation, a learnable Gabor-aware texture enhancement module has been proposed that explicitly promotes joint spatial–frequency decoupling. Gabor filters [46] are of special interest in this area because they are capable of localizing image structures simultaneously in spatial and frequency domains, thereby selectively detecting the oriented and high-frequency patterns that are characteristic of fine facial features. Multiple 2D Gabor filters are embedded in shallow layers of the image-to-image translation network in the design. Instead of pre-set parameters that are handcrafted and specific for each individual Gabor kernel, the entire optimization process is carried out through backpropagation. A quadratic constraint is imposed on the trainable parameters in order to achieve numerical stability and to move the learning toward small-scale, high-frequency regimes as in Equation (12). This restriction limits the effective parameter range in terms of optimization to avert degenerate solutions and, instead, encourages the filters to converge toward the frequency bands that are most informative for fine-detail reconstruction. It should be noted that there is no defined list of functional roles for each layer in the six-layer Gabor. Instead, layer-level specialization reveals itself implicitly through learning, so that distinct kernels learn complementary spatial–frequency responses for different scales or orientations. In practice, a multi-kernel, multi-layered design strikes a fine balance between representational capacity and computational efficiency when applied in the ablation study. Gabor filters’ output is then combined with the UV texture representation with learnable weighting coefficients to allow the adaptability of texture cues extracted from both original UV maps as well as from frequency-selective responses. This fusion strategy improves the sensitivity of the network to fine-scale geometric differences while maintaining its fidelity to the underlying facial structure. Due to the entangled nature of medium- and fine-scale facial textures, networks tend to attend to medium-level variations when performing model architectures. To alleviate this problem, we propose a learnable Gabor-aware texture enhancement module that implicitly disentangles finer-scale features, are shown in Figure 3 Through a dual restriction mechanism on learnable Gabor parameters, this module allows joint decoupling in spatial and frequency space. Embedded in the shallow layers of the image-to-image network [43], the module includes multiple learnable 2D Gabor filters together with a feature-fusion mechanism.
g ( x , y ) = 1 2 π σ x σ y e 1 2 x 2 σ x 2 + y 2 σ y 2 e 2 π j ω x ,
with rotated coordinates:
x = x c o s θ   +   y s i n θ   y = x s i n θ   +   y c o s θ .
Trainable parameters σ x , σ y , ω are constrained within task-specific ranges via the following:
p = u p l p e P 1 + e P ,
providing numerical stability. Other quadratic constraints are used as restrictions on σ x , σ y , ω to reduce to small-scale and high-frequency regimes.
Combining N f Gabor filter outputs with results from the UV texture map is a fusion mechanism as follows:
U f u s i o n = ω t e x · U t e x + k = 1 N f ω k F g k ,
where ω t e x ,   ω k are learnable weights optimized end-to-end.
The six-layer Gabor configuration was selected based on an empirical trade-off between representational capacity and computational stability, as demonstrated in the ablation study. Rather than assigning predefined semantic roles to individual layers, specialization emerges implicitly during training, allowing complementary orientation–frequency responses to be learned across layers. This design avoids over-parameterization while maintaining sufficient flexibility for high-frequency facial detail modeling.

3.4.2. Loss Function

A composite loss function supervised the fine-resolution reconstruction phase to ensure that photometric fidelity and high-frequency details are maintained, while numerical stability is controlled. This loss equation integrates photometric loss, detail perception loss, and regularization as stated in Equation (14). A photometric loss maintains pixel-level consistency between the input image and the reconstructed image so that the recovered geometry remains visually aligned with the observed facial appearance. Yet photometric supervision is not enough to retrieve fine-scale geometric details, particularly in the presence of lighting changes and sensor noise typical of in-the-wild images. We address this limitation with a wavelet-based detail perception loss. Specifically, the Haar discrete wavelet transform is utilized to decompose both predicted and reference UV texture maps into directional high-frequency components. We employ Haar because it is computationally efficient, and sharp edge responses that are important for modeling fine facial structures, like wrinkles and skin creases, are preserved by it. It penalizes discrepancies in the high-frequency wavelet coefficients, motivating the network to adopt edge-aware texture semantics while remaining robust to low-frequency changes in illumination. Regularization is then performed to constrain the magnitude of predicted UV displacement maps, thereby minimizing noise amplification and preventing unstable surface oscillations during the process of optimization. These loss terms together offer complementary supervision, which encourages accurate fine-detail reconstruction while retaining geometric smoothness and numerical stability. Photometric loss, detail perception loss, and regularization supervise the fine-scale detail reconstruction:
L f i n e = ω p · L p i x e l + ω d p · L d p + ω r · L r e g ,
with weights {1.0, 1.0, 0.01}.
Detail Perception Loss: To retain edge-aware texture semantics, a Haar discrete wavelet transform [19] is used to compute LH and HL coefficients of both predicted UV texture and ground truth. Then, an L1 penalty is imposed:
L d p = i M V F x ( P i ) F x ( Y i ) 1 + F y ( P i ) F y ( Y i ) 1 ,
permitting the model to gain better sensitivity to high-frequency edge structures.
Regularization Loss: Noise sensitivity is reduced by using the following:
L r e g = i M V D i 2 ,
where D i indicates the UV displacement values.

4. Experiments

4.1. Implementation Details

The proposed framework is trained on the CelebA dataset, which provides a large collection of in-the-wild facial images with substantial variations in expression, illumination, and identity. CelebA is employed exclusively for training due to its scale and diversity, enabling stable self-supervised optimization without reliance on paired 3D ground truth. For evaluation, the FaceScape benchmark is adopted to assess geometric reconstruction accuracy under controlled and challenging conditions. The FaceScape-Wild subset contains synthetic renderings with varying pose angles, while the FaceScape-Lab subset provides high-resolution facial scans with accurate ground-truth geometry. This cross-dataset evaluation protocol is intentionally designed to assess the generalization capability of the proposed framework beyond the training domain.
The pose distribution in the evaluation sets spans frontal to large-angle views (up to 60°), allowing systematic analysis of reconstruction performance under increasing self-occlusion. Although occlusion and illumination variations are partially represented through pose changes and rendering conditions, it is acknowledged that the test sets do not fully cover all real-world scenarios. This limitation is explicitly discussed in Section 5.
A model was trained on the CelebA dataset [47]. A training subset containing 100,000 in-the-wild face images, of which 1000 were saved for validation, and a test one containing 19,000 images were then used. The geometry-prior reconstruction module was realized with a structure similar to VGG-Face [42]; the geometry-detail module was implemented with the image-to-image translation technique [43].
The initial learning rate in the geometry-prior reconstruction stage was equal to 0.0001 and was decayed by a factor of 0.9 for every 5000 steps. The learning rate for the geometry-detail modeling module and texture-detail modeling module was 0.00001 and decayed at a rate of 0.99 every 5000 steps. A batch size of 10 was adopted. Model optimization was performed using the Adam optimizer on a workstation with an NVIDIA GTX-5090 GPU. The geometry-detail modeling submodule was trained over 10,000 steps run period.

4.2. Qualitative Comparative Analysis

Comparisons were made using this method to other state-of-the-art ones, such as Deep3D [12], UDL [14], DECA [16], FaceVerse [48], EMOCA [49], and HRN [50]. Since HRN takes as ground-truth detail labels real deformation and displacement maps, it is considered a supervised method, while the other methods are weakly supervised or unsupervised. For the sake of fairness, pretrained models and official code releases were publicly available.
The CelebA test set was used for all comparison experiments. Qualitative results from single-image 3D face reconstruction, as representative, are shown in Figure 4. Deep3D generates smooth geometric surfaces and no fine-scale facial detail. UDL and FaceVerse do recover a few medium-scale geometric structures, but there are large differences to the input image, especially for large-scale texture variations, and both methods show limited sensitivity for fine textures. DECA reconstructs more local features but typically produces remarkably similar features in forehead wrinkles across various identities. By contrast, the proposed approach generates high-fidelity 3D faces, which preserve medium-scale and fine-scale geometric structure together, as well as fine-scale geometric structure that corresponds quite well to those of the images for which it was trained. In particular, significant advances in the fine quality of the fine-detail reconstruction are achieved by the method compared to previous unsupervised methods. Even against supervised HRN, the proposed method reaches similar fine-scale reconstruction but has significantly better robustness under occlusion, in which HRN is very ineffective.

4.3. Comparison with Other Geometric Reconstruction Methods

The accuracy of geometry reconstruction was assessed by employing Chamfer Distance (CD) and Mean Normal Error (MNE), using the FaceScape benchmarking process [15], and comparing it with current approaches. The quantitative responses are presented in Table 1 on the FaceScape-Wild dataset [15], comprising 400 synthetic images with pose angle. The proposed method shows better CD performance at pose angles between 0° and 30° and is also the only method to consistently come in within the top three ranks, evaluated consistently using all metrics. The comparison on the FaceScape-Lab dataset [15] is presented in Table 2, which contains 660 high-resolution images along with ground-truth scans. This has clear leading performance in CD and MNE at 0°. Performance drops marginally between angles 30° and 60° pose, but on average, the ranking is still ranked among the best among the competing approaches.
As no quantitative metric directly evaluates fine-scale 3D facial geometry, LPIPS became an appropriate metric of this concern, as it quantifies the semantic accuracy and detail consistency, as reported in Table 3. Although LPIPS was developed specifically for real-world RGB images, its use for assessing fine-scale 3D facial detail was validated for visual similarity and semantic coherence without pixel-level alignment [51].
They used FaceScape-Lab images with ground-truth meshes (100 images distributed randomly). Because few of the competing models output only unrendered meshes, all the meshes of the competing methods and the ground truth were orthographically projected to 2D, aligned with the original images, and masked to retain only visible facial patches. It encoded the masked results out to latent representations, decoded back to images, and also applied LPIPS for validation. As indicated in Table 4, the proposed method has the best semantic accuracy, followed by detail consistency, which aligns with the qualitative results described in Figure 5. Although a moderate performance degradation is observed at extreme pose angles (≈60°), the proposed framework consistently ranks among the top unsupervised methods across all evaluation metrics. This behavior is primarily attributed to self-occlusion and missing texture supervision in non-visible regions, a known limitation of monocular self-supervised reconstruction. Importantly, the results demonstrate that the hierarchical design preserves competitive accuracy under moderate poses while substantially improving fine-scale detail fidelity.

4.4. Ablation Study

An ablation study is performed to analyze the contribution of the suggested architectural components (multi-scale hierarchical strategy, the global–local MRF loss, Gabor-aware texture enhancement module, and detail perception loss). As shown in Table 4, the reconstruction quality is significantly enhanced as each building block is added incrementally to the baseline model. LPIPS is used as a performance measurement in the ablation, as no existing metric directly quantifies perceptual consistency of fine-scale 3D facial geometry. LPIPS measures semantic similarity between the rendered reconstructions and their reference images in a learned feature space, making it an appropriate approach to evaluate high-frequency detail quality, not necessarily needing precise pixel-level correlation. Although geometric data such as Chamfer Distance (CD) and Mean Normal Error (MNE) are the two main metrics to assess overall reconstruction accuracy, LPIPS is utilized to provide perceptual fidelity of information at a fine scale. The experimental results show that the hierarchical multi-resolution approach provides the greatest performance gain, reinforcing the ability to progressively model facial geometry at different scales. The Gabor-aware module and detail perception loss improve fine-detail consistency, indicating a complementary role of these two approaches for high-frequency reconstructions. A hierarchical modeling strategy with multiple scales was applied to characterize patterns at coarse, medium, and fine dimensions, respectively, so that the facial geometry reconstruction is efficient and accurate. Figure 6 shows the representative results. The coarse-scale reconstruction stage (Figure 6b) predicts the basic 3D facial contour. The medium-scale stage (Figure 6c) improves intermediate structures, including nasolabial folds and dimples. The fine-scale stage (Figure 6d) targets fine features—such as forehead lines and crow’s feet—leading to feasible high-frequency geometry. This hierarchical method retains global structural uniformity, while also allowing for adaptable, stage-wise tuning.
A quantitative ablation study was performed by LPIPS after the validation method in [51]. The input images and the reconstructed images were coded and recoded into latent codes, decoded, and then compared with an analogue process using LPIPS. 300 CelebA test images were used in the analysis. Table 4 shows that the multi-scale hierarchical strategy reduces LPIPS error by 9% based on the baseline 3DMM. Global–local MRF loss, learnable Gabor module, and detail perception loss can be added which have an additional value of up to 0.3%, 0.7%, and 1.4%, respectively. The proposed components are confirmed to be effective with an overall 11.4% improvement from the baseline 3DMM.
Finally, to explore the effect of Gabor filter depth and kernel dimensions, further ablation experiments were completed. As summarized in Table 5, it was found that multi-kernel structures (1 × 1, 3 × 3, 5 × 5) surpass single-kernel designs using multi-kernel configurations. When six layers of Gabor filters are fitted, model performance converges; as the number of layers increases, the gains increase gradually. Thus, as the optimal computational cost vs. reconstruction accuracy trade-off, we opt for a six-layer Gabor configuration.

5. Discussion

In this study, the two-stage reconstruction framework provided in the baseline method [14] was generalized to a three-stage structure. For a more detailed exploration of the computational efficiency of the expanded model, a broader series of comparative experiments was performed. More precisely, 300 images were randomly sampled from the CelebA test set for inference, and the preprocessing time was omitted from this measure. Table 6 summarizes the final average computation times. Since UDL requires just two stages to obtain the final reconstruction, its computational time is reasonably low. The proposed approach only marginally differs in computation time, in comparison to HRN, which follows a three-stage system as well. These results show that a three-stage reconstruction method is somewhat less computationally efficient, although the additional computational cost remains within limits based on significant improvement in reconstruction quality. Knowledge distillation and model compression will be explored in future research to increase computational speed of the formulated model.

5.1. Analysis of Large-Pose Degradation

Quantitative results in Table 1 and Table 2 show a significant degradation in performance at larger pose angles (approximately 60°) where the proposed framework has higher reconstruction errors than frontal or moderate poses. This is essentially due to self-occlusion and texture omission in the non-visible parts of the facial region. When the pose becomes large, part of the face will not be visible in the input image, leading to ambiguous geometric cues and incomplete texture supervision during self-supervised training. Notably, while fully supervised approaches leverage dense ground-truth displacement maps, we propose the concept of occluded geometry to be inferred implicitly from visible cues and can introduce uncertainty near occlusion borders. This is especially visible in high-frequency areas, where texture loss and inconsistent detail propagation might exist. Such difficulties have been reported in previous monocular 3D face reconstruction work. Remedies may involve symmetry constraints, explicit UV texture completion modules, or priors based on multi-view data to increase robustness under extreme poses. Such directions are seen as promising avenues for future work.

5.2. Computational Efficiency and Trade-Offs

As demonstrated through the computational efficiency analysis presented in Table 6, the proposed framework takes around 6 s per inference, similar to other three-stage architectures like HRN, but slower than two-stage approaches. The increase in the run time of that process is mainly a result of the extra fine-resolution reconstruction stage and the incorporation of frequency-aware processing. It incurs these additional costs to execute, but the hierarchical architecture results in significant improvements in reconstruction accuracy, especially for fine-scale details. Both of these show a well-defined trade-off between accuracy and efficiency, with moderate increases in run time producing clear improvements in perceptual and geometric quality. Importantly, reconstruction accuracy is given priority over real-time performance in this implementation. Thus, future optimization methods such as model compression, knowledge distillation, and reduced Gabor filter depth will be expected for a greater speed of inference while not degrading the quality of the reconstruction. They will also explore the extent of GPU memory usage and parallelization efficiency to contribute to practical deplorability.
Moreover, Table 2 demonstrates a reduction in reconstruction accuracy for large-pose input. This restriction is possibly due to the lack of explicit constraints on these non-visible facial areas. Uncertainties typically occur in the gaps between the visible and occluded regions, and inadequate geometric supervision permits uneven detail generation. Future work will address this problem by including UV texture completion and prior geometric and textural cues to increase consistency regarding visual and texture detail.

6. Conclusions

In this work, we proposed a hierarchical multi-resolution framework supported by self-supervised learning (HMR-Framework) to model geometric facial features in different scales while also improving the recovery of fine-scale structures. The face model was divided into three hierarchical component scales, allowing structured representations of coarse, medium, and fine spatial details. A learnable Gabor-aware texture enhancement module was integrated and used to decompose fine-scale textures across spatial and frequency domains, enabling deep fine-scale processing of high-frequency facial information. Results were shown in experimental studies, which suggest that the proposed method attained competitive performance with state-of-the-art 3D face reconstruction techniques, especially for detailed fine-scale recovery. Further studies will be requiredired to improve on realistic 3D face reconstruction for occlusion but also to improve the robustness and accuracy of the proposed framework.
Experimental results on the FaceScape benchmarks (Table 1, Table 2 and Table 3) and comprehensive ablation studies (Table 4) confirm that the hierarchical multi-resolution strategy contributes most significantly to reconstruction accuracy, while the Gabor-aware texture enhancement and wavelet-based detail perception losses provide complementary gains in fine-scale detail recovery.

Author Contributions

Conceptualization, P.M. and R.F.; Methodology, R.F.; Writing—original draft, R.F.; Writing—review and editing, R.F.; Supervision, R.F.; Project administration, P.M.; Funding acquisition, R.F. All authors have read and agreed to the published version of the manuscript.

Funding

The APC was funded by Innovative Education and Lifelong Learning Institute, Rajamangala University of Technology Tawan-ok. No specific grant number is associated with this funding.

Institutional Review Board Statement

Ethical review and approval were waived for this study due to the present study exclusively utilized the CelebA dataset, a publicly available benchmark dataset released for non-commercial research purposes. No new data collection, human subject recruitment, or interaction with individuals was conducted. All images were processed in an automated and algorithmic manner without attempting to identify individuals or infer sensitive personal information. Accordingly, this study did not require ethical approval.

Informed Consent Statement

Patient consent was waived due to the informed consent was obtained by the original creators of the CelebA dataset from all individuals included in the dataset. The present study exclusively used the dataset in accordance with its stated research-only usage policy. No direct human participation, subject recruitment, or additional data collection was conducted by the authors.

Data Availability Statement

The data presented in this study are openly available in CelebA dataset at https://mmlab.ie.cuhk.edu.hk/projects/CelebA.html, (accessed on 16 March 2023). The FaceScape dataset used for evaluation is available at https://facescape.nju.edu.cn/, (accessed on 16 March 2023) subject to the dataset’s license agreement.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yu, C.; Zhang, Z.; Li, H.; Liu, C. Enhancing 3D Face Recognition: Achieving Significant Gains by Leveraging Widely Available Face Images. Sensors 2025, 25, 5049. [Google Scholar] [CrossRef]
  2. Kim, S.; An, B. S.; Lee, E.C. Comparative Analysis of AI-Based Facial Identification and Expression Recognition Using Upper and Lower Facial Regions. Appl. Sci. 2023, 13, 6070. [Google Scholar] [CrossRef]
  3. Shi, Y.; Zhang, H.; Guo, W.; Zhou, M.; Li, S.; Li, J.; Ding, Y. LighterFace Model for Community Face Detection and Recognition. Information 2024, 15, 215. [Google Scholar] [CrossRef]
  4. El Fadel, N. Facial Recognition Algorithms: A Systematic Literature Review. J. Imaging 2025, 11, 58. [Google Scholar] [CrossRef]
  5. Zhao, X.; Zhang, K.; Huang, Q.; Jiang, S.; Pan, Y. Human–Computer Interaction in Healthcare: A Bibliometric and Visual Analysis. Behav. Sci. 2024, 15, 2467. [Google Scholar] [CrossRef]
  6. Kim, S.; An, B.S.; Lee, E.C. Exploring Technology Acceptance of Healthcare Devices: The Moderating Role of Device Type and Generation. Sensors 2024, 24, 7921. [Google Scholar] [CrossRef]
  7. Jiang, D.; Chang, J.; You, L.; Bian, S.; Kosk, R.; Maguire, G. Audio-Driven Facial Animation with Deep Learning: A Survey. Information 2024, 15, 675. [Google Scholar] [CrossRef]
  8. Wang, X.; Huo, Y.; Liu, Y.; Guo, X.; Yan, F.; Zhao, G. Multimodal Feature-Guided Audio-Driven Emotional Facial Animation for Talking-Face Generation. Electronics 2025, 14, 2684. [Google Scholar] [CrossRef]
  9. Blanz, V.; Vetter, T. A morphable model for the synthesis of 3-D faces. In Seminal Graphics Papers: Pushing the Boundaries; Association for Computing Machinery: New York, NY, USA, 2023; Volume 2, pp. 157–164. [Google Scholar] [CrossRef]
  10. Yang, W.; Chen, Z.; Chen, C.; Chen, G.; Wong, K.-Y.K. Deep face video inpainting via UV mapping. IEEE Trans. Image Process. 2023, 32, 1145–1157. [Google Scholar] [CrossRef]
  11. Lou, J.; Cai, X.; Dong, J.; Yu, H. Real-time 3-D facial tracking via cascaded compositional learning. IEEE Trans. Image Process. 2021, 30, 3844–3857. [Google Scholar] [CrossRef]
  12. Deng, Y.; Yang, J.; Xu, S.; Chen, D.; Jia, Y.; Tong, X. Accurate 3-D face reconstruction with weakly supervised learning: From single image to image set. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019; pp. 285–295. [Google Scholar] [CrossRef]
  13. Sanyal, S.; Bolkart, T.; Feng, H.; Black, M.J. Learning to regress 3-D face shape and expression from an image without 3-D supervision. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 7763–7772. [Google Scholar] [CrossRef]
  14. Chen, Y.; Wu, F.; Wang, Z.; Song, Y.; Ling, Y.; Bao, L. Self-supervised learning of detailed 3-D face reconstruction. IEEE Trans. Image Process. 2020, 29, 8696–8705. [Google Scholar] [CrossRef] [PubMed]
  15. Zhu, H.; Yang, H.; Guo, L.; Zhang, Y.; Wang, Y.; Huang, M.; Wu, M.; Shen, Q.; Yang, R.; Cao, X. FaceScape: 3D Facial Dataset and Benchmark for Single-View 3D Face Reconstruction. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 14528–14545. [Google Scholar] [CrossRef]
  16. Feng, Y.; Feng, H.; Black, M.J.; Bolkart, T. Learning an animatable detailed 3-D face model from in-the-wild images. ACM Trans. Graph. 2021, 40, 1–13. [Google Scholar] [CrossRef]
  17. Lai, J.-Y.; Wu, T.-C.; Phothong, W.; Wang, D.W.; Liao, C.-Y.; Lee, J.-Y. A High-Resolution Texture Mapping Technique for 3D Textured Model. Appl. Sci. 2018, 8, 2228. [Google Scholar] [CrossRef]
  18. Zhu, L.; Chen, T.; Yin, J.; See, S.; Liu, J. Learning Gabor texture features for fine-grained recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 1621–1631. [Google Scholar] [CrossRef]
  19. Kanagaraj, H.; Muneeswaran, V. Image compression using Haar discrete wavelet transform. In Proceedings of the 2020 5th International Conference on Devices, Circuits and Systems (ICDCS), Coimbatore, India, 5–6 March 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 271–274. [Google Scholar] [CrossRef]
  20. Xia, Y.; Zheng, W.; Wang, Y.; Yu, H.; Dong, J.; Wang, F.-Y. Local and global perception generative adversarial network for facial expression synthesis. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 1443–1452. [Google Scholar] [CrossRef]
  21. Wang, Y.; Tao, X.; Qi, X.; Shen, X.; Jia, J. Image inpainting via generative multi-column convolutional neural networks. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; pp. 329–338. [Google Scholar]
  22. Deng, N.; Xu, Z.; Li, X.; Gao, C.; Wang, X. Deep Learning and Face Recognition: Face Recognition Approach Based on the DS-CDCN Algorithm. Appl. Sci. 2024, 14, 5739. [Google Scholar] [CrossRef]
  23. Genova, K.; Cole, F.; Maschinot, A.; Sarna, A.; Vlasic, D.; Freeman, W.T. Unsupervised training for 3D morphable model regression. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8377–8386. [Google Scholar] [CrossRef]
  24. Lee, Y.; Jang, S.; Bae, H.; Jeon, T.; Lee, S. Multitask Learning Strategy with Pseudo-Labeling: Face Recognition, Facial Landmark Detection, and Head Pose Estimation. Sensors 2024, 24, 3212. [Google Scholar] [CrossRef]
  25. Huang, Y.; Fang, L.; Hu, S. TED-Face: Texture-Enhanced Deep Face Reconstruction in the Wild. Sensors 2023, 23, 6525. [Google Scholar] [CrossRef]
  26. Dou, P.; Shah, S.K.; Kakadiaris, I.A. End-to-end 3D face reconstruction with deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5908–5917. [Google Scholar] [CrossRef]
  27. Li, Y.; Hao, Q.; Hu, J.; Pan, X.; Li, Z.; Cui, Z. 3D3M: 3D modulated morphable model for monocular face reconstruction. IEEE Trans. Multimedia 2022, 25, 6642–6652. [Google Scholar] [CrossRef]
  28. Tu, X.; Zhao, J.; Xie, M.; Jiang, Z.; Balamurugan, A.; Luo, Y.; Zhao, Y.; He, L.; Ma, Z.; Feng, J. 3D face reconstruction from a single image assisted by 2D face images in the wild. IEEE Trans. Multimed. 2020, 23, 1160–1172. [Google Scholar] [CrossRef]
  29. Yang, W.; Zhao, Y.; Yang, B.; Shen, J. Learning 3D face reconstruction from the cycle-consistency of dynamic faces. IEEE Trans. Multimed. 2023, 26, 3663–3675. [Google Scholar] [CrossRef]
  30. Chai, X.; Chen, J.; Liang, C.; Xu, D.; Lin, C.-W. Expression-aware face reconstruction via a dual-stream network. IEEE Trans. Multimed. 2021, 23, 2998–3012. [Google Scholar] [CrossRef]
  31. Fathallah, M.; Eletriby, S.; Alsabaan, M.; Ibrahem, M.I.; Farok, G. Advanced 3D Face Reconstruction from Single 2D Images Using Enhanced Adversarial Neural Networks and Graph Neural Networks. Sensors 2024, 24, 6280. [Google Scholar] [CrossRef] [PubMed]
  32. Dib, A.; Ahn, J.; Thebault, C.; Gosselin, P.-H.; Chevallier, L. S2F2: Self-supervised high fidelity face reconstruction from monocular image. In Proceedings of the 2023 IEEE 17th International Conference on Automatic Face and Gesture Recognition (FG), Waikoloa Beach, HI, USA, 5–8 January 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–8. [Google Scholar] [CrossRef]
  33. Petkov, N. Biologically motivated computationally intensive approaches to image pattern recognition. Future Gener. Comput. Syst. 1995, 11, 451–465. [Google Scholar] [CrossRef]
  34. Martins, N.; Silva, J.S.; Bernardino, A. Fingerprint Recognition in Forensic Scenarios. Sensors 2024, 24, 664. [Google Scholar] [CrossRef] [PubMed]
  35. Oh, B.-S.; Toh, K.-A.; Teoh, A.B.J.; Lin, Z. An analytic Gabor feedforward network for single-sample and pose-invariant face recognition. IEEE Trans. Image Process. 2018, 27, 2791–2805. [Google Scholar] [CrossRef]
  36. Hosseini, S.; Cho, N.I. GF-CapsNet: Using Gabor jet and capsule networks for facial age, gender, and expression recognition. In Proceedings of the 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), Lille, France, 14–18 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–8. [Google Scholar] [CrossRef]
  37. Hosseini, S.; Lee, S.H.; Kwon, H.J.; Koo, H.I.; Cho, N.I. Age and gender classification using wide convolutional neural network and Gabor filter. In Proceedings of the 2018 International Workshop on Advanced Image Technology (IWAIT), Chiang Mai, Thailand, 7–9 January 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–3. [Google Scholar] [CrossRef]
  38. Hu, X.-D.; Wang, X.-Q.; Meng, F.-J.; Hua, X.; Yan, Y.-J.; Li, Y.-Y.; Huang, J.; Jiang, X.-L. Gabor-CNN for object detection based on small samples. Defence Technol. 2020, 16, 1116–1129. [Google Scholar] [CrossRef]
  39. Yuan, Y.; Wang, L.-N.; Zhong, G.; Gao, W.; Jiao, W.; Dong, J.; Shen, B.; Xia, D.; Xiang, W. Adaptive Gabor convolutional networks. Pattern Recognit. 2022, 124, 108495. [Google Scholar] [CrossRef]
  40. Zhao, X.; Tao, R.; Li, W.; Philips, W.; Liao, W. Fractional Gabor convolutional network for multisource remote sensing data classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–18. [Google Scholar] [CrossRef]
  41. Le, H.T.; Phung, S.L.; Chapple, P.B.; Bouzerdoum, A.; Ritz, C.H.; Tran, L.C. Deep Gabor neural network for automatic detection of mine-like objects in sonar imagery. IEEE Access 2020, 8, 94126–94139. [Google Scholar] [CrossRef]
  42. Parkhi, O.; Vedaldi, A.; Zisserman, A. Deep face recognition. In Proceedings of the BMVC 2015—British Machine Vision Conference, Swansea, UK, 7–10 September 2015. [Google Scholar] [CrossRef]
  43. Isola, P.; Zhu, J.-Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar] [CrossRef]
  44. Wang, C.; Shao, M.; Meng, D.; Zuo, W. Dual-pyramidal image inpainting with dynamic normalization. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 5975–5988. [Google Scholar] [CrossRef]
  45. Huo, J.; Liu, X.; Li, W.; Gao, Y.; Yin, H.; Luo, J. CAST: Learning both geometric and texture style transfers for effective caricature generation. IEEE Trans. Image Process. 2022, 31, 3347–3358. [Google Scholar] [CrossRef]
  46. Li, Y.; Bi, Y.; Zhang, W.; Ren, J.; Chen, J. M2GF: Multi-Scale and Multi-Directional Gabor Filters for Image Edge Detection. Appl. Sci. 2023, 13, 9409. [Google Scholar] [CrossRef]
  47. Zhou, X.; Gao, Z.; Gong, H.; Li, S. DeFFace: Deep Face Recognition Unlocked by Illumination Attributes. Electronics 2024, 13, 4566. [Google Scholar] [CrossRef]
  48. Wang, L.; Chen, Z.; Yu, T.; Ma, C.; Li, L.; Liu, Y. FaceVerse: A fine-grained and detail-controllable 3D face morphable model from a hybrid dataset. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 20333–20342. [Google Scholar] [CrossRef]
  49. Daneček, R.; Black, M.J.; Bolkart, T. EMOCA: Emotion-driven monocular face capture and animation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 8–24 June 2022; pp. 20311–20322. [Google Scholar] [CrossRef]
  50. Lei, B.; Ren, J.; Feng, M.; Cui, M.; Xie, X. A hierarchical representation network for accurate and detailed face reconstruction from in-the-wild images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 394–403. [Google Scholar] [CrossRef]
  51. Ling, J.; Wang, Z.; Lu, M.; Wang, Q.; Qian, C.; Xu, F. Structure-aware editable morphable model for 3D facial detail animation and manipulation. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 23–27 October 2022; Springer: Cham, Switzerland, 2022; pp. 249–267. [Google Scholar] [CrossRef]
Figure 1. Conceptual illustration comparing Gabor-based spatial–frequency decoupling with alternative multi-scale representations, including wavelet transforms and Laplacian pyramids. Unlike fixed basis decompositions, the learnable Gabor-aware module provides adaptive orientation and frequency selectivity, enabling joint spatial–frequency modeling tailored to fine-scale facial detail reconstruction.
Figure 1. Conceptual illustration comparing Gabor-based spatial–frequency decoupling with alternative multi-scale representations, including wavelet transforms and Laplacian pyramids. Unlike fixed basis decompositions, the learnable Gabor-aware module provides adaptive orientation and frequency selectivity, enabling joint spatial–frequency modeling tailored to fine-scale facial detail reconstruction.
Jimaging 12 00026 g001
Figure 2. Overview of the proposed hierarchical multi-resolution self-supervised framework (HMR-Framework) for single-image 3D face reconstruction.
Figure 2. Overview of the proposed hierarchical multi-resolution self-supervised framework (HMR-Framework) for single-image 3D face reconstruction.
Jimaging 12 00026 g002
Figure 3. Architecture of the Gabor-aware texture enhancement module. Multiple learnable two-dimensional Gabor filters are embedded in the shallow layers of the image-to-image translation network to selectively capture high-frequency facial textures. Quadratic constraints are imposed on the Gabor parameters to ensure numerical stability and to bias learning toward small-scale, high-frequency regimes. The Gabor responses are adaptively fused with the UV texture representation using learnable weights, enabling effective integration of spatial–frequency cues for fine-detail reconstruction.
Figure 3. Architecture of the Gabor-aware texture enhancement module. Multiple learnable two-dimensional Gabor filters are embedded in the shallow layers of the image-to-image translation network to selectively capture high-frequency facial textures. Quadratic constraints are imposed on the Gabor parameters to ensure numerical stability and to bias learning toward small-scale, high-frequency regimes. The Gabor responses are adaptively fused with the UV texture representation using learnable weights, enabling effective integration of spatial–frequency cues for fine-detail reconstruction.
Jimaging 12 00026 g003
Figure 4. Qualitative comparison on the image test set, demonstrating fine-level detail reconstruction in the HMR-Framework, highlighted by red dashed circles.
Figure 4. Qualitative comparison on the image test set, demonstrating fine-level detail reconstruction in the HMR-Framework, highlighted by red dashed circles.
Jimaging 12 00026 g004
Figure 5. Visualization of detail consistency analysis results.
Figure 5. Visualization of detail consistency analysis results.
Jimaging 12 00026 g005
Figure 6. Visual comparison of the results of the framework. Column (a) is the original image, (b) is the large-scale geometry prior reconstruction result, (c) is the medium-scale geometry detail reconstruction result, and (d) is the fine-scale geometry detail reconstruction result. The red boxes emphasize regions where additional geometric details are recovered by the proposed framework at the medium-scale and fine-scale stages.
Figure 6. Visual comparison of the results of the framework. Column (a) is the original image, (b) is the large-scale geometry prior reconstruction result, (c) is the medium-scale geometry detail reconstruction result, and (d) is the fine-scale geometry detail reconstruction result. The red boxes emphasize regions where additional geometric details are recovered by the proposed framework at the medium-scale and fine-scale stages.
Jimaging 12 00026 g006
Table 1. Quantitative evaluation of FaceScape-Wild dataset.
Table 1. Quantitative evaluation of FaceScape-Wild dataset.
Methods0–5°5–30°30–60°
CDMNECDMNECDMNE
DFDN3.7020.0913.3070.0927.3130.130
DF2Net2.9530.1222.4410.1296.6250.159
UDL2.3530.0923.2870.0944.2940.109
FaceScape2.8420.0873.1780.0944.0450.109
SADRNet3.2680.1143.6170.0746.4880.120
LAP4.2380.0934.5240.0826.0100.100
DECA2.9130.0812.6640.0802.9120.093
EMOCA2.7090.0902.7140.0992.9430.101
HRN2.5290.0862.6120.1152.150.080
HMR-Framework2.2250.0872.4880.0833.3430.103
Table 2. Quantitative analysis of FaceScape-Wild dataset.
Table 2. Quantitative analysis of FaceScape-Wild dataset.
Methods30°60°
CDMNECDMNECDMNE
DFDN5.3500.1388.3900.16529.5400.350
DF2Net5.6000.1909.5500.250N/AN/A
UDL2.7600.1156.6800.1547.0400.209
FaceScape4.0100.1136.0900.1495.8500.182
SADRNet5.3200.1368.8400.1718.8600.185
LAP5.3400.1409.2600.18610.8800.244
DECA4.1300.1165.1800.1255.2500.134
EMOCA3.0900.1084.0600.1175.3800.124
HRN2.9500.1064.7000.1185.2900.118
HMR-Framework2.6600.1126.2900.1336.0400.188
Table 3. Detailed consistency analysis.
Table 3. Detailed consistency analysis.
MethodsDeep3DUDLDECADECAHRNHMR-Framework
LPIPS0.3853750.2436250.4556250.4231250.301250.22425
Table 4. Ablation study reporting LPIPS results for different component configurations in the HMR-Framework.
Table 4. Ablation study reporting LPIPS results for different component configurations in the HMR-Framework.
Base ModelMulHi L g l _ m r f Gabor L d p LPIPS
0.1037
0.0944
0.0941
0.0934
0.0921
Table 5. Ablation study comparing the number of Gabor filters and model accuracy.
Table 5. Ablation study comparing the number of Gabor filters and model accuracy.
Convolution Kernel SizeNumber of Gabor FiltersLPIPS
1 × 1 ,   1 × 1 ,   1 × 1   × 1 0.0941
3 × 3 ,   3 × 3 ,   3 × 3 × 1 0.0943
5 × 5 ,   5 × 5 ,   5 × 5 × 1 0.0938
1 × 1 ,   3 × 3 ,   5 × 5 × 1 0.0934
1 × 1 ,   3 × 3 ,   5 × 5 × 2 0.0921
1 × 1 ,   3 × 3 ,   5 × 5 × 3 0.0925
1 × 1 ,   3 × 3 ,   5 × 5 × 4 0.0919
1 × 1 ,   3 × 3 ,   5 × 5 × 5 0.0922
⇨ We selected the most optimized solution.
Table 6. Computational efficiency analysis.
Table 6. Computational efficiency analysis.
MethodsUDLHRNHMR-Framework
Times (s)5.38085.86646.0192
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mareo, P.; Fooprateepsiri, R. A Hierarchical Multi-Resolution Self-Supervised Framework for High-Fidelity 3D Face Reconstruction Using Learnable Gabor-Aware Texture Modeling. J. Imaging 2026, 12, 26. https://doi.org/10.3390/jimaging12010026

AMA Style

Mareo P, Fooprateepsiri R. A Hierarchical Multi-Resolution Self-Supervised Framework for High-Fidelity 3D Face Reconstruction Using Learnable Gabor-Aware Texture Modeling. Journal of Imaging. 2026; 12(1):26. https://doi.org/10.3390/jimaging12010026

Chicago/Turabian Style

Mareo, Pichet, and Rerkchai Fooprateepsiri. 2026. "A Hierarchical Multi-Resolution Self-Supervised Framework for High-Fidelity 3D Face Reconstruction Using Learnable Gabor-Aware Texture Modeling" Journal of Imaging 12, no. 1: 26. https://doi.org/10.3390/jimaging12010026

APA Style

Mareo, P., & Fooprateepsiri, R. (2026). A Hierarchical Multi-Resolution Self-Supervised Framework for High-Fidelity 3D Face Reconstruction Using Learnable Gabor-Aware Texture Modeling. Journal of Imaging, 12(1), 26. https://doi.org/10.3390/jimaging12010026

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop