Next Article in Journal
Implicit Quiescent Optical Soliton Perturbation with Nonlinear Chromatic Dispersion and Kudryashov’s Self-Phase Modulation Structures for the Complex Ginzburg–Landau Equation Using Lie Symmetry: Linear Temporal Evolution
Previous Article in Journal
A Refined Inertial-like Subgradient Method for Split Equality Problems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Distance Measurement Between a Camera and a Human Subject Using Statistically Determined Interpupillary Distance

Faculty of Electrical Engineering, “Gheorghe Asachi” Technical University of Iași, 700050 Iași, Romania
*
Author to whom correspondence should be addressed.
AppliedMath 2025, 5(3), 118; https://doi.org/10.3390/appliedmath5030118
Submission received: 21 July 2025 / Revised: 25 August 2025 / Accepted: 1 September 2025 / Published: 3 September 2025

Abstract

This paper presents a non-intrusive method for estimating the distance between a camera and a human subject using a monocular vision system and statistically derived interpupillary distance (IPD) values. The proposed approach eliminates the need for individual calibration by utilizing average IPD values based on biological sex, enabling accurate, scalable distance estimation for diverse users. The algorithm, implemented in Python 3.12.11 using the MediaPipe Face Mesh framework, extracts pupil coordinates from facial images and calculates IPD in pixels. A sixth-degree polynomial calibration function, derived from controlled experiments using a uniaxial displacement system, maps pixel-based IPD to real-world distances across three intervals (20–80 cm, 80–160 cm, and 160–240 cm). Additionally, a geometric correction is applied to compensate for in-plane facial rotation. Experimental validation with 26 participants (15 males, 11 females) demonstrates the method’s robustness and accuracy, as confirmed by relative error analysis against ground truth measurements obtained with a Bosch GLM120C laser distance meter. Males exhibited lower relative errors across the intervals (3.87%, 4.75%, and 5.53%), while females recorded higher mean relative errors (6.0%, 6.7%, and 7.27%). The results confirm the feasibility of the proposed method for real-time applications in human–computer interaction, augmented reality, and camera-based proximity sensing.

1. Introduction

A particularly significant application area is human-to-camera distance estimation, which underpins the development of responsive graphical user interfaces (GUIs). For instance, intelligent document viewers and adaptive content delivery systems rely heavily on accurate proximity measurements to tailor user experience in real time [1]. Furthermore, distance estimation is integral to autofocus algorithms in imaging devices, allowing for the automatic optimization of image quality without requiring manual adjustments [2]. Beyond user-centric applications, this capability extends its influence into safety-critical and autonomous domains, such as autonomous driving (ADAS), where vehicles must continuously and precisely gauge the distance to surrounding objects to ensure safe navigation [3]. Similarly, in robotics, effective spatial perception is essential for tasks ranging from navigation to manipulation within complex environments [4]. Moreover, emerging technologies in Augmented Reality (AR) and Virtual Reality (VR) demand real-time and precise distance measurements to create immersive, responsive virtual environments that closely mimic real-world spatial interactions [5,6].
Traditionally, binocular vision systems, which rely on stereo image pairs to triangulate distances, have been widely employed for such purposes due to their ability to resolve depth directly. Furthermore, effective stereo vision requires dual-camera setups with precise extrinsic and intrinsic calibration, as well as additional computational resources for stereo correspondence and rectification. These constraints limit their practicality in embedded or resource-constrained scenarios. Consequently, monocular methods—based on single-camera inputs—have garnered increasing attention due to their hardware simplicity and ease of integration into existing imaging platforms [7]. In sensitive applications, where velocity of the visual task and the placement is critical, binocular measures are relatively robust [8,9].
Despite the growing importance of distance estimation, the development of methods that are both non-intrusive and calibration-free remains a significant technical challenge [10]. This paper introduces a novel approach leveraging average anatomical features—specifically, interpupillary distance (IPD)—as a dimensional reference for estimating subject distance using a monocular camera, without requiring personalized calibration.
Monocular distance estimation typically necessitates prior knowledge of a dimensional reference to resolve scale ambiguity inherent to single-camera systems [11]. Depending on the application domain, various types of reference features have been employed. For example, smartphone applications designed for video-based distance measurement, such as Smart Distance and Distance Measure, use either known physical landmarks or complementary smartphone sensors, including accelerometers, to facilitate accurate measurement [12,13]. In automotive contexts, distance estimation methods often assume a priori knowledge of object dimensions, such as the average width of vehicles [14]; however, this assumption introduces considerable variability, as both vehicle and lane dimensions can differ significantly across geographic regions, limiting the generalizability and robustness of such approaches. Other methods employ identifiable features like license plate dimensions or artificial markers as scale references. Nonetheless, small physical markers pose challenges in the case of long-range detection, as beyond 30 m it becomes increasingly difficult. Furthermore, marker dimensions vary globally, limiting their applicability across regions [15]. Additional techniques for monocular distance estimation include leveraging camera motion across a known displacement to infer scale through pixel variation [16,17], placing physical reference markers at predetermined distances [18], exploiting defocus blur characteristics of target objects [19], or using lens focus adjustments to estimate subject distance [20].
When estimating the distance between a camera and a human subject, anatomical features such as interpupillary distance or iris diameter are frequently employed as intrinsic dimensional references. Prior work has demonstrated the feasibility of using IPD for distance estimation [21]; however, such methods generally require individual calibration for each subject and depend on real-time facial recognition [22], limiting their practical applicability. Alternative methods based on iris diameter as a reference have also been proposed [23], but they often suffer from higher inter-subject variability and measurement challenges.
In contrast, the method presented in this paper exploits the average interpupillary distance, which exhibits relatively low variability across the general population, to enable accurate and generalizable distance estimation without the need for individualized calibration. By leveraging this stable anatomical reference, the proposed approach enhances estimation accuracy and broadens applicability, particularly in scenarios where subject-specific calibration is impractical or impossible [24,25].

2. Methods

This paper presents a novel monocular distance estimation method that eliminates the need for individual user calibration by leveraging a statistically established average interpupillary distance (IPD), which varies according to biological sex [26]. Conventional approaches for estimating the distance between a camera and a user typically rely on either physical reference markers within the scene or personalized biometric calibration for each individual, which limits scalability and practical deployment in real-world applications. The method proposed herein aims to achieve a balance between simplicity, generalizability, and robustness, making it particularly suitable for human–computer interaction scenarios.
The proposed approach exploits the relatively stable and measurable nature of the IPD across adult populations. The IPD, defined as the distance between the centers of the pupils, constitutes a consistent anatomical feature observable in frontal facial images. Biometric data reported in [27] indicate the following mean IPD values derived from a statistically significant sample:
  • 61.53 mm for adult females, with a standard deviation (SD) of ±2.66 mm;
  • 65.32 mm for adult males, with a standard deviation (SD) of ±1.50 mm.
These population-based averages serve as biometric references for inferring real-world scale from image pixel measurements. By detecting the pixel distance between the pupil centers in a captured image—using computer vision frameworks such as MediaPipe FaceMesh—and associating this measurement with the known average IPD, the method estimates the physical distance between the subject and the camera based on the principles of pinhole camera geometry.
A key innovation of this approach is the removal of the dependency on per-user calibration, which is traditionally required to accommodate individual facial dimension variability. Instead, the method assumes that subjects conform to a normal distribution centered on the population mean.
Although this assumption introduces some approximation, the relatively low standard deviation of IPD—especially among male subjects—ensures that the resultant error remains within acceptable limits for many practical applications, including interface adaptation and gaze correction.
To enhance accuracy, the system optionally incorporates a sex classification module—based on facial features or external inputs—to select the appropriate average IPD for the subject. In the present implementation, the user’s biological sex is manually selectable via keyboard input. For real-world deployment, this manual selection could be replaced with automated classification modules based on facial features or integration with user profile data in authenticated systems. Such automation could improve usability, particularly in consumer or industrial applications, but must be balanced against considerations of algorithmic bias, user privacy and informed consent.
This distance estimation approach offers several advantages:
  • It is non-intrusive, requiring no physical markers or auxiliary devices;
  • It is scalable and suitable for real-time deployment across diverse populations;
  • It is hardware-agnostic, adaptable to different camera models through a one-time calibration procedure specific to the camera’s optics;
  • It is computationally efficient, enabling integration with mobile and embedded platforms.

Calibration Procedure

The calibration process utilizes a printed frontal facial image with known IPD values as a reference standard (Figure 1), specifically 61.53 mm assigned for female calibration and 65.32 mm assigned for male calibration. These values are incorporated as reference inputs within the distance measurement algorithm.
During calibration, the printed image is positioned in front of the camera, which is mounted on a uniaxial displacement system to precisely control its distance from the image (Figure 2). At multiple predetermined distances (d), images are captured, and the pixel coordinates of the pupil centers are extracted through graphical analysis. The interpupillary distance in pixels is calculated for each frame as the Euclidean distance between the detected pupil centers. This technique is performed at multiple established distances to generate a dataset that correlates pixel-based inter-pupillary distance measurements with actual distances. From this dataset, a calibration function d = f(IPD) is derived, characterizing the relationship between the real-world distance and the interpupillary pixel distance. This calibration curve encapsulates the optical parameters of the camera and its imaging characteristics.
The calibration is performed once per camera model, assuming fixed resolution and focal length. Subsequently, real-time distance estimation is accomplished by detecting the interpupillary pixel distance of the subject’s face in video frames and applying the calibrated function to estimate the physical distance from the camera.
Pupil detection and IPD estimation were implemented in Python, utilizing the MediaPipe Face Mesh framework. Image data was acquired using the camera integrated into the calibration stand. The imaging device is a consumer-grade webcam with a native resolution of 1280 × 720 pixels.
The uniaxial displacement system was electrodynamically actuated and servo-controlled using a closed-loop PID control architecture. The system was designed for precise and repeatable displacement, and its main components are described as follows:
  • Structural frame: a fixed steel frame served as the support base. The frame was equipped with mounting holes to facilitate modular attachment of system components and external fixtures.
  • The cart, actuated via a rack-and-pinion mechanism (Ø6.35 mm, 24 teeth), moved along a stainless-steel shaft supported by linear bearings. The cart had a mass of 500 g and was designed to carry an additional 370 g of load. The increased total mass provided enhanced inertia, contributing to effective vibration damping of the structure during motion.
  • Actuation system: the system was powered by a high-performance DC servomotor, capable of achieving a nominal speed of 4160 rpm and an angular acceleration of 40 × 103 rad/s2. Key motor parameters included a back electromotive force (EMF) constant of 0.804 mV/rpm, a torque constant of 7.67 × 10−3 Nm/A, and a mechanical time constant of 17 ms, enabling robust and accurate actuation performance.
  • Gearbox: a single-stage gearbox (23/1 series) with a 3.71:1 reduction ratio and 100% efficiency was coupled to the motor. The gearbox reduced the load on the motor and enhanced torque transmission to the cart.
  • Power amplifier brush DC servomotor: the amplifier was powered by a 230 V AC input and provided a continuous output of ±24 V, controlled via a ±10 V input signal.
  • Position feedback: a high-resolution optical encoder was integrated into the system for accurate position sensing. The encoder featured a resolution of 0.0235 mm (4096 counts/rev in quadrature mode, equivalent to 1024 lines/rev) and could track rotational speeds up to 10,000 rpm. It operated as a rotary-to-digital converter based on phased array detector technology. The encoder was attached to a 148 mm diameter, 56-tooth pinion gear.
  • Data Acquisition and Control Interface (DAQ): a board with eight configurable digital input/output channels interfaced with the encoder. It provided real-time conversion of shaft angle, speed, and direction into TTL-compatible quadrature signals. The board was equipped with two 5-pin DIN encoder input connectors and supported 16-bit count values. It allowed initialization of the encoder count and optional reloading on index pulses. The entire system was controlled via a LabVIEW 2024 Q3-based PID control loop.
The control system was implemented using a closed-loop PID controller, continuously adjusting the cart’s position based on encoder feedback. The PID parameters—proportional gain (Kp), integral time (Ti), and derivative time (Td)—were determined using the Ziegler–Nichols tuning method. Initially, the integral and derivative terms were set to zero, and the proportional gain was incrementally increased until sustained oscillations were observed at the ultimate gain (Ku = 0.116) and ultimate period (Tu = 0.01 min).
The final controller parameters were computed as:
  • Kp = 0.6 × Ku = 0.07
  • Ti = 0.5 × Tu = 0.005 min
  • Td = 0.125 × Tu = 0.0013 min
The control algorithm was implemented in LabVIEW using the Control Design and Simulation Toolkit. The control loop consisted of four main functional blocks:
  • HIL Initialize—initializes the hardware-in-the-loop (HIL) interface;
  • HIL Read—acquires position feedback from the encoder in real-time;
  • HIL Write—transmits control signals to the power amplifier and DC motor;
  • PID Controller—calculates the control signal based on error input and feedback.
This configuration enabled real-time, closed-loop control of the cart’s displacement with high precision and stability.
To investigate the dependency between the IPD in pixels and the distance of the image from the camera, three measurement intervals were established: 20–80 cm, 80–160 cm, and 160–240 cm. The positioning of the calibration stand was carried out with high precision using a Bosch GLM 120C laser (Iași, România) distance meter to measure the offset from the camera. Within each interval, the cart holding the web camera was moved in 1 cm increments. At each position, an image was captured and the corresponding IPD in pixels was computed.
The function that provides the best approximation of the calibration data was a sixth-degree polynomial, applicable to both male and female printed face targets.
Figure 3 shows the activity diagram summarizing the sequence of phases in the proposed method.

3. Results

Accurate estimation of IPD from 2D facial images is often compromised by the subject’s head orientation relative to the camera. Specifically, in-plane facial rotation (i.e., rotation around the camera’s Z-axis) can cause an apparent reduction in the measured IPD due to perspective distortion. To mitigate this effect, a geometric correction method based on facial landmark analysis was proposed.
Using the MediaPipe Face Mesh model, the 2D coordinates of the pupils were extracted along with three additional facial landmarks: the nose tip, and landmarks on the left and right cheeks. These three points define a local facial triangle within the image coordinate space. Two vectors were computed from the nose tip to each cheek point ( v L , v R ) and their average direction ( v m ) was used to estimate the in-plane orientation of the face (Figure 4).
A 2D normal vector n is perpendicular to the average direction of the facial plane ( v m ). The direction of this normal vector gives the frontal orientation of the face in the image.
The raw IPD was calculated as the Euclidean distance between the detected positions of the left and right pupils.
I P D r a w = P R P L ,
where PR and PL are defined by 2D coordinates of the right and left pupils, as detected in the image. IPDraw is a straight-line distance between the pupils in the 2D image, which is affected by in-plane head rotation.
To correct for in-plane rotation, the angle between the unit pupil vector ( v e y e s ) and the unit normal to the estimated facial direction ( n ) was computed.
c o s = n ^ v e y e s ^ ,
where n ^ = n n   is the unit normal vector pointing in the estimated frontal direction of the face in the 2D image and v e y e s ^ = v e y e s v e y e s is the unit pupil vector. v e y e s and n are the Euclidean norms of the corresponding vectors, calculated from the coordinates of the pupils, respectively from the coordinates of v m . As n is perpendicular to v m (90° counterclockwise) its coordinates are inversed and negative for y.
The corrected IPD was obtained by projecting the pupil vector onto the estimated normal direction, which involves multiplying the raw IPD by the cosine of the angle between the two vectors.
I P D c o r r e c t e d = I P D r a w · c o s α
This method effectively compensates for distortions introduced by in-plane rotation and yields an IPD measurement that more closely approximates the value observed under frontal facial alignment.
As illustrated in Figure 5, in the first case the subject was oriented directly towards the camera, resulting in the same IPD value for both the uncorrected and corrected versions. In the other two cases, the subject’s face was rotated away from the camera’s frontal axis; however, the proposed correction method successfully adjusts the raw IPD value to compensate for this deviation.
The application was tested with the participation of 26 volunteers, all of whom provided informed consent. The test group consisted of 15 male participants aged between 20 and 68 years, and 11 female participants aged between 22 and 67 years. For each individual, the distance to the camera was measured within each of the three calibration intervals. Ten measurements were performed per interval.
Each measurement was carried out using both the proposed method and a Bosch GLM 120C laser distance meter, which served as the ground truth. Relative errors were computed based on the difference between the estimated and reference distances. For each subject and each interval, the minimum, maximum and mean relative errors were calculated from the set of ten individual measurements. The results are presented in Table 1 and Table 2.
Confidence intervals were calculated for mean errors to assess precision, but not for maximum and minimum errors, as these extreme values do not support straightforward interval estimation. They were computed as:
C I = t α / 2 , d f · s n
With tα/2,df the t-value for 95% confidence and degrees of freedom df = n − 1, s is the sample standard deviation (Std dev), n is the sample size (15 males and 11 females).
The results obtained from the experimental evaluation highlight the effectiveness and practical viability of using interpupillary distance (IPD) as a printed reference for monocular distance estimation. Across all measurement intervals, the proposed method demonstrated consistent performance, with relative errors remaining within acceptable bounds, particularly for male subjects, whose biological IPD variability is statistically lower.
The error distributions presented in Table 1 and Table 2 indicate that inter-subject variability in interpupillary distance (IPD) contributes significantly to estimation accuracy. Female participants exhibited higher relative errors across all distance ranges, which aligns with biometric data reporting a larger standard deviation of IPD in adult females (61.53 mm ± 2.66 mm) compared to males (65.32 mm ± 1.50 mm) [27]. The current implementation, based on fixed sex-specific mean values, does not account for such deviations; consequently, individuals with atypical or outlier IPDs introduce additional estimation errors. A more general formulation would treat IPD as a random variable following a normal distribution with population-derived parameters and incorporate this variability directly into the inference process. Within a Bayesian framework, the observed pixel-based IPD can be interpreted as the likelihood, while the prior distribution encodes anatomical variability. The resulting posterior distribution would provide both point estimates and credible intervals, thereby quantifying uncertainty and reducing sensitivity to extreme values.
In line with the methodological aim of validating estimation accuracy rather than testing group-level hypotheses, no inferential statistics (e.g., ANOVA, t-test) were applied. Instead, descriptive metrics together with 95% confidence intervals were reported as the most appropriate indicators of accuracy for this proof-of-concept study.

4. Discussion

The present study focused exclusively on adult participants. The choice to use sex-based average IPD rather than age- or ethnicity-adjusted values was motivated by two factors: (1) statistical stability of IPD within adult populations, where variability is relatively low (±2.66 mm for females, ±1.50 mm for males), and (2) the practical aim of minimizing required user data, avoiding the complexity and privacy concerns associated with collecting or inferring additional demographic details. While demographic variation can influence IPD, its impact is smaller than the gain in simplicity and scalability for non-intrusive applications. This limitation has been acknowledged as a potential area for future work involving adaptive demographic scaling.
The correction mechanism applied to account for in-plane facial rotation significantly contributed to improving measurement accuracy, especially in scenarios where the subject’s head was not perfectly aligned with the camera’s optical axis. As observed in the experimental figures, in the absence of such correction, the raw IPD tends to be underestimated due to geometric distortion. By projecting the raw pupil vector onto the estimated facial normal direction, the algorithm successfully compensates for orientation-induced errors, thereby stabilizing IPD-based distance estimation.
The calibration procedure, grounded in a statistically determined average IPD (61.53 mm for females and 64.32 mm for males), proved to be both scalable and repeatable. The use of a printed facial reference in conjunction with a precision-controlled uniaxial displacement system enabled the derivation of a high-fidelity sixth-degree polynomial model mapping pixel-based IPD measurements to physical distance. This function encapsulates the camera’s optical behavior under fixed resolution and focal length, enabling robust real-time distance estimation once calibration is complete. Camera resolution plays a critical role in preserving accuracy, particularly at longer ranges where facial features occupy fewer pixels. In our tests, a 1280 × 720 resolution webcam provided acceptable performance for distances up to 2.4 m. For optimal results, cameras should provide at least 720p resolution, maintain >30 fps and have adequate low-light performance to ensure robust facial landmark detection. The proposed method achieved ~40 ms per frame processing time on a mid-range CPU without GPU acceleration, corresponding to ~25 fps in real time. This suggests that implementation of embedded vision hardware, such as ARM-based processors (e.g., Google Coral, NVIDIA Jetson Nano) with dedicated AI accelerators, is feasible.
In a fixed optical configuration, the pixel-IPD-distance relationship is smooth but nonlinear due to projection geometry, landmark rounding, head-pose correction, and lens distortion remnants. A sixth-degree polynomial provides sufficient flexibility to capture this curvature across the calibrated ranges without oscillatory artifacts. The choice is supported by cross-validated error comparisons against lower/higher degrees and by residual diagnostics. Additionally, the polynomial mapping is computationally lightweight and fully deterministic, which favors embedded real-time implementations. We note that for heterogeneous, multi-camera deployments, machine-learning regressors (e.g., SVR, Random Forest) may improve cross-device generalization and constitute a natural extension of this work.
An analysis of the error distributions presented in the two tables—corresponding to male and female participants—reveals consistent trends with respect to both distance intervals and biological sex. Male participants exhibited lower relative errors across all three distance ranges, with average mean relative errors of approximately 3.87%, 4.75%, and 5.53% for the 20–80 cm, 80–160 cm, and 160–240 cm intervals, respectively. In contrast, female participants recorded higher mean relative errors of 6.0%, 6.7%, and 7.27% across the same intervals. This discrepancy can be attributed in part to the higher standard deviation of interpupillary distance within the female population (±2.66 mm) compared to males (±1.50 mm), as reported in the biometric reference data. Since the method relies on fixed average IPD values, increased inter-subject variability introduces greater estimation error, particularly when individual IPDs deviate from the population mean. Moreover, the relatively smaller average IPD in female subjects results in a lower number of pixels per millimeter in the image space, compounding the sensitivity of measurements to pixel-level inaccuracies. These findings underscore the importance of accounting for biological variation in the design of generalized distance estimation models and suggest that error bounds may be further reduced through dynamic sex estimation or adaptive reference scaling.
However, one critical observation from the results is the degradation in measurement precision at larger distances. As the subject moves farther from the camera, the apparent size of facial features in the image diminishes, resulting in a lower pixel resolution for the IPD measurement. This reduction in pixel density limits the sensitivity of the method, as even small variations in pixel measurement translate to larger absolute errors in real-world distance estimation. This effect is particularly evident in the 160–240 cm interval, where both male and female groups exhibited slightly higher mean relative errors compared to shorter ranges. Despite this limitation, the method maintains its reliability due to the stability of the printed reference and the geometric correction process. For practical applications, the observed resolution loss suggests the importance of selecting appropriate camera resolution and positioning for the target measurement range. Future implementations may benefit from higher-resolution sensors or multi-frame averaging techniques to enhance pixel-level accuracy at greater distances.
Beyond human–camera interactions, the proposed approach could be adapted to other contexts where stable anatomical or morphological features are known. Potential examples include monitoring animals in veterinary or agricultural environments by exploiting average interocular distances or applying similar geometric principles to structured objects in security and traffic monitoring. These extensions illustrate the broader applicability of the method as a non-intrusive monocular distance estimation tool.

5. Conclusions

This study proposed and validated a monocular distance estimation method based on the statistically averaged interpupillary distance (IPD), eliminating the need for subject-specific calibration. Through a structured calibration process using printed face templates and a precisely controlled displacement system, a sixth-degree polynomial function was derived to map pixel-based IPD measurements to real-world distances. The implementation of a geometric correction algorithm for in-plane facial rotation further improved the robustness and accuracy of the method.
Experimental validation involving 26 participants demonstrated that the method achieves consistent and reliable performance across a broad range of distances. The lowest relative errors were observed at shorter distances, with a moderate increase in error at longer ranges due to reduced pixel resolution of facial features. This highlights a key trade-off between camera resolution, subject distance, and measurement precision, suggesting that optimal performance is achieved within the mid-range calibration interval (80–160 cm).
While the method demonstrated consistent performance within the tested range, certain limitations should be noted. The current implementation was validated only up to 2.4 m, beyond this distance, the reduced pixel resolution of facial landmarks may increase estimation error. In addition, the approach relies on sex-based average IPD values, which do not fully capture demographic variability across age groups, ethnicities or individuals with atypical facial proportions. These factors may affect generalization and should be addressed in larger studies through extended calibration datasets and adaptive modeling strategies. Future large-scale studies with balanced cohorts could incorporate formal hypothesis testing to further characterize subgroup effects, complementing the descriptive statistics and confidence intervals reported here.
Future work will explore integrating this method into application OptoGuard [28], where continuous monitoring of user-screen distance can provide alerts to maintain the recommended 40–63 cm range, helping prevent digital eye strain and other computer-related visual disorders.

Author Contributions

Conceptualization, C.D. and E.S.; methodology, M.C.T.; software, C.D.; validation, C.D., M.C.T. and E.S.; formal analysis, M.C.T.; investigation, C.D.; resources, M.C.T.; data curation, C.D., M.C.T. and E.S.; writing—original draft preparation, C.D.; writing—review and editing, E.S.; visualization, M.C.T.; project administration, M.C.T.; funding acquisition, M.C.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by North-East Regional Programme 2021–2027, PR/NE/2024/P1/RSO1.1_RSO1.3/1, through grant “Multifunctional video intercom system connected to the internet with cyber-attacks protection” code 338210.

Institutional Review Board Statement

Ethical review and approval were waived for this study due to its non-invasive nature and the absence of sensitive personal data, in accordance with the institutional guidelines of Gheorghe Asachi Technical University of Iasi. The study involved only image-based analysis of adult volunteers and did not include any medical procedures, interventions, or collection of sensitive data.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Written consent was also obtained for the use of any images or data that may be considered identifiable.

Data Availability Statement

The datasets generated during the current study are available from the corresponding author upon reasonable request. All data have been anonymized to ensure participant privacy.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Qiu, L. Internet of Things and Cloud Computing-Based Adaptive Content Delivery in E-Learning Platforms. Int. J. Adv. Comput. Sci. Appl. 2024, 15, 11. [Google Scholar] [CrossRef]
  2. Karaoglu, A. Optimal Sample Lens Positioning in Digital Camera Systems. arXiv 2024, arXiv:2407.01789. [Google Scholar] [CrossRef]
  3. Rachidi, O.; Ed-Dahmani, C.; Bououlid Idrissi, B. Advanced Pedestrian Distance Estimation for ADAS with Canny Edge Detection and Stereo Vision. In Proceedings of the E3S Web of Conferences; EDP Sciences: Les Ulis, France, 2025; Volume 601, p. 02012. [Google Scholar]
  4. Song, C.H.; Blukis, V.; Tremblay, J.; Tyree, S.; Su, Y.; Birchfield, S. Robospatial: Teaching Spatial Understanding to 2D and 3D Vision-Language Models for Robotics. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: New York, NY, USA, 2025; pp. 15768–15780. [Google Scholar]
  5. Turner, J.; Weyers, B. Towards Formalisation of 3D User Interaction: Making Spatial Interaction Verifiable. In Proceedings of the 17th ACM SIGCHI Symposium on Engineering Interactive Computing Systems (EICS ‘25), Companion Proceedings; ACM: New York, NY, USA, 2025. [Google Scholar]
  6. del Campo, G.; Saavedra, E.; Piovano, L.; Luque, F.; Santamaria, A. Virtual Reality and Internet of Things Based Digital Twin for Smart City Cross-Domain Interoperability. Appl. Sci. 2024, 14, 2747. [Google Scholar] [CrossRef]
  7. Wade, N.J. Revealing the Concealed in Monocular and Binocular Vision. Vision 2025, 9, 47. [Google Scholar] [CrossRef] [PubMed]
  8. Leivers, H.K.; Thomas, C.; Van den Tillaar, R.; Bampouras, T.M. The Effects of Simulated Monocular and Binocular Vision Impairment on Football Penalty Kick Performance. Eur. J. Sport Sci. 2024, 24, 918–929. [Google Scholar] [CrossRef]
  9. Zhang, Z.; Xiang, H.; Li, D.; Leng, C. Measurement Method of Interpupillary Distance and Pupil Height Based on Ensemble of Regression Trees and the BlendMask Algorithm. Appl. Sci. 2023, 13, 8628. [Google Scholar] [CrossRef]
  10. Zaremba, A.; Nitkiewicz, S. Distance Estimation with a Stereo Camera and Accuracy Determination. Appl. Sci. 2024, 14, 11444. [Google Scholar] [CrossRef]
  11. Wang, Y.; Liu, X.; Zhang, M.; Chen, H.; Li, Y. Multi-Camera Visual Odometry. In Advances in Guidance, Navigation and Control: Proceedings of the 2024 International Conference on Guidance, Navigation and Control; Springer Nature: Cham, Switzerland, 2025; Volume 11, p. 1347. [Google Scholar]
  12. Li, Y.; Li, Y.; Wang, Z. FaceCloseup: Enhancing Mobile Facial Authentication with Perspective Distortion-Based Liveness Detection. Computers 2025, 14, 254. [Google Scholar] [CrossRef]
  13. Xing, H.; Tan, S.Y.; Qamar, F.; Jiao, Y. Face Anti-Spoofing Based on Deep Learning: A Comprehensive Survey. Appl. Sci. 2025, 15, 6891. [Google Scholar] [CrossRef]
  14. Vodichev, N.; Leus, A.; Gavrilov, D.; Efremov, V.; Zuev, V.; Kholodnyak, I.; Dastgerdi, M.V. Distance and Relative Speed Estimation of Vehicles Moving in the Same Direction. In Proceedings of the 2024 8th Scientific School Dynamics of Complex Networks and Their Applications (DCNA), Kaliningrad, Russian, 19–21 September 2024; IEEE: New York, NY, USA, 2024; pp. 262–265. [Google Scholar]
  15. George, R.; Clancy, J.; Brophy, T.; Sistu, G.; O’Grady, W.; Chandra, S.; Glavin, M. Infrastructure Assisted Autonomous Driving: Research, Challenges, and Opportunities. IEEE Open J. Veh. Technol. 2025, 6, 662–716. [Google Scholar] [CrossRef]
  16. Ni, T.; Wang, L.; Yin, X.; Cai, Z.; Yang, Y.; Kong, D.; Liu, J. Experimental Study on Measuring and Tracking Structural Displacement Based on Surveillance Video Image Analysis. Sensors 2024, 24, 601. [Google Scholar] [CrossRef] [PubMed]
  17. Yao, B.; Liu, J.; Luo, H. Real-Time Safety Distance Detection and Warning for Elevator Inspection Based on Adaptive Monocular Depth Estimation. Measurement 2024, 235, 114960. [Google Scholar] [CrossRef]
  18. Kim, J.; An, K.; Lee, D. RoSA Dataset: Road Construct Zone Segmentation for Autonomous Driving. In Proceedings of the European Conference on Computer Vision (ECCV); Springer Nature: Cham, Switzerland, 2024; pp. 322–338. [Google Scholar]
  19. Liu, Y.; Zhang, F.; Chen, L.; Jiang, T. A 3D Blur Suppression Method for High-Resolution and Wide-Swath Blurred Images Based on Estimating and Eliminating Defocused Point Clouds. Remote Sens. 2025, 17, 928. [Google Scholar] [CrossRef]
  20. Nasibov, H. Evaluation of Focus Measures for Hyperspectral Imaging Microscopy Using Principal Component Analysis. J. Imaging 2024, 10, 240. [Google Scholar] [CrossRef] [PubMed]
  21. Westermeier, F.; Murmu, C.; Kohm, K.; Pagano, C.; Wienrich, C.; Babu, S.V.; Latoschik, M.E. Interpupillary to Inter-Camera Distance of Video See-Through AR and Its Impact on Depth Perception. In Proceedings of the 2025 IEEE Conference on Virtual Reality and 3D User Interfaces (VR), Saint Malo, France, 8–12 March 2025; IEEE: Piscataway, NJ, USA, 2025; pp. 537–547. [Google Scholar]
  22. Merino, X.; Langborgh, S.; Pangelinan, G.; King, M. Reading between the Eyes: Assessing the Impact of Interpupillary Distance on Face Recognition. In Proceedings of the SoutheastCon 2025, Concord, NC, USA, 22–30 March 2025; IEEE: Piscataway, NJ, USA, 2025; pp. 267–272. [Google Scholar]
  23. Pirayesh, Z.; Hassanzadeh-Samani, S.; Farzan, A.; Rohban, M.H.; Ghorbanimehr, M.S.; Mohammad-Rahimi, H.; Motamedian, S.R. A Deep Learning Framework to Scale Linear Facial Measurements to Actual Size Using Horizontal Visible Iris Diameter: A Study on an Iranian Population. Sci. Rep. 2023, 13, 13755. [Google Scholar] [CrossRef] [PubMed]
  24. Wu, Y.; Wang, Y.; Qu, S.; Li, Z.; Chen, G. ELF-UA: Efficient Label-Free User Adaptation in Gaze Estimation. In Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI), Macao, China, 3–9 August 2024; pp. 1452–1461. [Google Scholar]
  25. Slapničar, G.; Wang, W.; Luštrek, M. Generalized Channel Separation Algorithms for Accurate Camera-Based Multi-Wavelength PTT and BP Estimation. Biomed. Opt. Express 2024, 15, 3128–3146. [Google Scholar] [CrossRef] [PubMed]
  26. Frisch, F.; Mehringer, W.; Ring, M.; Michelson, G.; Eskofier, B.M. Physical Exertion Does Not Lead to a Change in the Time Taken to Correctly Recognize Stereoscopic Stimuli in Changing Horizontal Disparity Levels. Front. Physiol. 2025, 16, 1602503. [Google Scholar] [CrossRef]
  27. Murray, N.P.; Hunfalvay, M.; Bolte, T. The reliability, validity, and normative data of interpupillary distance and pupil diameter using eye-tracking technology. Transl. Vis. Sci. Technol. 2017, 6, 2. [Google Scholar] [CrossRef]
  28. Serea, E.; Temneanu, M.C.; Donciu, C. Opto-Guard: Non-invasive System of Continuous User Assurance of Keeping the Optimal Distance of the Eyes from the Computer Screen. In Proceedings of the 2022 International Conference and Exposition on Electrical And Power Engineering (EPE), Iasi, Romania, 20–22 October 2022; pp. 1–5. [Google Scholar]
Figure 1. Facial test targets used for calibration, male and female references.
Figure 1. Facial test targets used for calibration, male and female references.
Appliedmath 05 00118 g001
Figure 2. Experimental setup for distance calibration using printed facial targets, for male reference with IPD = 65.32 mm and female reference with IPD = 61.53 mm.
Figure 2. Experimental setup for distance calibration using printed facial targets, for male reference with IPD = 65.32 mm and female reference with IPD = 61.53 mm.
Appliedmath 05 00118 g002
Figure 3. Flow diagram of the proposed distance estimation method.
Figure 3. Flow diagram of the proposed distance estimation method.
Appliedmath 05 00118 g003
Figure 4. Facial landmarks and corresponding vectors for IPDcorrected calculation (abstracted case).
Figure 4. Facial landmarks and corresponding vectors for IPDcorrected calculation (abstracted case).
Appliedmath 05 00118 g004
Figure 5. Real IPD achieved using angle correction with the proposed method.
Figure 5. Real IPD achieved using angle correction with the proposed method.
Appliedmath 05 00118 g005
Table 1. Relative error statistics for male participants.
Table 1. Relative error statistics for male participants.
Males
Intervals20–80 cm80–160 cm160–240 cm
ErrorsMaximum Relative Errors [%]Minimum Relative Errors [%]Mean Relative Errors
[%]
Maximum Relative Errors [%]Minimum Relative Errors [%]Mean Relative Errors
[%]
Maximum
Relative Errors [%]
Minimum Relative Errors [%]Mean Relative Errors
[%]
14.94.44.65.14.24.86.25.45.8
24.13.63.95.24.65.05.55.15.4
33.63.13.54.94.24.75.74.85.3
44.23.53.84.84.54.66.15.35.9
53.83.23.75.14.34.85.44.64.9
63.93.33.85.24.75.15.34.85.1
74.13.63.94.94.44.76.15.45.8
84.13.43.74.84.34.45.85.15.5
94.63.94.34.84.64.75.54.95.3
104.23.54.15.14.14.56.35.25.9
114.13.63.95.04.24.95.45.15.3
123.83.13.74.94.54.85.75.05.6
133.93.33.55.14.14.45.95.15.5
144.83.94.05.34.64.96.15.55.9
154.13.63.85.24.44.95.85.35.7
Mean4.153.533.875.034.384.755.795.115.53
Std dev0.380.360.340.160.140.270.330.290.31
95% CI--±0.19--±0.19--±0.14
Table 2. Relative error statistics for female participants.
Table 2. Relative error statistics for female participants.
Females
Intervals20–80 cm80–160 cm160–240 cm
ErrorsMaximum Relative Errors [%]Minimum Relative Errors [%]Mean Relative Errors
[%]
Maximum Relative Errors [%]Minimum Relative Errors [%]Mean Relative Errors
[%]
Maximum Relative Errors [%]Minimum Relative Errors [%]Mean Relative Errors
[%]
16.55.96.36.96.36.77.56.87.1
26.65.86.46.86.16.57.87.17.5
36.35.55.96.86.06.67.67.07.3
46.45.45.77.06.46.87.26.77.0
56.95.86.27.16.26.97.16.46.9
65.95.65.86.66.06.47.86.37.5
76.35.45.76.96.36.67.56.37.2
85.45.95.86.46.16.37.67.17.5
96.76.46.67.36.67.17.87.37.5
106.15.75.97.26.77.07.77.17.3
116.25.35.77.06.56.87.47.07.2
Mean6.35.76.0766.96.77.556.827.27
Std dev0.470.380.290.260.260.290.250.30.21
95% CI--±0.20--±0.20--±0.14
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Temneanu, M.C.; Donciu, C.; Serea, E. Distance Measurement Between a Camera and a Human Subject Using Statistically Determined Interpupillary Distance. AppliedMath 2025, 5, 118. https://doi.org/10.3390/appliedmath5030118

AMA Style

Temneanu MC, Donciu C, Serea E. Distance Measurement Between a Camera and a Human Subject Using Statistically Determined Interpupillary Distance. AppliedMath. 2025; 5(3):118. https://doi.org/10.3390/appliedmath5030118

Chicago/Turabian Style

Temneanu, Marinel Costel, Codrin Donciu, and Elena Serea. 2025. "Distance Measurement Between a Camera and a Human Subject Using Statistically Determined Interpupillary Distance" AppliedMath 5, no. 3: 118. https://doi.org/10.3390/appliedmath5030118

APA Style

Temneanu, M. C., Donciu, C., & Serea, E. (2025). Distance Measurement Between a Camera and a Human Subject Using Statistically Determined Interpupillary Distance. AppliedMath, 5(3), 118. https://doi.org/10.3390/appliedmath5030118

Article Metrics

Back to TopTop