Next Article in Journal
Energy Evaluation and Passive Damage Detection for Structural Health Monitoring in Aerospace Structures Using Machine Learning Models
Previous Article in Journal
Analysis of Corrosion by Speckle Polarimetry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Blur Feature-Guided Cascaded Calibration Method for Plenoptic Cameras

1
College of Resource Environment and Tourism, Capital Normal University, Beijing 100048, China
2
Engineering Research Center of Spatial Information Technology, Beijing 100048, China
3
Chinese Academy of Surveying and Mapping, Beijing 100830, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(16), 4940; https://doi.org/10.3390/s25164940 (registering DOI)
Submission received: 9 July 2025 / Revised: 6 August 2025 / Accepted: 8 August 2025 / Published: 10 August 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

Accurate and robust calibration of multifocal plenoptic cameras is essential for high-precision 3D light field reconstruction. In this work, we propose a blur feature-guided cascaded calibration for the plenoptic camera. First, white images at different aperture values are used to estimate the high-confidence center point and radius of micro-images, and the defocus theory is used to estimate the initial values of the intrinsic parameters. Second, the gradient value is introduced to quantify the degree of blurring of the corner points, which are then divided into three types: clear, semi-clear, and blurred. Furthermore, a joint geometric constraint model of epipolar lines and virtual depth is constructed, and the coordinates of the semi-clear and blurred corner points are optimized in a step-by-step manner by using the clear corner point coordinates. The micro-image center ray projection equation is then devised to assist in the optimization of the microlens array core parameters and establish blur-adaptive credibility weights, thereby constructing a global nonlinear optimization. Finally, the proposed method is tested on both simulated and captured datasets, and the results exhibit superior performance when compared with the established methods described by Labussière, Nousias, and Liu. The proposed method excels in corner feature extraction, calibration accuracy of both internal and external parameters, and calibration sensitivity when applied to multifocal-length light field cameras, highlighting its advantages and robustness.

1. Introduction

Plenoptic (light field) imaging can record the position and direction of light in a scene, forming a light field image containing four-dimensional information. By using this information, sub-aperture images at different angles can be obtained, or refocusing can be used to obtain refocused images with different depths of field [1,2,3]. These characteristics give light field imaging technology unique advantages in terms of depth estimation and 3D reconstruction, and it is gradually being incorporated into autonomous driving, digital twins, augmented reality, industrial inspection, and medical imaging techniques [3,4,5].
Camera calibration is a process that determines the intrinsic and extrinsic parameters of a camera and establishes an accurate correspondence between image pixels and real-world coordinates. Defining this correspondence is a fundamental problem in computer vision, an important basis for reconstruction and positioning, and an important bridge for establishing connections between 2D images and 3D space [5]. On the basis of camera design structures, the existing calibration methods can be roughly divided into those involving camera arrays [4,5,6,7], unfocused light field cameras [8,9,10,11], and focused light field cameras [12,13,14,15]. Among them, the core research on light field camera calibration has focused on imaging models and the accuracy and robustness of the calibration parameters. However, for more advanced multifocal focused light field cameras designed for varying working distances (such as small scenes at close range and large scenes at long range), the existing methods present very significant challenges in terms of the accuracy and sensitivity of the calibration parameters.
To address these problems, we propose a blur feature-guided cascaded calibration method for plenoptic cameras (BFC-PC). The main contributions of this paper are as follows:
  • For the first time, a geometric observation model of the virtual depth domain is proposed. The geometric relationship between the epipolar lines in the micro-image (MI) and the virtual depth space is used to construct a geometric observation model of the virtual depth domain that is unique to plenoptic cameras.
  • A quantification and grading rule for micro-image clarity that accounts for blur-perceived plenoptic features is constructed. The geometric observation mode of the virtual depth domain is used to iteratively filter the outliers and optimize the corner point coordinates in a step-by-step diffusion manner, effectively improving the extraction quality of the corner point features.
  • A decoupled and cascaded optimization strategy is proposed for the first time. The central light projection equation of the target MI is constructed to assist in optimizing the parameters of the microlens array (MLA), and redundant observations of the same corner point and adaptive confidence weights that account for blur are established to construct a global nonlinear optimization model. The complexity of the nonlinear function can thus be effectively reduced, avoiding sensitivity to the initial estimated parameters and making it easier to converge to the optimal solution.
The components of this work are as follows: Section 2 introduces the existing focused light field camera calibration methods and their shortcomings, Section 3 details the methods proposed in this paper, Section 4 presents the experiments and analyses of the results, and Section 5 presents the conclusions.

2. Related Work

The related work in this field has centered primarily on focused light field camera-based approaches. According to the different types of images used during the calibration process, the existing focused light field camera calibration methods can be roughly divided into two categories: calibration methods based on interpreted images and calibration methods based on original images.
Calibration methods based on interpreted images rely on reconstructed images (such as sub-aperture images and all-focus images). Johansend et al. [12] first proposed a focused light field camera calibration method and introduced a new model for correcting radial and depth distortions; however, this approach easily falls into local optima and requires a good initial value to converge. Heinze et al. [16] further improved the model and incorporated the main lens tilt and displacement into the model, and the results were incorporated into the RxLive software from Raytrix GmbH. Zeller et al. [13,17] proposed two methods for calibrating focused plenoptic cameras and depth maps. They extracted features from the input fully focused image and equivalently treated the focused light field camera as a multi-eye pinhole camera array for camera calibration purposes.
Some scholars have proposed the direct use of original light field images to calibrate camera parameters. Noury et al. [15] defined the projection process from a spatial point to a raw light field image; at the same time, a new corner detection algorithm was developed to obtain the corners from the raw image with subpixel accuracy. O’Brien et al. [18] proposed the concept of circular-domain light field features, which were used to determine the correspondences between corners, and Liu et al. [19] used a step-by-step method to detect the corner features of a raw image and complete the calibration process. Nousias et al. [14] established a projection model from spatial points to original images and developed a new corner detection algorithm to extract checkerboard corners. Furthermore, Labussière et al. [20,21] calibrated the focal length of the MLA; however, the accuracy of the blur radius was low. The calibration method used for a focused light field camera usually includes three key steps: initial estimation of the calibration parameters, feature extraction, and nonlinear optimization.
In recent years, deep learning-based approaches for corner detection and adaptive weighting have shown promising advances in camera calibration tasks. For example, CCDN (Checkerboard Corner Detection Network) [22] introduces a fully convolutional network tailored for detecting corner points under distortion, noise, and blur, achieving high robustness across challenging imaging conditions. Although such methods perform well on conventional cameras, they are not specifically designed for the multi-aperture and defocus characteristics of light field imaging. Additionally, several learning-based camera calibration frameworks have been proposed. LiFCal [23] presents an online calibration strategy for light field cameras by leveraging bundle adjustment and dynamic sequence analysis, enabling real-time parameter estimation. Moreover, Liao et al. [24] provide a comprehensive survey of deep learning approaches for camera calibration, including techniques involving uncertainty modeling and adaptive confidence weighting.
While these learning-based strategies offer automation and robustness, most of them focus on standard image modalities and lack physical interpretability in plenoptic camera geometry. In contrast, our method proposes a blur-aware corner modeling and cascaded optimization strategy that explicitly leverages the multi-focus nature of plenoptic images, offering better integration with the physical imaging model and higher controllability for calibration tasks.
When calibrating a multifocal focused light field camera, the feature extraction and nonlinear optimization steps still present problems such as feature points with uneven quality levels, highly complex optimization problems, and high sensitivity to the initial estimation value.
(1)
Feature point quality issues
Currently, the corner detectors used in more advanced methods include saddle point detection methods [14] and template matching methods [15,20]. The primary conception behind saddle point detection methods is that when a corner point is present in an MI, that corner point is found at the “saddle point/maximum and minimum axes” in the intensity domain. These two axes are not necessarily orthogonal, but the direction of the saddle is determined by using the corner point as the center, where these are the maximum and minimum cumulative intensity lines passing through the image. These lines determine two potential saddle axes, and the corner point is located at their intersection. Moreover, the main idea motivating template matching methods involves describing the appearance characteristics of a corner point (such as its intensity distribution or geometric shape) via a predefined template, adjusting the template parameters and calculating similarity values (such as the cross-correlation and mean squared error metrics) to best fit the target corner point and calculate its coordinate value.
For a certain corner point in the real 3D world, the MIs formed by a multifocal focused light field camera will exhibit different degrees of blur. The existing corner detectors all use the characteristics of the image intensity changes to extract corner points, and when an MI is relatively blurred, the intensity change induced at the corner point is distributed in a pixel area within a certain range centered on the real corner point. Therefore, the corner point coordinates extracted by the existing corner detectors are inaccurate or even wrong, which leads to poor accuracy for the calculated camera parameters. Notably, the greater the blur of the MI is, the larger the intensity change distribution area, and the worse the accuracy of the extracted corner point coordinates may be.
(2)
Sensitivity of nonlinear optimization to initial values
Nonlinear optimization involves optimizing the internal and external camera parameters of the initial estimation process to minimize the reprojection error. The existing methods typically use nonlinear optimization algorithms such as the Levenberg–Marquardt algorithm [25] or the Powell method [26] to optimize all of the parameters at once. These nonlinear optimization algorithms usually rely on the initial value to gradually adjust their parameters and iteratively calculate them to approach the optimal solution of the objective function. However, this single process for optimizing all of the parameters may be sensitive to the selected initial value. The main reasons for these findings include the following three points: First, due to the complexity of nonlinear functions, different initial values may cause the employed nonlinear optimization algorithm to iterate along different paths and finally converge to different solutions. Second, multiple extreme points may be present in a nonlinear optimization problem, including local optimal and global optimal values. If the initial point is not selected properly, the algorithm may fall into a local optimal solution and fail to find the global optimal solution because the nonlinear function may have a flat gradient or steep curvature near a local optimal solution, making it difficult for the algorithm to jump out of the local optimal area. Finally, if the initial point is not chosen properly, a nonlinear optimization algorithm may not converge to the optimal solution during the iteration process because of the accumulation of factors such as calculation errors and rounding errors.

3. Methods

When calibrating multifocal plenoptic cameras, we overcome accuracy and sensitivity issues via three streamlined stages: (1) Robust Initialization of the Intrinsic Parameters: Multiple white images corresponding to different aperture values are used to estimate the center and radius of the target MI. The linear relationship between the aperture and radius is fitted according to defocusing theory to solve for the initial values of the intrinsic parameters. (2) Blur-Guided Corner Extraction: In this process, the raw checkerboard corners are detected and clustered and then sorted by the estimated blur. The outliers are then filtered via epipolar and virtual depth constraints, and the coordinates are iteratively refined from sharp to blurred by employing a diffusion strategy. (3) Cascaded, Decoupled Optimization: First, the microlens array geometry is refined with a central-ray projection model. Then the intrinsic parameters are fixed, and the extrinsic parameters are estimated via a pinhole-lens model by using the clearest corners. Finally, gross corner errors across frames are eliminated, and global blur-weighted nonlinear optimization is performed. An overview of the proposed method is shown in Figure 1.

3.1. Micro-Image Central Grid and Initial Intrinsic Parameter Estimation

The high precision of micro-image central grid parameters and initial estimated intrinsic parameter values are related to the accuracy of the geometric observation data, namely, the original white image (generally obtained via light diffuser photography), and existing methods usually use only a single white image at different apertures to extract the center point and micro-image radius. However, since it is impossible to strictly control photographic conditions when shooting white images, there may be overexposure or underexposure in addition to large errors due to improper operation when the camera aperture scale value is manually adjusted. Ultimately, the center point and micro-image radius extracted from the original white image may have large errors. The proposed method uses white images taken at different apertures to extract the coordinates of the center point and estimate the initial values of the intrinsic parameters. Its accuracy directly affects the geometric constraint reliability of the virtual depth observation mode and the convergence efficiency of the decoupled stacking optimization (reasonable initial parameter values are required).
Extraction of the center point of the micro-image. Due to the vignetting effect at the edge of the microlens, the brightest position of the micro-image in the white image can be regarded as the center of the microlens above it. The white image is processed via brightness correlation and non-maximum suppression, and the position of the maximum light intensity in each micro-image is determined by finding the micro-image center with the subpixel precision intensity centroid method [27,28]. Figure 2 shows the estimated distribution of the micro-image center.
First, by using the different focal lengths of the microlenses and changing the aperture of the main lens, the camera aperture is adjusted to different scale values N m ,   m { 0,1 , , M } via a light diffuser mounted on the main lens, and multiple original white images I j ,   j   { 0 ,   1 , , J } are taken at the same scale value. Second, the subpixel precision intensity centroid method [15,27,28] is used to detect the micro-image center C k , l , ( k , l ) { ( 0 ,   0 ) , ( 0 ,   1 ) ,   , ( K , L ) } in all original white images at a certain aperture scale value N m . To avoid outliers at the center point of the extracted micro-image caused by the randomness of the above-mentioned errors, the median of the center of the micro-image under the aperture scale value N m is obtained by calculating the median of the relationship between the multiple original white images and the micro-image center C k , l ~ , ( k , l ) { ( 0 ,   0 ) , ( 0 ,   1 ) ,   , ( K , L ) } . The pixel translation ( T x , T y ) and rotation angle φ of the micro-image array are calculated via a nonlinear optimization method that uses the extracted median coordinates of all of the micro-images.
After the coordinates of the micro-image center are obtained, the micro-image center grid parameters are then fitted. First, the initial side length of the micro-image array is determined by calculating the distance between adjacent micro-image centers, and then, according to the MLA arrangement rule, the coordinate position C k , l = u ,   v ,   1 T of the grid vertex on the image plane can be expressed as
R o w a l i g n e d : τ o f f s e t = 1 2   ( i f   l   i s   e v e n   ,   0   o t h e r w i s e ) u = δ i · ( k + τ o f f s e t ) v = δ i · l · s i n π 3 C o l u m n a l i g n e d : τ o f f s e t = 1 2   ( i f   k   i s   o d d ,   0   o t h e r w i s e ) u = δ i · k · c o s π 6 v = δ i · ( l + τ o f f s e t )  
where ( k ,   l ) represents the row and column number of the vertex in the grid and where τ o f f s e t is the offset.
Finally, the Levenberg–Marquardt algorithm is used for nonlinear optimization to compare the distances between the mesh vertices and the detected micro-image center to determine the pixel offset ( T x ,   T y ) in the image coordinates and the rotation angle ϑ z around the Z-axis:
arg m i n τ x , τ y , ϑ z , δ i   k , l   T c k , l c k , l 2 w i t h   T = c o s ϑ z s i n ϑ z τ x s i n ϑ z c o s ϑ z τ y 0 0 1
Estimation of the micro-image radius. In a white image, different types of MLAs produce micro-image features with specific sizes and light intensity distributions, as shown in Figure 3. For all original white images of size N m at the same aperture scale, the Euclidean distance D ( k ,   l ) between the micro-image center C k , l and the median C k , l ~ corresponding to each MLA k ,   l is calculated, and the MLA is degenerated into a grid image to generate an error map, where one pixel represents one MLA and the pixel value is D ( k ,   l ) in the normalized range of 0 to 255. The structural similarity index proposed is then used to evaluate the similarity between the error maps by comparing the pixel values, brightness, contrast, and structural information. The error maps with lower similarity are selected, and the corresponding original white images are regarded as unreliable data sources in terms of calculating the micro-image radius. To avoid the influence of the vignetting effect, the micro-image radius is estimated only for the micro-image in the center area of the white image. For the white image I j of a certain aperture scale value N m , the second-order central moment is calculated by the median of the micro-image center, the covariance matrix is constructed, and the micro-image radius is calculated, according to reference [20].
Estimation of the initial intrinsic parameters. By using the correspondence N = 2 A V between the lens aperture value A V and the actual aperture value N , the actual aperture value N corresponding to each white image can be calculated. Then, according to the linear relationship between the micro-image radius R and the actual aperture value N , the linear parameters are calculated as follows:
R i ( N 1 ) = m ·   N 1 + q i
where i represents the type of MLA, R i represents the micro-image radius of type i , and m and q i are linear parameters. The intrinsic parameter relationships are defined in Formulas (4) and (5):
m = d · F 2 D
q i = 1 f i ( c · D d + D ) · d 2 c 2
where d represents the distance from the MLA to the sensor, F represents the focal length of the main lens, D represents the distance from the MLA to the main lens, f i represents the focal length of different types of MLAs, and c represents the diameter of the micro-image.
The initial intrinsic parameters of the camera are further calculated by using the m and q i values calculated above as well as the factory parameters of the camera (focal length F of the main lens and sensor size). The calculation formulas for parameters d and D are as follows:
d = 2 m F + 4 m · h 2 1 1 4 F h
D = h 2 1 1 4 F h 2 d
We set the distance between the main lens plane and the sensor plane to ( D + d ) and the MLA plane to be parallel to the main lens plane, with a distance of D . Furthermore, we set the principal point to the center of the image and initialized all distortion coefficients of the main lens to zero. The translation ( t x , t y ) and rotation angle θ of the MLA are initialized by the pixel offset ( T x , T y ) and rotation angle φ of the acquired micro-image parameters, as shown in Figure 4. Among them, θ can be decomposed into three components, the rotation angles θ x , θ y , and θ z around the X-axis, Y-axis, and Z-axis. The rotation angles around the X-axis and Y-axis are initialized to zero. The MLA focal length f i is initialized to
f i = d 2 q i D D + d c = d 2 q i μ  

3.2. Corner Feature Extraction Based on Blur Quantization Under the Virtual Depth Geometric Constraint Model

The intensity change at the corner point in the micro-image in the original frame is distributed within a certain range of pixels centered on the real corner point. Notably, the greater the blur in the micro-image is, the larger the intensity change distribution area, and thus, the corner point coordinates extracted by the existing method will have errors or even mistakes.
To effectively improve the accuracy of the corner point coordinates, this paper proposes a corner point feature extraction strategy guided by neighboring blur information. This approach uses the geometric relationship between the epipolar lines between the micro-images and the virtual depth space to build a geometric observation model of the virtual depth domain; then, by using this constructed model, the degree of the blurred imaging at multiple corner points in the cluster can be determined, and the corner point coordinates can be optimized from clear to blurred in a step-by-step diffusion manner. This process proceeds as follows:
Determination of the correspondence between corner points. When a traditional optical camera performs imaging, a 3D point projected into a single-frame image will only form a 2D point. In contrast, since the focused light field camera adds an MLA between the main lens and the sensor, this optical path design decomposes the light of the 3D point into multiple light beams with different directions. These light beams form multiple 2D points in multiple micro-images, which are generally called points of the same name. Since the subsequent ambiguity calculations, joint geometric constraint model construction, and corner point optimization are applicable only to points of the same name, it is necessary to first determine the correspondence between these 2D corner points.
When a focused light field camera is used to capture the corners of a chessboard, the light emitted from the same corner in the object space is projected onto multiple adjacent image points on the sensor; the spacing between these image points is very close, whereas the corners of the image space projected onto the sensor from different corners in the object space are far apart, as shown in Figure 5.
Based on the distribution law for projected 2D corner points, to effectively identify and process these observations, the Density-Based Spatial Clustering of Applications with Noise (DBSCAN) [29] spatial clustering algorithm is used to cluster the corner point coordinates detected within a certain distance. All corner point coordinates in the cluster are regarded as the projection point coordinates for the same 3D point in the image. This clustering algorithm can not only effectively process the spatial distribution relationship between corner points but also reduce the mismatch caused by image blur or noise interference to a certain extent. The corner point clustering result is shown in Figure 6, where different colors represent different corner point clusters.
Geometric observation mode in the virtual depth domain. Since MLAs of different focal lengths image the same corner point on a black and white chessboard, they will appear as corner points with different blur levels in the image space. Therefore, the coordinates of the corner points detected by existing methods are inaccurate. We analyze and utilize the intrinsic structure and imaging characteristics of multifocal focused light field cameras and propose a geometric observation mode in the virtual depth domain to improve the accuracy of the corner point coordinates and ensure the accuracy of camera parameter estimation, as shown in Figure 7.
In our model, both the Main Lens and the Microlens Array (MLA) are treated as thin lenses, consistent with classical geometric optics assumptions. As described by Zeller et al. [13], in both Galilean and Keplerian configurations of focused plenoptic cameras, the light rays refracted by the Main Lens tend to converge toward a specific spatial location. Although this location does not physically coincide with the sensor plane and cannot form a real image, it can be regarded as a virtual image point, which serves as a geometric reference for modeling. In this framework, a single object-space corner point is projected by the Main Lens onto a virtual image point, which is then observed by multiple microlenses from different viewpoints. These microlens observations produce multiple image-space corner projections that are constrained by a shared geometric relationship in the virtual depth domain.
Since all corner points in a cluster correspond to the same virtual image point, if the coordinates I i and I j of two or more corner points are known, the relative distance (virtual depth) v between the MLA and the virtual image can be calculated via triangulation.
v = M i M j M i M j I i I j
where M i M j represents the Euclidean distance (baseline) between the centers of MLA M i and M j , I i I j represents the Euclidean distance between corner points I i and I j in the micro-image, M i M j I i I j is the parallax between corner points I i and I j , and i and j represent the index numbers of the MLA and its micro-image, respectively.
Once the virtual depth v is estimated, a certain corner point I i can be used to backproject the virtual image point P v into the micro-image, where any corner point in the cluster is located by the following formula to obtain the back-projected coordinate I j :
I j = I i + C j C i · ( M j M i M j M i v )    
where C j C i represents the unit direction vector from the micro-image center C j to the micro-image center C i , that is, the epipolar direction.
Optimization of the coordinates of the corner points in the cluster via step-by-step diffusion. Based on the geometric observation mode for the virtual depth domain, a corner point blur feature quantization strategy based on the dual features of light field light propagation and the image gradient is proposed. Then, the corner point coordinates are refined via multilevel iterative processing according to the quantized blur values, as shown in Figure 8.
(1)
Blur feature quantization of corner points based on the dual features of light field ray propagation and the image gradient.
Multiple corner points { I 0 ,   I 1 , , I j } within the image-space cluster correspond to a unique corner point O on the black and white checkerboard in the object space, and according to the thin lens imaging principle, the blur degree of { I 0 ,   I 1 , , I j } is related only to the type of MLA. Based on geometric optics and image gradient information, the degree of blur of corner { I 0 ,   I 1 , , I j } can be quantified as follows:
S = α λ c ¯ i 2 · v 1 + q i   λ c ¯ i 2 + β 1 n · x y S x , y 2  
where S represents the blur score, where the higher the value, the clearer the corner points. The weighting parameters are set as α = 0.4 and β = 0.6, based on empirical evaluation across multiple datasets. The first term, weighted by α , is derived from a geometric optics perspective and reflects the spatial consistency between the virtual projection and the expected corner location, as inspired by Labussière et al. [20]. The second term, weighted by β , measures the visual blur degree of the corner in the micro-image. Since corner detection ultimately relies on image content, the image-based term is assigned a slightly higher weight to reflect its greater reliability in practical feature localization; c ¯ i represents the mean diameter of the micro-image of the i type; λ = D D + d represents the proportional coefficient between the micro-image and the MLA; v represents the virtual depth; q i = q i + c ¯ i 2 , S x , y = G x I x , y + G y I x , y represents the horizontal and vertical gradient value at coordinate ( x , y ) as calculated by the Sobel operator, and n represents the total number of pixels in the micro-image.
(2)
Calculation of the initial virtual depth.
Grading is based on the blur quantization value of the corner points. For each corner point cluster, the k-means method is applied to divide the corner point set P = { I 0 ,     I 1 ,   , I j } into multiple corner point subsets with different blur levels according to the fuzziness score. Assuming that there are three types of MLA, the corner point set can be divided into three mutually exclusive subsets: P A = I i P   S ( I i ) = A } , P B = I i P   S ( I i ) = B } , and P C = I i P   S ( I i ) = C } , where A, B, and C represent the three levels of clarity, sub-clarity, and blur, respectively.
To avoid the accuracy loss caused by MLA distortion and vignetting effects, we select at least three corner points close to the center of the micro-image from the corner point set P A of the clarity level and use the virtual depth geometric relationship (Formula (9)) to perform pairwise forward intersection to calculate the virtual depth v . The average is taken as the initial virtual depth value of the cluster, that is, the position of the virtual image point, as shown in P v .
(3)
Optimization of the corner point coordinates via a diffusion method.
By using each corner point I i in the set of clear corner points P A in the cluster, the virtual image point P v and Formula (10) form the rear intersection. The virtual image point P v is back-projected into the micro-image, where the target corner point I j is in the secondary definition corner point P B , and the median is considered to be the optimized corner point coordinate.
To avoid errors, the pixel blocks around the corner points are fitted, and the local maximum is used to represent the refined corner point coordinates. This optimization process is applied to all corner points in P B . Then, the clear corner point set P A and the optimized sub-clear corner point set P B are used as inputs, and more than six corner points close to the center of the micro-image are selected. Finally, the new virtual depth value is calculated and averaged to obtain the virtual depth v of the new cluster. On the basis of the clear corner point set P A and the updated sub-clear corner point set P B , the same optimization strategy is adopted to optimize the coordinates of the corner points in the fuzzy corner point set P C .

3.3. Nonlinear Optimization Strategy with Hierarchical Decoupling

Nonlinear optimization is the last step in camera calibration. Its function is to minimize the reprojection error or maximize a certain likelihood function by adjusting the initially estimated intrinsic and extrinsic parameters of the camera, light direction, scene depth, and other information to obtain accurate light restoration results. The existing methods usually use nonlinear optimization algorithms to optimize all of the parameters at once; however, this single-process optimization of all parameters may be sensitive to the initial values, resulting in an inability to converge to the optimal solution.
The intrinsic structure of the plenoptic camera 2.0 is composed of a main lens, an MLA, and a sensor element. In a scenario in which a white image and a black and white checkerboard are used as calibration objects, this structure enables the light field camera to capture two types of light: central light (as shown in Figure 9a) and the light of the corner feature (as shown in Figure 9b). Following the principle of “simplifying the beam solution problem and avoiding error accumulation”, we use these two types of light to construct the beam equation, introduce the progressive idea of multilayer grading, and propose a decoupled stacked nonlinear optimization, which primarily includes the construction of the projection model, the optimization of the MLA parameters assisted by the central light, the initial estimation and optimization of the extrinsic parameters, and the construction of the global nonlinear optimization function, which accounts for the credibility weight.
Construction of the projection model. To make better use of the feature that micro-images with different blurs are represented by MLAs with different focal lengths, the feature points in the micro-image space are represented by the center and radius, that is, p =   [ u , v , ρ , 1 ] T , and follow the projection model proposed by Labussière et al. [20]., as defined in Formula (12).
u v ρ 1 P ( i , k , l ) · T u ( k , l ) · ( K ( F ) · T c · p w )  
where P ( i , k , l ) represents the projection matrix of the MLA numbered ( k , l ) of type i, which accounts for the degree of blur. T u ( k , l ) represents the pose of MLA k , l expressed in the camera coordinate system, the function ( · ) models the lateral distortion, T c is the pose of the main lens relative to the world coordinate system, and p w =   [ x , y , z , 1 ] T represents the three-dimensional point coordinates of the real world. The intrinsic parameters of the light field camera include 16 + i parameters, including the focal length F of the main lens and its five lateral distortion coefficients Q 1 , Q 2 , Q 3 , P 1 , and   P 2 ; the sensor translations d and ( u 0 ,   v 0 ); the pose of the MLA, including three rotations θ x , θ y , and θ z ; three translations t x , t y , and D ; the spacing of the MLA μ ; and the focal length of the i MLA f i ,   i = 0,1 , 2 .
Center-ray-assisted MLA parameter optimization. By using the extracted high-confidence center point, the initially estimated MLA center and the main lens center, the light reflected by the light diffuser is inversely constructed, as shown by the dotted line in Figure 9a. In other words, a center ray can be regarded to emit from the light diffuser, passing through the center of the main lens and then passing through the center of MLA (k, l) to reach the sensor element for imaging. Based on this type of light, a center ray projection model is established to calculate the cost function of the observed value and the reprojected value.
Θ ( M I ) = C k , l k , l ( p w 0 ) 2
where M I represents the set of parameters related to the MLAs that need to be optimized, C k , l represents the center of the micro-image, and p w 0 represents the object point on the light diffuse reflector in the real world.
The Levenberg–Marquardt algorithm is used for iterative optimization to obtain the optimized values of the parameters related to the MLA.   M I includes the distance from the MLA to the main lens (D), the distance from the MLA to the sensor (d), and the rotation and translation of the MLA relative to the main lens ( θ , t ).
Initial estimation and optimization of the extrinsic parameters. For each corner point cluster, the weight of the corner point is calculated according to the blur radius, and the weighted average center of gravity is calculated. The center of gravity of the corner point cluster can be regarded as the image point of the object-point projected through the pinhole model. For each frame, the intrinsic parameters are fixed, and the initial extrinsic parameters are estimated via the P3P method. A corner point projection model is established on the basis of the intrinsic parameters and initial extrinsic parameters of the MLA parameter optimization assisted by the center ray. The intrinsic parameters are still fixed, and the cost function is constructed by using the clearest corner points to optimize the extrinsic parameters.
Θ ( E ) = p k , l n k , l ( p w n ) 2
where E represents the set of extrinsic parameters to be optimized, p k , l n represents the coordinates of the corner points detected in the n frame image, and p w n represents the coordinates of the checkerboard corner points in the real world.
Corner point removal and blur radius optimization. After the optimization is complete, the object point is reprojected into the micro-image, and the reprojection error of each corner point is calculated. If the reprojection error of the corner point is greater than a certain threshold (generally twice the root mean square value of the reprojection error in the previous step), the corner point is directly removed; otherwise, it is retained. According to Formulas (3) and (9), the blur radius r is related to v , D, and d. When these values change after central ray-assisted MLA parameter optimization and the initial estimation and optimization steps for the extrinsic parameters, the blur radius can be recalculated for the corner points that meet the threshold to solve for the intrinsic and extrinsic parameters of the camera with higher accuracy during global nonlinear optimization.
Global nonlinear optimization. Finally, we propose a new cost function for global nonlinear optimization. Both the intrinsic and extrinsic parameters are optimized in one iteration. The new cost function is composed of the corner reprojection error, blur radius reprojection error, and micro-image center reprojection error.
Θ ( T ) = ( p k , l n k , l ( p w n ) ) · w k , l n 2 + r k , l n k , l ( p w n ) 2 + Θ ( M I )
where T represents the set of all intrinsic and extrinsic parameters to be optimized, ( p k , l n k , l ( p w n ) ) · w k , l n 2 represents the corner cost function considering the blur perception characteristics, and w k , l n refers to the normalized weight corresponding to the quantized blur degree of the corner point. The weighting function is as follows:
w k , l n = α   + ( 1 α ) · e ( r k , l n m i n ( r k , l n ) )  
where r k , l n represents the blur radius in the n frame image; α is the lowest weight value, which is generally set to 0.4. This value was empirically determined based on a series of calibration experiments conducted on multiple focused light field camera models. It provides a good trade-off between sensitivity to blur variations and robustness against noise. This threshold can be adjusted according to specific hardware configurations or imaging conditions; and m i n ( r k , l n ) represents the minimum value of the blur radius in the n frame.

4. Experiments and Analysis

4.1. Experiment Setup

To validate the effectiveness of our proposed method, we evaluated it in a controlled environment using both real and simulated data acquired with a multi-focus plenoptic camera. Our experimental setup is shown in Figure 10. The camera is mounted on a linear motion stage with micron-level precision. The target plane is orthogonal to the translation axis, and the camera optical axis is aligned with this axis. Table 1 lists the approximate absolute distances for images captured with corresponding step sizes.
1.
Hardware Environment
The light field cameras used in the experiments primarily included two multifocal plenoptic cameras, namely, the Raytrix R12 and HR260. The Raytrix R12 is produced by the German company Raytrix, and the HR260 is independently developed by our research group. The detailed specifications of the three light field cameras are as follows:
(1)
The main lens used by Raytrix R12 is a Nikon AF Nikkor F/1.8D with a focal length of 50 mm. The MLA consists of 176 × 152 microlenses with three focal lengths arranged crosswise. The sensor is a Basler boA4000-KC with a pixel size of 0.0055 mm, a resolution of 4080 × 3068, and a working distance of approximately 0.1~5 m.
(2)
The main lens for the HR260 has a focal length of 105 mm, and the MLA consists of 65 × 50 microlenses with three focal lengths arranged crosswise. The sensor pixel size is 0.0037 mm, the resolution is 6240 × 4168, and the working distance is approximately 0.1~100 m.
2.
Software Environment
When capturing light field images, we set the shutter speed to 5 ms. When capturing white images, we set the gain to maximum. For the Raytrix data, we used their proprietary software, RxLive (v4.0.50.2), to calibrate the camera and calculate the depth maps used in the evaluation. For the HR260 data, we used our own imaging software to generate the light field images.
These data include publicly available datasets acquired with the Raytrix R12 camera by Labussière et al. [20,21], alongside self-collected datasets from HR260 cameras. Specifically, the sequences R12-A~R12-C and HR260-A~HR260-B correspond to these respective devices. To comprehensively validate the accuracy of the calibrated intrinsic and extrinsic parameters, the datasets cover multi-view and controlled translational sequences captured at a fixed focus. A detailed introduction to the dataset is provided in Table 1.
3.
Datasets
We establish three datasets with different focus distances H, obtained from two different light field cameras, with detailed parameter settings as shown in Table 1. Each dataset consists of the following:
(1)
White raw plenoptic images captured at different apertures ( N 5.66 , 8.0 , 11.31 , 16.0 ) using a light diffuser mounted on the main objective lens.
(2)
Target images captured at different poses (distances and directions), divided into two subsets, one for the calibration process (16 images) and the other for reprojection error evaluation (15 images).
(3)
White raw plenoptic images acquired under the same illumination conditions and the same aperture are used for calibration target acquisition for dehalation, as well as calibration targets acquired through controlled translation motion for quantitative evaluation.
Figure 11 shows some of the captured light field images, and Table 1 gives the detailed parameter settings for each dataset. Our dataset can be downloaded from public repository: https://github.com/LightFieldVision/LightField2.0-datasets-for-calibration, accessed on 6 August 2025.
4.
Simulation Environment
Based on the method proposed in reference [30], three datasets were generated using its open source code. These datasets provide ground truth corner positions, enabling direct accuracy assessment. The camera parameters are set as follows: the MLA structure is arranged in a hexagonal shape and consists of three types microlenses. It should be noted that F represents the focal length of the main lens in millimeters, F is set to 105 mm, and s x y represents the pixel size of the sensor in micrometers where s x y is set to 5.5 µm. The detailed simulation parameters are listed in Table 2, and Figure 12 shows some of the simulation light field images

4.2. Ablation Study

To evaluate the effectiveness and performance of the proposed method, we conducted a comprehensive ablation study. This study systematically isolates and assesses the contributions of three key modules in our approach: corner feature extraction, principal ray-assisted optimization of microlens array parameters, and nonlinear optimization. The goal is to evaluate the individual impact of each module on the overall calibration performance and to gain a deeper understanding of their respective roles.

4.2.1. Evaluation Metrics

We use the reprojection error as a key metric to evaluate performance.
Mean Reprojection Error: This metric is a critical and commonly used error evaluation criterion in pose estimation and 3D reconstruction. It can also be used to determine whether the pose estimation has converged. It represents the average distance, measured in pixels, between the reprojected position of all 3D points onto the images and their corresponding observed positions in the actual images across all matched points. The formula is as follows:
M S E = 1 M j = 1 M p j p j 2
Here, M denotes the number of observations for each 3D point; p j = u j , v j T represents the ground truth image point of the j-th observation; and p j = u j , v j T denotes the reprojected point of the corresponding 3D point onto the same image.

4.2.2. Experimental Setup

To evaluate the contribution of each key module, we conducted ablation experiments using the R12-A, HR260-A, and HR260-B datasets described in Section 4.1. The specific experimental configurations are as follows:
(1)
Without neighboring blur-aware corner feature extraction: This configuration evaluates the system’s performance across different datasets when the blur-aware corner feature extraction guided by neighboring views is disabled.
(2)
Without principal ray-assisted microlens array parameter optimization: This setting assesses the system’s accuracy on various datasets when the optimization of microlens array parameters using principal rays is not employed.
(3)
Without decoupled hierarchical nonlinear optimization: This test examines the system’s accuracy on different datasets when the decoupled hierarchical nonlinear optimization module is omitted.
(4)
Full system: This configuration includes all modules, namely the neighboring blur-aware corner feature extraction, principal ray-assisted microlens array parameter optimization, and decoupled hierarchical nonlinear optimization.

4.2.3. Ablation Results and Contribution Impact

Table 3 summarizes the calibration accuracy under different module configurations. The average corner reprojection error (RMSE) across three datasets—R12-A, HR260-A, and HR260-B—was used as the evaluation metric. The following observations can be drawn:
(1)
Effect of blur-aware corner extraction
When the blur-aware corner feature extraction is disabled, the reprojection error increases significantly on all datasets. For instance, the RMSE on HR260-B increases from 1.250 pixels (full system) to 1.945 pixels. This confirms that the proposed corner refinement strategy—guided by virtual depth and blur quantization—is effective in improving corner localization accuracy.
(2)
Effect of center-ray-assisted MLA optimization
Disabling the center-ray-based optimization leads to a moderate accuracy drop on R12-A (from 0.527 to 0.610 pixels) and HR260-A (from 0.572 to 1.171 pixels), but a drastic drop on HR260-B (from 1.250 to 14.421 pixels). This indicates that the optimization of microlens geometry using center rays is particularly important in scenes with long-range depth variation or strong blur, such as HR260-B.
(3)
Effect of decoupled nonlinear optimization:
Removing the staged nonlinear optimization leads to a consistent performance drop, especially in the HR260-B dataset, where RMSE increases to 17.125 pixels. This suggests that direct joint optimization without prior parameter refinement may suffer from poor convergence in complex configurations.
(4)
Overall effectiveness:
The full system consistently achieves the lowest reprojection errors across all datasets. This demonstrates that each of the proposed modules contributes to improved calibration accuracy and that their integration produces the most stable and accurate results.
These results validate the importance of each innovation in our pipeline and highlight their complementary roles in enhancing calibration robustness and precision.

4.3. Corner Feature Extraction Accuracy Evaluation

The quality of corner feature extraction from micro-images with different blur levels in multifocal focused light field images is one of the main problems addressed in this paper. To verify the effectiveness of the proposed method in terms of the feature corner accuracy, we used three sets of light field simulation image datasets to quantitatively compare and analyze the proposed method with classic methods such as those from Nousias [14] and Noury [15]. The evaluation indicators include the mean error (Mean) and standard deviation (Std).
Table 4 lists the performance of the extracted corner points and the true corner points in terms of the error mean, error standard deviation, maximum error, minimum error, and other indicators on the R1~R3 datasets. The following conclusions can be drawn from the experimental results:
(1)
As the image blur changes, the Nousias method and Noury method exhibit large fluctuations. Furthermore, the error gradually increases with increasing blur, indicating that the accuracy of these two methods is strongly affected by blur, which also verifies the problem raised in this paper. However, the proposed method performs smoothly, and the accuracies of the three types of corner points are almost the same.
(2)
Noury’s method and the proposed method have multiple identical numerical values in terms of the minimum error because the first step of the corner feature extraction strategy proposed in this paper, which is guided by neighboring blur information, uses the template matching algorithm introduced by Noury et al. [15] to detect the corner coordinates of the micro-image. The corner coordinates with the minimum error are the detection results corresponding to the clearest corners. The corner feature extraction strategy uses the coordinates of these clearest corners to construct geometric observations in the virtual depth domain to optimize the coordinates of other corners.
(3)
The overall performance of the proposed method is the best in terms of the two indicators of the error mean and standard deviation, as the experimental results for the proposed method are the smallest. Compared with the second-ranked method, the accuracy is improved by 11%, 57%, 60%, 44%, and 43%, respectively, with an average improvement of more than 30%. These improved results can be obtained due to the corner feature extraction strategy guided by neighboring blur information proposed in this paper. This approach uses the geometric relationship between the epipolar lines between micro-images and the virtual depth space to construct a multi constraint model and then, on this basis, constructs a geometric observation model of the virtual depth domain. It next quantifies and grades the degree of blur imaging for multiple corner points in the cluster, in turn iteratively filtering the noise in the cluster according to the step-by-step diffusion method and optimizes the coordinates of the corner points from clear to blur. Therefore, the proposed method refines relatively blurred corner coordinates and effectively improves the extraction quality of corner features.
To more intuitively show the error between the extracted corner point coordinates and the real corner point coordinates, we randomly selected local areas in the two groups of simulated light field images R2 for detailed display, as shown in Figure 13. The figure shows that the corner point coordinates extracted by the proposed method are basically the closest to the true corner point coordinates, whether in clear or blurred micro-images.

4.4. Accuracy Assessment of the Intrinsic Calibration Parameters

Existing methods usually use the reprojection error (RMSE) as the main indicator to measure the accuracy of the intrinsic calibration parameters. We selected five public datasets and field-collected datasets, including R12-A~R12-C and HR260-A~HR260-B, as experimental data and quantitatively compared and analyzed the proposed method with well-known methods such as those from Labussière [20], Nousias [14], and Liu [19]. The light field image data used in the experiment were all taken by a multifocal focused light field camera composed of three different types of microlenses.
The reprojection errors of each method are shown in Table 5, and the estimated intrinsic parameters of the camera can be found in Table 6, Table 7, Table 8, Table 9 and Table 10. The following conclusions can be drawn from the experimental results:
(1)
When the given camera parameters are limited, only the focal length of the main lens is considered. Although the reprojection errors of the Nousias and Liu methods are different, especially the Liu method, which has a better reprojection error performance, the estimated focal length of the main lens is significantly different from the given value, so the accuracy of the estimated intrinsic parameters is questionable.
(2)
The focal length of the main lens estimated by the proposed method and the Labussière method is closest to the given value, and the reprojection error of the proposed method is the smallest at approximately 0.5 pixels; however, the reprojection error of the Labussière method can reach 2.080 pixels. Therefore, the proposed method has the best comprehensive performance in terms of calibrating the intrinsic parameters because the proposed method constructs the quantification and classification rules related to the micro-image clarity to account for the blur-perceived full-light characteristics and uses the geometric observation mode of the virtual depth domain to effectively improve the extraction quality for corner features.
(3)
Compared with the given main lens focal length parameters, the main lens focal length values estimated by the Nousias and Liu methods are still significantly different from the given values, and the reprojection error is large, even reaching 6 pixels. The main lens focal length values estimated by the Labussière method are all close to the given values, but the reprojection error, an indicator that characterizes the accuracy of the intrinsic parameters, is too large, with a maximum of 20.049 pixels. Therefore, the robust performance of the proposed method is the best because it introduces blur-aware plenoptic features, constructs a step-by-step diffusion corner point optimization model to achieve high-quality extraction of corner point coordinates, and establishes a decoupled stacked intrinsic and extrinsic parameter optimization strategy to ensure the accuracy and robustness of the calibration.

4.5. Comparative Analysis of the Z-Axis Translation Error

To further validate the effectiveness of the calibration parameters, the Z-axis displacement error index was employed to assess the accuracy of the extrinsic parameters. This approach is similar to those used by Labussière [20], Nousias [14], and Liu [19]. A total of 6 datasets, comprising both public and field-collected data, were selected and three different cameras were used: the Raytrix R12 and HR260. The datasets were acquired on a controlled acquisition platform, where the translation distance was precisely managed.
We compared the estimated displacement of the extrinsic parameter along the Z-axis with the ground truth displacement value to calculate the translation error. The corresponding calculation formula is as follows:
ε z   ( δ z ) = η 1   ( T i , T j ) | z i z j = δ z δ z   δ z ^ / δ z  
In this context, ε z denotes the translation error, while δ z signifies the ground truth displacement. η is the normalization constantly associated with the number of frame pairs; ( T i , T j ) represents a frame pair, with indices i and j indicating the specific frame numbers; z i and z j correspond to the ground truth distance values for the two images within the frame pair; and δ z ^ represents the estimated displacement value, δ z ^ = z i ^ z j ^ . Note: Equation (17) and the associated index notation are directly adopted from the original definition of the baseline method, Labussière [20], in order to ensure consistency in implementation and comparison.
We used four methods, namely, the BFC-PC, Labussière, Nousias, and Liu, to calibrate 5 sets of data from three different types of cameras and calculated the extrinsic parameter errors between frame pairs in addition to the error mean and standard deviation. For each camera model, a representative set of extrinsic parameter error distributions was randomly selected, along with a comprehensive error bar chart illustrating the overall calibration performance.
Figure 14 and Figure 15 show the statistical results of the extrinsic parameter errors for the three cameras. Figure 14a and Figure 15a present the statistics of the extrinsic parameter errors between frame pairs under different displacement conditions contained in the dataset, while Figure 14b and Figure 15b present the means and standard deviations of the extrinsic parameter errors. The bar graph represents the mean, and the straight-line segment represents the standard deviation. The following conclusions can be made:
(1)
Comparative Performance in terms of the Extrinsic Parameter Error
The proposed method demonstrates superior performance in terms of the extrinsic parameter error across all three camera models. Specifically, on the R12-A dataset (Figure 14a), our method achieves an optimal extrinsic parameter error for 7 out of 10 datasets, with the remaining 3 datasets resulting in marginally greater errors than the Nousias method. Notably, the Nousias method exhibits significant performance variability across different microlens types. Furthermore, on the HR260-A dataset (Figure 15a), while all four methods show substantial error fluctuations, our method maintains superior stability, with error values ranging between 4.77% and 15.85%; these values are significantly lower than those of the competing methods. These results collectively demonstrate the robustness of our method across varying intrinsic parameter configurations in different light field cameras.
(2)
Consistency Analysis via the Standard Deviation
The standard deviation serves as a critical metric for evaluating calibration consistency across datasets and method generalizability to different camera configurations. As evidenced by the error bar graphs in Figure 14b and Figure 15b, our method consistently results in the smallest standard deviation, indicating minimal performance variation across datasets. This consistency not only reflects the reliability of our calibration method but also confirms the accuracy of the intrinsic parameter estimation. Consequently, our method demonstrates both accuracy and consistency when applied to diverse camera configurations.
(3)
Distance-dependent Performance Analysis
The proposed method maintains stable performance across varying shooting distances, with no observable correlation between the extrinsic parameter error and an increased shooting distance. This distance-independent performance enables reliable pose readjustment and precise estimation regardless of the shooting conditions.

4.6. Runtime Efficiency Comparison

To further evaluate the practicality of the proposed method, we measured the runtime of each stage in our calibration pipeline on four representative datasets (R12-A, R12-B, HR260-A, and HR260-B) and compared the results with those of the method proposed by Labussière et al. [20]. Since the methods by NOUS [14] and Liu [19] are implemented in MATLAB (MATLAB 2020b), and our method and Labussière’s are implemented in C++ (C++ 11), a direct runtime comparison across different programming languages would be misleading. Therefore, we excluded NOUS and Liu from the runtime comparison.
The runtime statistics are summarized in Table 11. Despite incorporating several additional modules (e.g., center-ray-assisted microlens array optimization and blur-aware corner refinement), the overall computational time of our method is comparable to that of Labussière’s method, and in some cases, superior—particularly in the global nonlinear optimization stage. This improvement stems from better parameter initialization and the cascaded optimization structure, which accelerates convergence.
Moreover, the newly introduced modules are computationally lightweight. The center ray optimization and extrinsic parameter refinement typically require less than 20 s per dataset, which demonstrates that the performance gains do not come at the cost of significantly increased runtime. These results indicate that the proposed method offers a favorable trade-off between accuracy and efficiency and is scalable for high-resolution plenoptic datasets.
In addition, the convergence trends of different optimization stages are visualized in Figure 16. The figure illustrates the reduction in the reprojection error over iterations for both the decoupled nonlinear optimization and global optimization stages. It can be observed that the proposed method exhibits stable convergence behavior, with rapid error reduction in the early iterations and smooth convergence in later stages. The decoupled structure allows intermediate parameters to be progressively refined, which helps improve the convergence rate and stability of the subsequent global optimization. These curves further confirm the effectiveness of the cascaded design in accelerating convergence while maintaining robustness.

5. Conclusions

We propose a blur feature-guided cascaded calibration method for plenoptic cameras. The blur-aware plenoptic feature is introduced to construct a step-by-step diffusion corner optimization model and a decoupled cascade intrinsic and extrinsic parameter optimization strategy to achieve high-quality extraction of corner coordinates and robust calibration. First, high-precision center point extraction and initial estimation of intrinsic parameters are performed, and then corner feature extraction guided by neighboring blur information is performed. Based on the Fermat principle and Gaussian imaging principle, a geometric observation model of the virtual depth domain unique to the plenoptic camera is constructed; then, a decoupled cascade nonlinear optimization strategy is proposed. Experimental verification and analysis were conducted by using public datasets and captured datasets. The results reveal that the proposed method excels in corner feature extraction, calibration accuracy of both intrinsic and extrinsic parameters, and calibration sensitivity when applied to multifocal focused plenoptic cameras, highlighting its advancement and robustness.
Moreover, despite the multi-stage optimization design, the proposed method maintains high computational efficiency, striking a favorable balance between accuracy and runtime cost, and proving its practical applicability to real-world engineering scenarios.
While the proposed method demonstrates high accuracy and robustness in the calibration of multifocal plenoptic cameras, several limitations still exist. Currently, the method is designed for offline calibration scenarios and has not yet been extended to real-time or online environments. Additionally, the performance of the blur-aware corner extraction may be affected under extremely low-texture or high-noise conditions, where blur estimation becomes less stable. Future work will focus on enhancing the method’s adaptability to online applications and improving robustness under challenging imaging conditions.

Author Contributions

Conceptualization, H.G.; Methodology, Z.L.; Software, Z.L.; Validation, Z.L.; Resources, H.G.; Data curation, Q.N.; Writing—original draft, Z.L.; Writing—review & editing, Z.L. and Q.N.; Visualization, Q.N.; Supervision, H.G.; Project administration, H.G.; Funding acquisition, H.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program of China, grant number 2024YFC3015600.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, Y.; Ji, X.; Dai, Q. Key Technologies of Light Field Capture for 3D Reconstruction in Microscopic Scene. Sci. China Inf. Sci. 2010, 53, 1917–1930. [Google Scholar] [CrossRef]
  2. Kim, C. 3D Reconstruction and Rendering from High Resolution Light Fields. Ph.D. Thesis, ETH Zurich, Zürich, Switzerland, 2015. [Google Scholar]
  3. Feng, W.; Qu, T.; Gao, J.; Wang, H.; Li, X.; Zhai, Z.; Zhao, D. 3D Reconstruction of Structured Light Fields Based on Point Cloud Adaptive Repair for Highly Reflective Surfaces. Appl. Opt. 2021, 60, 7086–7093. [Google Scholar] [CrossRef] [PubMed]
  4. Koch, R.; Pollefeys, M.; Heigl, B.; Van Gool, L.; Niemann, H. Calibration of Hand-Held Camera Sequences for Plenoptic Modeling. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Corfu, Greece, 20–25 September 1999; IEEE: Piscataway, NJ, USA, 1999; Volume 1, pp. 585–591. [Google Scholar]
  5. Zhang, Z. A Flexible New Technique for Camera Calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 22, 1330–1334. [Google Scholar] [CrossRef]
  6. Vaish, V.; Wilburn, B.; Joshi, N.; Levoy, M. Using Plane+ Parallax for Calibrating Dense Camera Arrays. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2004, Washington, DC, USA, 27 June–2 July 2004; IEEE: Piscataway, NJ, USA, 2004; Volume 1, p. I–I. [Google Scholar]
  7. Georgiev, T.; Lumsdaine, A.; Goma, S. Plenoptic Principal Planes. In Proceedings of the Computational Optical Sensing and Imaging, Toronto, ON, Canada, 10–14 July 2011; Optica Publishing Group: Washington, DC, USA, 2011; p. JTuD3. [Google Scholar]
  8. Ng, R.; Levoy, M.; Brédif, M.; Duval, G.; Horowitz, M.; Hanrahan, P. Light Field Photography with a Hand-Held Plenoptic Camera. Ph.D. Thesis, Stanford University, Stanford, CA, USA, 2005. [Google Scholar]
  9. Dansereau, D.G.; Pizarro, O.; Williams, S.B. Decoding, Calibration and Rectification for Lenselet-Based Plenoptic Cameras. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 1027–1034. [Google Scholar]
  10. Bok, Y.; Jeon, H.-G.; Kweon, I.S. Geometric Calibration of Micro-Lens-Based Light Field Cameras Using Line Features. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 287–300. [Google Scholar] [CrossRef] [PubMed]
  11. Wan, J.; Zhang, X.; Yang, W.; Zhang, C.; Lei, M.; Dong, Z. A Calibration Method for Defocused Cameras Based on Defocus Blur Estimation. Measurement 2024, 235, 115045. [Google Scholar] [CrossRef]
  12. Johannsen, O.; Heinze, C.; Goldluecke, B.; Perwaß, C. On the Calibration of Focused Plenoptic Cameras. In Time-of-Flight and Depth Imaging. Sensors, Algorithms, and Applications; Grzegorzek, M., Theobalt, C., Koch, R., Kolb, A., Eds.; Springer: Berlin/Heidelberg, Germany, 2013; Volume 8200, pp. 302–317. ISBN 978-3-642-44963-5. [Google Scholar]
  13. Zeller, N.; Quint, F.; Stilla, U. Calibration and Accuracy Analysis of a Focused Plenoptic Camera. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2014, 2, 205–212. [Google Scholar] [CrossRef]
  14. Nousias, S.; Chadebecq, F.; Pichat, J.; Keane, P.; Ourselin, S.; Bergeles, C. Corner-Based Geometric Calibration of Multi-Focus Plenoptic Cameras. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 957–965. [Google Scholar]
  15. Noury, C.-A.; Teulière, C.; Dhome, M. Light-Field Camera Calibration from Raw Images. In Proceedings of the 2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA), Sydney, Australia, 29 November–1 December 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–8. [Google Scholar]
  16. Heinze, C.; Spyropoulos, S.; Hussmann, S.; Perwaß, C. Automated Robust Metric Calibration Algorithm for Multifocus Plenoptic Cameras. IEEE Trans. Instrum. Meas. 2016, 65, 1197–1205. [Google Scholar] [CrossRef]
  17. Zeller, N.; Noury, C.; Quint, F.; Teulière, C.; Stilla, U.; Dhome, M. Metric Calibration of a Focused Plenoptic Camera Based on a 3D Calibration Target. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, 3, 449–456. [Google Scholar]
  18. O’brien, S.; Trumpf, J.; Ila, V.; Mahony, R. Calibrating Light-Field Cameras Using Plenoptic Disc Features. In Proceedings of the 2018 International Conference on 3D Vision (3DV), Verona, Italy, 5–8 September 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 286–294. [Google Scholar]
  19. Liu, Q.; Xie, X.; Zhang, X.; Tian, Y.; Li, J.; Wang, Y.; Xu, X. Stepwise Calibration of Plenoptic Cameras Based on Corner Features of Raw Images. Appl. Opt. 2020, 59, 4209–4219. [Google Scholar] [CrossRef] [PubMed]
  20. Labussière, M.; Teulière, C.; Bernardin, F.; Ait-Aider, O. Leveraging Blur Information for Plenoptic Camera Calibration. Int. J. Comput. Vis. 2022, 130, 1655–1677. [Google Scholar] [CrossRef]
  21. Labussière, M.; Teulière, C.; Bernardin, F.; Ait-Aider, O. Blur Aware Calibration of Multi-Focus Plenoptic Camera. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2020; pp. 2545–2554. [Google Scholar]
  22. Chen, B.; Xiong, C.; Zhang, Q. CCDN: Checkerboard Corner Detection Network for Robust Camera Calibration. arXiv 2023, arXiv:2302.05097. [Google Scholar] [CrossRef]
  23. Fleith, A.; Ahmed, D.; Cremers, D.; Zeller, N. LiFCal: Online Light Field Camera Calibration via Bundle Adjustment. In Proceedings of the DAGM German Conference on Pattern Recognition, Heidelberg, Germany, 19–22 September 2023; Springer: Berlin/Heidelberg, Germany, 2024; pp. 120–136. [Google Scholar]
  24. Liao, K.; Nie, L.; Huang, S.; Lin, C.; Zhang, J.; Zhao, Y.; Gabbouj, M.; Tao, D. Deep Learning for Camera Calibration and beyond: A Survey. arXiv 2023, arXiv:2303.10559. [Google Scholar] [CrossRef]
  25. Pan, L.; Baráth, D.; Pollefeys, M.; Schönberger, J.L. Global Structure-from-Motion Revisited. In Computer Vision—ECCV 2024; Leonardis, A., Ricci, E., Roth, S., Russakovsky, O., Sattler, T., Varol, G., Eds.; Springer Nature: Cham, Switzerland, 2025; Volume 15098, pp. 58–77. ISBN 978-3-031-73660-5. [Google Scholar]
  26. Ragonneau, T.M.; Zhang, Z. PDFO: A Cross-Platform Package for Powell’s Derivative-Free Optimization Solvers. Math. Program. Comput. 2024, 16, 535–559. [Google Scholar] [CrossRef]
  27. Thomason, C.M.; Thurow, B.S.; Fahringer, T.W. Calibration of a Microlens Array for a Plenoptic Camera. In Proceedings of the 52nd Aerospace Sciences Meeting, National Harbor, MD, USA, 13–17 January 2014; American Institute of Aeronautics and Astronautics: National Harbor, MD, USA, 2014. [Google Scholar]
  28. Suliga, P.; Wrona, T. Microlens Array Calibration Method for a Light Field Camera. In Proceedings of the 2018 19th International Carpathian Control Conference (ICCC), Szilvásvárad, Hungary, 28–30 May 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 19–22. [Google Scholar]
  29. Ester, M.; Kriegel, H.-P.; Sander, J.; Xu, X. Density-Based Spatial Clustering of Applications with Noise. In Proceedings of the International Conference on Knowledge Discovery and Data Mining, Portland, OR, USA, 2–4 August 1996; Volume 240. [Google Scholar]
  30. Michels, T.; Petersen, A.; Koch, R. Creating Realistic Ground Truth Data for the Evaluation of Calibration Methods for Plenoptic and Conventional Cameras. In Proceedings of the 2019 International Conference on 3D Vision (3DV), Québec City, QC, Canada, 16–19 September 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 434–442. [Google Scholar]
Figure 1. Overview of the methodology proposed in this paper.
Figure 1. Overview of the methodology proposed in this paper.
Sensors 25 04940 g001
Figure 2. Results of center extraction. (a) White image and (b) center of the micro-image.
Figure 2. Results of center extraction. (a) White image and (b) center of the micro-image.
Sensors 25 04940 g002
Figure 3. White image captured by a multifocal light field camera.
Figure 3. White image captured by a multifocal light field camera.
Sensors 25 04940 g003
Figure 4. Initial values for estimating the intrinsic camera parameters. F represents the focal length of the main lens, D represents the distance from the MLA to the main lens, d represents the distance from the MLA to the sensor, μ represents the spacing of the MLA, ( θ , t ) represents the rotation and translation of the MLA relative to the main lens, and f 0 , f 1 , f 2 represent three types of MLAs with different focal lengths.
Figure 4. Initial values for estimating the intrinsic camera parameters. F represents the focal length of the main lens, D represents the distance from the MLA to the main lens, d represents the distance from the MLA to the sensor, μ represents the spacing of the MLA, ( θ , t ) represents the rotation and translation of the MLA relative to the main lens, and f 0 , f 1 , f 2 represent three types of MLAs with different focal lengths.
Sensors 25 04940 g004
Figure 5. Correspondence between corner points in the object space and image space of a chessboard.
Figure 5. Correspondence between corner points in the object space and image space of a chessboard.
Sensors 25 04940 g005
Figure 6. Corner clustering results, different colors represent different clusters.
Figure 6. Corner clustering results, different colors represent different clusters.
Sensors 25 04940 g006
Figure 7. Geometric relationships between image points and virtual image points in the virtual depth domain.
Figure 7. Geometric relationships between image points and virtual image points in the virtual depth domain.
Sensors 25 04940 g007
Figure 8. Geometric observation in the virtual depth domain. The geometric observation model for corners in clusters in the virtual depth domain is depicted, where I′1 is the corner of I1, corrected by the geometric observations of clear corners such as I0 and I3 in the virtual depth domain, and I′2 is the corner of I2, corrected by the geometric observations for corners such as I0, I3, and I′1 in the virtual depth domain.
Figure 8. Geometric observation in the virtual depth domain. The geometric observation model for corners in clusters in the virtual depth domain is depicted, where I′1 is the corner of I1, corrected by the geometric observations of clear corners such as I0 and I3 in the virtual depth domain, and I′2 is the corner of I2, corrected by the geometric observations for corners such as I0, I3, and I′1 in the virtual depth domain.
Sensors 25 04940 g008
Figure 9. Light beams reconstructed by a light field camera. (a) represents the light beam reconstructed from the center of the micro-image and (b) represents the light beam reconstructed from the corner features.
Figure 9. Light beams reconstructed by a light field camera. (a) represents the light beam reconstructed from the center of the micro-image and (b) represents the light beam reconstructed from the corner features.
Sensors 25 04940 g009
Figure 10. Original checkerboard image acquired with fixed step length.
Figure 10. Original checkerboard image acquired with fixed step length.
Sensors 25 04940 g010
Figure 11. Checkerboard images captured by the Raytrix R12.
Figure 11. Checkerboard images captured by the Raytrix R12.
Sensors 25 04940 g011
Figure 12. The simulation light field checkerboard images.
Figure 12. The simulation light field checkerboard images.
Sensors 25 04940 g012
Figure 13. Corner point coordinates extracted via different methods and the true corner point coordinates. The true values of the corner point coordinates and the corner point coordinates extracted via the Nousias method, the Noury method, and the proposed method are marked with different symbols. In addition, there are four types of corner point coordinates in R2.
Figure 13. Corner point coordinates extracted via different methods and the true corner point coordinates. The true values of the corner point coordinates and the corner point coordinates extracted via the Nousias method, the Noury method, and the proposed method are marked with different symbols. In addition, there are four types of corner point coordinates in R2.
Sensors 25 04940 g013
Figure 14. Statistics of the extrinsic parameter errors produced by the Raytrix R12 camera. (a) presents the extrinsic parameter errors under 10 different displacements in R12-A and (b) depicts the means and standard deviations of the extrinsic parameter errors for R12-A, R12-B, and R12-C.
Figure 14. Statistics of the extrinsic parameter errors produced by the Raytrix R12 camera. (a) presents the extrinsic parameter errors under 10 different displacements in R12-A and (b) depicts the means and standard deviations of the extrinsic parameter errors for R12-A, R12-B, and R12-C.
Sensors 25 04940 g014
Figure 15. Statistics of the extrinsic parameter errors for the HR260 camera. (a) presents the extrinsic parameter errors under 8 different displacements in HR260-A and (b) depicts the means and standard deviations of the extrinsic parameter errors for HR260-A and HR260-B.
Figure 15. Statistics of the extrinsic parameter errors for the HR260 camera. (a) presents the extrinsic parameter errors under 8 different displacements in HR260-A and (b) depicts the means and standard deviations of the extrinsic parameter errors for HR260-A and HR260-B.
Sensors 25 04940 g015
Figure 16. Iterative Convergence Graph. (a) represents the number of iterations and convergence trend of R12-A; (b) represents the number of iterations and convergence trend of R12-B; (c) represents the number of iterations and convergence trend of HR260-A; (d) represents the number of iterations and convergence trend of HR260-B.
Figure 16. Iterative Convergence Graph. (a) represents the number of iterations and convergence trend of R12-A; (b) represents the number of iterations and convergence trend of R12-B; (c) represents the number of iterations and convergence trend of HR260-A; (d) represents the number of iterations and convergence trend of HR260-B.
Sensors 25 04940 g016
Table 1. Raw image datasets collected by the three light field cameras. H refers to the focus distance, where all distances are in millimeters; Scale refers to the square size of the checkerboard; and Depth Distance refers to the setting of the linear motion workbench.
Table 1. Raw image datasets collected by the three light field cameras. H refers to the focus distance, where all distances are in millimeters; Scale refers to the square size of the checkerboard; and Depth Distance refers to the setting of the linear motion workbench.
DatasetH (mm)CheckerboardDepth Distance
Size
(Rows × Cols)
Scale (mm)Min (mm)Max (mm)Step (mm)
R12-A4509 × 51026538510
R12-B10008 × 52045090050
R12-C6 × 430400125050
HR260-A15007 × 53512002200100
HR260-B40006 × 44010002600200
Table 2. Simulated light field image dataset. Resolution represents the resolution of the image; D m i represents the radius of the micro-image in pixels; D represents the distance from the main lens to the MLA in millimeters; d represents the distance from the MLA to the sensor in millimeters; and f i represents the focal length of the different types of MLAs contained in the MLA, where i = 0, 1, 2, in millimeters.
Table 2. Simulated light field image dataset. Resolution represents the resolution of the image; D m i represents the radius of the micro-image in pixels; D represents the distance from the main lens to the MLA in millimeters; d represents the distance from the MLA to the sensor in millimeters; and f i represents the focal length of the different types of MLAs contained in the MLA, where i = 0, 1, 2, in millimeters.
DatasetResolution D m i
(Pixels)
D
(mm)
D
(mm)
f i
(mm)
R12000 × 300023.4101.431.321.35/1.62/1.95
R22000 × 30006099.621.321.34/1.75/2.25
R36000 × 800013099.411.301.35/1.75/2.25
Table 3. Ablation study results.
Table 3. Ablation study results.
ConfigurationRMSE (Pixel)
R12-AHR260-AHR260-B
Without neighboring blur-aware corner extraction0.9361.5411.945
Without center-ray-assisted microlens array optimization0.6101.17114.421
Without decoupled nonlinear optimization0.6861.23517.125
Full system (all modules included)0.5270.5721.250
Table 4. Statistics of the pixel errors between the extracted corner points and the true values. The mean error (mean), standard deviation error (std), maximum error (max), and minimum error (min) are shown in turn. Type1 represents the corner points in the clearest image; Type2 represents the corner points in the second-clearest image; Type3 represents the corner points in the most blurred image; and All represents the corner points in the three types of images.
Table 4. Statistics of the pixel errors between the extracted corner points and the true values. The mean error (mean), standard deviation error (std), maximum error (max), and minimum error (min) are shown in turn. Type1 represents the corner points in the clearest image; Type2 represents the corner points in the second-clearest image; Type3 represents the corner points in the most blurred image; and All represents the corner points in the three types of images.
DatasetMethodType1Type2Type3All
R1Nousias(0.83, 0.42)
(2.28, 0.04)
(0.79, 0.41)
(2.22, 0.02)
(0.95, 0.45)
(2.56, 0.10)
(0.84,0.43)
(2.56,0.02)
Noury(0.81, 0.37)
(2.78, 0.03)
(0.82, 0.37)
(2.06, 0.01)
(0.90, 0.47)
(2.87, 0.03)
(0.84, 0.40)
(2.87, 0.01)
BFC-PC(0.81, 0.37)
(2.78, 0.03)
(0.71, 0.24)
(1.04, 0.07)
(0.71, 0.23)
(1.24, 0.03)
(0.74, 0.29)
(2.78, 0.03)
R2Nousias(1.37, 0.69)
(3.22, 0.13)
(1.55, 0.76)
(3.63, 0.16)
(2.32, 0.88)
(3.89, 0.41)
(1.53, 0.78)
(3.89, 0.13)
Noury(0.51, 0.33)
(2.18, 0.03)
(0.73, 0.39)
(2.37, 0.03)
(2.25, 0.80)
(2.99, 0.11)
(0.73, 0.59)
(2.99, 0.02)
BFC-PC(0.51, 0.33)
(1.39, 0.03)
(0.43, 0.23)
(1.12, 0.03)
(0.43, 0.23)
(1.35, 0.01)
(0.46, 0.27)
(2.18, 0.01)
R3Nousias(1.82, 0.92)
(3.93, 0.08)
(2.91, 1.38)
(6.26, 0.88)
(3.17, 1.54)
(6.01, 0.23)
(2.43, 1.38)
(6.26, 0.08)
Noury(1.16, 0.63)
(3.15, 0.04)
(2.17, 1.43)
(4.83, 0.28)
(2.72, 1.64)
(4.95, 0.06)
(1.71, 1.31)
(4.95, 0.04)
BFC-PC(1.16, 0.63)
(3.01, 0.04)
(1.11, 0.61)
(3.48, 0.06)
(1.16, 0.62)
(2.77, 0.10)
(1.14, 0.63)
(3.48, 0.04)
Table 5. Statistics of the corner reprojection error (RMSE) in the pixels. In addition to the reprojection error for the corner points, the results for our method and the Labussière method also include the reprojection error of the blur radius, which is the value in parentheses. Underlined values indicate larger reprojection errors.
Table 5. Statistics of the corner reprojection error (RMSE) in the pixels. In addition to the reprojection error for the corner points, the results for our method and the Labussière method also include the reprojection error of the blur radius, which is the value in parentheses. Underlined values indicate larger reprojection errors.
DatasetsBFC-PCLabussièreNOUS1NOUS2NOUS3Liu
R12-A0.527 (0.080)0.856 (0.083)0.7730.6670.9580.691
R12-B0.421 (0.086)0.674 (0.183)0.5380.5190.5930.330
R12-C0.336 (0.040)0.738 (0.041)1.2870.6810.4110.338
HR260-A0.572 (0.257)1.827 (0.293)1.6241.4321.3160.681
HR260-B1.250 (0.371)20.049 (12.866)3.29619.1753.5145.849
Table 6. Initial internal parameters for the R12-A dataset and the optimized parameters obtained via the proposed method, Labussière method, Nousias method (for each microlens type: NOUS1, NOUS2, and NOUS3), and Liu method. The dashes indicate solutions that did not consider the parameters.
Table 6. Initial internal parameters for the R12-A dataset and the optimized parameters obtained via the proposed method, Labussière method, Nousias method (for each microlens type: NOUS1, NOUS2, and NOUS3), and Liu method. The dashes indicate solutions that did not consider the parameters.
R12-A (F = 50 mm, H = 450 mm)
OursLabussièreNOUS1NOUS2NOUS3Liu
F  [ m m ]50.47149.71461.30562.47663.32858.233
Q 1  [ × 10 5 ]16.8224.66
Q 2   [ × 10 6 ]2.0442.998
Q 3  [ × 10 8 ]0.7171.063
P 1  [ × 10 5 ]−6.8−14.6
P 2  [ × 10 5 ]0.5566.340
D   [ m m ]57.52056.70171.13172.54173.53066.985
t x   [ m m ]10.9010.97
t y   [ m m ]8.4197.887
θ x   [ μ r a d ]463.1843. 1
θ y   [ μ r a d ]389.2637.1
θ z   [ μ r a d ]29.131.5
u   [ μ m ]127.46127.46
f 1   [ μ m ]766.26578.18
f 2   [ μ m ]689.95505.42
f 3   [ μ m ]855.03552.08
u 0   [ p i x ]2083.92070.91984.92034.51973.72085.8
v 0   [ p i x ]1513.81610.91482.11481.01495.21590.3
d   [ μ m ]3,36.48324.77585.16527.59561.93455.30
Table 7. Initial internal parameters for the R12-B dataset and the optimized parameters obtained via the proposed method, Labussière method, Nousias method (for each microlens type: NOUS1, NOUS2, and NOUS3), and Liu method. The dashes indicate solutions that did not consider the parameters.
Table 7. Initial internal parameters for the R12-B dataset and the optimized parameters obtained via the proposed method, Labussière method, Nousias method (for each microlens type: NOUS1, NOUS2, and NOUS3), and Liu method. The dashes indicate solutions that did not consider the parameters.
R12-B (F = 50 mm, H = 450 mm)
OursLabussièreNOUS1NOUS2NOUS3Liu
F  [ m m ]50.17050.04753.91352.98852.97748.949
Q 1  [ × 10 5 ]−0.4192.900
Q 2   [ × 10 6 ]0.0860.300
Q 3  [ × 10 8 ]0.0360.064
P 1  [ × 10 5 ]8.4614.13
P 2  [ × 10 5 ]17.65521.540
D   [ m m ]52.21752.12556.06255.12855.12450.830
t x   [ m m ]12.0312.44
t y   [ m m ]6.4475.988
θ x   [ μ r a d ]436.9607.2
θ y   [ μ r a d ]477.5514.5
θ z   [ μ r a d ]38.546.0
u   [ μ m ]127.46127.45
f 1   [ μ m ]708.39580.49
f 2   [ μ m ]646.18504.31
f 3   [ μ m ]795.08546.36
u 0   [ p i x ]1877.21958.32074.72094.71837.01946.9
v 0   [ p i x ]1874.41802.91640.21649.11620.41668.7
d   [ μ m ]337.83336.38447.81401.93414.32324.70
Table 8. Initial internal parameters for the R12-C dataset and the optimized parameters obtained via the proposed method, Labussière method, Nousias method (for each microlens type: NOUS1, NOUS2, and NOUS3), and Liu method. The dashes indicate solutions that did not consider the parameters.
Table 8. Initial internal parameters for the R12-C dataset and the optimized parameters obtained via the proposed method, Labussière method, Nousias method (for each microlens type: NOUS1, NOUS2, and NOUS3), and Liu method. The dashes indicate solutions that did not consider the parameters.
R12-C (F = 50 mm, H = ∞)
OursLabussièreNOUS1NOUS2NOUS3Liu
F  [ m m ]50.19750.01351.11349.91950.81244.523
Q 1  [ × 10 5 ]17.2118.61
Q 2   [ × 10 6 ]3.0162.646
Q 3  [ × 10 8 ]1.4061.038
P 1  [ × 10 5 ]9.1619.11
P 2  [ × 10 5 ]7.2457.311
D   [ m m ]49.42449.36250.33149.06749.88243.819
t x   [ m m ]12.7613.12
t y   [ m m ]8.1537.446
θ x   [ μ r a d ]417.8490.9
θ y   [ μ r a d ]389.7388.9
θ z   [ μ r a d ]35.141.1
u   [ μ m ]127.46127.48
f 1   [ μ m ]650.32569.88
f 2   [ μ m ]598.01491.71
f 3   [ μ m ]712.68535.28
u 0   [ p i x ]1743.71692.11966.31913.82052.51845.3
v 0   [ p i x ]1562.31677.81484.61487.21492.71514.4
d   [ μ m ]328.63319.53357.80349.99353.26274.00
Table 9. Initial internal parameters for the HR260-A dataset and the optimized parameters obtained via the proposed method, Labussière method, Nousias method (for each microlens type: NOUS1, NOUS2, and NOUS3), and Liu method. The dashes indicate solutions that did not consider the parameters.
Table 9. Initial internal parameters for the HR260-A dataset and the optimized parameters obtained via the proposed method, Labussière method, Nousias method (for each microlens type: NOUS1, NOUS2, and NOUS3), and Liu method. The dashes indicate solutions that did not consider the parameters.
HR260-A (F = 105 mm, H = 1500 mm)
OursLabussièreNOUS1NOUS2NOUS3Liu
F  [ m m ]105.658112.12987.11552.48652.22757.274
Q 1  [ × 10 5 ]−45.0611.82
Q 2   [ × 10 6 ]−14.7942.066
Q 3  [ × 10 8 ]−9.5761.143
P 1  [ × 10 5 ]12.08−2.21
P 2  [ × 10 5 ]−24.3206.325
D   [ m m ]109.929109.08689.09750.41649.49754.867
t x   [ m m ]11.3610.92
t y   [ m m ]7.6907.206
θ x   [ μ r a d ]352.7834.6
θ y   [ μ r a d ]4752.33274.3
θ z   [ μ r a d ]120.8237.5
u   [ μ m ]351.22349.55
f 1   [ μ m ]2812.022365.11
f 2   [ μ m ]2778.642333.44
f 3   [ μ m ]2729.262297.09
u 0   [ p i x ]3066.23186.13119.23117.73098.72171.7
v 0   [ p i x ]2015.72147.12084.32083.72056.31608.3
d   [ μ m ]1584.161486.54392.60522.90583.80668.50
Table 10. Initial internal parameters for the HR260-B dataset and the optimized parameters obtained via the proposed method, Labussière method, Nousias method (for each microlens type: NOUS1, NOUS2, and NOUS3), and Liu method. The oblique line indicates that the calibration failed, and the dashes indicate solutions that did not consider the parameters.
Table 10. Initial internal parameters for the HR260-B dataset and the optimized parameters obtained via the proposed method, Labussière method, Nousias method (for each microlens type: NOUS1, NOUS2, and NOUS3), and Liu method. The oblique line indicates that the calibration failed, and the dashes indicate solutions that did not consider the parameters.
HR260-B (F = 105 mm, H = 2000 mm)
OursLabussièreNOUS1NOUS2NOUS3Liu
F  [ m m ]110.766111.88858.72842.71242.23357.919
Q 1  [ × 10 5 ]−7.784.38
Q 2   [ × 10 6 ]−1.2240.065
Q 3  [ × 10 8 ]−0.7300.002
P 1  [ × 10 5 ]−9.03−11.29
P 2  [ × 10 5 ]2.250−8.907
D   [ m m ]107.737106.40356.6920.66758.12454.857
t x   [ m m ]11.13−7.25
t y   [ m m ]8.86316.043
θ x   [ μ r a d ]689.2824.7
θ y   [ μ r a d ]3087.13425.9
θ z   [ μ r a d ]276.7381.2
u   [ μ m ]350.27349.40
f 1   [ μ m ]2155.542588.68
f 2   [ μ m ]2190.222546.87
f 3   [ μ m ]2225.472506.37
u 0   [ p i x ]3134.58168.13121.93119.53122.11909.2
v 0   [ p i x ]1696.6−280.72086.72088.62085.21687.9
d   [ μ m ]1354.741639.79526.201.102203.80686.00
Table 11. Runtime comparison (in seconds) between the proposed method and Labussière et al. [20] on four datasets.
Table 11. Runtime comparison (in seconds) between the proposed method and Labussière et al. [20] on four datasets.
DatasetMethodCorner ExtractionCenter Ray Optimization
(Iterations, Time)
Extrinsic Optimization
(Iterations, Time)
Global Nonlinear Optimization
(Iterations, Time)
R12-ALabussière1027//(160, 451.26)
Ours1041(6, 8.88)(98, 20.69)(152, 378.89)
R12-BLabussière982//(100, 284.58)
Ours1045(6, 8.82)(74, 8.78)(160, 197.75)
HR260-ALabussière4824//(77, 101.68)
Ours5069(5, 0.92)(250, 19.86)(93, 64.45)
HR260-BLabussière4809//(42, 52.70)
Ours4970(5, 0.91)(180, 14.45)(57, 65.48)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, Z.; Guan, H.; Ni, Q. A Blur Feature-Guided Cascaded Calibration Method for Plenoptic Cameras. Sensors 2025, 25, 4940. https://doi.org/10.3390/s25164940

AMA Style

Liu Z, Guan H, Ni Q. A Blur Feature-Guided Cascaded Calibration Method for Plenoptic Cameras. Sensors. 2025; 25(16):4940. https://doi.org/10.3390/s25164940

Chicago/Turabian Style

Liu, Zhendong, Hongliang Guan, and Qingyang Ni. 2025. "A Blur Feature-Guided Cascaded Calibration Method for Plenoptic Cameras" Sensors 25, no. 16: 4940. https://doi.org/10.3390/s25164940

APA Style

Liu, Z., Guan, H., & Ni, Q. (2025). A Blur Feature-Guided Cascaded Calibration Method for Plenoptic Cameras. Sensors, 25(16), 4940. https://doi.org/10.3390/s25164940

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop