1. Introduction
Stochastic frontier analysis (SFA) [
1] is a widely used method for efficiency measurement that accounts for the difference between actual output and an efficient “frontier” by modeling technical inefficiency and random noise [
2,
3]. Traditionally, SFA employs parametric models [
4,
5] estimated via maximum likelihood or Bayesian methods [
6]. Recent advances leverage machine learning to capture non-linear relationships, e.g., Kumbhakar [
7] uses neural networks for flexible frontier estimation, Fan et al. [
8] applies support vector regression to model production functions, and Tsionas [
9] combines Bayesian deep learning with SFA for dynamic efficiency analysis.
However, modern SFA applications increasingly involve highly sensitive datasets—patient medical records, proprietary financial data, and confidential agricultural information—that pose severe privacy risks when conducting statistical inference or variable selection in high-dimensional settings [
10]. This creates a fundamental challenge, i.e., how to perform accurate efficiency analysis while protecting individual privacy through rigorous mathematical processes.
Differential privacy (DP) [
11] is the standard framework for rigorous privacy protection. While DP has been successfully applied to linear and logistic regression, as well as deep learning [
12,
13], its application to SFA is fundamentally challenging. The main difficulty lies in the non-convexity of the SFA log-likelihood and the complex sensitivity of its efficiency parameters, which violate the convexity assumptions required by most existing DP optimization methods.
For example, DP-stochastic gradient descent (DP-SGD) [
13]—a widely adopted approach—adds Gaussian noise to clipped gradients at each iteration. While effective in convex or moderately non-convex problems, DP-SGD becomes unstable for SFA due to cumulative noise and gradient clipping interactions that undermine convergence. Other methods, such as output perturbation [
14] or Frank–Wolfe-based mechanisms [
15], introduce less noise but require substantial adaptation for non-convex objectives, which have not been developed for SFA.
Recent studies exploring DP alternatives, including one-step mechanisms [
16,
17] for convex empirical risk minimization and differentially private variance reduction methods [
18], remain fundamentally limited by convexity assumptions that make them inapplicable to SFA. Non-gradient DP methods such as Bayesian DP [
19] can theoretically handle non-linear efficiency parameters but are computationally prohibitive for large-scale applications. DP-LASSO methods for convex problems [
20,
21,
22] introduce excessive noise when adapted to non-convex objectives, completely compromising accuracy and variable selection performance.
The Frank–Wolfe (FW) algorithm [
23], originally developed for constrained convex optimization, offers theoretical promise due to its projection-free nature and ability to provide sparse solutions [
24]. DP-FW methods enhance this framework by selectively adding calibrated noise to linear objectives [
25,
26,
27]. Acharya et al. [
28] introduce personalized DP for ridge regression, but their focus on
-regularization does not address sparsity or non-convexity. However, these methods cannot be directly applied to SFA due to the non-convexity of its objective function; no existing work has analyzed the sensitivity properties of SFA’s efficiency parameters or developed appropriate noise calibration for the complex parameter interactions inherent in frontier models. Non-convex DP optimization methods, such as DP-projected gradient descent and DP-FW [
29,
30], provide frameworks for complex losses, but their application to SFA’s unique efficiency decomposition and sparse variable selection remains completely unaddressed.
The fundamental gap is that no existing method adequately addresses the intersection of non-convexity, efficiency parameter estimation, sparsity, and privacy, which defines the SFA privacy problem. Existing DP-SGD-based approaches fail due to instability and noise accumulation, while DP-Bayesian and convex DP-LASSO methods are either computationally intensive or fundamentally ill-suited for sparse variable selection under privacy constraints in non-convex problems. To address these critical limitations, we develop the first differentially private optimization framework specifically designed for stochastic frontier models. Our approach is a novel paradigm that combines a constrained Frank–Wolfe optimization framework with LASSO regularization to induce sparsity. This allows us to operate within a carefully constructed bounded convex parameter space that, for the first time, enables us to bound gradient sensitivity for SFA’s non-convex parameters. By adding calibrated Laplace noise at the linear oracle level, we fundamentally mitigate the noise accumulation that plagues gradient-based methods like DP-SGD. This provides rigorous -differential privacy guarantees while preserving accuracy and interpretability in high dimensions—capabilities that existing methods cannot achieve.
The key innovations of this paper are as follows:
- (1)
For non-convex stochastic frontier models, we establish a differentially private estimation framework by deriving bounded gradient sensitivity.
- (2)
Based on this foundation, we design a constrained Frank–Wolfe optimization algorithm that enables efficient privacy-preserving estimation with calibrated noise.
- (3)
For high-dimensional settings, we integrate -regularization into the framework to achieve sparse and interpretable variable selection under differential privacy.
The rest of the paper is organized as follows.
Section 2 reviews the stochastic frontier model and
-regularized variable selection, as well as introducing our bounded parameter space, gradient sensitivity bounds, noise calibration strategy, and a Frank–Wolfe algorithm with an
-privacy guarantee.
Section 3 presents numerical results, including simulations and real-world validation on the California Housing and FADN datasets.
Section 4 concludes the paper and outlines future research directions.
2. Methodology
This section develops our differentially private framework for SFA. It begins with a concise overview of the necessary background concepts before presenting our core methodological contributions—the design of a novel constrained parameter space, a formal sensitivity analysis, and a custom Frank–Wolfe algorithm with rigorous privacy guarantees.
2.1. Stochastic Frontier Model and Maximum Likelihood Estimation
The stochastic frontier model (SFM) assumes that the production process is influenced by both random noise and technical inefficiency. For
N production units, the model can be expressed as follows:
where
,
, and
are the production frontier parameters.
is the symmetric random error, capturing measurement errors and other random factors, where is the N-dimensional identity matrix. is the one-sided error term, representing technical inefficiency, following a zero-truncated half-normal distribution.
We assume and u to be uncorrelated and independent of covariates X. The model’s parameter set is . For estimation convenience, the total variance is defined as , while the proportion of inefficiency’s relative contribution is defined as .
Common approaches to estimating the stochastic frontier model include corrected least squares, the generalized method of moments, maximum likelihood estimation, and Bayesian methods. Among these, MLE stands out for its simplicity and efficiency. Below, we describe the MLE for the SFM, which forms the foundation for the variable selection techniques that are subsequently addressed.
The probability density function of the composite error term
is obtained through convolution techniques. Given the distributional properties of
and
, the density of
incorporates the standard normal cumulative distribution function
. The log-likelihood function for the model is expressed as follows:
where
represents the residual for the
i-th observation, and
denotes the cumulative distribution function of the standard normal distribution. MLE seeks to estimate the
parameters by maximizing the log-likelihood function
.
2.2. Variable Selection via LASSO and Adaptive LASSO
In practical applications, when the number of covariates p is substantial compared to the sample size N (i.e., high-dimensional data scenarios), traditional MLE tends to cause overfitting and reduce model interpretability. Therefore, we introduce L1 regularization to identify the most explanatory features for output through variable selection, enhancing model robustness and practicality.
The penalized log-likelihood is as follows:
where
, and
r controls the sparsity level. The estimator
is obtained via gradient-based optimization.
Using initial estimates
, adaptive weights are calculated as follows:
(with the convention
).
The Adaptive Lasso Refinement is as follows:
This two-step approach enhances variable selection consistency while mitigating the overpenalization of significant coefficients.
2.3. Differential Privacy
We start by introducing the concept of differential privacy.
Definition 1 (Differential Privacy [
11])
. A randomized mechanism is said to be (ϵ, δ)-differentially private for positive parameters ϵ, δ. If, for any pair of adjacent datasets differs by a single record, and for every measurable subset under , the following holds:where the probability arises solely from the randomness of the mechanism M. In this definition, the datasets X and are considered fixed, with the probability capturing the stochastic nature of . The privacy level against potential adversaries is determined by the parameters and . Smaller values of these parameters indicate stronger privacy protections.
A fundamental principle in differential privacy is sensitivity, which measures the maximum impact on an algorithm’s output resulting from the modification of a single record in the dataset.
Definition 2 (Sensitivity)
. For a deterministic, vector-valued function , the sensitivity of is given as follows:where X and differ by only one entry. We proceed to describe the Laplace mechanism, which is a well-known technique that guarantees (, )-differential privacy for functions exhibiting constrained sensitivity.
Lemma 1 (Laplace Mechanism [
31])
. For a function with -sensitivity , the mechanism , where , ensures ϵ-differential privacy, provided . Lemma 2 (Advanced Composition [
32])
. For a sequence of T mechanisms, each satisfying -differential privacy, their composition satisfies (ϵ, δ)-differential privacy, where the following is true: 2.4. Constrained Parameter Space and Sensitivity Analysis
Our first methodological innovation is the formulation of a convex constrained parameter space for the SFM. This design plays a dual role—(i) it guarantees numerical stability during optimization by preventing extreme parameter realizations, and (ii) it ensures bounded gradient sensitivity, which is indispensable for the enforcement of differential privacy in a non-convex setting.
We define the optimization objective via the negative log-likelihood, as follows:
where
. By introducing parameter constraints, the feasible set is restricted to a convex region, thereby avoiding unstable or ill-posed solutions.
Definition 3 (Constrained Parameter Space)
. The feasible parameter set is defined as a convex polytope , as follows:where are fixed constants. These bounds simultaneously guarantee numerical stability and finite sensitivity. Unlike convex optimization problems, where the Lipschitz continuity of the gradient is sufficient for privacy guarantees, the non-convexity induced by the term in SFM complicates such analysis. Our approach circumvents this difficulty by focusing directly on bounding the sensitivity of the gradient.
Gradient Sensitivity Analysis
The sensitivity of the gradient is highly dependent on the bounds of
and
. From the derivations provided in
Appendix A, the partial derivative with respect to
exhibits the following scaling:
where
and
. This behavior implies that the gradient grows nearly linearly when
is small, but increases superlinearly as
becomes larger, primarily due to the Mills ratio.
Similarly, the partial derivative with respect to
satisfies the following:
indicating that excessively small values of
cause gradient explosion. This effect not only destabilizes optimization but also amplifies the noise required by differentially private mechanisms.
Taken together, these analyses reveal a sharp trade-off, whereby smaller and larger improve model flexibility but substantially increase gradient sensitivity. Our convex constrained parameter design directly addresses this issue by imposing explicit bounds on both parameters. This constitutes the first methodological contribution of the paper, as it establishes a feasible optimization domain that ensures bounded sensitivity—a prerequisite for developing a differentially private stochastic frontier estimator.
2.5. Private SFM
This subsection presents our primary algorithmic contribution—a differentially private estimation framework for SFM, based on the Frank–Wolfe procedure within the previously defined constrained parameter space. The method leverages convex–analytic techniques to balance statistical efficiency with privacy guarantees.
The Frank–Wolfe algorithm [
33] is a first-order method for constrained convex optimization. At each iteration, the algorithm linearizes the objective function around the current iterate and then updates the solution by moving toward the minimizer of the linearized problem over the convex domain. Formally, given the empirical loss function
defined in (
9), the optimization problem can be written as follows:
Due to the presence of the inefficiency term
, the objective
is generally non-convex, as the cumulative distribution term
introduces nonlinearities [
2]. Nonetheless, the Frank–Wolfe procedure (Algorithm 1) remains a suitable vehicle for our private adaptation.
Algorithm 1 Frank–Wolfe Procedure |
Input:
Feasible domain , objective function , learning rate schedule |
1:
Initialize |
2:
for to do |
3:
Solve linear subproblem: |
4:
Update iterate: |
5:
end for |
6:
return Final output |
The convergence rate of Frank–Wolfe depends on the curvature of . Intuitively, this curvature quantifies how much deviates from its linear approximation over the domain .
Definition 4 (Curvature Constant)
. The curvature constant of over is defined as follows: To safeguard individual-level data during SFM estimation, we embed differential privacy into the Frank–Wolfe framework by adapting the differentially private Frank–Wolfe (DP-FW) algorithm proposed by Bassily et al. [
34]. In each iteration
t, the gradient direction is selected using the exponential mechanism, with Laplace noise being calibrated to the
sensitivity of the gradient. This ensures that private information about any single observation is obfuscated.
Although is non-convex, DP-FW remains an effective heuristic, since privacy guarantees do not rely on convexity. Furthermore, the restricted parameter space , being convex and polyhedral, enables efficient vertex-based updates.
The privacy budget depends critically on bounding the sensitivity of the gradient , where denotes the dataset.
Theorem 1 (Gradient Sensitivity)
. The sensitivity of is as follows:where . Proof. Let
D and
be adjacent datasets differing in one record. Since the following is true:
the difference between gradients under
D and
reduces to the following:
Thus,
With bounded covariates (
), bounded outputs (
), and constraints on
, a finite constant
G exists such that
. Hence,
. □
Incorporating the sensitivity bound, we adapt the DP-FW algorithm to the SFM setting (Algorithm 2). At each iteration, noisy gradient scores are computed for all candidate vertices using the exponential mechanism, and the update is taken toward the vertex with the lowest perturbed score.
Algorithm 2 Differentially Private Frank–Wolfe for SFM. |
Input:
Data , loss , privacy budget , feasible domain , steps T |
Output:
Private estimate |
1:
Initialize |
2:
for to do |
3:
for all do |
4:
Compute noisy score: |
5:
Select vertex: |
6:
Update: , with |
7:
return
|
Here, , and the noise calibration ensures -differential privacy. This integration of Frank–Wolfe with differential privacy constitutes the second methodological contribution of this work.
2.6. Privacy Guarantee
This subsection establishes the formal privacy properties of our proposed algorithm, proving that it satisfies -differential privacy. This constitutes a central theoretical contribution of the paper.
Theorem 2 (Laplace Noise Scale)
. In Algorithm 2, the required Laplace noise scale λ is given as follows:where . Proof. Consider the score function
. Its sensitivity is as follows:
According to the Laplace mechanism (Lemma 1), ensuring
-differential privacy per iteration requires the following:
To guarantee an overall
-differential privacy budget over
T iterations, we invoke the advanced composition theorem (Lemma 2), setting the following:
Substituting yields the claimed result. □
We now establish the main theoretical guarantee for Algorithm 2.
Theorem 3. Algorithm 2 satisfies -differential privacy.
Proof. The proof builds on the analysis of the private Frank–Wolfe algorithm proposed by Talwar et al. [
15]. At each iteration
t, the algorithm selects a direction
using the exponential mechanism with the following score function:
where
.
We first verify per-iteration privacy. For adjacent datasets
D and
, the sensitivity of the score function is as follows:
According to the Cauchy–Schwarz inequality in the
-norm, the following is true:
From Theorem 1, the following holds true:
and since
, we have
. Thus, the overall sensitivity is bounded as follows:
To ensure
-differential privacy per iteration, the Laplace noise must satisfy the following:
Setting
gives the following:
matching the chosen
in Algorithm 2.
Finally, applying the advanced composition theorem, the cumulative privacy loss after
T iterations is as follows:
Substituting
, we obtain the following:
For small
,
, implying the following:
Hence,
when
is sufficiently small and
is moderate. Therefore, Algorithm 2 satisfies
-differential privacy. Importantly, this guarantee holds irrespective of the potential non-convexity of
, as it relies solely on bounded sensitivity and the Laplace mechanism. □
3. Numerical Experiments
This section evaluates the proposed DP-FW framework through sensitivity analysis, simulation studies, and real-world data applications, contributing original empirical evidence to validate its effectiveness and robustness in SFA.
3.1. Gradient Sensitivity Analysis
To validate our theoretical gradient bounds and demonstrate the practical impact of parameter constraints, we conduct a comprehensive sensitivity analysis examining how and affect gradient magnitude and numerical stability.
Specifically, we fix
and
, while varying
and
. We first examine the influence curves of
and
on gradient components (
Figure 1), and then provide a spatial heatmap visualization of sensitivity and stability across the parameter space (
Figure 2; see Appendix Stability Metrics for stability score definitions).
Figure 1 illustrates that the gradient component associated with
grows exponentially when
, confirming the Mills ratio effect. For
, explosive sensitivity occurs near zero, with the gradient magnitude increasing by up to
when
. Overall,
dominates the gradient dynamics across most practical regimes.
The heatmaps in
Figure 2 further highlight the interplay between
and
, visually confirming regions of instability. These results empirically validate our theoretical framework and provide practical guidance for parameter selection in differentially private implementations.
3.2. Simulation Study
We next investigate the statistical inference and variable selection capabilities of DP-FW under controlled simulation settings.
Our study addresses two fundamental questions—(1) How effectively does the proposed method estimate parameters across different sample sizes and privacy constraints? (2) Can the framework maintain variable selection accuracy in high-dimensional sparse settings while preserving privacy?
We benchmark DP-FW against maximum likelihood estimation (MLE), classical Frank–Wolfe (FW), stochastic gradient descent (SGD), and differentially private variants (DP-SGD and DP-Bayesian). The initial estimation focuses on DP-FW and DP-SGD to contrast Frank–Wolfe with gradient-based optimization, while DP-Bayesian is introduced later due to its computational cost and greater relevance in small-sample or high-dimensional settings. Evaluation metrics include Mean Squared Error (MSE), True Positive Rate (TPR), False Positive Rate (FPR), and Sum of Squared Errors (SSE).
3.2.1. Parameter Estimation
We adopt the following linear stochastic frontier model:
where
represents symmetric noise, and
denotes non-negative inefficiency, both independent across
N observations.
In low-dimensional settings, we fix
non-zero coefficients and
, with variance parameters
and
, as well as true coefficients
. The sample size
N is varied from 500 to 5000 in increments of 500, while privacy parameters are set to
and
. Hyperparameters are tuned via 5-fold cross-validation, with MSE as the criterion, using Optuna. Each experiment is repeated 100 times to mitigate randomness. For brevity,
Table 1 reports representative results for
.
Table 1 shows that while all methods accurately estimate the structural coefficients (
), MLE consistently achieves the lowest bias and variance, especially for the error parameters (
and
). Among the private methods, our proposed framework not only estimates
almost as well as the MLE baseline, but also decisively outperforms unconstrained private gradient methods in overall accuracy.
Sample Size and Privacy Effects
We further analyze sensitivity to data availability and privacy strength.
Figure 3 reports MSE trends as
N increases under
.
Figure 4 explores the privacy–utility trade-off by varying
from
to 1 at fixed
.
Figure 3 shows that increasing
N consistently reduces MSE across all methods. While non-private estimators converge stably, private gradient-based methods exhibit volatility at moderate sample sizes. In contrast, DP-FW scales smoothly with
N, confirming that bounded optimization effectively suppresses gradient sensitivity and facilitates the efficient use of additional data.
Figure 4 demonstrates the privacy–utility trade-off. As
increases, all private methods benefit from reduced noise, but DP-FW shows a markedly smoother improvement curve. This stability arises from the controlled gradient domain, which ensures that injected perturbations scale predictably with privacy level. The result is better utility per unit of privacy budget—a critical advantage in practical applications.
Small-Sample Robustness
Small-sample scenarios are common in specialized efficiency analysis applications where data availability is inherently limited, posing a greater challenge for differentially private methods due to the amplified effect of injected noise.
To examine robustness, we conduct experiments with
under
, including DP-Bayesian methods for comparison. The results are summarized in
Figure 5 and
Table 2.
Table 2 and
Figure 5 reveal several key insights. As anticipated, all methods show improved accuracy with increasing sample size. However, our proposed framework demonstrates superior robustness. Notably, at
, DP-FW achieves a good performance for structural parameters, which is comparable to the non-private SGD method.
This counterintuitive result—where private methods occasionally outperform non-private alternatives—arises from two complementary mechanisms. First, gradient descent methods exhibit a higher variance in small-sample settings, potentially leading to overfitting or unstable convergence paths. Second, bounded-domain optimization combined with calibrated perturbation provides implicit regularization effects that can improve generalization performance when data are limited.
Privacy–Utility Trade-Off
Building upon the previous analysis of sample size effects under a fixed privacy budget, we now examine how model performance changes with varying privacy parameters .
Figure 6 reports the MSE of differentially private estimators as the privacy budget
varies. Unconstrained DP-SGD exhibits inflated MSE and instability under tight privacy constraints, reflecting the difficulty of controlling gradient sensitivity in high-dimensional settings. In contrast, both Bayesian approaches and DP-FW remain substantially more stable across privacy levels. Importantly, DP-FW consistently achieves a lower MSE than Bayesian alternatives, as well as displaying smoother improvements as
increases.
Taken together, these results highlight how bounding the optimization domain translates into predictable sensitivity and improved variance control. Although differentially private estimators inevitably lose some efficiency relative to their non-private counterparts due to noise injection, DP-FW mitigates much of this loss by converting perturbation into a form of implicit regularization. The benefits are most pronounced in challenging regimes—small sample sizes, high-dimensional covariates, and stringent privacy budgets—where domain constraints and calibrated noise jointly stabilize estimation, narrow the gap to non-private baselines, and consistently outperform alternative private methods.
3.2.2. Variable Selection
This subsection evaluates the variable selection performance of our proposed DP-FW-Lasso in high-dimensional sparse settings, providing original evidence of its effectiveness under privacy constraints.
We consider a design with predictors, of which only coefficients are non-zero; these are randomly set to . To promote sparsity, we incorporate LASSO regularization into both the non-private Frank–Wolfe estimator and its private counterpart (DP-FW-Lasso). This integration is natural, as Frank–Wolfe directly accommodates -type constraints. As a non-private benchmark, we include Adaptive LASSO (ALasso) with MLE, which is well known for its variable selection consistency. We deliberately exclude a private version of ALasso, as its reliance on accurate initial estimates and iterative reweighting amplifies the effects of privacy-induced noise, reducing stability and complicating privacy budget allocation. In contrast, standard LASSO is more amenable to privacy-preserving optimization, ensuring both computational tractability and stability.
Hyperparameters were tuned using the Bayesian Information Criterion (BIC) and 5-fold cross-validation. Performance was assessed using TPR, FPR, SSE, and MSE.
Table 3 reports the results for
and
, with graphical comparisons provided in
Figure 7.
Table 3 shows that all methods recover the true variables (TPR = 1.00) but differ in false positive control and estimation accuracy. Among non-private estimators, FW-LASSO produces the most parsimonious models. Under privacy constraints, DP-FW-Lasso attains the lowest FPR and estimation error, outperforming DP-SGD-Lasso and DP-Bayesian while avoiding their scalability and stability issues.
Figure 7 further shows that DP-FW-Lasso better preserves sparsity and interpretability than other DP methods. These findings confirm the design objectives of our framework—by leveraging a convex constraint space, controlling gradient sensitivity, and integrating LASSO regularization, DP-FW-Lasso achieves a robust and superior balance between privacy, sparsity, and estimation accuracy in high-dimensional stochastic frontier models.
3.3. Real Data Analysis
We evaluate our differentially private stochastic frontier inference framework using two complementary datasets. The California Housing dataset serves as a methodological validation platform to assess algorithm performance in high-dimensional sparse regression under privacy constraints. In contrast, the Farm Accountancy Data Network (FADN) dataset demonstrates a genuine application of SFA for measuring farm production efficiency.
The California Housing dataset (
https://www.dcc.fc.up.pt/~ltorgo/Regression/cal_housing.html, accessed on 1 March 2025) comprises 20,640 observations with 8 features from the 1990 U.S. census. We expand these through nonlinear transformations, interactions, and spatial indicators to exceed 100 dimensions, creating a challenging setting for differentially private inference. The FADN dataset (
http://ec.europa.eu/agriculture/rica/, accessed on 1 March 2025) tracks farm income and productivity across EU member states. Using a 2013 cross-section, we constructed a dataset with 51 features after handling missing values, outliers, and feature engineering. The model regresses total output on total input, agricultural land, and labor, with pronounced multicollinearity posing a rigorous test for inference methods.
We compare the following six estimators:
MLE-ALasso: Non-private maximum likelihood estimation with adaptive LASSO.
FW-Lasso: Non-private Frank–Wolfe estimator with constraint.
SGD-Lasso: Stochastic gradient descent with penalty—non-private.
DP-SGD-Lasso: Standard differentially private SGD with penalty.
DP-FW-Lasso: Proposed differentially private Frank–Wolfe estimator with constraint.
DP-Bayesian: Bayesian differentially private approach via posterior sampling.
All private methods ensure -differential privacy with . Each model is evaluated over five independent trials to assess variance, and prediction accuracy is measured via MAE and RMSE.
On the California Housing dataset, we assess methods under
.
Figure 8 shows that DP-FW-Lasso consistently outperforms other private methods, with its advantage growing under stricter privacy constraints where gradient-based approaches exhibit pronounced instability. Bayesian methods remain competitive at moderate privacy levels but are limited in high-dimensional scenarios due to scalability and computational demands. These results motivate the selection of
for subsequent real-world evaluation.
Under the strictest privacy constraint (
), we apply all methods to the FADN dataset.
Figure 9 and
Table 4 summarize the MAE and RMSE results. DP-FW-Lasso achieves a 35% lower MAE than DP-SGD-Lasso and a 12% improvement over DP-Bayesian. These gains reflect the superior handling of the privacy–utility trade-off, whereby Bayesian approaches are hindered by computational cost and limited scalability, while gradient descent suffers from optimization instability exacerbated by tight privacy constraints.
Privacy-preserving methods inherently incur higher errors than non-private baselines, as noise injection for -differential privacy increases estimation uncertainty in non-convex SFM. Frank–Wolfe’s slightly higher errors relative to maximum likelihood reflect its focus on sparsity and variable selection over pure likelihood maximization, along with extra variance from mini-batch optimization.
Overall, across both datasets and privacy regimes, DP-FW-Lasso demonstrates consistent robustness, outperforming alternative private approaches in high-dimensional and structured econometric settings, while providing more scalable inference than computationally intensive Bayesian methods. These findings underscore the practical advantages of our framework for deploying differentially private stochastic frontier models in real-world applications.
4. Summary
This paper develops a novel framework for differentially private stochastic frontier analysis that enables reliable efficiency estimation under strict privacy protection. By bounding gradient sensitivity in non-convex frontier models and embedding these constraints into a Frank–Wolfe optimization scheme, we provide the first tractable approach for rigorous privacy-preserving efficiency analysis. The integration with regularization further allows consistent variable selection in high-dimensional settings.
Empirical studies confirm the advantages of the proposed framework. Compared with existing private methods, it achieves 15–35% less prediction errors, maintains robustness under small samples and strong privacy constraints, and perfectly recovers sparsity in high-dimensional cases. Application to agricultural data further demonstrates that the method preserves both accuracy and interpretability, making it suitable for practical economic efficiency evaluation under privacy restrictions.
While achieving substantial advances, our framework relies on convex relaxations and bounded parameter domains, which may limit applicability to highly flexible frontier specifications. The current cross-sectional focus also constrains application to dynamic efficiency tracking. Future research directions include extending the approach to panel data settings, developing adaptive noise calibration mechanisms for improved privacy–utility trade-offs, and exploring hybrid Bayesian–private inference for richer uncertainty quantification.