Next Article in Journal
A Priori Uniform Bounds as Measure-Theoretic Tools: Long-Term Analysis via Classical-Enhanced Synthesis
Previous Article in Journal
One Model for Many Fakes: Detecting GAN and Diffusion-Generated Forgeries in Faces, Invoices, and Medical Heterogeneous Data
Previous Article in Special Issue
Variable Dose-Constraints Method Based on Multiplicative Dynamical Systems for High-Precision Intensity-Modulated Radiation Therapy Planning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bayesian Analysis of Nonlinear Quantile Structural Equation Model with Possible Non-Ignorable Missingness

by
Lu Zhang
and
Mulati Tuerde
*
College of Mathematics and System Sciences, Xinjiang University, Urumqi 830017, China
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(19), 3094; https://doi.org/10.3390/math13193094
Submission received: 24 August 2025 / Revised: 14 September 2025 / Accepted: 17 September 2025 / Published: 26 September 2025
(This article belongs to the Special Issue Research on Dynamical Systems and Differential Equations, 2nd Edition)

Abstract

This paper develops a nonlinear quantile structural equation model via the Bayesian approach, aiming to more accurately analyze the relationships between latent variables, with special attention paid to the issue of non-ignorable missing data in the model. The model not only incorporates quantile regression to examine the relationships between latent variables at different quantile levels but also features a specially designed mechanism for handling missing data. The non-ignorable missing mechanism is specified through a logistic regression model, and a combined method of Gibbs sampling and Metropolis–Hastings sampling is adopted for missing value imputation, while simultaneously estimating unknown parameters, latent variables, and parameters in the missing data model. To verify the effectiveness of the proposed method, simulation studies are conducted under conditions of different sample sizes and missing rates. The results of these simulation studies indicate that the developed method performs excellently in handling complex data structures and missing data. Furthermore, this paper demonstrates the practical application value of the nonlinear quantile structural equation model through a case study on the growth of listed companies, providing researchers in related fields with a new analytical tool.

1. Introduction

Structural equation modeling (SEM) is a statistical method used to analyze relationships among latent variables, consisting of two components: measurement equations composed of multiple related observed indicators, and structural equations that assess the mutual influences between latent variables. SEM has found extensive utilization across diverse domains, including behavioral science, education, medicine, psychology, and social sciences. Despite its many advantages in data processing, SEM, like traditional regression models, is sensitive to outliers and typically focuses on analyzing conditional means. To address these limitations, researchers have begun integrating quantile regression techniques into SEM to provide a more comprehensive analysis of relationships between variables.
Since its introduction by Koenker and Bassett [1], quantile regression has emerged as an effective tool to overcome the limitations of traditional regression models. By estimating at different quantile levels, it more comprehensively reflects data characteristics, exhibiting strong robustness, especially when dealing with non-normal distributions and outliers, without requiring normality assumptions for error distributions. The wide application of quantile regression across various fields has significantly expanded the scope of traditional regression models. Against this backdrop, Wang et al. [2] were the first to integrate quantile regression into SEM, proposing the quantile structural equation model (QSEM). This model effectively addresses the sensitivity of traditional models to non-normal distributions and outliers by introducing the Asymmetric Laplace Distribution (ALD) and Markov Chain Monte Carlo (MCMC) algorithms, and has played an important role in research areas such as chronic kidney disease.
In the extended research on QSEM, Feng et al. [3] proposed a Bayesian regularized quantile structural equation model, introducing Bayesian Lasso and adaptive Lasso methods for variable selection and parameter estimation to enhance model flexibility and interpretability. Wang Zhiqiang [4] put forward a composite quantile structural equation model, which combines Gibbs sampling and Metropolis–Hastings algorithms for parameter and latent variable estimation. Xue Jiao et al. [5] proposed a variational approximation inference for Bayesian quantile structural equation models based on Mean Field Variational Bayes (MFVB) and adopted a non-parametric bootstrap method to improve the performance of MFVB. Cheng [6] developed a quantile varying-coefficient structural equation model and a new estimation method based on local polynomials (LPs) for flexible model estimation. However, relevant studies have mostly focused on linear frameworks and not involved nonlinear relationships between latent variables. Therefore, we propose a new nonlinear quantile structural equation model that can capture nonlinear interactions between variables and enhance the interpretability and prediction accuracy for complex phenomena.
In practice, missing data are prevalent in various research scenarios. Regarding missing data issues in SEM, previous studies have covered both linear and nonlinear frameworks, involving two mechanisms: Missing At Random (MAR) and Missing Not At Random (MNAR). In early research, Lee and Tang [7] developed a Bayesian method to analyze nonlinear structural equation models with non-ignorable missing data, which specifies the non-ignorable missing mechanism through a logistic regression model and combines Gibbs sampling and Metropolis–Hastings algorithms for parameter estimation. Cai et al. [8] proposed a nonlinear structural equation model capable of handling mixed data types (continuous, ordered, and unordered categorical) and non-ignorable missing data, exploring its Bayesian estimation and model comparison methods. The integrated SEM proposed by Lee and Song [9] realizes estimation and model comparison of mixed data under MAR through Bayesian methods combined with MCMC and path sampling. Cai and Song [10] focused on the situation where response variables have non-ignorable missing data in mixed structural equation models, also adopting Bayesian methods for analysis. Cai et al. [11] addressed the problem of both response variables and covariates having non-ignorable missing data in mixed structural equation models, achieving parameter estimation and model selection through Bayesian methods. Existing studies have mostly focused on traditional mean-structured SEM, with insufficient attention to quantile structural equation models, especially lacking systematic research on non-ignorable missing data in nonlinear quantile structural equation models. Therefore, this paper considers a nonlinear quantile structural equation model with non-ignorable missing data.
This paper focuses on the influencing factors of the growth of listed companies in Xinjiang. Corporate growth refers to a company’s ability to achieve sustained development and expansion in the future, and this study mainly examines it from two dimensions: profitability and solvency. Since these influencing factors cannot be directly measured by a single observed variable and there are complex nonlinear relationships between them, this study aims to construct a nonlinear quantile structural equation model to conduct an in-depth analysis of the influencing factors and their interaction mechanisms of corporate growth among 61 listed companies in Xinjiang, providing a scientific decision-making basis for corporate management and contributing to the sustainable development of the regional economy.
The structure of this paper is arranged as follows: Section 2 introduces the nonlinear quantile structural equation model with non-ignorable missing data. Section 3 presents a Bayesian analysis of the above model. Section 4 conducts simulation studies with different sample sizes and missing rates to evaluate the performance of the proposed method. Section 5 illustrates the proposed method through a practical case study. Section 6 includes a discussion. Appendix A presents some technical details.

2. Model and Notation

2.1. Nonlinear Quantile Structural Equation Model

The quantile structural equation model (QSEM) typically consists of a measurement equation in the form of median regression and a structural equation in the form of quantile regression. The measurement equation characterizes the features of latent variables through multiple observed variables, while the structural equation evaluates the associations between latent variables and reveals the mechanism by which explanatory latent variables influence outcome latent variables. In previous QSEM studies, both the measurement equation and the structural equation were mostly assumed to have linear relationships. Based on this, we extend the “double-linear” specification of traditional QSEM and propose a new definition of the nonlinear quantile structural equation model.
This model takes “linear measurement equation + nonlinear structural equation” as its core feature. It not only inherits the classical framework of nonlinear structural equation models defined by Lee [12] and Song and Lee [13], where the linear measurement equation ensures the effective reflection of latent variables by observed variables and the nonlinear form characterizes the associations between latent variables, but also introduces innovations within the quantile regression framework. Specifically, the structural equation allows nonlinear relationships in the form of differentiable functions between latent variables. Meanwhile, by means of the quantile regression-type equation, it overcomes the limitation that traditional models can only characterize the average associations between variables, realizes the accurate characterization of variable relationships at different quantiles, and thus better conforms to the actual characteristics of variable associations in complex data scenarios.
First, to investigate the association between observed and latent variables, we begin by assuming that the observed variables y 1 ,   ,   y n are independently distributed, where each y i = ( y i 1 ,   ,   y i p ) T represents a p-dimensional vector of measurements. Correspondingly, the latent variables ω 1 ,   ,   ω n are also independent, with each ω i = ( ω i 1 ,   ,   ω i q ) T being a q-dimensional random vector ( q < p ).
The linear relationship between y i and ω i can then be defined by the following measurement equation:
y i = A c i + Λ ω i + ε i , i = 1 , , n ,
where A ( p × r 1 ) is an unknown coefficient matrix; Λ ( p × q ) is an unknown factor loading matrix; c i ( r 1 × 1 ) is a vector of fixed covariates; and ε i ( p × 1 ) is a random error vector. Within this model, a linear regression model connects the conditional median of y i with c i and ω i .
Next, to study the relationships between latent variables, we decompose ω i into ( η i T , ξ i T ) T , where η i ( q 1 × 1 ) denotes the outcome latent variable, ξ i ( q 2 × 1 ) denotes the explanatory latent variable, and q 1 + q 2 = q . For simplicity, we assume q 1 = 1 . The nonlinear relationship between η i and ξ i can then be defined by the following structural equation:
η i = B τ d i + Γ τ H ( ξ i ) + δ i , i = 1 , , n ,
where B τ ( q 1 × r 2 ) and Γ τ ( q 1 × q 3 ) are unknown coefficient matrices to be estimated; the subscript τ indicates that these matrices may differ across quantiles; d i ( r 2 × 1 ) is a vector of fixed covariates; H ( ξ i ) = h 1 ( ξ i ) ,   ,   h q 3 ( ξ i ) T is a q 3 × 1 -dimensional nonlinear function of ξ i , whose components h 1 ,   ,   h q 3 are differentiable functions; and δ i ( q 1 × 1 ) is a random error vector. In the present model, a nonlinear regression model connects the conditional quantile of η i with d i and H ( ξ i ) .
Equations (1) and (2) together constitute the nonlinear quantile structural equation model (NQSEM).

2.2. Asymmetric Laplace Distribution

In the Bayesian quantile regression method based on the Asymmetric Laplace Distribution (ALD) proposed by Yu and Moyeed [14], the “true underlying distribution” of the error term is unknown. This is not only an objective constraint in real-world data scenarios but also the starting point of the core design logic of this method. In practical research, data often exhibit complex characteristics such as non-normality, heavy tails, and heteroscedasticity, which makes it difficult for researchers to pre-determine or describe the true underlying distribution of the error term using a single fixed distribution.
To address this issue, the key innovation of Yu and Moyeed [14] is to avoid direct modeling of the “unknown true underlying distribution”. Instead, they revealed the intrinsic theoretical connection between quantile regression and ALD: the quantiles of ALD correspond exactly to the conditional quantiles of the regression model. On this basis, they designated ALD as a “working distribution”. By assuming that the error term follows ALD, they constructed the likelihood function. This assumption is not intended to reconstruct the true distribution of the error term; instead, it uses ALD as a bridge to enable Bayesian estimation of quantile regression parameters without relying on the true underlying distribution of the error term.
The core that supports the validity of this approach is the in-depth compatibility between the check loss function of quantile regression and the probability density function of ALD. The check loss function of quantile regression is defined as ρ τ ( x ) = x ( τ I ( x < 0 ) ) , where x denotes the residual (i.e., the difference between the model’s predicted values and the actual observed values), τ is the preset quantile level, and I ( · ) is an indicator function that takes the value of 1 when x < 0 and 0 otherwise. This function assigns asymmetric weights to positive and negative residuals: when x > 0 , the loss is τ · x ; when x 0 , the loss is ( 1 τ ) · | x | , which precisely matches the need of quantile regression to “characterize the relationship between variables at different quantiles”. The probability density function of ALD is expressed as f ( y | μ , σ , τ ) = τ ( 1 τ ) σ exp ρ τ y μ σ , where y represents a random variable following ALD, μ is the location parameter, σ is the scale parameter, and τ is the skewness parameter. This probability density function directly embeds the aforementioned check loss function, and the equivalence between the two means that maximizing the ALD likelihood function is essentially equivalent to minimizing the quantile regression loss function—providing a theoretically consistent quantitative basis for Bayesian inference. At the same time, this modeling approach based on the check loss function and ALD can effectively handle complex data characteristics such as non-normality and heteroscedasticity, and ultimately yield robust parameter estimation results.
Building on this core finding of Yu and Moyeed [14], Wang et al. [2] further extended the application of ALD to quantile structural equation models (SEMs). They continued the approach of “constructing the likelihood function using ALD as the working distribution” and derived a form of likelihood function suitable for Bayesian analysis of quantile SEM, which provides effective support for Bayesian inference in such models.
Assume that the probability density function of the k-th component ε i k ( k = 1 ,   ,   p ) of the error term ε i follows an Asymmetric Laplace Distribution, with the form
f ε i k | μ , σ , τ = τ ( 1 τ ) σ exp ρ τ ε i k μ σ , ε i k ( , + ) ,
where μ R is the location parameter, σ > 0 is the scale parameter, τ ( 0 < τ < 1 ) is the skewness parameter, and ρ τ ( x ) = x τ I ( x < 0 ) is the check function. Thus, we denote ε i k ALD ( μ , σ , τ ) . Similarly, this paper assumes q 1 = 1 , δ i is a one-dimensional real number, and its probability density function is given by
f δ i | μ , σ , τ = τ ( 1 τ ) σ exp ρ τ δ i μ σ , δ i ( , + ) ,
Thus, we denote δ i ALD ( μ , σ , τ ) .
In statistics, “standard distributions” generally refer to distribution types (such as the normal distribution and gamma distribution) that are widely used in theoretical research and practical applications, with clear probability density forms and mature supporting analysis tools (e.g., having tractable conjugate priors). From this definition, ALD does not belong to such standard distributions. It is not only not a commonly used distribution type in statistics but, more importantly, lacks a tractable conjugate prior. Coupled with the inherent complexity of its likelihood function, this directly makes the posterior density function of the model difficult to handle through analytical methods, thereby inevitably increasing the computational burden in the process of model parameter estimation. Nevertheless, according to studies by Reed and Yu [15] and Kozumi and Kobayashi [16], ALD can be expressed as a mixture of exponential and normal distributions. Specifically, if the random variables ε i k and δ i follow ALD ( 0 , σ , τ ) , they can be expressed as
ε i k = k 1 ( τ ) e y i k + k 2 ( τ ) σ y k e y i k ς i ,
δ i = k 1 ( τ ) e η i + k 2 ( τ ) σ η e η i ς i ,
where k 1 ( τ ) = 1 2 τ τ 1 τ , k 2 ( τ ) = 2 τ 1 τ , e y i k exp 1 σ y k , e η i exp 1 σ η , and ς i N 0 , 1 ; e y i k is the k-th component of e y i , σ y k is the k-th component of σ y , and ς i is independent of e y i k and e η i , respectively.
Given the expressions for ε i k and δ i , and noting that ε i = ( ε i 1 , ε i 2 ,   ,   ε i p ) T , we can derive the forms of both y i and η i in the nonlinear quantile structural equation model as follows:
y i = A c i + Λ ω i + k 1 ( τ ) e y i + k 2 ( τ ) σ y e y i ς i ,
η i = B τ d i + Γ τ H ( ξ i ) + k 1 ( τ ) e η i + k 2 ( τ ) σ η e η i ς i ,
where y i follows ALD ( A c i + Λ ω i , σ , τ ) , and η i follows ALD ( B τ d i + Γ τ H ( ξ i ) , σ , τ ) . By introducing latent variables e y i and e η i to augment y i and η i , the conditional distributions of y i and η i become normal distributions. Their conditional means are A c i + Λ ω i + k 1 ( τ ) e y i and B τ d i + Γ τ H ( ξ i ) + k 1 ( τ ) e η i , respectively, and their variances are k 2 ( τ ) σ y e y i and k 2 ( τ ) σ η e η i , respectively. This data augmentation facilitates subsequent Bayesian analysis, allowing normal distributions to be used as prior distributions for unknown coefficients.

2.3. Non-Ignorable Missing Data

Wang et al. [2] performed Bayesian statistical analyses on linear QSEM using completely observed data. In the present study, we extend this framework to account for incomplete observations of the vector y i , where missingness arises from a non-ignorable mechanism. Specifically, we decompose y i into two components: y i = y o i T , y m i T T , where y o i ( p 1 i × 1 ) represents the observed portion of the manifest variables, and y m i ( p 2 i × 1 ) denotes the missing portion, with the dimension constraint p 1 i + p 2 i = p .
We assume that the missingness in y i occurs in an arbitrary manner. Thus, y i = y o i T , y m i T T can be seen as a reordering of the elements from the original y i . No matter how the elements are reordered, the observed vector y o i and the unobserved vector y m i together form the complete manifest variable vector y i with a non-ignorable missing mechanism.
In the study of missing data problems, if the missing pattern is correlated with the missing data itself, the mechanism is referred to as a non-ignorable missing mechanism. In this case, p r i y i , ω i , φ is a key conditional probability distribution, which describes the probability distribution of the missing indicator variable r i given the observed data y i and latent variable ω i . It is crucial to select an appropriate model to describe this distribution, as the model needs to balance complexity and identifiability to avoid ineffective parameter estimation or computational difficulties caused by excessive model complexity.
Before introducing the model, it is necessary to clarify the definition of the missing indicator variable r i :
r i = 1 , if y i is missing data , 0 , if y i is observed data ,
According to the properties of the indicator function, r i is a binary vector and follows a 0–1 distribution with probability p r i y i , ω i , φ .
In the nonlinear quantile structural equation model, given that the observation vectors y 1 , y 2 ,   ,   y n are mutually independent, it is reasonable to assume that the missing indicator variables r 1 , r 2 , , r n are also mutually independent, and their joint probability is the product of the respective independent probabilities. Since the components of y i are independent given ω i , when j l , r i j and r i l are conditionally independent given y i and ω i , and thus the joint probability can be further decomposed into the product of the probabilities of each component.
Drawing on the non-ignorable missingness mechanism model proposed by Ibrahim et al. [17], we model the binary missing missingness indicator r i j by specifying the probability p ( r i j = 1 y i , ω i , φ ) . We employ a logistic regression model for this purpose, as it reduces the number of parameters in the missing data mechanism and facilitates efficient sampling from the conditional distribution given the observed data. The link function for this model is denoted by G ( · ) , specifically, G ( x ) = logit 1 ( x ) = ( 1 + e x ) 1 . The linear predictor is given by
logit p ( r i j = 1 y i , ω i , φ ) = φ 0 + φ 1 y i 1 + + φ p y i p + φ p + 1 ω i 1 + + φ p + q ω i q = φ T F i ,
Consequently, the distribution of the missingness indicator is
r i j y i , ω i , φ Bernoulli G ( φ T F i ) ,
where logit p = log p 1 p , F i = ( 1 , y i 1 ,   ,   y i p , ω i 1 ,   ,   ω i q ) T is the design vector, and φ = ( φ 0 , φ 1 ,   ,   φ p + q ) T is the coefficient vector; both are ( p + q + 1 ) -dimensional.
In addressing nonlinear structural equation models (NSEMs) with non-ignorable missing data, Lee and Tang [7] made pioneering contributions. After in-depth analysis of the complex model structure and data characteristics, they adopted a specific form, as shown in Equation (12):
logit p ( r i j = 1 y i , ω i , φ ) = φ 0 + φ 1 y i 1 + + φ p y i p = φ T F i * ,
where F i * = ( 1 , y i 1 ,   ,   y i p ) T (i.e., excluding the latent variable ω i from the design vector), and this definition of F i * is consistent with the notation used for F i * in subsequent parts of the manuscript.
A prominent feature of this form is that it no longer depends on the latent variables ω i . They argued that the characteristics of latent variables ω can be reflected by observed variables y. This insight offers dual advantages: from a computational perspective, it significantly reduces the dimension of parameter estimation and complex matrix operations during model computation, effectively alleviating the computational burden; from the perspective of model application, it simplifies the originally obscure and complex model structure, enhancing the model’s operability and interpretability in practical scenarios.
In subsequent studies, they further conducted large-scale simulation studies to comprehensively verify the effectiveness of this simplified model. During the simulation process, multiple sets of different data generation mechanisms and model parameter combinations were set up, covering common complex data scenarios. The results of the simulation studies verified the superiority of the simplified model from multiple dimensions: in terms of parameter estimation accuracy, its estimation bias (Bias) and root mean square error (RMS) were both at a level similar to those of the full model, indicating that this simplification did not lead to a significant loss of estimation precision; more importantly, the computational efficiency of the simplified model was significantly improved—it not only shortened the time required for parameter iteration convergence but also reduced the computational complexity in the high-dimensional sampling process.
Given that the research findings of Lee and Tang [7] possess both reliability and practicality, considering that the research scenario and data characteristics of this paper are highly consistent with those of their study, and for the further purposes of ensuring model identifiability and improving computational tractability, this paper decides to adopt Equation (12) for modeling. It is expected that this will also enable efficient and accurate model construction and analysis when dealing with nonlinear quantile structural equation models with non-ignorable missing data.
It can be seen from the above missing data mechanism model that the missingness of data y i is non-ignorable. Thus, p r i j y i , ω i , φ describes such a data mechanism where the response variable has non-ignorable missingness.

3. Bayesian Inference for the Proposed Model

Let Y = ( y 1 , , y n ) , Y m = { y m 1 , , y m n } denote the set of missing values related to observed variables, Y o = { y o 1 , , y o n } denote the set of observed response variables, Ω = ( ω 1 , , ω n ) , r = ( r 1 , , r n ) denote the vector of missing indicators, and θ = A , Λ , B τ , Γ τ , Φ , σ y k , e y i k , σ η , e η i represent all unknown parameters in Equations (1) and (2). Our primary focus is to perform posterior inference on the unknown parameters of interest θ and the missing mechanism parameter φ by utilizing the missing data indicators r and the observed dataset Y o .
The joint posterior distribution of the unknown parameters θ and φ conditioned on Y o and r is derived as follows:
p ( θ , φ Y o , r ) p ( Y o , r θ , φ ) · p ( θ , φ ) p ( Y o , Y m , r , Ω θ , φ ) d Ω d Y m · p ( θ , φ ) i = 1 n p ( y i , r i , ω i θ , φ ) d ω i d y m i · p ( θ , φ ) i = 1 n p ( y i ω i , θ ) p ( r i y i , ω i , φ ) p ( ω i θ ) d ω i d y m i · p ( θ , φ ) .
where p · · denotes a conditional probability density, and p θ , φ denotes the combined prior distribution assigned to θ and φ . It should be noted that the integral contained in Equation (13) is a high-dimensional integral, whose dimension equals the sum of the dimensions of ω i and y m i . Since such high-dimensional integrals are difficult to solve directly and have no closed-form expression, directly conducting posterior inference poses numerous challenges and is quite difficult. Therefore, numerical methods (such as Markov Chain Monte Carlo methods) need to be used for approximate calculation. In this study, we adopt a hybrid algorithm combining Gibbs sampling and the Metropolis–Hastings algorithm to implement posterior Bayesian analysis of the model.
By drawing on the data augmentation method proposed by Tanner and Wong [18], we extend the original joint posterior distribution p ( θ , φ Y o , r ) to p ( Ω , Y m , θ , φ Y o , r ) . Subsequently, based on this extended joint posterior distribution, we use the Gibbs sampling and Metropolis–Hastings algorithms to iteratively sample from the following conditional distributions in sequence, thereby obtaining a random observation sequence of { Ω , Y m , θ , φ } . The specific implementation steps can be divided into two main parts.
1. Generating Observed Data { Y o , r }
This part constructs the observed data and missing indicator matrix required for subsequent sampling by simulating complete data and the missing mechanism. The specific steps are as follows:
Step 1: Generate the Complete Dataset Y
Generate the complete dataset { y i j : i = 1 , , n ; j = 1 , , p } according to Equations (1) and (2). This dataset serves as the “ground truth dataset” for simulating missing status, which is used to determine whether data are missing and define the scope of observed data in subsequent steps, acting as a reference benchmark.
Step 2: Determine the Missing Status of Observations and Construct { Y o , r }
Based on the missing mechanism (12) and a pre-specified value of φ , determine whether each observation y i j is missing. Specifically, generate a random number u from a uniform distribution. If u p r i j = 1 y i , ω i , φ , mark y i j as missing (denoted as r i j = 1 ); otherwise, mark it as non-missing (denoted as r i j = 0 ). Finally, extract non-missing values from the complete dataset Y to form the observed data Y o , and construct the missing indicator matrix r from all r i j . Thus, { Y o , r } required for subsequent sampling is obtained.
2. Posterior Sampling Based on { Y o , r } to Obtain { Ω , Y m , θ , φ }
With the generated { Y o , r } as input, iterative sampling is performed to estimate the target parameters and missing data. The specific steps are as follows:
Step 1: Specify Initial Values
Specify the initial values as Ω ( 0 ) , Y m ( 0 ) , θ ( 0 ) , φ ( 0 ) .
Step 2: Iterative Sampling by Traversing Each Component
Traverse each component of Ω , Y m , θ , and φ in sequence. Each component is sampled based on the current values of all other components. Denote the values at the t-th iteration as Ω ( t ) , Y m ( t ) , θ ( t ) , φ ( t ) ; the hybrid sampling for the ( t + 1 ) -th iteration is defined as follows:
(a)
Sample Ω ( t + 1 ) from the conditional distribution p Ω Y o , Y m ( t ) , r ( t ) , θ ( t ) , φ ( t ) ;
(b)
Sample Y m ( t + 1 ) from the conditional distribution p Y m Y o , Ω ( t ) , r ( t ) , θ ( t ) , φ ( t ) ;
(c)
Sample θ ( t + 1 ) from the conditional distribution p θ Y o , Y m ( t ) , Ω ( t ) ;
(d)
Sample φ ( t + 1 ) from the conditional distribution p φ Y o , Y m ( t ) , Ω ( t ) , r ( t ) , θ ( t ) .
Repeat the iterative process in Step 2 (Posterior Sampling Iteration Step), which is the process of traversing components to sample and update Ω , Y m , θ , φ , until the algorithm converges. Among them, the distributions corresponding to (a), (b), and (d) are non-standard, uncommon, and relatively complex; thus, the Metropolis–Hastings algorithm is required to address the sampling challenge. In contrast, the distribution corresponding to (c) is a standard and common distribution, and its sampling process is relatively straightforward and simple, so the Gibbs sampler can be directly used for sampling.

3.1. Posterior Distributions

To perform hybrid sampling, we need to specify the four posterior distributions involved in the sampling process. Drawing on relevant derivations in Lee and Tang [7], we can obtain the following conclusions:
First, consider the posterior distribution p Ω Y o , Y m , r , θ , φ of the latent variable Ω . When both the missing data Y m and observed data Y o are known, the posterior distribution can be simplified to p Ω Y , r , θ , φ , which depends on the complete dataset Y .
Since ω i are mutually independent, and y i are also mutually independent given ω i , it follows that
p Ω Y , r , θ , φ = i = 1 n p ω i y i , r i , θ , φ i = 1 n p y i ω i , θ p η i ξ i , θ p ξ i θ p r i y i , ω i , φ .
For the reasons stated earlier, the ALD is used to model the error terms regardless of their true latent distribution. Specifically, let the k-th component ε i k of the error term ε i follow ALD 0 , σ y k , 0.5 to model the median-based regression specified in Equation (1); let the error term δ i follow ALD 0 , σ η , τ to model the τ -th quantile regression in Equation (2). This is denoted as y i k ALD μ , σ y k , 0.5 and η i ALD μ , σ η , τ .
Building on this ALD specification, we first define the covariance matrices for the conditional distributions: the covariance matrix of y i is Ψ ε i = diag k 2 ( τ ) σ y 1 e y i 1 , , k 2 ( τ ) σ y p e y i p , and the variance of η i is Ψ δ i = k 2 ( τ ) σ η e η i . As the parameter k 2 ( τ ) = 2 τ ( 1 τ ) was defined previously, substituting τ = 0.5 (the median case, consistent with ε i ’s ALD setting) yields k 2 ( 0.5 ) = 2 0.5 × ( 1 0.5 ) = 8 ; therefore, Ψ ε i can also be simply expressed as diag 8 σ y 1 e y i 1 , , 8 σ y p e y i p .
Based on these settings, we can further derive the following conditional distributions:
y i ω i , θ N p A c i + Λ ω i , Ψ ε i ,
η i ξ i , θ N B τ d i + Γ τ H ξ i + k 1 ( τ ) e η i , Ψ δ i ,
ξ i θ N q 2 0 , Φ ,
Thus, it follows that
p ω i y i , θ exp 1 2 ξ i T Φ 1 ξ i 1 2 y i A c i Λ ω i T Ψ ε i 1 y i A c i Λ ω i 1 2 η i B τ d i Γ τ H ξ i k 1 ( τ ) e η i T Ψ δ i 1 η i B τ d i Γ τ H ξ i k 1 ( τ ) e η i + j = 1 p r i j φ T F i * p log 1 + exp φ T F i * ,
Second, consider the posterior distribution p Y m Y o , Ω , r , θ , φ of the missing observed variables Y m . It follows that
p Y m Y o , Ω , r , θ , φ = i = 1 n p y m i y o i , ω i , r i , θ , φ i = 1 n p y m i ω i , θ p r i y i , ω i , φ ,
Moreover, p y m i y o i , ω i , r i , θ , φ
exp 1 2 y m i A m i c i Λ m i ω i T Ψ m ε i 1 y m i A m i c i Λ m i ω i + j = 1 p r i j φ T F i * p log 1 + exp φ T F i * ,
where A m i denotes a p 2 i × 1 subvector extracted from A , whose entries correspond to the missing values in y i ; Λ m i refers to a p 2 i × q submatrix derived from Λ , with rows that map to the missing entries in y i ; and Ψ m ε i represents a p 2 i × p 2 i submatrix of Ψ , featuring both rows and columns that align with the missing components of y i .
Third, consider the posterior distribution p θ Y o , Y m , Ω of the unknown parameter θ . Similarly, the posterior distribution can be simplified to p θ Y , Ω .
Let Λ y = A , Λ , where Λ y k T is the k-th row of Λ y for k = 1 , , p ; and let Λ ω τ = B τ , Γ τ . Given Y and Ω , the posterior distribution of θ can be obtained using the following conjugate prior distributions:
Λ y k N r 1 + q Λ 0 y k , H 0 y k ,
σ y k 1 Gamma α 0 y k , β 0 y k ,
Φ 1 Wishart R 0 , ρ 0 ,
Λ ω τ N r 2 + q 3 Λ 0 ω , H 0 ω ,
σ η 1 Gamma α 0 σ , β 0 σ ,
where α 0 y k , β 0 y k , α 0 σ , β 0 σ , ρ 0 , Λ 0 y k , Λ 0 ω , and the positive definite matrices H 0 y k , H 0 ω , R 0 are hyperparameters, whose specific values need to be set based on prior information or expert knowledge.
Finally, consider the posterior distribution p φ Y o , Y m , Ω , r , θ of the missing mechanism parameter φ . This can be simplified to p φ Y , Ω , r , θ . Assume that p φ denotes the prior probability density function of parameter φ , which follows a p + 1 -dimensional multivariate normal distribution N p + 1 φ 0 , V , where φ 0 and V are hyperparameters whose values are predetermined based on prior information. Given that the distribution of r depends exclusively on Y , Ω , and φ , coupled with our assumption that the prior for φ is statistically independent of that for θ , the following result emerges:
p φ Y , Ω , r , θ p r Y , Ω , φ p φ .
From Equations (11) and (12), we have
p r i j = 1 y i , ω i , φ = exp φ T F i * 1 + exp φ T F i * ,
and substituting the value of p r i j = 1 y i , ω i , φ into p r Y , Ω , φ , we get
p r Y , Ω , φ = i = 1 n j = 1 p p r i j = 1 y i , ω i , φ r i j 1 p r i j = 1 y i , ω i , φ 1 r i j = i = 1 n j = 1 p exp φ T F i * 1 + exp φ T F i * r i j 1 exp φ T F i * 1 + exp φ T F i * 1 r i j = exp i = 1 n j = 1 p r i j φ T F i * i = 1 n p log 1 + exp φ T F i * .
where the posterior distribution of φ can be further derived as follows:
p φ Y , Ω , r , θ exp i = 1 n j = 1 p r i j φ T F i * 1 2 φ φ 0 T V 1 φ φ 0 i = 1 n p log 1 + exp φ T F i * .
The specific implementation steps of the aforementioned hybrid sampling algorithm, as well as the complete derivation of the relevant posterior distributions, are all presented in Appendix A.

3.2. Bayesian Estimation

By applying the hybrid algorithm described in Section 3.1, we draw a random sample from the posterior distribution p Ω , Y m , θ , φ Y o , r , denoted as Ω ( t ) , θ ( t ) , φ ( t ) , Y m ( t ) :   t = 1 , , T . Based on these samples, we can calculate the joint Bayesian estimates of Ω , Y m , θ , and φ as follows:
Ω ^ = T 1 t = 1 T Ω ( t ) , Y ^ m = T 1 t = 1 T Y m ( t ) , θ ^ = T 1 t = 1 T θ ( t ) , φ ^ = T 1 t = 1 T φ ( t ) ,
As documented in Geyer’s [19] study, these combined Bayesian joint estimates serve as consistent approximations for the posterior means they each correspond to.
In Bayesian structural equation modeling, the evaluation of the overall model fit needs to comprehensively examine both its measurement and structural components. To ensure the reliability of statistical inference, it is insufficient to only focus on the accuracy of parameter estimation; a systematic assessment of the overall fit between the model and the data is also required. Regarding this issue, the existing literature has developed various diagnostic methods based on posterior distributions. For example, Gelman et al. [20] proposed the method of posterior predictive p-values. To address the issue of anomalous posterior p-value calculations that may arise from repeated use of observed data, Bayarri and Berger [21] further proposed the method of partial posterior predictive p-values. The most intuitive approach is to perform residual analysis. We can obtain the residual estimates of the measurement equations by calculating ε ^ i = y i A ^ c i Λ ^ ω ^ i , where A ^ , Λ ^ , and ω ^ i are the Bayesian estimates of A , Λ , and ω i , respectively. Residual plots are generated based on the values of ε ^ i , and scatter plots of ε ^ i against ω ^ i are constructed. If the plots of ε ^ i and the scatter plots of ε ^ i versus ω ^ i all lie between two parallel horizontal lines that are narrowly spaced and centered at zero, this indicates a good model fit. Similarly, the residual estimate of the structural equation is given by δ ^ i = η ^ i B ^ τ d i Γ ^ τ H ( ξ ^ i ) , where η ^ i , B ^ τ , Γ ^ τ , and ξ ^ i are all posterior estimates. The plotting and interpretation of δ ^ i are the same as those for ε ^ i .

4. Simulation

In this section, we verify the validity of the estimation methods for the Bayesian quantile structural equation model in handling complete and missing data through two simulation experiments. Simulated data are generated using the following nonlinear quantile structural equation model:
y i = A c i + Λ ω i + ε i ,
η i = b 1 τ d i + γ 1 τ ξ i 1 + γ 2 τ ξ i 2 + γ 3 τ ξ i 1 ξ i 2 + γ 4 τ ξ i 1 2 + γ 5 τ ξ i 2 2 + δ i .
Let p = 9 , q = 3 , q 1 = 1 , q 2 = 2 , q 3 = 5 , and r 1 = r 2 = 1 . We consider two cases of sample sizes, n = 100 and n = 300 , with quantiles τ set to 0.1, 0.5, and 0.9, respectively. The covariate coefficient A = ( 1 , , 1 ) T . Fixed covariates c i k and d i are both sampled from the standard normal distribution N ( 0 , 1 ) . The non-overlapping factor loading matrix Λ is shown as follows:
Λ T = 1 * λ 21 λ 31 0 * 0 * 0 * 0 * 0 * 0 * 0 * 0 * 0 * 1 * λ 52 λ 62 0 * 0 * 0 * 0 * 0 * 0 * 0 * 0 * 0 * 1 * λ 83 λ 93 ,
where the zeros and ones marked with ∗ are pre-specified to clearly explain the identification of latent variables and the model, while other λ j k are unknown parameters to be estimated. Let λ 21 = λ 31 = λ 52 = λ 62 = λ 83 = λ 93 = 0.8 , b 1 τ = 1 , and γ 1 τ = γ 2 τ = γ 3 τ = γ 4 τ = γ 5 τ = 0.6 . The explanatory latent variables ξ i N ( 0 , Φ ) , where ϕ 11 = ϕ 22 = 1 and ϕ 12 = 0.2 .
In practical research, the true underlying distribution of error terms is often difficult to know in advance. To fully verify the adaptability of the proposed model under different data distribution characteristics, we refer to the setting of error scenarios in quantile structural equation models by Wang et al. [2] and the simulation idea of complex distributions in composite quantile regression by Wang Zhiqiang [4]. We select the following four representative distributions as the generation distributions for the error terms ε i k and δ i , thereby constructing simulated data y i and η i with different characteristics. The specific distribution settings are as follows:
Case 1: Normal distribution N ( 0 , 0.3 ) .
Case 2: Skewed log-normal distribution ln N ( 0 , 0.25 ) .
Case 3: U-shaped distribution Beta ( 0.5 , 0.5 ) .
Case 4: Heavy-tailed skewed distribution 0.3 χ 2 ( 3 ) .
To simplify the simulation design and avoid cumbersome permutations and combinations, this study only considers the scenario where ε i k and δ i follow the same distribution. In the subsequent Bayesian analysis, given that the true underlying distribution of error terms is unknown, we select the Asymmetric Laplace Distribution (ALD) as a “working distribution” to replace the true distribution for modeling. By comparing the parameter estimation effects (such as bias and root mean squared error) of the ALD substitution assumption under the above four distribution scenarios, we corely verify the applicability of ALD when the true distribution is unknown, providing a reliable basis for the practical application of the model.
Based on the given true values and values sampled from specific distributions, η i and y i are calculated sequentially to obtain the complete dataset y i j : i = 1 , , n , j = 1 , , p . Missing data are then generated according to the pre-specified values of φ and the missing mechanism Equation (12). In this process, we investigate three scenarios with missing rates of 20%, 30%, and 40%. Specifically, different values of φ are adopted for each scenario, and these varying φ values ultimately lead to the corresponding differences in missing rates. The true values of φ applied in these scenarios are presented in Table A1 and Table A2 of Appendix A.
Hyperparameters are set as follows: free elements in Λ 0 y k are set to their true values, and the covariance matrix H 0 y k is taken as a diagonal matrix with diagonal elements 10 3 ; Λ 0 ω = ( 1 , 0.6 , 0.6 , 0.6 , 0.6 , 0.6 ) , and the covariance matrix H 0 ω is set the same as H 0 y k ; α 0 y k = α 0 σ = 9 , β 0 y k = β 0 σ = 4 ; ρ 0 = 1 , R 0 = 0.2 I 2 ; φ 0 is set to its corresponding true values; and V = I 10 .
Accurate computation of Bayesian estimates requires determining how many iterations are required to achieve convergence. Therefore, we set three different sets of initial values, ran three chains separately, and calculated the EPSR values. To intuitively show the variation process of each parameter’s EPSR value with the number of iterations, we plotted the EPSR convergence trend graph for all parameters (see Figure 1). The results indicate that after around 3000 iterations, the EPSR values for all parameters fall below 1.2. Thus, we performed 5000 iterations, discarded the first 2000 iterations as the burn-in period, and conducted Bayesian parameter estimation based on the subsequent 3000 iterations. Residual plots and scatter plots were generated using the obtained parameter estimates to evaluate the model fit.
Figure 2 presents partial estimated residual plots of the model under the scenario with non-ignorable missing data, showing the changes in ε ^ i 1 , ε ^ i 2 , ε ^ i 3 , and δ ^ i with observation indices. Figure 3 displays scatter plots of the measurement equation residuals ε ^ i 1 against the latent variables ξ ^ i 1 , ξ ^ i 2 , and η ^ i . Figure 4 shows scatter plots of the structural equation residuals δ ^ i against the latent variables ξ ^ i 1 and ξ ^ i 2 . From the overall performance of the residual plots and scatter plots, all residuals are uniformly distributed around zero, and most fall between two parallel horizontal lines with a narrow spacing, without showing obvious systematic bias or abnormal fluctuations. This indicates that the constructed model has a good fit to the data and can effectively capture the underlying relationships between variables. The residual plots and scatter plots for the model without missing data are similar, and thus are not repeated here.
We repeated the sampling 100 times. To evaluate the accuracy of the model estimates, we used bias and root mean square error (RMSE), with the specific formulas as follows:
Bias θ ^ = E θ ^ θ , RMSE θ ^ = 1 n i = 1 n E θ i ^ θ 2 1 / 2 .
Table 1 presents the Bayesian estimation results of the core regression coefficients ( b 1 τ , γ 1 τ γ 5 τ ) in the structural equation under the scenario of missing data with a sample size of n = 100 . It covers three error distributions, namely N ( 0 , 0.3 ) , Beta ( 0.5 , 0.5 ) , and 0.3 χ 2 ( 3 ) , as well as three missing rates (M1, M2, M3). The table focuses on reporting the bias and root mean square error (RMS) of each coefficient at the quantiles of τ = 0.1 , 0.5 , and 0.9 .
As can be seen from the data in the table, regardless of whether the error distribution is normal or not and how the missing rate changes, the estimation performance of the core regression coefficients remains stable: the absolute values of bias are generally low, and those of all parameters are less than 0.1; the overall RMS maintains a low fluctuation range without significant abnormal values, and there is no obvious fluctuation across different quantiles. This fully demonstrates the reliable estimation ability of the proposed model for core parameters under the conditions of small sample size and missing data.
It is worth noting that the estimation results of other parameters in the measurement equation (such as a 1 a 9 , λ 21 , λ 31 , λ 52 etc., and ϕ 11 ϕ 22 ) are similar to or even better than those of the core regression coefficients. To save space, the detailed estimation results of the above-mentioned other parameters, as well as those under different sample sizes, complete data, and more error distribution combinations (Table A3, Table A4, Table A5, Table A6, Table A7, Table A8 and Table A9), have been uniformly organized in Appendix A, which can be further referred to by readers to verify the estimation performance of the model in a wider range of scenarios.
All simulation experiments in this study were computed using a device with an Intel (R) Core (TM) i5-10210U CPU (1.60GHz, 2112 MHz, 4 cores, and 8 logical processors) and implemented in R software (Version 4.4.2). The total runtime for completing both Simulation 1 and Simulation 2 was approximately 364 h, and the relevant codes are provided in Appendix A.

5. A Real Example

In this section, we illustrate the proposed model by analyzing the factors influencing the growth of Chinese listed companies. We selected annual financial report data of Xinjiang listed companies from the Shanghai, Shenzhen, and Beijing Stock Exchanges during the period 2020–2024 from CNINFO (China Securities Information Co., Ltd., Beijing, China). These data include nine observed variables ( y 1 to y 9 ). After data preprocessing, we obtained continuous 5-year observation records of 61 companies, with an overall data missing rate of 8.6%.
Different from the simulation study, in which it was necessary to preset four types of distributions for the error term (such as the normal distribution and skewed log-normal distribution) and simulate the generation of missing data, in practical application, the real data itself already contains inherent distribution characteristics and has natural missing situations. Therefore, there is no need to additionally preset the distribution type of the error term or simulate the generation of missing data. It is only required to generate the missing indicator matrix r based on the preprocessed real data, then substitute the real data into the model, and conduct Bayesian inference using the Asymmetric Laplace Distribution (ALD) as the working distribution (for the specific process, refer to the Section 2 “Posterior Sampling Based on { Y o , r } to Obtain { Ω , Y m , θ , ϕ } ” in Section 3).
We consider an NQSEM with parameters n = 305 , p = 9 , q = 3 , q 1 = 1 , q 2 = 2 , q 3 = 5 , r 1 = r 2 = 0 . Its measurement equation is as follows:
y i = Λ ω i + ε i ,
where ω i = ( η i , ξ i 1 , ξ i 2 , ξ i 1 ξ i 2 , ξ i 1 2 , ξ i 2 2 ) T . The non-overlapping factor loading matrix Λ has the following form:
Λ T = 1 λ 21 λ 31 0 0 0 0 0 0 0 0 0 1 λ 52 λ 62 λ 72 0 0 0 0 0 0 0 0 0 1 λ 93 ,
The measurement equation expresses the explicit relationship between each latent variable and its corresponding indicators. Based on this, we propose the following structural equation to evaluate the impact of potential determinants on growth:
η i = γ 1 τ ξ i 1 + γ 2 τ ξ i 2 + γ 3 τ ξ i 1 ξ i 2 + γ 4 τ ξ i 1 2 + γ 5 τ ξ i 2 2 + δ i ,
Inspired by the existing literature such as Zhi Zheng [22], we set “profitability ( ξ 1 )” and “solvency ( ξ 2 )” as latent variables, and selected five quantiles (0.1, 0.3, 0.5, 0.7, 0.9) to more comprehensively characterize the impact of various factors on development capability under different quantiles. Neither of the above equations considers covariates c i and d i . The observation indicators of the two types of latent variables are calculated according to the formulas listed in Table 2, and the specific path relationships are shown in Figure 5.
The hyperparameter values of the model are set as follows: The elements to be estimated in Λ 0 y k are set to 0.4, and the covariance matrix H 0 y k is a diagonal matrix with diagonal elements of 10 2 ; Λ 0 ω = ( 0.5 , 0.5 , 0.5 , 0.5 , 0.5 ) , and the covariance matrix H 0 ω is a diagonal matrix with diagonal elements of 10 4 ; α 0 y k = α 0 σ = 9 , β 0 y k = β 0 σ = 4 ; ρ 0 = 1 and R 0 = I 2 ; φ 0 = ( 1 , 0.3 , . . . , 0.3 ) ; and V = I 10 .
To examine convergence, we calculated the EPSR values for all parameters based on two parallel Markov chains. The results indicate that the algorithm converges within 5000 iterations. Therefore, we generated a total of 7000 iterations, discarded the results of the first 5000 iterations, and performed Bayesian estimation of the parameters based on the last 2000 iterations. The specific estimation results are shown in Table 3.
First, we set the quantile τ = 0.5 to obtain the estimation results of the factor loading matrix Λ . Since factor loadings may differ across quantiles, which would lead to inconsistent measurement scales of latent variables across quantiles, we fixed the factor loadings across all quantiles to the above-estimated values Λ ^ . Subsequent analyses under other quantile scenarios were conducted based on this setting to enhance the robustness of the model.
As shown in Table 3, these factors all affect corporate growth, and the magnitudes of the coefficients vary across quantiles, indicating that their impacts on the growth of Xinjiang-listed companies change with quantiles.
  • Profitability has a significant positive impact on corporate growth, which is consistent with the findings of Wang and He [23] and Kou [24]. Quantile regression analysis reveals that the profitability coefficient γ 1 τ exhibits a clear monotonically increasing characteristic: for low-growth companies, the marginal contribution of profitability is relatively limited because their growth is mainly constrained by factors such as market environment or management efficiency; for high-growth companies, the driving effect of profitability is significantly enhanced, which is consistent with the research conclusion of Du [25]. Specifically, as a company’s weighted return on equity, operating profit margin, net profit margin, and gross profit margin improve, its comprehensive growth gradually strengthens, with high-growth companies significantly outperforming low-growth ones.
  • Solvency also has a positive effect on corporate growth—the stronger a company’s solvency, the higher its growth. This is consistent with the findings of Shen and Wu [26], but its impact is slightly weaker than that of profitability, which aligns with the conclusion drawn by Xu and Guo [27]. The coefficient γ 2 τ shows an overall downward trend and stabilizes in the high quantile range: solvency has a relatively stronger impact on low-growth companies; its impact on high-growth companies is relatively weaker. This may be because low-growth companies often face issues such as tight capital chains and financing constraints, and strong solvency (e.g., low debt ratio, sufficient cash flow) can effectively reduce financial risks, provide a foundation for their basic growth, and serve as an important support to break through growth bottlenecks. In contrast, the growth of high-growth companies relies more on endogenous funds generated by profits or efficient external financing to support expansion. Although solvency still provides necessary financial stability for growth, its marginal driving effect on growth decreases.
  • The impact coefficient of the interaction term between profitability and solvency on growth ( γ 3 τ ) is significantly positive at all quantiles and shows an overall upward trend, indicating that their synergistic effect positively drives growth, with such effect strengthening as corporate growth improves. Specifically, the coefficient of the interaction term for low-growth companies, though positive, is relatively small, suggesting that the synergistic pull on growth is weak in this case. This may be because the core constraints on the growth of such companies lie in unresolved fundamental issues such as insufficient business expansion capabilities and market competitiveness, rather than inadequate coordination between profitability and solvency, making it difficult for synergistic effects to take hold. For high-growth companies, the coefficient of the interaction term increases significantly, and the synergistic effect is notably enhanced. This is because high-growth companies not only need profitability to support endogenous expansion but also require strong solvency to ensure smooth external financing, forming a positive cycle of “profit laying the foundation for expansion + solvency alleviating financing constraints” and thus providing stronger impetus for high growth.
  • The impact coefficient of the squared term of profitability on growth ( γ 4 τ ) gradually shifts from negative to positive with increasing values: for low-growth companies, the negative coefficient indicates that increased profitability may inhibit growth, i.e., an “excessive pursuit of profits” can have adverse effects. This is likely because such companies have limited profit scales, with profits mostly used to cover operational gaps; an excessive focus on profitability would squeeze investments in growth-oriented activities such as research and development and market development, leading to a diminishing marginal positive effect of profitability on growth. For high-growth companies, the positive and increasing coefficient reflects an increasing marginal positive effect of profitability on growth. At this stage, companies have a solid profit foundation, and profit growth can be more efficiently converted into growth momentum by supporting large-scale expansion, enhancing risk resistance, and improving financing advantages.
  • The impact coefficient of the squared term of solvency on growth ( γ 5 τ ) changes from a slightly negative value to a significantly positive one with a monotonic increasing trend: for low-growth companies, the coefficient of the squared term is close to zero, indicating that marginal changes in solvency have little differential impact on growth. This may be because the growth bottleneck for these companies lies in business fundamentals (e.g., insufficient market demand and low efficiency), and improved solvency can only maintain financial stability without generating additional impetus for growth. For high-growth companies, the significantly positive and increasing coefficient of the squared term indicates an increasing marginal positive effect of improved solvency. This is probably because their expansion requires substantial capital, and enhanced solvency can reduce financing costs, broaden financing channels, and boost investor confidence, supporting aggressive growth strategies. This forms a cycle of “strengthened solvency → expanded financing advantages → accelerated growth,” where the leverage effect of marginal improvements on growth increases with the enhancement of solvency.
In summary, this study identifies profitability, solvency, their squared terms, and their interaction term as key factors influencing corporate growth, and quantifies their specific impact intensities at different quantile levels. This provides more detailed empirical evidence for unraveling the driving logic behind enterprise growth.

6. Discussion

In this paper, we constructed a nonlinear quantile structural equation model with non-ignorable missing data. This model can not only comprehensively analyze the latent variable relationships between corporate financial variables and growth but also account for the non-ignorable missing mechanism through a linear logistic regression model. Meanwhile, we propose a Bayesian method based on ALD theory to carry out posterior inference. Simulation studies show that the model performs well in both computational efficiency and parameter estimation accuracy. After applying it to the data analysis of growth and its determinants of listed companies in Xinjiang, the results not only confirm previous research conclusions on the impact of factors such as profitability and solvency on growth but also provide new insights into the driving mechanisms of corporate growth.
Future research can proceed in the following four directions: First, in nonlinear quantile structural equation models, variable selection is also of great significance, especially when latent variables in the model exhibit significant impacts only at certain quantiles but not at others. Regularization methods show obvious advantages in this model framework because they can simultaneously realize parameter estimation and variable selection. A feasible research idea is to introduce the method proposed by Feng et al. [3] into nonlinear quantile structural equation models and conduct further research on its adaptability and extensibility under nonlinear settings. Second, the current model only incorporates continuous indicators in the measurement equations, while real-world data often include various types such as ordinal data (e.g., satisfaction scores, rating assessments), count data (e.g., number of event occurrences, quantity statistics), and unordered categorical data (e.g., attribute classification, group division). Therefore, we can further construct a generalized nonlinear quantile structural equation model compatible with multiple data types. By extending the distributional assumptions of measurement equations (e.g., introducing discrete response models within the quantile regression framework), the model can handle continuous, ordinal, count, and unordered categorical indicators simultaneously, thus better fitting the complexity of actual data and enhancing its applicability and explanatory power. Furthermore, regarding missing data, we can explore the Bayesian analysis of nonlinear quantile structural equation models under scenarios where covariates have non-ignorable missingness, or where both response variables and covariates have non-ignorable missingness. Finally, similar to Dang and Maestrini’s [28] exploration of variational approximate inference in conventional structural equation models (CSEMs), introducing variational methods into nonlinear quantile structural equation models with non-ignorable missing data is expected to significantly improve computational efficiency while providing reliable inference results, offering a more efficient solution for the practical application of such complex models.

Author Contributions

Conceptualization, L.Z.; methodology, M.T.; software, L.Z.; validation, M.T.; formal analysis, L.Z.; investigation, M.T.; resources, M.T.; data curation, L.Z.; writing—original draft preparation, L.Z.; writing—review and editing, M.T.; visualization, L.Z.; supervision, M.T.; project administration, M.T.; funding acquisition, M.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by 2025 Xinjiang University Faculty Excellence Program under the Autonomous Region’s Double First-Class Initiative (grant no. 51172500101).

Data Availability Statement

These data were derived from the following resources available in the public domain: [annual financial report data of Xinjiang 399 listed companies from the Shanghai, Shenzhen, and Beijing Stock Exchanges during the 400 period 2020–2024 from CNINFO] [https://www.cninfo.com.cn (accessed on 18 June 2025)].

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

When performing Metropolis–Hastings sampling, the required proposal distributions and their variances are as follows, with reference to the methods proposed by Lee and Song [13] and Lee and Tang [7].
First, samples ω i are drawn from p ω i y i , r i , θ , φ . Specifically, N 0 , σ ω 2 Ω ω i is chosen as its proposal distribution, where Ω ω i = Σ ω i 1 + Λ T Ψ ε i 1 Λ 1 , and Σ ω i 1 is given by
Σ ω i 1 = Ψ δ i 1 Ψ δ i 1 Γ τ Δ H Δ H T Γ τ T Ψ δ i 1 Φ 1 + Δ H T Γ τ T Ψ δ i 1 Γ τ Δ H ,
with Δ H = H ( ξ i ) / ξ i T | ξ i = 0 . The implementation of the MH algorithm is as follows:
In the ( t + 1 ) -th iteration, based on the current ω i ( t ) , a new candidate value ω i is generated from the proposal distribution N ω i ( t ) , σ ω 2 Ω ω i . The probability of accepting this new candidate is
min 1 , p ω i y i , r i , θ , φ p ω i ( j ) y i , r i , θ , φ ,
where the variance σ ω 2 is adjusted to ensure the mean acceptance rate reaches roughly 0.25 or greater (Gelman et al. [29]).
Second, samples y m i are drawn from p y m i y o i , ω i , r i , θ , φ . N 0 , σ y 2 Ω y m i is selected as its proposal distribution, with variance Ω y m i = Ψ m ε i 1 + φ m i p exp φ 0 + l D ¯ φ l y i l 1 + exp φ 0 + l D ¯ φ l y i l 2 φ m i T 1 , where Ψ m ε i is a submatrix of Ψ ε i , and φ m i is a subvector of φ i , both containing rows and columns corresponding to y m i . D ¯ { t 1 , , t k } is the set of indices corresponding to the observed data y o i . The selection criteria for σ y 2 and the sampling process similarly apply.
Finally, samples φ are drawn from p φ Y , Ω , r , θ . N 0 , σ φ 2 Ω φ is chosen as its proposal distribution, with variance Ω φ = p 4 i = 1 n F i * F i * T + V 1 1 . The selection criteria for σ φ 2 and the sampling process similarly apply.
The selection of proposal distributions for the above three parameters and the calculation of their covariances both refer to the article by Lee and Tang [7], and the average acceptance rates of the three parameters are all maintained between 0.25 and 0.4 by adjusting σ ω 2 , σ y 2 , and σ φ 2 .
The Gibbs sampling process and the posterior distribution p θ Y , Ω of the unknown parameter θ are presented below, where the notation used follows that in Song and Lee [13].
Let Ω 2 = ξ 1 , , ξ n , Y = y 1 , , y n , C = c 1 , , c n , D = d 1 , , d n , σ y = σ y 1 ,   ,   σ y p T , e η = e η 1 ,   ,   e η n T , e y i = e y i 1 ,   ,   e y i p T , e y = e y 1 , , e y n , and v i = d i T , H ( ξ i ) T T , and Ψ ε i k is the k-th diagonal element of the matrix Ψ ε i .
Given Λ y = A , Λ , let λ y k j denote the corresponding element of the matrix, i.e., Λ y = λ y k j , where j = 1 , , r 1 + q and k = 1 , , p . The positions of fixed elements in Λ y are identified by the index matrix L y = l y k j , whose elements are defined as follows:
l y k j = 0 , if λ y k j is fixed , 1 , if λ y k j is free .
Let u i = c i T , ω i T T , U = u 1 , , u n , and U k be a submatrix of U where U k = u k 1 * , , u k n * . Rows corresponding to l y k j = 0 are set to zero vectors, while rows with l y k j = 1 are retained. Meanwhile, let Y k * = y 1 k * ,   ,   y n k * T , where y i k * = y i k j = 1 r 1 + q λ y k j u i j 1 l y k j , and u i j is the j-th element of u i . The Gibbs sampling process is as follows:
1. Updating σ y k :
p σ y k 1 Y , U , Λ y k p σ y k 1 p Y σ y k 1 , Λ y k , U σ y k 1 α 0 y k 1 exp β 0 y k σ y k 1 · i = 1 n 0 . 5 2 σ y k exp σ y k 1 ρ τ y i k Λ y k u i = σ y k 1 α 0 y k 1 exp β 0 y k σ y k 1 · i = 1 n σ y k 1 4 exp 1 2 σ y k 1 y i k Λ y k u i σ y k 1 α 0 y k + n 1 exp β 0 y k + 1 2 i = 1 n y i k Λ y k u i σ y k 1 ,
For k = 1 , , p , it holds that
p σ y k 1 Y , U , Λ y k Gamma α 0 y k + n , β 0 y k + 1 2 i = 1 n y i k Λ y k u i .
2. Updating e y i k :
p e y i k y i k , u i , Λ y k , σ y k p e y i k p y i k Λ y k , u i , σ y k , e y i k σ y k 1 exp σ y k 1 e y i k · 1 2 π Ψ ε i k 1 2 exp 1 2 Ψ ε i k y i k Λ y k u i 2 ,
Let e y i k * = e y i k 1 . Then,
p e y i k * y i k , u i , Λ y k , σ y k σ y k 1 exp σ y k 1 e y i k * 1 · 1 2 π · 8 1 2 σ y k 1 2 e y i k * 1 2 exp 1 2 Ψ ε i k y i k Λ y k u i 2 e y i k * 2 σ y k 1 2 e y i k * 3 2 · 1 4 π exp y i k Λ y k u i 2 e y i k * 16 σ y k σ y k 1 e y i k * 1 2 σ y k 1 2 π e y i k * 3 exp y i k Λ y k u i 2 e y i k * 16 σ y k σ y k 1 e y i k * 1 ,
For i = 1 , , n and k = 1 , , p , it holds that
p e y i k 1 y i k , u i , Λ y k , σ y k Inverse Gaussian e y i k 1 ; μ , λ ,
where μ = 4 y i k Λ y k u i 1 and λ = 2 σ y k 1 .
3. Updating Λ y :
p Λ y k Y , e y i k , σ y k p Λ y k p Y Λ y k , e y i k , σ y k ,
To derive p Λ y k Y , e y i k , σ y k , we can refer to the method proposed by Lindley and Smith [30] for deriving posterior distributions. Specifically, let y N ( μ , D ) denote that the column vector y follows a multivariate normal distribution with mean vector μ and positive semi-definite variance matrix D . Given θ 1 ,
y N A 1 θ 1 , C 1 ,
and given θ 2 ,
θ 1 N A 2 θ 2 , C 2 ,
then the distribution of θ 1 given y is N B b , B , where B 1 = A 1 T C 1 1 A 1 + C 2 1 and b = A 1 T C 1 1 y + C 2 1 A 2 θ 2 . Similarly, we can derive the posterior distribution of Λ y k . Specifically, given Λ y k , e y i k , and σ y k , we have
y i Λ y k , e y i k , σ y k N p Λ y u i , Ψ ε i ,
Λ y k N r 1 + q Λ 0 y k , H 0 y k ,
Thus, for k = 1 , , p , it holds that
p Λ y k Y , e y i k , σ y k N r 1 + q B Λ k b Λ k , B Λ k ,
where A 1 θ 1 = Λ y u i , C 1 = Ψ ε i , A 2 θ 2 = Λ 0 y k , and C 2 = H 0 y k ; the variance of the multivariate normal distribution is B Λ k = i = 1 n u k i * u k i * T 8 σ y k e y i k + H 0 y k 1 1 , and the mean B Λ k b Λ k = B Λ k i = 1 n y i k * u k i * 8 σ y k e y i k + H 0 y k 1 Λ 0 y k .
4. Updating Φ :
p Φ Ω 2 P Φ i = 1 n P ξ i Φ Φ 1 ρ 0 q 2 1 2 exp 1 2 tr R 0 1 Φ 1 · i = 1 n Φ 1 2 exp 1 2 ξ i T Φ 1 ξ i = Φ ( ρ 0 + n ) + q 2 + 1 2 · exp 1 2 tr R 0 1 Φ 1 + i = 1 n ξ i T Φ 1 ξ i = Φ ( ρ 0 + n ) + q 2 + 1 2 · exp 1 2 tr R 0 1 Φ 1 + tr Ω 2 T Φ 1 Ω 2 = Φ ( ρ 0 + n ) + q 2 + 1 2 · exp 1 2 tr Φ 1 R 0 1 + Ω 2 Ω 2 T ,
Thus,
p Φ Ω 2 Inverse Wishart R 0 1 + Ω 2 Ω 2 T , ρ 0 + n .
5. Updating σ η :
p σ η 1 η , Λ ω τ , V p σ η 1 i = 1 n p η i σ η 1 , Λ ω τ , V σ η 1 α 0 σ 1 exp β 0 σ σ η 1 · i = 1 n τ 1 τ σ η exp σ η 1 ρ τ η i Λ ω τ v i σ η 1 α 0 σ + n 1 exp β 0 σ + i = 1 n ρ τ η i Λ ω τ v i σ η 1 ,
Thus,
p σ η 1 η , Λ ω τ , V Gamma α 0 σ + n , β 0 σ + i = 1 n ρ τ η i Λ ω τ v i .
6. Updating e η i :
p e η i η i , v i , Λ ω τ , σ η p e η i p η i v i , Λ ω τ , σ η , e η i σ η 1 e σ η 1 e η i · 1 2 π k 2 ( τ ) σ η e η i 1 2 exp η i Λ ω τ v i k 1 ( τ ) e η i 2 2 k 2 ( τ ) σ η e η i ,
Let e η i * = e η i 1 . Then,
p e η i * η i , v i , Λ ω τ , σ η σ η 1 exp σ η 1 e η i * 1 · 1 2 π k 2 ( τ ) σ η e η i * 1 1 2 exp η i Λ ω τ v i k 1 ( τ ) e η i * 1 2 e η i * 2 k 2 ( τ ) σ η e η i * 2 k 1 ( τ ) 2 + 2 k 2 ( τ ) 2 π k 2 ( τ ) σ η e η i * 3 exp 1 2 η i Λ ω τ v i 2 e η i * 2 η i Λ ω τ v i k 1 ( τ ) + k 1 ( τ ) 2 e η i * 1 + 2 k 2 ( τ ) e η i * 1 k 2 ( τ ) σ η ,
For i = 1 , , n , it holds that
p e η i 1 η i , v i , Λ ω τ , σ η Inverse Gaussian k 1 ( τ ) 2 + 2 k 2 ( τ ) η i Λ ω τ v i , k 1 ( τ ) 2 + 2 k 2 ( τ ) k 2 ( τ ) σ η .
7. Updating Λ ω τ :
p Λ ω τ Ω , e η , σ η p Λ ω τ i = 1 n p η i ξ i , e η i , σ η , Λ ω τ ,
Similar to the posterior derivation of Λ y k , given ξ i , e η i , and σ η , we have
η i k 1 ( τ ) e η i Λ ω τ , e η i , σ η N Λ ω τ v i , Ψ δ i ,
Λ ω τ N r 2 + q 3 Λ 0 ω , H 0 ω ,
Thus, we can conclude that
p Λ ω τ Ω , e η , σ η N r 2 + q 3 B ω τ b ω τ , B ω τ ,
where A 1 θ 1 = Λ ω τ v i , C 1 = Ψ δ i , A 2 θ 2 = Λ 0 ω , and C 2 = H 0 ω ; the variance of the multivariate normal distribution is B ω τ = i = 1 n v i v i T k 2 ( τ ) σ η e η i + H 0 ω 1 1 , and the mean B ω τ b ω τ = B ω τ i = 1 n η i k 1 ( τ ) e η i v i k 2 ( τ ) σ η e η i + H 0 ω 1 Λ 0 ω .
Table A1. True values of parameter vector φ under different missing rates and cases ( n = 100 ).
Table A1. True values of parameter vector φ under different missing rates and cases ( n = 100 ).
Missing RateCase φ 0 φ 1 φ 2 φ 3 φ 4 φ 5 φ 6 φ 7 φ 8 φ 9
20%1−4.000.280.280.280.280.280.280.280.280.28
3−3.800.190.190.190.190.190.190.190.190.19
4−4.000.150.150.150.150.150.150.150.150.15
30%1−4.000.400.400.400.400.400.400.400.400.40
3−4.000.270.270.270.270.270.270.270.270.27
4−4.000.200.200.200.200.200.200.200.200.20
40%1−4.000.600.600.600.600.600.600.600.600.60
3−4.000.360.360.360.360.360.360.360.360.36
4−4.000.250.250.250.250.250.250.250.250.25
Table A2. True values of parameter vector φ under different missing rates and cases ( n = 300 ).
Table A2. True values of parameter vector φ under different missing rates and cases ( n = 300 ).
Missing RateCase φ 0 φ 1 φ 2 φ 3 φ 4 φ 5 φ 6 φ 7 φ 8 φ 9
20%1−4.000.280.280.280.280.280.280.280.280.28
3−3.800.180.180.180.180.180.180.180.180.18
4−4.000.150.150.150.150.150.150.150.150.15
30%1−4.000.400.400.400.400.400.400.400.400.40
3−3.800.250.250.250.250.250.250.250.250.25
4−4.000.200.200.200.200.200.200.200.200.20
40%1−4.000.650.650.650.650.650.650.650.650.65
3−3.800.340.340.340.340.340.340.340.340.34
4−4.000.250.250.250.250.250.250.250.250.25
Table A3. Bayesian estimates of regression coefficients in the structural equation with complete data.
Table A3. Bayesian estimates of regression coefficients in the structural equation with complete data.
N ( 0 , 0.3 ) ln N ( 0 , 0.25 ) Beta ( 0.5 , 0.5 ) 0.3 χ 2 ( 3 )
Par τ Bias RMS Bias RMS Bias RMS Bias RMS
b 1 τ 0.10.00900.01450.00110.00300.00460.0076−0.00060.0033
0.50.00020.00930.00000.00230.00100.00430.00310.0042
0.9−0.00480.01130.00020.0037−0.00350.00680.00900.0101
γ 1 τ 0.10.00330.0082−0.01910.0193−0.03450.0348−0.02170.0219
0.5−0.00140.0070−0.00340.0041−0.02030.0206−0.00870.0090
0.9−0.00080.00820.02220.0228−0.00850.01080.01040.0121
γ 2 τ 0.10.01650.0182−0.01930.0195−0.03750.0378−0.02420.0244
0.50.00120.0071−0.00350.0041−0.02260.0229−0.01050.0108
0.9−0.00720.01100.02190.0222−0.01600.01730.00520.0087
γ 3 τ 0.10.00170.0058−0.02610.0263−0.02380.0241−0.02110.0213
0.5−0.00040.0053−0.01860.0188−0.02530.0255−0.01650.0166
0.9−0.00100.0059−0.00270.0062−0.03160.0323−0.00630.0080
γ 4 τ 0.1−0.04010.0410−0.02780.0282−0.02900.0297−0.02470.0251
0.50.00360.0088−0.00670.00750.00190.0058−0.00500.0060
0.90.05020.05090.03780.03890.07820.07880.04860.0496
γ 5 τ 0.1−0.05120.0519−0.02950.0299−0.03600.0366−0.02520.0255
0.5−0.00270.0080−0.00860.0092−0.00570.0080−0.00450.0063
0.90.04840.04930.03170.03250.05960.06060.04620.0477
Note: Sample size n = 100 .
Table A4. Bayesian estimates of regression coefficients in the structural equation with complete data.
Table A4. Bayesian estimates of regression coefficients in the structural equation with complete data.
N ( 0 , 0.3 ) ln N ( 0 , 0.25 ) Beta ( 0.5 , 0.5 ) 0.3 χ 2 ( 3 )
Par τ Bias RMS Bias RMS Bias RMS Bias RMS
b 1 τ 0.10.01310.01960.00900.00990.02450.02610.00500.0076
0.50.00210.01280.00760.00820.01910.01990.00800.0090
0.90.00560.01730.01260.01450.01620.01860.01010.0147
γ 1 τ 0.10.04900.0510−0.04710.0473−0.03060.03140.02290.0237
0.50.00600.0144−0.01140.0119−0.06430.0645−0.02970.0300
0.9−0.02110.02630.04700.04890.05720.06200.00740.0176
γ 2 τ 0.10.03680.0396−0.04860.0489−0.03600.03680.02180.0225
0.50.00040.0131−0.01430.0147−0.07170.0719−0.02210.0224
0.9−0.01920.02460.03970.04180.03180.04500.02280.0278
γ 3 τ 0.10.00580.01410.03500.03540.01660.01820.03000.0304
0.50.00130.0111−0.05090.0511−0.05910.0593−0.05390.0541
0.9−0.00160.01460.00240.0122−0.02800.0311−0.03540.0389
γ 4 τ 0.1−0.02970.03320.03010.0312−0.01650.01960.02660.0280
0.50.00260.0105−0.02580.0265−0.01710.0190−0.01000.0116
0.90.04810.0501−0.00270.0208−0.04340.05280.05820.0630
γ 5 τ 0.1−0.03630.03900.03470.03570.00110.01090.01660.0188
0.50.00160.0114−0.02170.0224−0.00070.0079−0.00810.0099
0.90.04470.04680.00830.0191−0.00820.04590.05490.0605
Note: Sample size n = 300 .
Table A5. Bayesian estimates of other parameters with complete data when ε i k , δ i Beta ( 0.5 , 0.5 ) .
Table A5. Bayesian estimates of other parameters with complete data when ε i k , δ i Beta ( 0.5 , 0.5 ) .
τ = 0.1 τ = 0.5 τ = 0.9
Par Bias RMS Bias RMS Bias RMS
a 1 0.00280.00370.00280.00370.00290.0037
a 2 0.00130.0119−0.00070.0112−0.00240.0103
a 3 0.00000.0120−0.00200.0111−0.00380.0111
a 4 0.02850.02950.02850.02940.02810.0290
a 5 0.01580.02070.01560.02100.01420.0205
a 6 0.01690.02190.01690.02190.01510.0211
a 7 −0.01930.0211−0.01910.0209−0.01910.0209
a 8 −0.00630.0159−0.00550.0156−0.00100.0161
a 9 −0.00280.0140−0.00220.01420.00210.0145
λ 21 −0.05960.0609−0.04450.04580.03930.0448
λ 31 −0.06030.0616−0.04520.04610.03840.0438
λ 52 0.05710.05840.01790.02240.01920.0254
λ 62 0.05940.06080.01920.02370.02100.0268
λ 83 0.03690.03950.00220.0140−0.00340.0173
λ 93 0.03860.04100.00300.0145−0.00300.0173
ϕ 11 −0.03840.0496−0.01130.02030.03960.0603
ϕ 12 0.07310.07510.05120.05270.02810.0451
ϕ 22 0.01400.0299−0.03500.03970.01470.0388
Note: Sample size n = 300 .
Table A6. Bayesian estimates of other parameters with missing data when ε i k , δ i N ( 0 , 0.3 ) .
Table A6. Bayesian estimates of other parameters with missing data when ε i k , δ i N ( 0 , 0.3 ) .
τ = 0.1 τ = 0.5 τ = 0.9
Par Bias RMS Bias RMS Bias RMS
a 1 0.00140.00280.00110.00230.00100.0025
a 2 −0.00470.0121−0.00170.01070.00150.0106
a 3 −0.00100.01080.00250.01090.00380.0112
a 4 −0.01060.0119−0.01100.0124−0.01020.0113
a 5 −0.00420.0114−0.00340.0116−0.00160.0115
a 6 −0.00610.0115−0.00520.0113−0.00540.0111
a 7 −0.00550.0077−0.00440.0072−0.00480.0074
a 8 −0.00490.0119−0.00310.0111−0.00090.0106
a 9 −0.00390.0119−0.00320.0125−0.00080.0116
λ 21 −0.02480.0268−0.00310.01090.00450.0117
λ 31 −0.02620.0287−0.00470.01210.00340.0119
λ 52 0.01550.01800.00000.0104−0.01440.0179
λ 62 0.01720.0200−0.00200.0098−0.01280.0167
λ 83 0.01460.0174−0.00060.0103−0.01410.0182
λ 93 0.01150.0142−0.00200.0101−0.01860.0215
ϕ 11 0.01900.05870.07670.0947−0.00390.0591
ϕ 12 −0.03140.0445−0.04180.0563−0.03850.0546
ϕ 22 −0.04220.0610−0.02210.0561−0.04560.0655
Note: Sample size n = 100 ; missing rate = 30%.
Table A7. Bayesian estimates of regression coefficients in the structural equation with missing data.
Table A7. Bayesian estimates of regression coefficients in the structural equation with missing data.
M1M2M3
N ( 0 , 0 . 3 )
Par τ Bias RMS Bias RMS Bias RMS
b 1 τ 0.10.00980.01760.00990.01560.00950.0165
0.5−0.00230.0128−0.00280.0122−0.00350.0105
0.9−0.00070.0151−0.00610.0154−0.00640.0158
γ 1 τ 0.10.04880.05070.04690.04880.04560.0472
0.50.00630.01280.00460.01190.00470.0117
0.9−0.01300.0189−0.02300.0274−0.02150.0250
γ 2 τ 0.10.03890.04110.03640.03870.03570.0378
0.50.00090.01030.00120.01070.00110.0091
0.9−0.01210.0161−0.02030.0241−0.02040.0239
γ 3 τ 0.10.00720.01350.00670.01210.00480.0101
0.50.00200.01000.00220.00840.00200.0092
0.90.00390.0129-0.00260.0116−0.00220.0109
γ 4 τ 0.1−0.01970.0241−0.00830.0158−0.00470.0151
0.50.00320.01150.00420.01170.00360.0111
0.90.09310.09390.03200.03520.02900.0318
γ 5 τ 0.1−0.01550.0213−0.01320.0199−0.00740.0155
0.50.00140.01210.00050.01060.00050.0105
0.90.09070.09170.02540.02860.02260.0263
Table A8. Bayesian estimates of regression coefficients in the structural equation with missing data.
Table A8. Bayesian estimates of regression coefficients in the structural equation with missing data.
M1M2M3
Beta ( 0.5 , 0.5 )
Par τ BiasRMSBiasRMSBiasRMS
b 1 τ 0.10.01700.01900.01600.01790.01370.0157
0.50.01510.01620.01330.01460.01360.0146
0.90.01000.01450.00750.01290.00690.0137
γ 1 τ 0.1−0.01110.0132−0.00960.0134−0.00440.0109
0.5−0.06050.0607−0.05950.05970.03980.0402
0.90.05120.05370.05550.05850.05840.0608
γ 2 τ 0.1−0.01540.0178−0.01090.0146−0.00600.0119
0.5−0.06690.0671−0.06460.06500.03600.0364
0.90.02810.03280.04040.04390.05010.0525
γ 3 τ 0.10.04010.04070.04100.04170.04370.0448
0.5−0.04880.0490−0.04670.0470−0.04480.0453
0.9−0.03050.0336−0.02740.0307−0.02360.0275
γ 4 τ 0.10.01240.01730.01380.01930.01870.0243
0.5−0.00510.0099−0.00090.0089−0.00480.0112
0.9−0.02360.0343−0.03540.0427−0.04220.0509
γ 5 τ 0.10.02780.03020.03240.03450.03290.0353
0.50.01230.01500.01160.01480.00660.0122
0.90.00380.0291−0.01480.0298−0.03220.0402
0.3 χ 2 ( 3 )
Par τ BiasRMSBiasRMSBiasRMS
b 1 τ 0.10.00310.00610.00230.00620.00110.0052
0.50.00640.00760.00610.00710.00460.0059
0.90.01260.01550.00990.01300.00930.0135
γ 1 τ 0.10.03910.03960.04160.04210.04570.0462
0.5−0.02130.0218−0.02110.0216−0.01900.0194
0.90.02360.02800.02330.02630.02450.0270
γ 2 τ 0.10.03910.03950.04110.04160.04480.0452
0.5−0.01560.0161−0.01640.0169−0.01550.0162
0.90.03720.03930.03330.03500.03590.0383
γ 3 τ 0.10.05100.05120.05210.05240.05550.0558
0.5−0.04230.0425−0.04000.0403−0.03770.0381
0.9−0.01540.0205−0.00640.0131−0.00050.0124
γ 4 τ 0.10.05200.05270.05240.05330.05900.0601
0.50.00860.01100.01010.01260.01620.0178
0.90.04490.04980.02750.03400.02340.0316
γ 5 τ 0.10.04560.04640.04740.04850.05080.0517
0.50.00760.01060.00830.01150.01180.0140
0.90.04650.05100.03110.03790.02560.0321
Note: Sample size n = 300 .
Table A9. Bayesian estimates of other parameters with missing data when ε i k , δ i 0.3 χ 2 ( 3 ) .
Table A9. Bayesian estimates of other parameters with missing data when ε i k , δ i 0.3 χ 2 ( 3 ) .
τ = 0.1 τ = 0.5 τ = 0.9
Par Bias RMS Bias RMS Bias RMS
a 1 −0.00480.0051−0.00510.0054−0.00550.0058
a 2 −0.00700.0119−0.00860.0123−0.01460.0162
a 3 −0.00600.0109−0.00670.0111−0.01290.0151
a 4 −0.00040.0076−0.00070.00660.00070.0070
a 5 −0.00230.0093−0.00210.00990.00210.0104
a 6 −0.00120.0079−0.00080.00780.00200.0098
a 7 0.00420.00770.00430.00750.00500.0082
a 8 0.00150.00930.00160.01070.00510.0126
a 9 0.00150.00870.00120.00940.00460.0103
λ 21 0.02360.02870.00710.0148−0.03430.0380
λ 31 0.02240.02600.00360.0128−0.03700.0399
λ 52 −0.02850.0315−0.03100.03370.03400.0365
λ 62 −0.02850.0316−0.03050.03370.03410.0371
λ 83 −0.02080.0245−0.02580.02880.03580.0382
λ 93 −0.02140.0248−0.02320.02760.03420.0359
ϕ 11 −0.02380.0691−0.03070.0545−0.05970.0828
ϕ 12 −0.05380.0633−0.04350.04910.06820.0748
ϕ 22 0.00830.06460.01740.04570.00200.0517
Note: Sample size n = 300 ; missing rate = 40%.

References

  1. Koenker, R.; Bassett, G., Jr. Regression quantiles. Econometrica 1978, 46, 33–50. [Google Scholar] [CrossRef]
  2. Wang, Y.; Feng, X.N.; Song, X.Y. Bayesian quantile structural equation models. Struct. Equ. Model. 2016, 23, 246–258. [Google Scholar] [CrossRef]
  3. Feng, X.N.; Wang, Y.; Lu, B.; Song, X.Y. Bayesian regularized quantile structural equation models. J. Multivar. Anal. 2017, 154, 234–248. [Google Scholar] [CrossRef]
  4. Wang, Z.Q. Bayesian Statistical Inference for Quantile Regression Models. Ph.D. Thesis, Yunnan University, Kunming, China, 2019. [Google Scholar]
  5. Xue, J. Bayesian Quantile Factor Models and Their Extensions. Ph.D. Thesis, Lanzhou University of Finance and Economics, Lanzhou, China, 2023. [Google Scholar]
  6. Cheng, H. Quantile varying-coefficient structural equation model. Stat. Methods Appl. 2023, 32, 1439–1475. [Google Scholar] [CrossRef]
  7. Lee, S.Y.; Tang, N.S. Bayesian analysis of nonlinear structural equation models with nonignorable missing data. Psychometrika 2006, 71, 541–564. [Google Scholar] [CrossRef]
  8. Cai, J.H.; Lee, S.Y.; Song, X.Y. Bayesian analysis of nonlinear structural equation models with mixed continuous, ordered and unordered categorical, and nonignorable missing data. Stat. Interface 2008, 1, 99–114. [Google Scholar] [CrossRef]
  9. Lee, S.Y.; Song, X.Y. On Bayesian estimation and model comparison of an integrated structural equation model. Comput. Stat. Data Anal. 2008, 52, 4814–4827. [Google Scholar] [CrossRef]
  10. Cai, J.H.; Song, X.Y. Bayesian analysis of mixtures in structural equation models with non-ignorable missing data. Br. J. Math. Stat. Psychol. 2010, 63, 491–508. [Google Scholar] [CrossRef]
  11. Cai, J.H.; Song, X.Y.; Hser, Y.I. A Bayesian analysis of mixture structural equation models with non-ignorable missing responses and covariates. Stat. Med. 2010, 29, 1861–1874. [Google Scholar] [CrossRef] [PubMed]
  12. Lee, S.Y. Structural Equation Modeling: A Bayesian Approach; John Wiley & Sons: Hoboken, NJ, USA, 2007. [Google Scholar]
  13. Lee, S.Y.; Song, X.Y. Basic and Advanced Bayesian Structural Equation Modeling: With Applications in the Medical and Behavioral Sciences; John Wiley & Sons: Hoboken, NJ, USA, 2012. [Google Scholar]
  14. Yu, K.; Moyeed, R.A. Bayesian quantile regression. Stat. Probab. Lett. 2001, 54, 437–447. [Google Scholar] [CrossRef]
  15. Reed, C.; Yu, K. A Partially Collapsed Gibbs Sampler for Bayesian Quantile Regression; Technical report; Department of Mathematical Sciences, Brunel University: Uxbridge, UK, 2009. [Google Scholar]
  16. Kozumi, H.; Kobayashi, G. Gibbs sampling methods for Bayesian quantile regression. J. Stat. Comput. Simul. 2011, 81, 1565–1578. [Google Scholar] [CrossRef]
  17. Ibrahim, J.G.; Chen, M.H.; Lipsitz, S.R. Missing responses in generalised linear mixed models when the missing data mechanism is nonignorable. Biometrika 2001, 88, 551–564. [Google Scholar] [CrossRef]
  18. Tanner, M.A.; Wong, W.H. The calculation of posterior distributions by data augmentation. J. Am. Stat. Assoc. 1987, 82, 528–540. [Google Scholar] [CrossRef]
  19. Geyer, C.J. Practical markov chain monte carlo. Stat. Sci. 1992, 7, 473–483. [Google Scholar] [CrossRef]
  20. Gelman, A.; Meng, X.L.; Stern, H. Posterior predictive assessment of model fitness via realized discrepancies. Stat. Sin. 1996, 6, 733–760. [Google Scholar]
  21. Bayarri, M.; Berger, J.O. P values for composite null models. J. Am. Stat. Assoc. 2000, 95, 1127–1142. [Google Scholar]
  22. Zhi, Z. Research on the Growth of Chinese Agricultural Listed Companies. Master’s Thesis, Nanjing Agricultural University, Nanjing, China, 2023. [Google Scholar]
  23. Wang, Q.; He, Y. Analysis of main factors affecting the growth of Chinese listed companies. Stat. Decis. 2005, 61–63. [Google Scholar]
  24. Kou, P. Analysis of Corporate Growth Based on Nonlinear Structural Equation Modeling. Master’s Thesis, Kunming University of Science and Technology, Kunming, China, 2013. [Google Scholar]
  25. Du, L. Research on Enterprise Growth Evaluation Model and Its Application: An Empirical Test Based on 66 Listed Companies in Henan Province. Financ. Manag. Res. 2025, 97–102. [Google Scholar]
  26. Shen, H.; Wu, Q. Analysis of Factors Affecting the Growth of Small and Medium-Sized Enterprises: An Empirical Study Based on Panel Data of Small and Medium-Sized Board Listed Companies. J. Financ. Dev. Res. 2010, 66–70. [Google Scholar]
  27. Xu, Y.; Guo, H. Analysis on the Growth of Listed Companies on China’s GEM. Coop. Econ. Sci. 2016, 60–62. [Google Scholar]
  28. Dang, K.D.; Maestrini, L. Fitting structural equation models via variational approximations. Struct. Equ. Model. 2022, 29, 839–853. [Google Scholar] [CrossRef]
  29. Gelman, A.; Roberts, G.O.; Gilks, W.R. Efficient Metropolis jumping rules. Bayesian Stat. 5 1996, 5, 599–608. [Google Scholar]
  30. Lindley, D.V.; Smith, A.F. Bayes estimates for the linear model. J. R. Stat. Soc. B Stat. Methodol. 1972, 34, 1–18. [Google Scholar] [CrossRef]
Figure 1. EPSR convergence trend graph for all parameters.
Figure 1. EPSR convergence trend graph for all parameters.
Mathematics 13 03094 g001
Figure 2. Estimated residual plots: (a) ε ^ i 1 , (b) ε ^ i 2 , (c) ε ^ i 3 , and (d) δ ^ i .
Figure 2. Estimated residual plots: (a) ε ^ i 1 , (b) ε ^ i 2 , (c) ε ^ i 3 , and (d) δ ^ i .
Mathematics 13 03094 g002
Figure 3. Plots of estimated residuals ε ^ i 1 versus (a) ξ ^ i 1 , (b) ξ ^ i 2 , and (c) η ^ i .
Figure 3. Plots of estimated residuals ε ^ i 1 versus (a) ξ ^ i 1 , (b) ξ ^ i 2 , and (c) η ^ i .
Mathematics 13 03094 g003
Figure 4. Plots of estimated residuals δ ^ i versus (a) ξ ^ i 1 and (b) ξ ^ i 2 .
Figure 4. Plots of estimated residuals δ ^ i versus (a) ξ ^ i 1 and (b) ξ ^ i 2 .
Mathematics 13 03094 g004
Figure 5. Path diagram of the NQSEM model in financial data. 1 * indicates that this variable is fixed.
Figure 5. Path diagram of the NQSEM model in financial data. 1 * indicates that this variable is fixed.
Mathematics 13 03094 g005
Table 1. Bayesian estimates of regression coefficients in the structural equation with missing data.
Table 1. Bayesian estimates of regression coefficients in the structural equation with missing data.
M1M2M3
N ( 0 , 0.3 )
Par τ BiasRMSBiasRMSBiasRMS
b 1 τ 0.10.00910.01380.00850.01240.00740.0114
0.50.00070.00810.00040.0075−0.00030.0071
0.9−0.00410.0102−0.00410.0101−0.00450.0084
γ 1 τ 0.10.00520.00820.00350.00680.00430.0071
0.5−0.00020.0059−0.00140.0062−0.00120.0059
0.9−0.00110.0069−0.00210.0069−0.00200.0070
γ 2 τ 0.10.01370.01530.01300.01460.01260.0137
0.5−0.00030.0064−0.00040.00560.00090.0058
0.9−0.00840.0109−0.00800.0100−0.00650.0086
γ 3 τ 0.10.00120.00530.00260.00490.00190.0049
0.5−0.00040.0050−0.00030.00420.00020.0045
0.9−0.00160.0065−0.00170.0055−0.00010.0049
γ 4 τ 0.1−0.03670.0376−0.03470.0355−0.03160.0326
0.50.00210.00830.00250.00790.00460.0077
0.90.04410.04510.04490.04560.04440.0450
γ 5 τ 0.1−0.04320.0437−0.04140.0423−0.03820.0390
0.5−0.00190.0083−0.00180.0077−0.00060.0069
0.90.04280.04350.04210.04290.04170.0422
Beta ( 0.5 , 0.5 )
Par τ BiasRMSBiasRMSBiasRMS
b 1 τ 0.10.00190.00600.00210.00510.00180.0052
0.5−0.00080.0041−0.00080.0038−0.00080.0038
0.9−0.00480.0073−0.00410.0064−0.00390.0063
γ 1 τ 0.1−0.02870.0291−0.02770.0281−0.02700.0275
0.5−0.01890.0193−0.01780.0182−0.01700.0174
0.9−0.00760.0099−0.00580.0089−0.00510.0085
γ 2 τ 0.1−0.02980.0301−0.02930.0298−0.02820.0287
0.5−0.01890.0193−0.01870.0191−0.01850.0188
0.9−0.01080.0126−0.01020.0124−0.00780.0099
γ 3 τ 0.1−0.01790.0182−0.01840.0189−0.01720.0177
0.5−0.02080.0211−0.02070.0209−0.01990.0202
0.9−0.02400.0250−0.02200.0232−0.02120.0226
γ 4 τ 0.1−0.02080.0215−0.02090.0219−0.02090.0219
0.50.00660.00820.00650.00800.00680.0087
0.90.06620.06700.05810.05920.05360.0548
γ 5 τ 0.1−0.02540.0261−0.02530.0262−0.02520.0261
0.50.00020.00520.00040.00540.00090.0050
0.90.05350.05450.04850.04950.04530.0463
0.3 χ 2 ( 3 )
Par τ BiasRMSBiasRMSBiasRMS
b 1 τ 0.1−0.00110.0033−0.00050.0035−0.00080.0034
0.50.00280.00390.00320.00410.00290.0040
0.90.00700.00840.00650.00760.00620.0077
γ 1 τ 0.1−0.01720.0175−0.01600.0163−0.01380.0142
0.5−0.00620.0068−0.00610.0066−0.00510.0059
0.90.01030.01210.00950.01090.01040.0119
γ 2 τ 0.1−0.01880.0190−0.01780.0180−0.01610.0164
0.5−0.00780.0082−0.00760.0081−0.00600.0066
0.90.00700.00940.00750.00940.00720.0090
γ 3 τ 0.1−0.01560.0157−0.01450.0148−0.01330.0136
0.5−0.01300.0132−0.01270.0129−0.01050.0107
0.9−0.00390.0059−0.00260.0055−0.00290.0052
γ 4 τ 0.1−0.01760.0182−0.01550.0163−0.01270.0136
0.50.00160.00460.00130.00450.00350.0054
0.90.04360.04470.03810.03900.03710.0380
γ 5 τ 0.1−0.01790.0185−0.01660.0170−0.01430.0150
0.50.00040.00430.00130.00400.00380.0054
0.90.04050.04150.03600.03700.03390.0346
Note: Sample size n = 100 .
Table 2. Variables in the NQSEM for analyzing determinants of company growth.
Table 2. Variables in the NQSEM for analyzing determinants of company growth.
VariablesIndicatorsDetails
Growth ( η )Operating income growth rate ( y 1 )(Current turnover—Previous turnover)/Previous turnover
Operating profit growth rate ( y 2 )(Current operating profit—Previous operating profit)/Previous operating profit
Net profit growth rate ( y 3 )(Current net profit—Previous net profit)/Previous net profit
Profitability ( ξ i 1 )Weighted ROE ( y 4 )Net profit/Weighted average net assets
Main business profit margin ( y 5 )Main business profit/Main business income
Net profit margin ( y 6 )Net profit/Operating income
Gross profit margin ( y 7 )(Operating income—Operating cost)/Operating income
Solvency ( ξ i 2 )Quick ratio ( y 8 )(Current assets—Inventories)/Current liabilities
Asset–liability ratio ( y 9 )Total liabilities/Total assets
Table 3. Estimation results of parameters under different quantiles in the case study.
Table 3. Estimation results of parameters under different quantiles in the case study.
Parameter τ
0.1 0.3 0.5 0.7 0.9
Factor Loadings
λ 21 0.5131
λ 31 0.4750
λ 52 0.4960
λ 62 0.6187
λ 72 0.3723
λ 93 0.3738
Structural Equation Coefficients
γ 1 τ 0.39390.48160.61340.79850.9144
γ 2 τ 0.42710.35870.32980.31390.3154
γ 3 τ 0.24350.23950.28010.30620.4348
γ 4 τ −0.5631−0.21240.08350.36300.8079
γ 5 τ −0.03960.04730.30650.51370.7218
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, L.; Tuerde, M. Bayesian Analysis of Nonlinear Quantile Structural Equation Model with Possible Non-Ignorable Missingness. Mathematics 2025, 13, 3094. https://doi.org/10.3390/math13193094

AMA Style

Zhang L, Tuerde M. Bayesian Analysis of Nonlinear Quantile Structural Equation Model with Possible Non-Ignorable Missingness. Mathematics. 2025; 13(19):3094. https://doi.org/10.3390/math13193094

Chicago/Turabian Style

Zhang, Lu, and Mulati Tuerde. 2025. "Bayesian Analysis of Nonlinear Quantile Structural Equation Model with Possible Non-Ignorable Missingness" Mathematics 13, no. 19: 3094. https://doi.org/10.3390/math13193094

APA Style

Zhang, L., & Tuerde, M. (2025). Bayesian Analysis of Nonlinear Quantile Structural Equation Model with Possible Non-Ignorable Missingness. Mathematics, 13(19), 3094. https://doi.org/10.3390/math13193094

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop