Next Article in Journal
Switching Machine Improvisation Models by Latent Transfer Entropy Criteria
Previous Article in Journal
Is Quantum Tomography a Difficult Problem for Machine Learning?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Variational Bayesian Approximation (VBA): A Comparison between Three Optimization Algorithms †

by
Seyedeh Azadeh Fallah Mortezanejad
1 and
Ali Mohammad-Djafari
2,3,*
1
School of Automotive and Traffic Engineering, Jiangsu University, Zhenjiang 212013, China
2
International Science Consulting and Training (ISCT), 91440 Bures sur Yvette, France
3
Shanfeng Company, Shaoxing 312352, China
*
Author to whom correspondence should be addressed.
Presented at the 41st International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Engineering, Paris, France, 18–22 July 2022.
Phys. Sci. Forum 2022, 5(1), 48; https://doi.org/10.3390/psf2022005048
Published: 8 February 2023

Abstract

:
In many Bayesian computations, we first obtain the expression of the joint distribution of all the unknown variables given the observed data. In general, this expression is not separable in those variables. Thus, obtaining the marginals for each variable and computing the expectations is difficult and costly. This problem becomes even more difficult in high dimensional quandaries, which is an important issue in inverse problems. We may then try to propose a surrogate expression with which we can carry out approximate computations. Often, a separable expression approximation can be useful enough. The variational Bayesian approximation (VBA) is a technique that approximates the joint distribution p with an easier, for example separable, distribution q by minimizing the Kullback–Leibler divergence K L ( q | p ) . When q is separable in all the variables, the approximation is also called the mean field approximation (MFA), and so q is the product of the approximated marginals. A first standard and general algorithm is the alternate optimization of K L ( q | p ) with respect to q i . A second general approach is its optimization in the Riemannian manifold. However, in this paper, for practical reasons, we consider the case where p is in the exponential family and so is q. For this case, K L ( q | p ) becomes a function of the parameters θ of the exponential family. Then, we can use any other optimization algorithm to obtain those parameters. In this paper, we compare three optimization algorithms, namely a standard alternate optimization, a gradient-based algorithm and a natural gradient algorithm, and study their relative performances in three examples.

1. Introduction

In many applications, with direct or indirect observations, the use of the Bayesian computations starts with obtaining the expression of the joint distribution of all the unknown variables given the observed data. Then, we must use it for inference. In general, this expression is not separable in all the variables of the problem. So, the computations become hard and costly. For example, obtaining the marginals for each variable and computing the expectations are difficult and costly. This problem becomes even more crucial in high dimensional quandaries, which is an important issue in inverse problems. We may then need to propose a surrogate expression with which we can carry out approximate computations.
The variational Bayesian approximation (VBA) is a technique that approximates the joint distribution p with an easier, for example a separable one, q, by minimizing the Kullback–Leibler divergence K L ( q | p ) , which makes the marginal computations much easier. For example, in the case of two variables, p ( x , y ) is approximated by q ( x , y ) = q 1 ( x ) q 2 ( y ) via minimizing K L ( q 1 q 2 | p ) . When q is separable in all the variables of p, the approximation is also called mean field approximation (MFA).
To obtain the approximate marginals q 1 and q 2 , we have to minimize K L ( q 1 q 2 | p ) . The first standard and general algorithm is the alternate optimization of K L ( q 1 q 2 | p ) with respect to q 1 and q 2 . By finding the expression of the functional derivatives of K L ( q 1 q 2 | p ) with respect to q 1 and q 2 and then equating them to zero alternatively, we obtain an iterative optimization algorithm. A second general approach is its optimization in the Riemannian manifold. However, in this paper, for practical reasons, we consider the case where p is in the exponential family and so are q 1 and q 2 . For this case, K L ( q 1 q 2 | p ) becomes a function of the parameters θ of the exponential family. Then, we can use any other optimization algorithm to obtain those parameters.
In this paper, we compare three optimization algorithms: a standard alternate optimization (Algorithm 1), a gradient-based algorithm [1,2] (Algorithm 2) and a natural gradient algorithm [3,4,5] (Algorithm 3). The aim of this paper is to consider the first algorithm as the VBA method and compare it with the two other algorithms.
Of the main advantages of the VBA for inference problems, such as inverse problems and machine learning, we can mention the following:
-
First, VBA builds a sufficient model according to prior information and the final posterior distribution. Especially in the mean field approximation (MFA), the result ends in an explicit form for each unknown component using conjugate priors and works well for small sample sizes [6,7,8].
-
The second benefit is, for example in machine learning, that it is a robust way for classification based on the predictive posterior distribution and diminishes over-trained parameters [7].
-
The third privilege is that the target structure has uncertainty in the VBA recursive processes. This feature prevents further error propagation and increases the robustness of VBA [9].
Besides all these preponderances, the VBA has some weaknesses, such as difficulty regarding the solution of integrals and expectations in terms of obtaining a posterior distribution, and there is no evidence of finding an exact posterior [6]. Its most significant drawback arises when there are strong dependencies between unknown parameters, and the VBA ignores them. Then estimates, computed based on this approximation, may be very far from the exact values. However, it works well when the number of dependencies are low [8].
In this article, we examine three different estimating algorithms of the unknown parameters in a model concerning prior information. The first iterative algorithm is a standard alternate optimization based on VBA, which begins a certain initial points. Sometimes, the points are estimated from an available dataset, but most of the time, we do not have enough data on the parameters to make certain pre-estimations of them. To solve this obstacle, we can start the algorithm with certain desired points, and then by repeating the process, they approach the true values using the posterior distribution. The second two algorithms are gradient-based and natural gradient algorithms, whose base function is Kullback–Leibler divergence. First, the gradient of Kullback–Leibler for all unknown parameters must be found, then must start from points either estimated from data or desired choices. Then, we repeat the iterative algorithm until it converges to certain points. If we denote the unknown parameter space with θ , then the recursive formula is θ ˜ ( k + 1 ) = θ ˜ ( k ) γ K L ( θ ˜ ( k ) ) for gradient-based and natural gradient algorithms with different values of γ .
Additionally, we consider three examples, normal-inverse-gamma, multivariate normal and linear inverse problem for checking the performance and convergence speed of the algorithms.
We propose the following organization of this paper: In Section 2, we present a brief explanation of the basic analytical aspect of VBA. In Section 3, we explain our first example related to normal-inverse-gamma distribution analytically and, in practice, explain the outcomes of three algorithms to estimate the unknown parameters. In Section 5, we study a more complex example of a multivariate normal distribution whose means and variance–covariance matrix are unknown and have normal-inverse-Wishart distribution. The aim of this section is to demonstrate the marginal distributions of μ ˜ and Σ ˜ using a set of multivariate normal observations using the mean and variance. In Section 6, the example is closer to realistic situations and is a linear inverse problem. In Section 7, we present a summary of the work carried out in the article and compare the three recursive algorithms through three different examples.

2. Variational Bayesian Approach (VBA)

As we mentioned previously, VBA uses Kullback–Leibler divergence. Kullback–Leibler divergence [10] K L ( q | p ) is an information measure of discrepancy between two probability functions defined as follows. Let p ( x ) and q ( x ) be two density functions of a continuous random variable x with respect to support set S X . K L ( q | p ) function is introduced as:
K L ( q | p ) = x S X q ( x ) ln q ( x ) p ( x ) d x .
For simplicity, we assume a bivariate case of distribution p ( x , y ) and want to assess it via VBA; therefore, we have:
K L ( q | p ) = H ( q 1 ) H ( q 2 ) ln p ( x , y ) q 1 q 2 ,
where
H ( q 1 ) = x S X q 1 ( x ) ln q 1 ( x ) d x and H ( q 2 ) = y S Y q 2 ( y ) ln q 2 ( y ) d y
are, respectively, the Shannon entropies of x and of y, and
ln p ( x , y ) q 1 q 2 = ( x , y ) S X Y q 1 ( x ) q 2 ( y ) ln p ( x , y ) d x d y .
Now, differentiating the Equation (2) with respect to q 1 and then with respect to q 2 and equating them to zero, we obtain:
q 1 ( x ) exp ln p ( x , y ) q 2 ( y ) and q 2 ( y ) exp ln p ( x , y ) q 1 ( x )
These results can be easily extended to more dimensions [11]. They do not have any closed form because they depend on the expression of p ( x , y ) and that of q 1 and q 2 . An interesting case is that of exponential families and conjugate priors, where writing
p ( x , y ) = p ( x | y ) p ( y ) , and p ( y | x ) = p ( x , y ) p ( x ) = p ( x | y ) p ( y ) p ( x ) ,
we can consider p ( y ) as prior, p ( x | y ) as the likelihood, and p ( y | x ) as the posterior distributions. Then, if p ( y ) is a conjugate prior for the likelihood p ( x | y ) , then the posterior p ( y | x ) will be in the same family as the prior p ( y ) . To illustrate all these properties, we provide details of these expressions for a first simple example of normal-inverse-gamma p ( x , y ) = N ( x | μ , y ) I G ( y | α , β ) with q 1 ( x ) = N ( x | μ , v ) and q 2 = I G ( y | α , β ) . For this simple case, first we give the expression of K L ( q | p ) with q 1 ( x ) = N ( x | μ ˜ , v ˜ ) and q 2 ( y ) = I G ( y | α ˜ , β ˜ ) as a function of the parameters θ = ( μ ˜ , v ˜ , α ˜ , β ˜ ) and then the expressions of the three above-mentioned algorithms; after which, we study their convergence.

3. Normal-Inverse-Gamma Distribution Example

The purpose of this section is to explain in detail the process of performing calculations in VBA. For this we consider a simple case for which we have all the necessary expressions. The objective here is to compare the three different algorithms mentioned above. Additionally, its practical application can be explained as follows:
We have a sensor which measures a quantity X, N times x 1 , , x N . We want to model these data. In a first step, we model them as N ( x | μ , v ) with fixed μ and v. Then, it is easy to estimate the parameters ( μ , v ) either by maximal likelihood or Bayesian strategy. If we assume that the model is Gaussian with unknown variance and call this variance y and assign an I G prior to it, then we have a model N I G for p ( x , y ) . The N I G priors were applied to the wavelet context with correlated structures because they were conjugated with normal priors [12]. We chose the normal-inverse-gamma distribution because of this conjugated property and ease of handling.
We showed that the margins are S t and I G . Working directly with S t is difficult. So, we want to approximate it with a Gaussian q 1 ( x ) . This is equivalent to approximating p ( x , y ) with q 1 ( x ) q 2 ( y ) . Now, we want to find the parameters μ , v, α , and β , which minimize K L ( q 1 q 2 | p ) . This process is called VBA. Then, we want to compare three algorithms to obtain the parameters which minimize K L ( · | · ) . K L ( · | · ) is convex with respect to q 1 if q 2 is fixed and is convex with respect to q 2 if q 1 is fixed. So, we hope that the iterative algorithm converges. However, K L ( · | · ) may not be convex in the space of parameters. So, we have to study the shape of this criterion concerning the parameters v ˜ , α ˜ and β ˜ .
The practical problem considered here is the following: A sensors delivers a few samples x = { x 1 , x 2 , , x N } of a physical quantity X. We want to find p ( x ) . For this process, we assume a simple Gaussian model but with unknown variance y. Thus, the forward model can be written as p ( x , y ) = N ( x | μ , y ) I G ( y | α , β ) . In this simple example, we know that p ( x ) is Student’s t-distribution obtained by:
S ( x | μ , α , β ) = N ( x | μ , y ) I G ( y | α , β ) d y
Our objective is to find the three parameters θ = ( μ , α , β ) from the data x and an approximate marginal q ( x ) for p ( x ) .
The main idea is to find such q 1 ( x ) q 2 ( y ) as an approximation of p ( x , y ) . Here, we show the VBA, step by step. For this, we start by choosing the conjugate families q 1 ( x ) = N ( x | μ ˜ , v ˜ ) and q 2 ( y ) = I G ( y | α ˜ , β ˜ ) .
In the first step, we have to calculate ln p ( x , y )
ln p ( x , y ) = c 1 2 ln y 1 2 y ( x μ ˜ ) 2 ( α ˜ + 1 2 ) ln y β ˜ y .
where c is a constant value term independent of x and y. First of all, to use the iterative algorithm given in (3), starting by q 1 = N ( x | μ , v ) we have to find q 2 ( y ) , so we have to start by finding q 2 ( y ) . The integration of ln p ( x , y ) concerns q 1 ( x )
ln p ( x , y ) q 1 = c 1 2 y ( x μ ˜ ) 2 q 1 ( α ˜ + 1 ) ln y β ˜ y .
Since the mean of x is the same in prior and posterior distribution, μ ˜ = μ ˜ then ( x μ ˜ ) 2 q 1 = v ˜ , otherwise ( x μ ˜ ) 2 q 1 = v ˜ + ( μ ˜ μ ˜ ) 2 . Thus when μ ˜ = μ ˜
q 2 ( y ) exp [ ( α ˜ + 1 ) ln y ( v ˜ 2 + β ˜ ) 1 y ] .
Thus, the function q 2 ( y ) is equivalent to an inverse gamma distribution I G ( α ˜ , v ˜ 2 + β ˜ ) . Similarly, q 2 ( y ) is I G ( α ˜ , v ˜ + ( μ ˜ μ ˜ ) 2 2 + β ˜ ) when μ ˜ μ ˜ . We have to take integral of ln p ( x , y ) over q 2 to find q 1
ln p ( x , y ) q 2 = c ( α ˜ + 1 ) ln y q 2 ( β ˜ + 1 2 ( x μ ˜ ) 2 ) 1 y q 2
Note that the first term does not depend on x and 1 y q 2 = 2 α ˜ 2 β ˜ + v ˜ , so
q 1 ( x ) exp [ 2 α ˜ 2 β ˜ + v ˜ ( β ˜ + 1 2 ( x μ ˜ ) 2 ) ] exp [ ( x μ ˜ ) 2 2 2 β ˜ + v ˜ 2 α ˜ ] .
We see that q 1 is, again, a normal distribution but with updated parameters N ( μ ˜ , 2 β ˜ + v ˜ 2 α ˜ ) , so v ˜ = 2 β ˜ + v ˜ 2 α ˜ . Note that we obtained the conjugacy property: if p ( x | y ) = N ( x | μ , y ) and p ( y ) = I G ( y | α , β ) , then p ( y | x ) = I G ( y | α , β ) where μ , α and β are μ = μ , α = α , β = β + 2 β + v 2 α . In this case, we also know that p ( x | α , β ) = S t ( x | μ , α , β ) .
In standard alternate optimization based on VBA (Algorithm 1), there is no need for an iterative process for μ ˜ and α ˜ , which are approximated by μ ˜ = μ 0 and α ˜ = α 0 , respectively. The situation for β ˜ and v ˜ is different because there are circular dependencies among them. So, the approximation needs an iterative process, staring from μ ˜ ( 1 ) = μ 0 , v ˜ ( 1 ) = v 0 , α ( 1 ) = α 0 and β ( 1 ) = β 0 . As a conclusion for this case is that the values of α and μ do not change during the iterations, and so depend on the initial values. However, the values of β and v are interdependent and change during the iterations. This algorithm is summarized below.
Alternate optimization algorithm based on VBA (Algorithm 1):
α ˜ ( k + 1 ) = α ˜ ( k ) , β ˜ ( k + 1 ) = β ˜ ( k ) + v ˜ ( k ) 2 , μ ˜ ( k + 1 ) = μ ˜ ( k ) , v ˜ ( k + 1 ) = 2 β ˜ ( k ) + v ˜ ( k ) 2 α ˜ ( k ) .
This algorithm converges to v = ( 2 β ˜ + v ˜ ) / ( 2 α ˜ ) , which gives v ˜ = ( 2 β ˜ ) / ( 2 α ˜ 1 ) and β ˜ = 0 , so v ˜ = 0 , which is a degenrate solution.
The two other algorithms, gradient- and natural gradient-based, require to find the expression of K L ( q 1 q 2 : p ) as a function of the parameters θ = ( α , β , μ , v ) :
K L ( θ ˜ ) = 2 ln Γ ( α ˜ 1 2 ) ( 2 α ˜ + 3 2 ) ψ 0 ( α ˜ 1 2 ) + ( 2 α ˜ 1 ) ( v ˜ + 2 β ˜ ) 4 β ˜ + α ˜ + 5 2 ln β ˜ 1 2 ln v ˜ 1 .
We also need the gradient expression of K L ( θ ˜ ) for θ ˜ :
K L ( θ ˜ ) = v ˜ 2 β ˜ ( 2 α ˜ + 3 2 ) ψ 1 ( α ˜ 1 2 ) + 2 , ( 2 α ˜ 1 ) v ˜ 4 β ˜ 2 + 5 2 β ˜ , 0 , 2 α ˜ 1 4 β ˜ 1 2 v ˜ .
As we can see, these expressions do not depend on μ ˜ , so their derivatives with respect to μ ˜ are zero. The means are preserved.
Gradient and natural gradient algorithms (Algorithms 2 and 3):
α ˜ ( k + 1 ) = α ˜ ( k ) γ v ˜ ( k ) 2 β ˜ ( k ) ( 2 α ˜ ( k ) + 3 2 ) ψ 1 ( α ˜ ( k ) 1 2 ) + 2 , β ˜ ( k + 1 ) = β ˜ ( k ) γ ( 2 α ˜ ( k ) 1 ) v ˜ ( k ) 4 [ β ˜ ( k ) ] 2 + 5 2 β ˜ ( k ) , μ ˜ ( k + 1 ) = μ ˜ ( k ) , v ˜ ( k + 1 ) = v ˜ ( k + 1 ) γ 2 α ˜ ( k ) 1 4 β ˜ ( k ) 1 2 v ˜ ( k ) .
Here, γ is fixed for the gradient algorithm and is proportional to 1 / K L for the natural gradient algorithm.

4. Numerical Experimentations

To show the relative performances of these algorithms, we generate n = 100 samples from the model p ( x , y ) = N ( x | 1 , y ) I G ( y | 3 , 1 ) for the numerical computations. Thus, it should be noted that we know the exact values of the unknown parameters which can be used to show the performances of the proposed algorithms. The following results: θ ˜ 1 , θ ˜ 2 and θ ˜ 3 are the estimated parameters using, respectively, Algorithm 1, Algorithm 2 and Algorithm 3. The contour plots of the corresponding probability density functions are shown in Figure 1 compared with original model.
θ ˜ 1 = μ ˜ = 0.044724 , v ˜ = 0.590894 , α ˜ = 3.792337 , β ˜ = 1.765735 θ ˜ 2 = μ ˜ = 0.044724 , v ˜ = 7.910504 , α ˜ = 4.991621 , β ˜ = 3.002594 θ ˜ 3 = μ ˜ = 0.044724 , v ˜ = 0.423761 , α ˜ = 4.415239 , β ˜ = 0.706049 .
All three algorithms try to minimize the same criterion. So, the objectives are all the same, but the number of steps may differ. The requirements must reach the minimum K L ( · ) . In this simple example of the normal-inverse-gamma distribution, the convergence step numbers of VBA, gradient-based and natural gradient are 1, 2 and 1 using moment initializations. The overall performance of the standard alternate optimization (VBA) is more precise than any other. The poorest estimation is from a gradient-based algorithm. So, the algorithms are able to approximate the joint density function with a separable one but with different accuracy. In the following section, we will tackle a more complex model.

5. Multivariate Normal-Inverse-Wishart Example

In previous section, we explained how to preform VBA in order to approximate a complicated joint distribution function by tractable marginal factorials using a simple case study. In this section, a multivariate normal case p ( x ) = N ( x | μ ˜ , Σ ˜ ) is considered, which is approximated by q ( x ) = i N ( x i | μ ˜ i , v ˜ i ) for different shapes of the covariance matrix Σ ˜ .
We assume that the basic structure of an available dataset is multivariate normal with unknown mean vector μ ˜ and variance–covariance matrix Σ ˜ . Their joint prior distribution is a normal-inverse-Wishart distribution of N I W ( μ ˜ , Σ ˜ | μ ˜ 0 , κ ˜ , Ψ ˜ , ν ˜ ) , which is the generalized form of the classical N I G . The posteriors are multivariate normal for mean vector and inverse-Wishart in the variance–covariance matrix. Since the normal-inverse-Wishart distribution is a conjugate prior distribution for multivariate normal, the posterior distribution of μ ˜ and Σ ˜ again belongs to the same family, and their corresponding margins are
M N κ ˜ μ ˜ 0 + n x ¯ κ ˜ + n , 1 κ ˜ + n Λ ˜ , I W Λ ˜ + Ψ ˜ + i = 1 n ( x i x ¯ ) ( x i x ¯ ) T , ν ˜ + n ,
where n is the sample size. To present the performance of the three algorithms, we examined on a dataset based on x N I W ( x | μ , Σ ) , whose parameters have the following low-density structure:
μ M N μ | 2 1 , 1 2 3 1 1 1 , Σ I W Σ | 3 1 1 1 , 6 .
We used only the data of x in the estimation processes. The results of algorithms are presented in Figure 2 along with the true contour plot of the model. The VBA estimation is the most separable distribution compared with gradient and natural gradient methods. The next best case is the natural gradient algorithm, but its weakness is transferring the dependency slightly to the approximation. The results for the gradient-based algorithm show the dependency completely, along with its inability to obtain a separable model.

6. Simple Linear Inverse Problem

Finally the third example is the case of linear inverse problems with g = H f + ϵ with priors p ( ϵ ) = N ( ϵ | 0 , v ϵ I ) , p ( f ) = N ( f | 0 , diag [ v ] ) and p ( v | α , β ) = j I G ( v j | α , β ) , where f = [ f 1 , f 2 , , f N ] and v = [ v f 1 , v f 2 , , v f N ] . Using these priors, we get p ( f , v | g ) p ( g | f , v ϵ ) p ( f | v ) p ( v ) with p ( g | f , v ϵ ) = N ( g | H f , v ϵ I ) . See [13] for details.
Thus, the joint distribution of g , f , and v :
p ( g , f , v ) p ( g | f , v ϵ ) p ( f | v ) p ( v ) .
is approximated by q ( g , f ˜ , v ˜ ) = q 1 ( g | f ˜ ) q 2 ( f ˜ ) q 3 ( v ˜ ) using the VBA method. Even if the main interest is the estimation of q 2 ( f ˜ ) , but in the recursive process, q 1 ( g | f ˜ ) and q 3 ( v ˜ ) are also updated. For simplicity, we suppose that the transfer matrix H is an identical matrix I . The final outputs are as follows:
f ˜ M N μ ˜ f ˜ 1 + 2 v ϵ ˜ α ˜ n ( v ˜ + μ ˜ f ˜ 2 + 2 β ˜ ) , diag [ v ϵ ˜ ( v ˜ + μ ˜ f ˜ 2 + 2 β ˜ ) n ( v ˜ + μ ˜ f ˜ 2 ) + 2 n β ˜ + 2 v ϵ ˜ α ˜ ] , g N ( μ ˜ f ˜ , v ϵ ˜ I ) , and v ˜ j I G α ˜ j , v ˜ j + μ ˜ f ˜ j 2 2 + β ˜ j , j = 1 , , N .
For simulations, we chose a model to examine the performance of these margins and compare them with gradient and natural gradient algorithms. The selected model is g = H f + ϵ with the following assumptions:
H = I , f M N ( f | 0 , diag [ v 1 , v 2 ] ) , v 1 I G ( v 1 | 3 , 2 ) , v 2 I G ( v 2 | 4 , 3 ) , ϵ M N ( ϵ | 0 , I ) .
In the assessment procedure, we do not apply the above information. The output of algorithms are shown in Figure 3, as well as the actual contour plot. In this example, the best diagnosis is from the natural gradient algorithm. The VBA by construction is separable and cannot be the same as the original.

7. Conclusions

This paper presents an approximation method of the unknown density functions for hidden variables called VBA. It is compared with gradient and natural gradient algorithms. We also consider three examples normal-inverse-gamma, normal-inverse-Wishart and linear inverse problem. We provided details of the first model and showed examples of two other examples throughout the whole paper. In all three models, the parameters are unexplored and need to be estimated by recursive algorithms. We attempted to approximate the joint complex distribution with a simpler version of the margin factorials that appeared to be independent cases. The VBA and natural gradient converged fairly early. The major discrepancy in algorithms comes from the accuracy of the results. They estimate the intricate joint distribution with separable ones. Here, the best overall performance of VBA is demonstrated.

Author Contributions

Both authors contributed equally. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Acerbi, L. Variational bayesian monte carlo. In Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, Montréal, QC, Canada, 3–8 December 2018; pp. 8223–8233. [Google Scholar]
  2. Kuusela, M.; Raiko, T.; Honkela, A.; Karhunen, J. A gradient-based algorithm competitive with variational Bayesian EM for mixture of Gaussians. In Proceedings of the 2009 International Joint Conference on Neural Networks, Atlanta, GA, USA, 14–19 June 2009; pp. 1688–1695. [Google Scholar]
  3. Gharsalli, L.; Duchêne, B.; Mohammad-Djafari, A.; Ayasso, H. A gradient-like variational Bayesian approach: Application to microwave imaging for breast tumor detection. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 1708–1712. [Google Scholar]
  4. Zhang, G.; Sun, S.; Duvenaud, D.; Grosse, R. Noisy natural gradient as variational inference. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 5852–5861. [Google Scholar]
  5. Lin, W.; Khan, M.E.; Schmidt, M. Fast and simple natural-gradient variational inference with mixture of exponential-family approximations. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 3992–4002. [Google Scholar]
  6. Ghahramani, Z.; Beal, M. Variational inference for Bayesian mixtures of factor analysers. Adv. Neural Inf. Process. Syst. 1999, 12, 449–455. [Google Scholar]
  7. Watanabe, S.; Minami, Y.; Nakamura, A.; Ueda, N. Variational Bayesian estimation and clustering for speech recognition. IEEE Trans. Speech Audio Process. 2004, 12, 365–381. [Google Scholar] [CrossRef]
  8. Montesinos-López, O.A.; Montesinos-López, A.; Crossa, J.; Montesinos-López, J.C.; Luna-Vázquez, F.J.; Salinas-Ruiz, J.; Herrera-Morales, J.R.; Buenrostro-Mariscal, R. A variational Bayes genomic-enabled prediction model with genotype× environment interaction. G3 Genes Genomes Genet. 2017, 7, 1833–1853. [Google Scholar] [CrossRef] [PubMed]
  9. Babacan, S.D.; Molina, R.; Katsaggelos, A.K. Variational Bayesian super resolution. IEEE Trans. Image Process. 2010, 20, 984–999. [Google Scholar] [PubMed]
  10. Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  11. Fox, C.W.; Roberts, S.J. A tutorial on variational Bayesian inference. Artif. Intell. Rev. 2012, 38, 85–95. [Google Scholar] [CrossRef]
  12. De Canditiis, D.; Vidakovic, B. Wavelet Bayesian block shrinkage via mixtures of normal-inverse gamma priors. J. Comput. Graph. Stat. 2004, 13, 383–398. [Google Scholar] [CrossRef]
  13. Ayasso, H.; Mohammad-djafari, A. Joint image restoration and segmentation using Gauss-Markov-Potts prior models and variational Bayesian computation. In Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 1297–1300. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The true model is N ( x | 0 , y ) I G ( y | 3 , 1 ) . The numbers of iterations until the convergence are different in the algorithms.
Figure 1. The true model is N ( x | 0 , y ) I G ( y | 3 , 1 ) . The numbers of iterations until the convergence are different in the algorithms.
Psf 05 00048 g001
Figure 2. In this example, the numbers of iterations to obtain the convergence are different between the three algorithms. These numbers are, respectively, 2, 280 and 4. The VBA and natural gradient algorithms estimate the distribution with separable functions in fewer steps than the gradient which seems not converged even up to 280 iterations.
Figure 2. In this example, the numbers of iterations to obtain the convergence are different between the three algorithms. These numbers are, respectively, 2, 280 and 4. The VBA and natural gradient algorithms estimate the distribution with separable functions in fewer steps than the gradient which seems not converged even up to 280 iterations.
Psf 05 00048 g002
Figure 3. The true model is almost separable. The natural gradient algorithm works better in this example, and the poorest approximation is that of the gradient-based algorithm.
Figure 3. The true model is almost separable. The natural gradient algorithm works better in this example, and the poorest approximation is that of the gradient-based algorithm.
Psf 05 00048 g003
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fallah Mortezanejad, S.A.; Mohammad-Djafari, A. Variational Bayesian Approximation (VBA): A Comparison between Three Optimization Algorithms. Phys. Sci. Forum 2022, 5, 48. https://doi.org/10.3390/psf2022005048

AMA Style

Fallah Mortezanejad SA, Mohammad-Djafari A. Variational Bayesian Approximation (VBA): A Comparison between Three Optimization Algorithms. Physical Sciences Forum. 2022; 5(1):48. https://doi.org/10.3390/psf2022005048

Chicago/Turabian Style

Fallah Mortezanejad, Seyedeh Azadeh, and Ali Mohammad-Djafari. 2022. "Variational Bayesian Approximation (VBA): A Comparison between Three Optimization Algorithms" Physical Sciences Forum 5, no. 1: 48. https://doi.org/10.3390/psf2022005048

APA Style

Fallah Mortezanejad, S. A., & Mohammad-Djafari, A. (2022). Variational Bayesian Approximation (VBA): A Comparison between Three Optimization Algorithms. Physical Sciences Forum, 5(1), 48. https://doi.org/10.3390/psf2022005048

Article Metrics

Back to TopTop