Next Article in Journal
Information Properties of a Random Variable Decomposition through Lattices
Previous Article in Journal
SEIR Modeling, Simulation, Parameter Estimation, and Their Application for COVID-19 Epidemic Prediction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Adaptive Importance Sampling for Equivariant Group-Convolution Computation †

by
Pierre-Yves Lagrave
1,* and
Frédéric Barbaresco
2
1
Thales Research and Technology, 91767 Palaiseau, France
2
Thales Land and Air Systems, 92190 Meudon, France
*
Author to whom correspondence should be addressed.
Presented at the 41st International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Engineering, Paris, France, 18–22 July 2022.
Phys. Sci. Forum 2022, 5(1), 17; https://doi.org/10.3390/psf2022005017
Published: 5 December 2022

Abstract

:
This paper introduces an adaptive importance sampling scheme for the computation of group-based convolutions, a key step in the implementation of equivariant neural networks. By leveraging information geometry to define the parameters update rule for inferring the optimal sampling distribution, we show promising results for our approach by working with the two-dimensional rotation group SO ( 2 ) and von Mises distributions. Finally, we position our AIS scheme with respect to quantum algorithms for computing Monte Carlo estimations.

1. Introduction and Motivations

Geometric deep learning [1] is an emerging field receiving more and more traction because of its successful application to a wide range of domains [2,3,4]. In this context, equivariant neural networks (ENN) [5] have been shown to be superior to conventional deep learning approaches from both accuracy and robustness standpoints and appear as a natural alternative to data augmentation techniques [6,7] to achieve geometrical robustness.
One key bottleneck for scaling ENN to industrial applications lies with the numerical computation of the associated equivariant operators. More precisely, two main approaches have been used in the literature, namely a Monte Carlo sampling method [2] (which can be made exhaustive for small finite groups) and a generalized Fourier-based method [4,8,9]. However, these approaches suffer from scalability issues as the complexity of the underlying group increases (e.g., handling non-compact groups such as SU(1,1) or large finite groups such as the symmetric group S n is challenging). Even for groups such as SO(2) for which previous works on the use of spherical harmonics can be leveraged on, the efficient computation of a reliable estimate of the convolution remains a challenge (convergence).
In this context, the authors of [10] have proposed an efficient method for building adequate kernel functions to be used within steerable neural networks [11] by leveraging on the knowledge of infinitesimal generators of the considered Lie group and on a Krylov approach for solving the linear constraints. We propose in this paper to cover the specific case of group-convolutional neural networks (G-CNN) [2,12], which in particular, rely on the computation of group-based convolution operators. By leveraging on information geometry as proposed in [13] for quantile estimation, we introduce here an adaptive importance sampling (AIS) variance reduction method based on information geometric optimization [14] to improve the convergence of Monte Carlo estimators for the numerical computation of group-based convolution feature maps, as used in several recent works [2,9,15]. We illustrate our approach on the two-dimensional rotation group SO(2) by regularizing with von Mises distributions [16], a set-up for which the Fisher information metric [17] can be computed using closed form formulas.
Finally, we shed some light on the benefits of working toward a quantum version of our proposed AIS scheme in order to reach a quadratic speed-up [18]. Improving quantum Monte Carlo integration schemes is indeed a very active topic of research [19], mainly driven by applications within the financial industry [20]. Benchmarking with group-Fourier transform-based approaches, such as [21], which are more theoretically involved but with a promise of an exponential speed-up, will be of particular interest in this context.

2. Group Convolution and Expectation

We consider in the following a compact group G with corresponding Haar measure μ G . As μ G G < , we can choose μ G so that G d μ G = 1 by using an adequate normalization.
We are interested in evaluating the group-based convolution operator ψ G defined below for functionals f , k : G R and g G :
ψ G g = G k h 1 g f h d μ G h
Using a probabilistic interpretation of (1), we can write
ψ G g = E μ G k H 1 g f H
where H is a G valued random variable distributed according to μ G . The convolution can therefore be estimated with a Monte Carlo method by using the following estimator
ψ ˜ n G g = 1 n i = 1 n k h i 1 g f h i
where h i μ G and for which the efficiency could be improved through variance reduction techniques [22]. By anchoring in [13], we describe in the following an adaptive importance sampling approach for the computation of (1). Similar ideas were also used in [23] for financial applications.

3. Adaptive Importance Sampling

We consider in the following a set Φ Θ of parametric probability density functions on G, where Θ represents the parameters space. Each density ϕ θ Φ Θ is assumed to be absolutely continuous with respect to the Haar measure μ G of the group G, so that the corresponding probability measure can be written as d μ θ = ϕ θ d μ G and the Radon–Nikodym derivative ω θ = d μ G d μ θ can be considered.
Using the conventional importance sampling approach, we can then write:
ψ G g = G k h 1 g f h d μ G d μ θ h d μ θ h
= E μ θ ω θ H k H 1 g f H
The idea is then to choose a measure μ θ * for which θ * minimizes the variance v k , f , g of the random variable k H 1 g ω θ H f H , which can be written as
v k , f , g θ = E μ θ ω θ H k H 1 g f H 2 ψ G g 2
= m 2 k , f , g θ ψ G g 2
where
m 2 k , f , g θ = E μ θ k H 1 g ω θ H f H 2
= E μ G ω θ H k H 1 g f H 2

3.1. Monte Carlo Estimator and Convergence

We assume that we can construct a sequence of parameters θ i i = 0 n 1 , together with realizations h i i = 1 n of the random variables H i i = 1 n such that H i μ θ i 1 and that θ n θ * Θ as n . We can then consider the following Monte Carlo estimator:
ψ ^ n G g = 1 n i = 1 n ω θ i 1 h i k h i 1 g f h i
Under usual integrability conditions, Theorem 3.1 of [13] states that ψ ^ n G g ψ G g almost surely as n . Furthermore, we have the following distributional convergence result,
n ψ ^ n G g ψ G g N 0 , v k , f , g θ *
where N 0 , σ 2 refers to the Gaussian distribution with 0 mean and variance σ 2 .

3.2. Natural Gradient Descent

We now discuss how to build the sequence of parameters θ i i = 0 n 1 and corresponding realizations h i i = 1 n as introduced in Section 3.1, reminding ourselves that we have
θ * = arg min θ Θ v k , f , g θ = arg min θ Θ m 2 k , f , g θ
Assuming that the parameter space Θ R m is a smooth manifold, we can consider the Fisher information metric g on the density space Φ Θ , which is defined as it follows [17]:
g i j = E μ θ log ϕ θ θ i log ϕ θ θ j
We then propose using a natural gradient descent strategy to minimize the quantity m 2 k , f , g , namely
θ k + 1 = θ k α k F k 1 m 2 k , f , g θ k
where F k is the Fisher information matrix, i.e., the representation of the Fisher metric as a m × m matrix and α k R + * . Assuming that the considered functions are smooth enough, it is possible to write:
m 2 k , f , g θ = E μ G ω θ H k H 1 g f H 2
= E μ G ω θ H log ϕ θ H k H 1 g f H 2
= E μ θ log ϕ θ H ω θ H k H 1 g f H 2
Using a stochastic approximation scheme such as the Robbins–Monro algorithm [24] then leads to consider the following update rule,
θ k + 1 = θ k + α k F k 1 Λ H k + 1 , θ k
where Λ H , θ = log ϕ θ H ω θ H k H 1 g f H 2 , H k μ θ k 1 and the sequence of numbers α k is such that k α k = and k α k 2 < .

3.3. About IGO Algorithms

Information geometric optimization (IGO) algorithms are introduced in [14] as a unified framework to solve black-box optimization problems. IGO algorithms can be seen as performing an estimation of a distribution over the considered search space X leading to small values of the target function Q when sampling according to it. More precisely, the idea is to maintain at each iteration t a parametric probability distribution P λ t on the search space X , for λ t Λ R p and to have the value λ t evolve over time as to shift P λ t toward giving more weight to points x X associated with a lower value of Q.
The IGO algorithms described [14] first transfer the function Q from X to Λ by using an adaptive quantile-based approach and then applying a natural gradient descent by leveraging on the Fisher information metric of the considered statistical model. The scheme described in the Definition 5 of [14] defines the following the update rule for the parameter λ t :
λ t + d t = λ t + d t × I 1 λ t i = 1 N ω i ln P λ x i : N λ λ = λ t
where I is the fisher matrix of the model, x 1 , . . . , x N are N samples drawn according to P λ t at step t, x i : N denotes the sample point ranked i t h according to Q (i.e., Q ( x 1 : N ) < < Q ( x N : N ) ) and ω i = 1 N ω i 1 2 , with ω q = 1 q < q 0 a quantile-based selection function of threshold q 0 .
IGO algorithms could therefore be used in our context by setting Q = m 2 k , f , g θ and X = Θ to infer the optimal value θ * Θ . Implementing the update rule (19) requires a priori a large number of evaluations of the term Q = m 2 k , f , g θ to derive the sorted samples x i : N , making this approach generally not well suited to our context.

4. Application to SO ( 2 ) -Convolutions

We give here an application of our AIS approach for the computation of SO ( 2 ) -convolutions by using von Mises densities [16] for the weighting. This type of computation is in particular relevant when working with SE ( 2 ) -ENN by exploiting the semi-direct product structure SE ( 2 ) = R 2 SO ( 2 ) , as performed in [3].

4.1. Fisher Information Metric

The Haar measure on SO ( 2 ) is simply the Lebesgue measure on the unit circle [ 0 , 2 π ] and will be denoted d α before normalization. The convolution operator (1) is therefore applied to functionals defined on [ 0 , 2 π ] , so that we are interested in evaluating the following quantity for α 0 [ 0 , 2 π ] :
ψ S O ( 2 ) α 0 = 1 2 π 0 2 π k α 0 α f α d α
We consider in the following a family of von Mises densities ϕ θ on [ 0 , 2 π ] , for θ = μ , κ , and for which α [ 0 , 2 π ] ,
ϕ θ α = e κ cos α μ 2 π I 0 κ
where for n N , I n κ = 1 π 0 π e κ cos θ cos n θ d θ refers to modified Bessel functions. Denoting in the following μ , κ = log ϕ θ , we have μ μ , κ α = κ sin α μ and κ μ , κ α = cos α μ I 0 κ I 0 κ and it is now possible to evaluate the metric tensor g . More precisely, we easily obtain:
g μ , μ = E μ θ μ μ , κ α 2
= 0 2 π κ 2 sin 2 α μ ϕ θ α d α
= κ I 1 κ I 0 κ
g κ , κ = E μ θ κ μ , κ α 2
= 0 2 π cos α μ I 0 κ I 0 κ 2 ϕ θ α d α
= 1 1 κ I 1 κ I 0 κ 2 I 0 κ I 1 κ I 0 κ 2 + I 0 κ 2 I 0 κ 2
g μ , κ = E μ θ μ μ , κ α κ μ , κ α
= 0 2 π κ sin α μ cos α μ I 0 κ I 0 κ ϕ θ α d α
= 0
The inverse of the Fisher information matrix in the scheme (18) can therefore be computed with closed-form formulas, making its evaluation very efficient for the considered case.

4.2. Numerical Experiments

To numerically validate our approach, we have considered von Mises type feature functions f κ 0 , α 0 : α e κ 0 cos α 0 μ and kernel functions k : [ 0 , 2 π ] R modeled as small fully connected neural networks with one hidden layer of 128 neurons with ReLu activation and uniform random weights initialization. To run our testing, we have used κ 0 = 3 and μ 0 = π 2 .
Figure 1 shows the comparison between the results obtained with the estimator (10) using the adaptive importance sampling scheme and those obtained with the conventional estimator (3). We can in particular see that the adaptive importance sampling scheme converges faster to the theoretical value (here computed by using (3) with n = 50 , 000 and displayed in black in Figure 1), while providing much narrower confidence intervals (because of lower variance) than the conventional Monte Carlo estimator. Figure 2 shows the evolution of the parameter θ = μ , κ as we iterate through the update rule (18), from which we can also observe a fast convergence.

4.3. Extension to SO ( 3 ) -Convolutions

Generalizing the above results to cover SO ( 3 ) -convolutions is of particular interest when using ENN for processing spherical data such as fish-eye images [4,25]. The Fisher–Bingham distribution [26], also known as the Kent distribution, can be leveraged in this context. More precisely, we have in this case, for x S 2 (the 2D-sphere in R 3 ):
ϕ θ x = c κ , β 1 exp κ γ 1 . x + β γ 2 . x 2 γ 3 . x 2
where γ i for i = 1 , 2 , 3 are vectors of R 3 so that the 3 × 3 matrix Γ = γ 1 , γ 2 , γ 3 is orthogonal and c κ , β is a normalizing constant.
Although we defer to further work the details of the derivation of the corresponding AIS estimator (10), we illustrate on Figure 3 that SO ( 3 ) -convolutions could also benefit from variance reduction methods by using a simple quasi-Monte Carlo scheme [27] with a three-dimensional Sobol sequence [28].

5. Monte Carlo Methods in the Quantum Set-Up

Monte Carlo computations can generally benefit from a quadratic speed-up in a quantum computing set-up [18] and improving quantum Monte Carlo integration schemes is a very active topic of research [19], mainly driven by applications within the financial industry [20,29].
A similar speed-up can therefore be expected in our context by estimating (1) by leveraging on the quantum amplitude estimation (QAE) algorithm [30]. For g G , we denote ϕ g f , k : G R the function such that h G , ϕ g f , k h = k h 1 g f h .
We first construct the operator U μ G to load a discretized version of μ G so that U μ G 0 = h G ϵ p h h , with p h = g B h , r ϵ μ G g , B x , r the ball of radius r > 0 centered in x G and G ϵ a discrete subset of G. We scale ϕ g f , k to ϕ ˜ g f , k : G 0 , 1 and build another unitary operator U ϕ to compute and load the values of ϕ g f , k taken on G ϵ , that we defined by U ϕ h = 1 ϕ ˜ g f , k h h 0 + ϕ ˜ g f , k h h 1 . Using the QAE algorithm on U ϕ U μ G gives us access to an estimate of (1) after proper rescaling, with a precision of δ in O V k , f , g δ queries, with V k , f , g = V μ G ϕ g f , k H .
As described in Section 3.1, the AIS estimator (10) leads to a precision of δ for n = O v k , f , g θ * δ 2 samples, which is asymptotically less efficient than the above quantum estimator. However, no quantum advantage has been evidenced on current hardware for general Monte Carlo estimations and further challenges with respect to the precision of the evaluation of the integrand ϕ g f , k are expected in our specific context. Keeping working on the optimization of the estimators in the classical set-up while keeping track of the progress made on the development of quantum hardware therefore appears a reasonable path to follow.

6. Conclusions and Further Work

By leveraging on the approach proposed in [13] for quantile estimation, we have introduced in this paper an AIS variance reduction method for the computation of group-based convolution operators, a key component of equivariant neural networks. We have in particular used information geometry concepts to define an efficient update rule to infer the optimal sampling parametric distribution and have also shown promising results when working with the two-dimensional rotation group SO ( 2 ) and von Mises distributions.
Further work will include the study of non-compact groups such as SU ( 1 , 1 ) as to improve the efficiency of the computations underlying to the ENN introduced in [9]. As shown in [31], Souriau Thermodynamics can be used to build Gaussian distributions over SU ( 1 , 1 ) , which appear as natural candidates for applying the AIS scheme presented in this paper.
We have also seen that Monte Carlo computations can generally benefit from a quadratic speed-up in a quantum computing set-up. Further work will include the study of using AIS in this context as to provide a generic and efficient quantum algorithm for group-convolution computation. Benchmarking with group-Fourier transform-based approaches such as [21], which are more theoretically involved but with a promise of exponential speed-up, will also be of high interest, as it will be the case for results coming from the emerging field of quantum geometric deep learning [32,33].

Author Contributions

All authors contributed equally to the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This paper is the result of some research work conducted by the authors at Thales Group.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bronstein, M.M.; Bruna, J.; Cohen, T.; Veličković, P. Geometric Deep Learning: Grids, Groups, Graphs, Geodesics, and Gauges. arXiv 2021, arXiv:2104.13478. [Google Scholar]
  2. Finzi, M.; Stanton, S.; Izmailov, P.; Wilson, A.G. Generalizing Convolutional Neural Networks for Equivariance to Lie Groups on Arbitrary Continuous Data. arXiv 2020, arXiv:2002.12880. [Google Scholar]
  3. Lafarge, M.W.; Bekkers, E.J.; Pluim, J.P.W.; Duits, R.; Veta, M. Roto-Translation Equivariant Convolutional Networks: Application to Histopathology Image Analysis. arXiv 2020, arXiv:2002.08725. [Google Scholar] [CrossRef] [PubMed]
  4. Cohen, T.S.; Geiger, M.; Köhler, J.; Welling, M. Spherical CNNs. arXiv 2018, arXiv:1801.10130. [Google Scholar]
  5. Gerken, J.E.; Aronsson, J.; Carlsson, O.; Linander, H.; Ohlsson, F.; Petersson, C.; Persson, D. Geometric Deep Learning and Equivariant Neural Networks. arXiv 2021, arXiv:2105.13926. [Google Scholar]
  6. Shorten, C.; Khoshgoftaar, T.M. A survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  7. Chen, S.; Dobriban, E.; Lee, J.H. A Group-Theoretic Framework for Data Augmentation. arXiv 2020, arXiv:1907.10905. [Google Scholar]
  8. Kondor, R.; Trivedi, S. On the Generalization of Equivariance and Convolution in Neural Networks to the Action of Compact Groups. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; Dy, J., Krause, A., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2018; Volume 80, pp. 2747–2755. [Google Scholar]
  9. Lagrave, P.Y.; Cabanes, Y.; Barbaresco, F. SU(1,1) Equivariant Neural Networks and Application to Robust Toeplitz Hermitian Positive Definite Matrix Classification. In Geometric Science of Information; Nielsen, F., Barbaresco, F., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 577–584. [Google Scholar]
  10. Finzi, M.; Welling, M.; Wilson, A.G. A Practical Method for Constructing Equivariant Multilayer Perceptrons for Arbitrary Matrix Groups. arXiv 2021, arXiv:2104.09459. [Google Scholar]
  11. Cohen, T.S.; Geiger, M.; Weiler, M. A General Theory of Equivariant CNNs on Homogeneous Spaces. In Advances in Neural Information Processing Systems; Wallach, H., Larochelle, H., Beygelzimer, A., Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: New York, NY, USA, 2019; Volume 32, pp. 9145–9156. [Google Scholar]
  12. Cohen, T.; Welling, M. Group Equivariant Convolutional Networks. In Proceedings of the 33rd International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016; Balcan, M.F., Weinberger, K.Q., Eds.; PMLR: New York, NY, USA, 2016; Volume 48, pp. 2990–2999. [Google Scholar]
  13. Egloff, D.; Leippold, M. Quantile estimation with adaptive importance sampling. Ann. Stat. 2010, 38, 1244–1278. [Google Scholar] [CrossRef] [Green Version]
  14. Ollivier, Y.; Arnold, L.; Auger, A.; Hansen, N. Information-Geometric Optimization Algorithms: A Unifying Picture via Invariance Principles. J. Mach. Learn. Res. 2011, 18, 1–65. [Google Scholar] [CrossRef]
  15. Lagrave, P.Y.; Barbaresco, F. Generalized SU(1,1) Equivariant Convolution on Fock-Bargmann Spaces for Robust Radar Doppler Signal Classification. 2021; working paper or preprint. [Google Scholar]
  16. von Mises, R. Uber die “Ganzzahligkeit” der Atomgewicht und verwandte Fragen. Phys. Z. 1981, 19, 490–500. [Google Scholar]
  17. Amari, S.I.; Barndorff-Nielsen, O.E.; Kass, R.E.; Lauritzen, S.L.; Rao, C.R. Differential Geometry in Statistical Inference; Lecture Notes–Monograph Series; Institute of Mathematical Statistics: Hayward, CA, USA, 1987; Volume 10, pp. 1–240. [Google Scholar]
  18. Montanaro, A. Quantum speedup of Monte Carlo methods. Proc. R. Soc. Math. Phys. Eng. Sci. 2015, 471, 20150301. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Herbert, S. Quantum Monte-Carlo Integration: The Full Advantage in Minimal Circuit Depth. arXiv 2021, arXiv:2105.09100. [Google Scholar] [CrossRef]
  20. An, D.; Linden, N.; Liu, J.P.; Montanaro, A.; Shao, C.; Wang, J. Quantum-accelerated multilevel Monte Carlo methods for stochastic differential equations in mathematical finance. Quantum 2021, 5, 481. [Google Scholar] [CrossRef]
  21. Castelazo, G.; Nguyen, Q.T.; Palma, G.D.; Englund, D.; Lloyd, S.; Kiani, B.T. Quantum algorithms for group convolution, cross-correlation, and equivariant transformations. arXiv 2021, arXiv:2109.11330. [Google Scholar] [CrossRef]
  22. Botev, Z.; Ridder, A. Variance Reduction. In Wiley StatsRef: Statistics Reference Online; Wiley: Hoboken, NJ, USA, 2017; pp. 1–6. [Google Scholar] [CrossRef]
  23. Jourdain, B. Adaptive variance reduction techniques in finance. Adv. Financ. Model. 2009, 8, 205. [Google Scholar]
  24. Robbins, H.; Monro, S. A Stochastic Approximation Method. Ann. Math. Stat. 1951, 22, 400–407. [Google Scholar] [CrossRef]
  25. Martin, S.; Lagrave, P.Y. On the Benefits of SO(3)-Equivariant Neural Networks for Spherical Image Processing. 2022, working paper or preprint. 2022; working paper or preprint. [Google Scholar]
  26. Kent, J.T. The Fisher-Bingham Distribution on the Sphere. J. R. Stat. Soc. Ser. (Methodol.) 1982, 44, 71–80. [Google Scholar] [CrossRef]
  27. Niederreiter, H. Random Number Generation and Quasi-Monte Carlo Methods; Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 1992. [Google Scholar] [CrossRef] [Green Version]
  28. Sobol’, I. On the distribution of points in a cube and the approx imate evaluation of integrals. USSR Comput. Math. Math. Phys. 1967, 7, 86–112. [Google Scholar] [CrossRef]
  29. Orús, R.; Mugel, S.; Lizaso, E. Quantum computing for finance: Overview and prospects. Rev. Phys. 2019, 4, 100028. [Google Scholar] [CrossRef]
  30. Brassard, G.; Høyer, P.; Mosca, M.; Tapp, A. Quantum amplitude amplification and estimation. Contemp. Math. 2002, 305, 53–74. [Google Scholar] [CrossRef] [Green Version]
  31. Barbaresco, F. Lie Group Statistics and Lie Group Machine Learning Based on Souriau Lie Groups Thermodynamics & Koszul-Souriau-Fisher Metric: New Entropy Definition as Generalized Casimir Invariant Function in Coadjoint Representation. Entropy 2020, 22, 642. [Google Scholar] [CrossRef] [PubMed]
  32. Larocca, M.; Sauvage, F.; Sbahi, F.M.; Verdon, G.; Coles, P.J.; Cerezo, M. Group-Invariant Quantum Machine Learning. arXiv 2022, arXiv:2205.02261. [Google Scholar] [CrossRef]
  33. Meyer, J.J.; Mularski, M.; Gil-Fuster, E.; Mele, A.A.; Arzani, F.; Wilms, A.; Eisert, J. Exploiting symmetry in variational quantum machine learning. arXiv 2022, arXiv:2205.06217. [Google Scholar]
Figure 1. Comparison between the convergence of the AIS scheme with von Mises densities (red) and the traditional Monte Carlo approach (blue) by representing the evolution of the estimated convolution ψ S O ( 2 ) at α 0 = 0 (left) and α 0 = π / 2 (right), as a function of the number n of simulated samples.
Figure 1. Comparison between the convergence of the AIS scheme with von Mises densities (red) and the traditional Monte Carlo approach (blue) by representing the evolution of the estimated convolution ψ S O ( 2 ) at α 0 = 0 (left) and α 0 = π / 2 (right), as a function of the number n of simulated samples.
Psf 05 00017 g001
Figure 2. Evolution of the components of the parameter θ = μ , κ as updated according to (18) when estimating ψ S O ( 2 ) ( 0 ) with the AIS scheme with von Mises densities.
Figure 2. Evolution of the components of the parameter θ = μ , κ as updated according to (18) when estimating ψ S O ( 2 ) ( 0 ) with the AIS scheme with von Mises densities.
Psf 05 00017 g002
Figure 3. Convergence comparison for a SO ( 3 ) -convolution computation between the classical Monte Carlo estimate (3) with numpy random numbers generator (blue) and the corresponding quasi-Monte Carlo scheme leveraging on a Sobol sequence (red).
Figure 3. Convergence comparison for a SO ( 3 ) -convolution computation between the classical Monte Carlo estimate (3) with numpy random numbers generator (blue) and the corresponding quasi-Monte Carlo scheme leveraging on a Sobol sequence (red).
Psf 05 00017 g003
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lagrave, P.-Y.; Barbaresco, F. Adaptive Importance Sampling for Equivariant Group-Convolution Computation. Phys. Sci. Forum 2022, 5, 17. https://doi.org/10.3390/psf2022005017

AMA Style

Lagrave P-Y, Barbaresco F. Adaptive Importance Sampling for Equivariant Group-Convolution Computation. Physical Sciences Forum. 2022; 5(1):17. https://doi.org/10.3390/psf2022005017

Chicago/Turabian Style

Lagrave, Pierre-Yves, and Frédéric Barbaresco. 2022. "Adaptive Importance Sampling for Equivariant Group-Convolution Computation" Physical Sciences Forum 5, no. 1: 17. https://doi.org/10.3390/psf2022005017

Article Metrics

Back to TopTop