Next Article in Journal
Three-Dimensional Analysis of Flow Velocities through a Large Pool Fish Pass in the Trzebuńka Stream
Previous Article in Journal
The Information Geometry of Space-Time
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Interaction between Model Based Signal and Image Processing, Machine Learning and Artificial Intelligence  †

by
Ali Mohammad-Djafari
1,2
1
Laboratoire des Signaux et Systèmes, CNRS, CentraleSupélec-Univ Paris Saclay, 91192 Gif-sur-Yvette, France
2
International Science Consulting and Training (ISCT), 91440 Bures sur Yvette, France
Presented at the 39th International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Engineering, Garching, Germany, 30 June–5 July 2019.
Proceedings 2019, 33(1), 16; https://doi.org/10.3390/proceedings2019033016
Published: 28 November 2019

Abstract

:
Signale and image processing has always been the main tools in many area and in particular in Medical and Biomedical applications. Nowadays, there are great number of toolboxes, general purpose and very specialized, in which classical techniques are implemented and can be used: all the transformation based methods (Fourier, Wavelets, …) as well as model based and iterative regularization methods. Statistical methods have also shown their success in some area when parametric models are available. Bayesian inference based methods had great success, in particular, when the data are noisy, uncertain, incomplete (missing values) or with outliers and where there is a need to quantify uncertainties. In some applications, nowadays, we have more and more data. To use these “Big Data” to extract more knowledge, the Machine Learning and Artificial Intelligence tools have shown success and became mandatory. However, even if in many domains of Machine Learning such as classification and clustering these methods have shown success, their use in real scientific problems are limited. The main reasons are twofold: First, the users of these tools cannot explain the reasons when the are successful and when they are not. The second is that, in general, these tools can not quantify the remaining uncertainties. Model based and Bayesian inference approach have been very successful in linear inverse problems. However, adjusting the hyper parameters is complex and the cost of the computation is high. The Convolutional Neural Networks (CNN) and Deep Learning (DL) tools can be useful for pushing farther these limits. At the other side, the Model based methods can be helpful for the selection of the structure of CNN and DL which are crucial in ML success. In this work, I first provide an overview and then a survey of the aforementioned methods and explore the possible interactions between them.

1. Introduction

Nowadays, there are great number of general purpose and very specialized toolboxes, in which, classical and advanced techniques of signal and image processing methods are implemented and can be used. Between them, we can mention all the transformation based methods (Fourier, Hilbert, Wavelets, Radon, Abel, … and much more) as well as all the Model Based and iterative regularization methods. Statistical methods have also shown their success in some areas when parametric models are available.
Bayesian inference based methods had great success, in particular, when the data are noisy, uncertain, some missing and some outliers and where there is a need to account and to quantify uncertainties.
Nowadays, we have more and more data. To use these “Big Data” to extract more knowledge, the Machine Learning and Artificial Intelligence tools have shown success and became mandatory. However, even if in many domains of Machine Learning such as classification and clustering these methods have shown success, their use in real scientific problems are limited. The main reasons are twofold: First, the users of these tools can not explain the reasons when they are successful and when they are not. The second is that, in general, these tools can not quantify the remaining uncertainties.
Model based and Bayesian inference approach have been very successful in linear inverse problems. However, adjusting the hyper parameters is complex and the cost of the computation is high. The Convolutional Neural Networks (CNN) and Deep Learning (DL) tools can be useful for pushing farther these limits. At the other side, the Model based methods can be helpful for the selection of the structure of CNN and DL which are crucial in ML success. In this work, first I give an overview and a survey of the aforementioned methods and explore the possible interactions between them.
The rest of the paper is organized as follows: First a classification of signal and image processing methods is proposed. Then, very briefly, the Machine Learning tools are introduced. Then, through the problem of Imaging inside the body, we see the different steps from acquisition of the data, reconstruction, post-processing such as segmentation and finally the decision and conclusion of the user are presented. After mentioning some successful case studies in which the ML tools have been successful, we arrive at the main part of this paper: Looking for the possible interactions between Model based and Machine Learning tools. Finally, we mention the Open problems and challenges in both classical, model based and the ML tool.

2. Classification of Signal and Image Processing Methods

Signal and image processing methods can be classified in the following categories:
  • Transform based methods
  • Model based and inverse problem approach
  • Regularisation methods
  • Bayesian inference methods
In the first category, the main idea is to use different ways the signal and images can be represented in time, frequency, space, spacial frequency, time-frequency, wavelets, etc.

3. Transform Domain Methods

Figure 1 shows the main idea behind the transform based methods. Mainly, first a linear transform (Fourier, Wavelet, Radon, etc.) is applied to the signal or the image, then some thresholding or windowing is applied in this transform domain and finally an inverse transform is applied to obtain the result. Appropriate choices of the transform and the threshold or the widow size and shape are important for the success of such methods [1,2].

4. Model Based and Inverse Problem Approach

The model based methods are related to the notions of forward model and inverse problems approach. Figure 2 shows the main idea:
Given the forward model and the sources f , the prediction of the data g can be done, either in a deterministic way: g = ( f ) or via a probabilistic model: p ( g | f , ) .
In the same way, given the forward model and the data g , the estimation of the unknown sources f can be done either via a deterministic method or probabilistic one. One of the deterministic method is the Generalized inversion: f = ( g ) . A more general method is the regularization: f ^ = arg min f { J ( f ) } w i t h J ( f ) = g ( f ) 2 + λ ( f ) [3].
As we will see later, the only probabilistic method which can be efficiently use for the inverse problems is the Bayesian approach.

5. Regularization Methods

Let consider the linear inverse problem:
g = H f + ϵ ,
Then the basic idea in regularization is to define a regularization criterion:
J ( f ) = 1 2 g H f 2 2 + λ R ( f )
and optimize it to obtain the solution [4]. The first main issue in such regularization method is the choice of the regularizer. The most common examples are:
R ( f ) = f 2 2 , f β β , D f 2 2 , D f β β , j ϕ ( [ D f ] j ) , 1 β 2
The second main issue in regularization is the the choice of appropriate optimization algorithm. Mainly, depending on the type of the criterion, we have:
  • R ( f ) quadratic: Gradient based, Conjugate Gradient algorithms are appropriate.
  • R ( f ) non quadratic, but convex and differentiable: Here too the Gradient based and Conjugate Gradient (CG) methods can be used, but there are also great number of convex criterion optimization algorithms.
  • R ( f ) convex but non-differentiable: Here, the notion of sub-gradient is used.
Specific cases are:
  • L2 or quadratic: J ( f ) = 1 2 g H f 2 2 + λ D f 2 2 .
    In this case we have an analytic solution: f ^ = ( H H + λ D D ) 1 H g . However, in practice this analytic solution is not usable in high dimensional problems. In general, as the gradient J ( f ) = H ( g H f ) + 2 λ D D f can be evaluated analytically, gradient based algorithms are used.
  • L1 (TV): convex but not differentiable at zero: J ( f ) = 1 2 g H f 2 2 + λ D f 1 .
    The algorithms in this case use the notions of Fenchel conjugate, Dual problem, sub gradient, proximal operator, …
  • Variable splitting and Augmented Lagrangian
    ( f , z ^ ) = arg min f , z { 1 2 g H f 2 2 + λ z 1 + q z 2 2 } s . t . z = D f
    A great number of optimization algorithms have been proposed: ADMM, ISTA, FISTA, etc. [5,6,7].
Main limitations of deterministic regularization methods are:
  • Limited choice of the regularization term. Mainly, we have: a) Smoothness (Tikhonov), b) Sparsity, Piecewise continuous (Total Variation).
  • Determination of the regularization parameter. Even if there are some classical methods such as L-Curve and Cross validation, there are still controversial discussions about this.
  • Quantification of the uncertainties: This is the main limitation of the deterministic methods, in particular in medical and biological applications where this point is important.
The best possible solution to push further all these limits is the Bayesian approach which has: (a) Many possibilities to choose prior models, (b) possibility of the estimation of the hyper-parameters, and most important (c) accounting for the uncertainties.

6. Bayesian Inference Methods

The simple case of the Bayes rule is:
p ( f | g , ) = p ( g | f , ) p ( f | ) p ( g | ) where p ( g | ) = p ( g | f , ) p ( f | ) d f
When there are some hyper parameters which have also to be estimated, we have:
p ( f , θ | g , ) = p ( g | f , θ , ) p ( f | θ , ) p ( θ | ) p ( g | ) where p ( g | ) = p ( g | f , θ , ) p ( f | θ , ) d θ d f
From that joint posterior distribution, we may also obtain the marginals:
p ( f | g , ) = p ( f , θ | g , ) d f and p ( θ | g , ) = p ( f , θ | g , ) d f
To be more specific, let consider the case of linear inverse problems g = H f + ϵ , . Then, assuming Gaussian noise, we have:
p ( g | f ) = 𝒩 ( g | H f , v ϵ I ) exp [ 1 2 v ϵ g H f 2 2 ]
Assuming a Gaussian prior:
p ( f ) exp [ 1 2 v f f 2 2 ] or exp [ 1 2 v f D f 2 2 ] ,
Then, we see that the posterior is also Gaussian and the MAP and Posterior Mean (PM) estimates become the same and can be computed as the minimizer of : J ( f ) = g H f 2 2 + λ R ( f ) :
p ( f | g ) exp [ 1 2 v ϵ J ( f ) ] f ^ M A P = arg max f { p ( f | g ) } = arg min f { J ( f ) }
In summary, we have:
p ( g | f ) = 𝒩 ( g | H f , v ϵ I ) p ( f ) = 𝒩 ( f | 0 , v f I ) p ( f | g ) = 𝒩 ( f | f ^ , Σ ^ ) f ^ = [ H H + λ I ] 1 H g Σ ^ = v ϵ [ H H + λ I ] 1 , λ = v ϵ v f
For the case where the hyper parameters v ϵ and v f are unknown (Unsupervised case), we can derive the following:
p ( g | f , v ϵ ) = 𝒩 ( g | H f , v ϵ I ) p ( f | v f ) = 𝒩 ( f | 0 , v f I ) p ( v ε ) = 𝒢 ( v f | α ϵ 0 , β ϵ 0 ) p ( v f ) = 𝒢 ( v f | α f 0 , β f 0 ) p ( f | g , v ϵ , v f ) = 𝒩 ( f | f ^ , Σ ^ ) f ^ = [ H H + λ ^ I ] 1 H g Σ ^ = v ^ ϵ [ H H + λ ^ I ] 1 , λ ^ = v ^ ϵ v ^ f p ( v ϵ | g , f ) = 𝒢 ( v ϵ | α ̃ ϵ , β ̃ ϵ ) p ( v f | g , f ) = 𝒢 ( v f | α ̃ f , β ̃ f ) α ̃ ϵ , β ̃ ϵ , α ̃ f , β ̃ f
where the expressions for α ̃ ϵ , β ̃ ϵ , α ̃ f , β ̃ f can be found in [8].
The joint posterior can be written as:
p ( f , v ϵ , v ζ | g ) exp [ J ( f , v ϵ , v ζ ) ]
From this expression, we have different expansion possibilities:
  • JMAP: Alternate optimization with respect to f , v ϵ , v f :
    J ( f , v ϵ , v f ) = 1 2 v ϵ g H f 2 2 + 1 2 v f f 2 2 + ( α ϵ 0 + 1 ) ln v ϵ + β ϵ 0 v ϵ + ( α f 0 + 1 ) ln v f + β f 0 v f
  • Gibbs sampling MCMC:
    f p ( f , v ϵ , v f | g ) v ϵ p ( v ϵ | g , f ) v f p ( v f | g , f )
  • Variational Bayesian Approximation: Approximate p ( f , v ϵ , v f | g ) by a separable one q ( f , v ϵ , v f ) = q 1 ( f ) q 2 ( v ϵ ) q 3 ( v f ) minimizing KL ( q | p ) [8].

7. Imaging inside the Body: From Acquisition to Decision

To introduce the link between the different model based methods and the Machine Learning tools, let consider the case of medical imaging, from the acquisition to the decision steps:
  • Data acquisition:
    Object f Scanner Data g
  • Reconstruction:
    Data g Reconstruction Image f ^
  • Post Processing (Segmentation):
    Image f ^ Segmentation z ^
  • Understanding and Decision:
    Image f ^ Segmentation z ^ Interpretation Decision Tumor or Not Tumor
The questions now are: Can we join any of these steps? Can we go directly from the image to the decision? For the first one, the Bayesian approach can provide a solution:
Data g Reconstruction Segmentation Reconstruction f ^ Segmentation z ^
The main tool here is to introduce a hidden variable which can represent the segmentation. A solution is to introduce a classification hidden variable z with z j = { 1 , 2 , , K } . Then in Figure 3, we have in summary:
A few comments for these relations:
  • p ( g | f , z ) does not depend on z , so it can be written as p ( g | f ) .
  • We may choose a Markovian Potts model for p ( z ) to obtain more compact homogeneous regions [8,9].
  • If we choose for p ( f | z ) a Gaussian law, then p ( f , z | g ) becomes a Gauss-Markov-Potts model [8].
  • We can use the joint posterior p ( f , z | g ) to infer on ( f , z ) : We may just do JMAP: ( f ^ , z ^ ) = arg max { p ( f , z | g ) } or trying to access to the expected posterior values by using the Variational Bayesian Approximation (VBA) techniques [8,10,11,12,13].
This scheme can be extended to consider the estimation of the hyper parameters too. Figure 4 shows this.
Again, here, we can use the joint posterior p ( f , z , θ | g ) to infer on all the unknowns [12].

8. Advantages of the Bayesian Framework

  • Large flexibility of Prior models prior
    -
    Smoothness (Gaussian, Gauss-Markov)
    -
    Direct Sparsity (Double Exp, Heavy-tailed distributions)
    -
    Sparsity in the Transform domain (Double Exp, Heavy-tailed distributions on the WT coefficients)
    -
    Piecewise continuous (DE or Student-t on the gradient)
    -
    Objects composed of only a few materials (Gauss-Markov-Potts), …
  • Possibility of estimating hyper-parameters via JMAP or VBA
  • Natural ways to take account for uncertainties and quantify the remaining uncertainties.

9. Imaging inside the Body for Tumor Detection

  • Reconstruction and Segmentation
    Data g Reconstruction Image f ^ Segmentation z ^
  • Understanding and Decision
    Image f ^ Segmentation z ^ Interpretation Decision Tumor or Not Tumor
  • Can we do all together in a more easily way?
  • Machine Learning and Artificial Intelligence tools may propose solutions
    Data g Machine Learning and Artificial Intelligence Tumor or Not Tumor
  • Learning from a great number of data
From here, we may just do JMAP: ( f ^ , z ^ , θ ^ ) = arg max { p ( f , z , θ | g ) ) , or better, use the Variational Bayesian Approximation (VBA) to do inference.

10. Machine Learning Basic Idea

The main idea in Machine Learning is to learn from a great number of data: ( g i , d i ) , i = 1 , N :
Data ( g i , d i ) i = 1 N Machine Learning and Artificial Intelligence Tumor or Not Tumor
Between the basic tasks we can mention: (a) Classification (Tumor/Not Tumor), (b) Regression (Continuos parameter estimation) and (c) Clustering when the data have not yet labels.
Between the existing ML tools we may mention: Support Vector Machines (SVM), Decision-Tree learning (DT), Artificial Neural Networks (ANN), Bayesian Networks (BN), HMM and Random Forest (RF), Mixture Models (GMM, SMM, …), KNN, Kmeans,…
The frontiers between Image processing, Computer vision, Machine Learning & Artificial intelligence are not very precise as it is shown in Figure 5.
Between the Machine learning tools, we may mention Neural Networks (NN), Artificial NN (ANN), Convolutional NN (CNN), Recurrent NN (RNN), Deep Learning (DL). This last one had shown great success in Speech Recognition, Computer Vision and specifically in Segmentation, Classification and Clustering and in Multi-modality and cross-domain information fusion [14,15,16,17]. However, there are still many limitations: Lack of interpretability, reliability and uncertainty and No reasoning and explaining capabilities [18]. To overcome, there still much to do with the Fundamentals.

11. Interaction between Model Based and Machine Learning Tools

To show the possibilities of the interaction between classical and machine learning, let consider a few examples. The first one is the case of linear inverse problems and quadratic regularization or the Bayesian with Gaussian priors. The solution has an analytic expression:
g = H f + ϵ f ^ = ( H H t + λ D D t ) 1 H t g = B H t g
which can be presented schematically as
g H t B f ^ or directly g CNN or DL f ^
As we can see, this induces directly a linear NN structure. In particular, if H represents a convolution operator, then H t and H t H are too and probably the operator B can also be well approximated by a convolution and the whole inversion can be modelled by a CNN [19].
The second example is the denoising g = f + ϵ with L1 regularizer, or equivalently, the MAP estimator with a double exponential prior, where the solution can be obtained by a convolution followed by a thresholding [20,21].
g = f + ϵ f ^ = 1 λ H t g followed by Thresholding
g H t Thresholding f ^ or directly g CNN or DL f ^
The third example is the Joint Reconstruction and Segmentation that was presented in previous sections. If we present the different steps of reconstruction, segmentation and parameter estimation, we can also compare it with some kind of NN (combination of CNN, RNN and GAN) [22,23].
g Reconstruction f ^ Segmentation z ^ Parameter estimation θ ^

12. Conclusions and Challenges

Signal and image processing, imaging systems and computer vision have made great progress in the last forty years. The first category of the methods was based on linear transformation followed by a thresholding or windowing and coming back. The second generation was model based: forward modeling and inverse problems approach. The main successful approach was based on regularization methods using a combined criterion. The third generation was model based but probabilistic and using the Bayes rule, the so called Bayesian approach. Nowadays, Machine Learning (ML), Neural Networks (NN), Convolutional NN (CNN), Deep Learning (DL) and Artificial Intelligence (AI) methods have obtained great success in classification, clustering, object detection, speech and face recognition, etc. But, they need great number of training data and lack still explanation and they may fail very easily. For signal and image processing or inverse problems, they need still progress. This progress is coming via their interaction with the model based methods. In fact, the successful of CNN and DL methods greatly depends on the appropriate choice of the network structure. This choice can be guided by the model based methods. For inverse problems, when the forward models are not available or too complex, NN and DL may be helpfull. However, we may still need to choose the structure of the NN via approximate forward model and approximate Bayesian inversion.

References

  1. Carre, P.; Andres, E. Discrete analytical ridgelet transform. Signal Process. 2004, 84, 2165–2173. [Google Scholar] [CrossRef]
  2. Dettori, L.; Semler, L. A comparison of wavelet, ridgelet, and curvelet-based texture classification algorithms in computed tomography. Comput. Biol. Med. 2007, 37, 486–498. [Google Scholar] [CrossRef] [PubMed]
  3. Mohammad-Djafari, A. Inverse Problems in Vision and 3D Tomography; John Wiley & Sons: Hoboken, NJ, USA, 2010. [Google Scholar]
  4. Idier, J. Bayesian Approach to Inverse Problems; John Wiley & Sons: New York, NY, USA, 2008. [Google Scholar]
  5. Afonso, M.V.; Bioucas-Dias, J.M.; Figueiredo, M.A. Fast image recovery using variable splitting and constrained optimization. IEEE Trans. Image Process. 2010, 19, 2345–2356. [Google Scholar] [CrossRef] [PubMed]
  6. Bioucas-Dias, J.M.; Figueiredo, M.A. An iterative algorithm for linear inverse problems with compound regularizers. In Porceedings of the 2008 15th IEEE International Conference on Image Processing, San Diego, CA, USA, 12–15 October 2008; pp. 685–688. [Google Scholar]
  7. Florea, M.I.; Vorobyov, S.A. A robust FISTA-like algorithm. In Porceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 4521–4525. [Google Scholar]
  8. Chapdelaine, C.; Mohammad-Djafari, A.; Gac, N.; Parra, E. A 3D Bayesian Computed Tomography Reconstruction Algorithm with Gauss-Markov-Potts Prior Model and its Application to Real Data. Fundam. Informaticae 2017, 155, 373–405. [Google Scholar] [CrossRef]
  9. Féron, O.; Duchêne, B.; Mohammad-Djafari, A. Microwave imaging of inhomogeneous objects made of a finite number of dielectric and conductive materials from experimental data. Inverse Probl. 2005, 21, S95. [Google Scholar] [CrossRef]
  10. Chapdelaine, C.; Mohammad-Djafari, A.; Gac, N.; Parra, E. A Joint Segmentation and Reconstruction Algorithm for 3D Bayesian Computed Tomography Using Gauss-Markov-Potts Prior Model. In Porceedings of the 42nd IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2017), New Orleans, LA, USA, 5–9 March 2017. [Google Scholar]
  11. Chapdelaine, C.; Gac, N.; Mohammad-Djafari, A.; Parra, E. New GPU implementation of Separable Footprint Projector and Backprojector: First results. In Porceedings of the 5th International Conference on Image Formation in X-ray Computed Tomography, Salt Lake City, UT, USA, 20–23 May 2018. [Google Scholar]
  12. Ayasso, H.; Mohammad-Djafari, A. Joint NDT image restoration and segmentation using Gauss–Markov–Potts prior models and variational bayesian computation. IEEE Trans. Image Process. 2010, 19, 2265–2277. [Google Scholar] [CrossRef] [PubMed]
  13. Chapdelaine, C. Variational Bayesian Approach and Gauss-Markov-Potts prior model. arXiv 2018, arXiv:1808.09552. [Google Scholar]
  14. Szegedy, C.; Toshev, A.; Erhan, D. Deep Neural Networks for Object Detection. In Proceedings of the Advances in Neural Information Processing Systems 26 (NIPS 2013), Lake Tahoe, NV, USA, 5–10 December 2013; pp. 2553–2561. [Google Scholar]
  15. Deng, L.; Yu, D. Deep Learning: Methods and Applications; NOW Publishers: Hanover, MA, USA, 2014. [Google Scholar]
  16. Schmidhuber, J. Deep Learning in Neural Networks: An Overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [PubMed]
  17. Schmidhuber, J. The New AI: General & Sound & Relevant for Physics. In Artificial General Intelligence; Goertzel, B., Pennachin, C., Eds.; Springer: Berlin/Heidelberg, Germany, 2006; pp. 175–198. [Google Scholar]
  18. Chellappa, R.; Fukushima, K.; Katsaggelos, A.; Kung, S.Y.; LeCun, Y.; Nasrabadi, N.M.; Poggio, T.A. Applications of Artificial Neural Networks to Image Processing (guest editorial). IEEE Trans. Image Process. 1998, 7, 1093–1097. [Google Scholar] [CrossRef]
  19. Ciresan, D.C.; Meier, U.; Masci, J.; Gambardella, L.M.; Schmidhuber, J. Flexible, High Performance Convolutional Neural Networks for Image Classification. In Proceedings of the Twenty-Second International joint conference on Artificial Intelligence, Barcelona, Spain, 16–22 July 2011; pp. 1237–1242. [Google Scholar]
  20. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems (NIPS 2012), Lake Tahoe, NV, USA, 3–6 December 2012; p. 4. [Google Scholar]
  21. Zeiler, M.D.; Fergus, R. Visualizing and Understanding Convolutional Networks. arXiv 2013, arXiv:1311.2901. [Google Scholar]
  22. Masci, J.; Meier, U.; Ciresan, D.; Schmidhuber, J. Stacked Convolutional Auto-Encoders for Hierarchical Feature Extraction. In Artificial Neural Networks and Machine Learning—ICANN 2011; Honkela, T., Duch, W., Girolami, M.A., Kaski, S., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2011; Volume 6791, pp. 52–59. [Google Scholar]
  23. LeCun, Y.; Bengio, Y. Convolutional Networks for Images, Speech, and Time-Series. In The Handbook of Brain Theory and Neural Networks; Arbib, M.A., Ed.; MIT Press: Cambridge, MA, USA, 1995. [Google Scholar]
Figure 1. Transform methods.
Figure 1. Transform methods.
Proceedings 33 00016 g001
Figure 2. Model based methods.
Figure 2. Model based methods.
Proceedings 33 00016 g002
Figure 3. Bayesian approach for joint reconstruction and segmentation.
Figure 3. Bayesian approach for joint reconstruction and segmentation.
Proceedings 33 00016 g003
Figure 4. Advanced Bayesian approach for joint reconstruction and segmentation.
Figure 4. Advanced Bayesian approach for joint reconstruction and segmentation.
Proceedings 33 00016 g004
Figure 5. Frontiers between Image processing (IP), Computer vision (CV), Machine Learning (ML) and Artificial intelligence (AI).
Figure 5. Frontiers between Image processing (IP), Computer vision (CV), Machine Learning (ML) and Artificial intelligence (AI).
Proceedings 33 00016 g005

Share and Cite

MDPI and ACS Style

Mohammad-Djafari, A. Interaction between Model Based Signal and Image Processing, Machine Learning and Artificial Intelligence . Proceedings 2019, 33, 16. https://doi.org/10.3390/proceedings2019033016

AMA Style

Mohammad-Djafari A. Interaction between Model Based Signal and Image Processing, Machine Learning and Artificial Intelligence . Proceedings. 2019; 33(1):16. https://doi.org/10.3390/proceedings2019033016

Chicago/Turabian Style

Mohammad-Djafari, Ali. 2019. "Interaction between Model Based Signal and Image Processing, Machine Learning and Artificial Intelligence " Proceedings 33, no. 1: 16. https://doi.org/10.3390/proceedings2019033016

APA Style

Mohammad-Djafari, A. (2019). Interaction between Model Based Signal and Image Processing, Machine Learning and Artificial Intelligence . Proceedings, 33(1), 16. https://doi.org/10.3390/proceedings2019033016

Article Metrics

Back to TopTop