Next Article in Journal
An Innovative Image Encryption Algorithm Based on the DNAS_box and Hyperchaos
Previous Article in Journal
In-Motion Initial Alignment Method Based on Multi-Source Information Fusion for Special Vehicles
Previous Article in Special Issue
Generative Bayesian Computation for Maximum Expected Utility
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Editorial

Deep Generative Modeling: From Probabilistic Framework to Generative AI

Chan Zuckerberg Initiative, Redwood City, CA 94063, USA
Entropy 2025, 27(3), 238; https://doi.org/10.3390/e27030238
Submission received: 25 February 2025 / Accepted: 25 February 2025 / Published: 26 February 2025
(This article belongs to the Special Issue Deep Generative Modeling: Theory and Applications)

1. Why Deep Generative Modeling?

Large Language Models (LLMs) have unlocked a new frontier in AI applications, significantly advancing the field of generative modeling. Beyond assisting with tasks such as preparing a presentation plan, LLMs have fundamentally reshaped human–computer interactions by enabling seamless communication between machines and agents through natural language. This paradigm shift has not only driven a technological leap but has also transformed how we conceptualize AI and its role in solving complex problems.
The recent advancements in Generative AI (GenAI), such as Vision-Language Models (VLMs), exemplified by systems like GPT-4V (GPT-4 with Vision) [1] and Google’s Gemini [2], have brought multimodal learning back into the spotlight. These developments underscore the versatility of LLMs, which now form the core of multimodal systems capable of processing and generating diverse data types—including text, images, audio, and video. The fusion of modalities within these systems has set the stage for a new generation of AI termed Generative AI Systems (GenAISys) [3].
GenAISys are holistic systems where natural language serves as the universal medium for communication, replacing predefined formal protocols. These systems integrate modality encoders as I/O interfaces to handle various data sources. Furthermore, they incorporate databases, knowledge graphs, and external specialized tools—such as calculators, navigation apps, or molecular drawing utilities—that communicate with the system through dedicated modules for information retrieval and storage. While GenAI models remain central to GenAISys, they function as components within a broader, more interconnected framework rather than standalone entities.
For instance, a tool-augmented Seq2Seq model proposed by [4] uses a fine-tuned T5 LLM as a backbone to interface with external tools. Similarly, ref. [5] illustrates a GenAISys where an LLM collaborates with tools and databases to answer chemistry-related questions and generate molecular diagrams. These examples highlight the emergent capabilities of GenAISys in addressing domain-specific challenges while leveraging the general-purpose strengths of LLMs.
However, the increasing complexity of these systems necessitates a shift in perspective. Instead of viewing GenAISys as collections of separate modules, we must analyze them as cohesive systems. This shift invites us to examine their compositionality [6,7,8,9], i.e., how individual components interact to produce coherent and reliable outcomes. Insights from compositionality studies can guide the design of robust and verifiable systems, drawing parallels to disciplines like control theory and systems engineering [10].
In addition to compositionality, grounding generative systems in robust theoretical foundations is essential. This involves leveraging modeling principles that capture the generative processes underlying observed phenomena. Such processes define a joint distribution over random variables and their stochastic interactions, specifying how events occur and in what sequence. To create flexible, efficient, and effective modeling frameworks, these distributions can be parameterized using deep neural networks capable of processing raw data.
By modeling stochastic dependencies through probability theory, we ensure a rigorous and principled approach, minimizing potential reasoning errors. Probability theory also offers a unified framework in which the likelihood function is central to quantifying uncertainty and defining objective functions. This approach, known as deep generative modeling [11], integrates these concepts into a cohesive framework, providing a powerful foundation for the development of Generative AI Systems.
Moreover, addressing practical concerns such as scalability, interpretability, and ethical considerations is also crucial. Applying these principles to GenAISys can ensure they meet the high expectations placed upon them while avoiding potential pitfalls.
Looking ahead, the exploration of GenAISys offers an opportunity to refine our understanding of AI systems as interconnected ecosystems. By investigating their compositional, functional, and ethical dimensions, we can design more effective, reliable, and human-centric AI solutions, pushing the boundaries of what is achievable in this exciting era of Generative AI.

2. Formulating and Applying Deep Generative Modeling

There are multiple classes of (deep) generative models, namely, mixture models, Probabilistic Circuits, Autoregressive Models (and their special example, Large Language Models), Flow-based Models, Latent Variable Models, GANs, Hybrid (or Joint) Models, Score-based Generative Models, Diffusion-based Models, and Energy-based Models. All these classes utilize and rely on various probabilistic principles, e.g., normalizing flows are built upon the change in variable formula. However, their formulation depends on deep neural network architectures. For instance, Large Language Models are Autoregressive Models parameterized by transformers or recurrent neural networks, while normalizing flows require invertible deep neural networks.
The objective of this Special Issue is to provide the latest advances in Deep Generative Modeling both in theory and applications. In the following articles, various deep generative models are presented (e.g., diffusion models, Variational Auto-Encoders, normalizing flows, etc.) and applied to applications like image and video generation, tabular data processing, and microRNA generation. Additionally, some studies consider the theoretical aspects of these models, such as the number of steps in diffusion-based models, a connection between diffusion-based models and associative memory, or the new Generative Bayesian Framework.
The future seems significantly dependent on Deep Generative Modeling and Generative AI Systems that comprise the inevitable next steps in the evolution of AI. They will assist in many jobs, ranging from office jobs, healthcare, and education to industries like manufacturing. Although there are many other aspects like embodied AI [12] or cyber–physical systems (with humans) [9], GenAISys will still be necessary for formulating an artificial “brain” and/or GenAI-based agents and apps.

3. About This Special Issue

This Special Issue “Deep Generative Modeling: Theory and Applications” presents a collection of research papers that explore advancements in deep generative models and their diverse applications. Notable contributions include the development of Generative Bayesian Computation methods for efficient maximum-expected utility calculations, the use of Variational Auto-Encoders for generating descriptions for precursor microRNA aiding in gene regulation studies, and the introduction of Dimma, a semi-supervised approach for enhancing low-light images with natural colors. This Special Issue also features NodeFlow, a framework combining Neural Oblivious Decision Ensembles and Conditional Continuous Normalizing Flows for probabilistic regression on tabular data, and studies on diffusion-based causal representation learning, highlighting the theoretical and practical advancements in the field of deep generative modeling. Moreover, this collection features other notable papers advancing the field of deep generative models. The study on diffusion models as associative memory networks explores the connection between generative diffusion models and associative memory networks, proposing a framework that unifies creative generation and memory recall processes. Other significant contributions include studies on multi-modal latent diffusion, semi-supervised variational autoencoders for out-of-distribution generation, diffusion probabilistic modeling for video generation, learning energy-based models in high-dimensional spaces with multiscale denoising-score matching, and investigations into optimal diffusion times in score-based generative models. These works collectively enhance our understanding and application of deep generative modeling techniques.

Acknowledgments

The author would like to express his great appreciation to Zdzisław Bubnicki (R.I.P) and Jerzy Świątek who made a great impact on the author and his systems-based perspective on AI.

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. OpenAI Team. GPT-4V(ision) System Card. 2023. Available online: https://openai.com/index/gpt-4v-system-card/ (accessed on 24 February 2025).
  2. Team Gemini; Anil, R.; Borgeaud, S.; Wu, Y.; Alayrac, J.B.; Yu, J.; Soricut, R.; Schalkwyk, J.; Dai, A.M.; Hauth, A.; et al. Gemini: A family of highly capable multimodal models. arXiv 2023, arXiv:2312.11805. [Google Scholar]
  3. Tomczak, J.M. Generative AI Systems: A Systems-based Perspective on Generative AI. arXiv 2024, arXiv:2407.11001. [Google Scholar]
  4. Parisi, A.; Zhao, Y.; Fiedel, N. Talm: Tool augmented language models. arXiv 2022, arXiv:2205.12255. [Google Scholar]
  5. Bran, A.M.; Cox, S.; Schilter, O.; Baldassari, C.; White, A.; Schwaller, P. Augmenting large language models with chemistry tools. In Proceedings of the NeurIPS 2023 AI for Science Workshop, New Orleans, LA, USA, 6 October 2023. [Google Scholar]
  6. Bereska, L.; Gavves, E. Mechanistic Interpretability for AI Safety–A Review. arXiv 2024, arXiv:2404.14082. [Google Scholar]
  7. Cohen, T. Towards a grounded theory of causation for embodied ai. arXiv 2022, arXiv:2206.13973. [Google Scholar]
  8. Swan, J.; Nivel, E.; Kant, N.; Hedges, J.; Atkinson, T.; Steunebrink, B. The Road to General Intelligence; Springer Nature: Cham, Switzerland, 2022. [Google Scholar]
  9. Tripakis, S. Compositionality in the science of system design. Proc. IEEE 2016, 104, 960–972. [Google Scholar] [CrossRef]
  10. Bubnicki, Z. Modern Control Theory; Springer: Berlin/Heidelberg, Germany, 2005; Volume 2005925392. [Google Scholar]
  11. Tomczak, J.M. Deep Generative Modeling; Springer: Cham, Switzerland, 2024. [Google Scholar]
  12. Roy, N.; Posner, I.; Barfoot, T.; Beaudoin, P.; Bengio, Y.; Bohg, J.; Brock, O.; Depatie, I.; Fox, D.; Koditschek, D.; et al. From machine learning to robotics: Challenges and opportunities for embodied intelligence. arXiv 2021, arXiv:2110.15245. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tomczak, J.M. Deep Generative Modeling: From Probabilistic Framework to Generative AI. Entropy 2025, 27, 238. https://doi.org/10.3390/e27030238

AMA Style

Tomczak JM. Deep Generative Modeling: From Probabilistic Framework to Generative AI. Entropy. 2025; 27(3):238. https://doi.org/10.3390/e27030238

Chicago/Turabian Style

Tomczak, Jakub M. 2025. "Deep Generative Modeling: From Probabilistic Framework to Generative AI" Entropy 27, no. 3: 238. https://doi.org/10.3390/e27030238

APA Style

Tomczak, J. M. (2025). Deep Generative Modeling: From Probabilistic Framework to Generative AI. Entropy, 27(3), 238. https://doi.org/10.3390/e27030238

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop