Bridging the Gap between Deep Learning and Probabilistic Inference for Advancements in Robotics

A special issue of Electronics (ISSN 2079-9292). This special issue belongs to the section "Systems & Control Engineering".

Deadline for manuscript submissions: 15 July 2025 | Viewed by 5037

Special Issue Editors

School of Computer Science, The University of Sydney, Sydney, NSW 2006, Australia
Interests: machine learning; robot learning; motion planning; multi-agent systems; probabilistic inference

E-Mail Website
Guest Editor
School of Computing, University of Portsmouth, Portsmouth PO1 3HE, UK
Interests: machine learning; pattern recognition; robotics
Special Issues, Collections and Topics in MDPI journals

Special Issue Information

Dear Colleagues,

In recent times, the field of robotics has witnessed remarkable advancements propelled by the integration of artificial intelligence (AI) and machine learning (ML), particularly deep learning, into robotic systems. These developments have propelled robots to accomplish tasks with unprecedented levels of performance. This has spurred a reconsideration of the traditional probabilistic inference algorithms that have long been relied upon for reliable operation in uncertain and unstructured environments. These probabilistic techniques offer a comprehensive framework that unifies perception, control, and learning in robotics.

Concurrently, the paradigm of robot learning has gained significant momentum. It holds the promise of enabling robots to generalize their capabilities across a spectrum of scenarios, mitigating the necessity for the meticulous engineering of task-specific models, which is a hallmark of classic probabilistic methods. Yet, a fundamental question persists: can we entrust robots with dependable and adaptive behaviors solely through data-driven learning approaches?

Addressing this question lies at the heart of this Special Issue: Bridging the Gap between Deep Learning and Probabilistic Inference in Robotics. As the robotics landscape evolves, there is a growing recognition that extracting the best elements from both deep learning and probabilistic inference could hold the key to unlocking new levels of robotic performance.

This Special Issue serves as a platform to delve into this juncture where robotics, deep learning, and probabilistic inference converge. It aims to foster dialogue among researchers who are navigating the complexities of merging these diverse methodologies. By uniting experts from these distinct fields, this issue aims to push the boundaries of what is possible in intelligent robotics.

Key Objectives and Themes:

  • Learned Components in Probabilistic Contexts: A recent breakthrough in this domain has been the introduction of learned components within probabilistic frameworks. This issue seeks to explore the potential of such integration to enhance system adaptability.
  • Unpacking Epistemic Uncertainty: Epistemic uncertainty, often referred to as model uncertainty, encapsulates what we do not know about the underlying data distribution. It arises when the available data are insufficient to fully characterize the intricacies of the environment. In the context of robot learning, this uncertainty could manifest as gaps in the training data, variations in sensor measurements, or unmodeled dynamics of the robot and its surroundings.
  • End-to-End Differentiable Algorithms: The emergence of end-to-end differentiable algorithms is a significant milestone. This issue will delve into the implications of these algorithms on robust and reliable learning systems.
  • Bayesian inference for Probabilistic Reasoning: Recent advancements have propelled Bayesian inference to new heights within the field of robotics. This paves the way for seamlessly integrating learning and reasoning, potentially bridging the gap between the adaptability of deep learning and the robustness of probabilistic reasoning.
  • Adaptability for Robust Control: Adaptability in an integrated robotic system is often the most crucial factor to ensure robustness, even in the mist of uncertainty. The ultimate goal is to cultivate robotic learning systems that not only learn from data but also adapt to novel situations effectively.
  • Learning to Plan and Planning to Learn: The integration of robot learning and planning is a two-way street. On the one hand, robots can utilize data-driven insights to make more informed decisions and optimize action sequences. On the other hand, planning can be employed to facilitate the learning process. Robots can devise deliberate actions to gather informative data, thereby accelerating the learning curve and refining their understanding of the environment.

Dr. Tin Lai
Prof. Dr. Zhaojie Ju
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Electronics is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2400 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • robot learning
  • deep learning in robotics
  • probabilistic robotics

Benefits of Publishing in a Special Issue

  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • Reprint: MDPI Books provides the opportunity to republish successful Special Issues in book format, both online and in print.

Further information on MDPI's Special Issue policies can be found here.

Published Papers (4 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

16 pages, 695 KiB  
Article
Hierarchical Early Wireless Forest Fire Prediction System Utilizing Virtual Sensors
by Ahshanul Haque and Hamdy Soliman
Electronics 2025, 14(8), 1634; https://doi.org/10.3390/electronics14081634 - 18 Apr 2025
Viewed by 274
Abstract
Deploying thousands of sensors across remote and challenging environments—such as the Amazon rainforest, Californian wilderness, or Australian bushlands—is a critical yet complex task for forest fire monitoring, while our backyard emulation confirmed the feasibility of small-scale deployment as a proof of concept, large-scale [...] Read more.
Deploying thousands of sensors across remote and challenging environments—such as the Amazon rainforest, Californian wilderness, or Australian bushlands—is a critical yet complex task for forest fire monitoring, while our backyard emulation confirmed the feasibility of small-scale deployment as a proof of concept, large-scale scenarios demand a scalable, efficient, and fault-tolerant network design. This paper proposes a Hierarchical Wireless Sensor Network (HWSN) deployment strategy with adaptive head node selection to maximize area coverage and energy efficiency. The network architecture follows a three-level hierarchy as follows: The first level incorporates cells of individual sensor nodes that connect to dynamically assigned cell heads. The second level involves the aggregated clusters of such cell heads, each with an assigned cluster head. Finally, dividing all cluster heads into regions, each with a region head, directly reports all the collected information from the forest floor to a central control sink room for decision making analysis. Unlike traditional centralized or uniformly distributed models, our adaptive approach leverages a greedy coverage maximization algorithm to dynamically select head nodes that contribute to the best forest sensed data coverage at each level. Through extensive simulations, the adaptive model achieved over 96.26% coverage, using significantly fewer nodes, while reducing node transmission distances and energy consumption. This facilitates the real-world deployment of our HWSN model in large-scale, remote forest regions, with very promising performance. Full article
Show Figures

Figure 1

23 pages, 2516 KiB  
Article
Knitting Robots: A Deep Learning Approach for Reverse-Engineering Fabric Patterns
by Haoliang Sheng, Songpu Cai, Xingyu Zheng and Mengcheng Lau
Electronics 2025, 14(8), 1605; https://doi.org/10.3390/electronics14081605 - 16 Apr 2025
Viewed by 1391
Abstract
Knitting, a cornerstone of textile manufacturing, is uniquely challenging to automate, particularly in terms of converting fabric designs into precise, machine-readable instructions. This research bridges the gap between textile production and robotic automation by proposing a novel deep learning-based pipeline for reverse knitting [...] Read more.
Knitting, a cornerstone of textile manufacturing, is uniquely challenging to automate, particularly in terms of converting fabric designs into precise, machine-readable instructions. This research bridges the gap between textile production and robotic automation by proposing a novel deep learning-based pipeline for reverse knitting to integrate vision-based robotic systems into textile manufacturing. The pipeline employs a two-stage architecture, enabling robots to first identify front labels before inferring complete labels, ensuring accurate, scalable pattern generation. By incorporating diverse yarn structures, including single-yarn (sj) and multi-yarn (mj) patterns, this study demonstrates how our system can adapt to varying material complexities. Critical challenges in robotic textile manipulation, such as label imbalance, underrepresented stitch types, and the need for fine-grained control, are addressed by leveraging specialized deep-learning architectures. This work establishes a foundation for fully automated robotic knitting systems, enabling customizable, flexible production processes that integrate perception, planning, and actuation, thereby advancing textile manufacturing through intelligent robotic automation. Full article
Show Figures

Figure 1

20 pages, 37875 KiB  
Article
Unsupervised Domain Adaptation Semantic Segmentation of Remote Sensing Imagery with Scene Covariance Alignment
by Kangjian Cao, Sheng Wang, Ziheng Wei, Kexin Chen, Runlong Chang and Fu Xu
Electronics 2024, 13(24), 5022; https://doi.org/10.3390/electronics13245022 - 20 Dec 2024
Viewed by 944
Abstract
Remote sensing imagery (RSI) segmentation plays a crucial role in environmental monitoring and geospatial analysis. However, in real-world practical applications, the domain shift problem between the source domain and target domain often leads to severe degradation of model performance. Most existing unsupervised domain [...] Read more.
Remote sensing imagery (RSI) segmentation plays a crucial role in environmental monitoring and geospatial analysis. However, in real-world practical applications, the domain shift problem between the source domain and target domain often leads to severe degradation of model performance. Most existing unsupervised domain adaptation methods focus on aligning global-local domain features or category features, neglecting the variations of ground object categories within local scenes. To capture these variations, we propose the scene covariance alignment (SCA) approach to guide the learning of scene-level features in the domain. Specifically, we propose a scene covariance alignment model to address the domain adaptation challenge in RSI segmentation. Unlike traditional global feature alignment methods, SCA incorporates a scene feature pooling (SFP) module and a covariance regularization (CR) mechanism to extract and align scene-level features effectively and focuses on aligning local regions with different scene characteristics between source and target domains. Experiments on both the LoveDA and Yanqing land cover datasets demonstrate that SCA exhibits excellent performance in cross-domain RSI segmentation tasks, particularly outperforming state-of-the-art baselines across various scenarios, including different noise levels, spatial resolutions, and environmental conditions. Full article
Show Figures

Figure 1

22 pages, 6193 KiB  
Article
Lightweight UAV Object-Detection Method Based on Efficient Multidimensional Global Feature Adaptive Fusion and Knowledge Distillation
by Jian Sun, Hongwei Gao, Zhiwen Yan, Xiangjing Qi, Jiahui Yu and Zhaojie Ju
Electronics 2024, 13(8), 1558; https://doi.org/10.3390/electronics13081558 - 19 Apr 2024
Cited by 6 | Viewed by 1757
Abstract
Unmanned aerial vehicles (UAVs) equipped with remote-sensing object-detection devices are increasingly employed across diverse domains. However, the detection of small, densely-packed objects against complex backgrounds and at various scales presents a formidable challenge to conventional detection algorithms, exacerbated by the computational constraints of [...] Read more.
Unmanned aerial vehicles (UAVs) equipped with remote-sensing object-detection devices are increasingly employed across diverse domains. However, the detection of small, densely-packed objects against complex backgrounds and at various scales presents a formidable challenge to conventional detection algorithms, exacerbated by the computational constraints of UAV-embedded systems that necessitate a delicate balance between detection speed and accuracy. To address these issues, this paper proposes the Efficient Multidimensional Global Feature Adaptive Fusion Network (MGFAFNET), an innovative detection method for UAV platforms. The novelties of our approach are threefold: Firstly, we introduce the Dual-Branch Multidimensional Aggregation Backbone Network (DBMA), an efficient architectural innovation that captures multidimensional global spatial interactions, significantly enhancing feature distinguishability for complex and occluded targets. Simultaneously, it reduces the computational burden typically associated with processing high-resolution imagery. Secondly, we construct the Dynamic Spatial Perception Feature Fusion Network (DSPF), which is tailored specifically to accommodate the notable scale variances encountered during UAV operation. By implementing a multi-layer dynamic spatial fusion coupled with feature-refinement modules, the network adeptly minimizes informational redundancy, leading to more efficient feature representation. Finally, our novel Localized Compensation Dual-Mask Distillation (LCDD) strategy is devised to adeptly translate the rich local and global features from the higher-capacity teacher network to the more resource-constrained student network, capturing both low-level spatial details and high-level semantic cues with unprecedented efficacy. The practicability and superior performance of our MGFAFNET are corroborated by a dedicated UAV detection platform, showcasing remarkable improvements over state-of-the-art object-detection methods, as demonstrated by rigorous evaluations conducted using the VisDrone2021 benchmark and a meticulously assembled proprietary dataset. Full article
Show Figures

Figure 1

Back to TopTop