Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (26)

Search Parameters:
Keywords = learning from demonstration (LfD)

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
24 pages, 4913 KiB  
Article
Region-Wise Recognition and Classification of Arabic Dialects and Vocabulary: A Deep Learning Approach
by Fawaz S. Al–Anzi and Bibin Shalini Sundaram Thankaleela
Appl. Sci. 2025, 15(12), 6516; https://doi.org/10.3390/app15126516 - 10 Jun 2025
Viewed by 632
Abstract
This article presents a unique approach to Arabic dialect identification using a pre-trained speech classification model. The system categorizes Arabic audio clips into their respective dialects by employing 1D and 2D convolutional neural network technologies built from diverse dialects from the Arab region [...] Read more.
This article presents a unique approach to Arabic dialect identification using a pre-trained speech classification model. The system categorizes Arabic audio clips into their respective dialects by employing 1D and 2D convolutional neural network technologies built from diverse dialects from the Arab region using deep learning models. Its objective is to enhance traditional linguistic handling and speech technology by accurately classifying Arabic audio clips into their corresponding dialects. The techniques involved include record gathering, preprocessing, feature extraction, prototypical architecture, and assessment metrics. The algorithm distinguishes various Arabic dialects, such as A (Arab nation authorized dialectal), EGY (Egyptian Arabic), GLF (Gulf Arabic), LAV and LF (Levantine Arabic, spoken in Syria, Lebanon, and Jordan), MSA (Modern Standard Arabic), NOR (North African Arabic), and SA (Saudi Arabic). Experimental results demonstrate the efficiency of the proposed approach in accurately determining diverse Arabic dialects, achieving a testing accuracy of 94.28% and a validation accuracy of 95.55%, surpassing traditional machine learning models such as Random Forest and SVM and advanced erudition models such as CNN and CNN2D. Full article
(This article belongs to the Special Issue Speech Recognition and Natural Language Processing)
Show Figures

Figure 1

16 pages, 8812 KiB  
Article
Trajectory Learning Using HMM: Towards Surgical Robotics Implementation
by Juliana Manrique-Cordoba, Carlos Martorell-Llobregat, Miguel Ángel de la Casa-Lillo and José María Sabater-Navarro
Sensors 2025, 25(11), 3487; https://doi.org/10.3390/s25113487 - 31 May 2025
Viewed by 448
Abstract
Autonomy represents one of the most promising directions in the future development of surgical robotics, and Learning from Demonstration (LfD) is a key methodology for advancing technologies in this field. The proposed approach extends the classical Douglas–Peucker algorithm by incorporating multidimensional trajectory data, [...] Read more.
Autonomy represents one of the most promising directions in the future development of surgical robotics, and Learning from Demonstration (LfD) is a key methodology for advancing technologies in this field. The proposed approach extends the classical Douglas–Peucker algorithm by incorporating multidimensional trajectory data, including both kinematic and dynamic information. This enhancement enables a more comprehensive representation of demonstrated trajectories, improving generalization in high-dimensional spaces. This representation allows clearer codification and interpretation of the information used in the learning process. A series of experiments were designed to validate this methodology. Motion data and force interaction data were collected, preprocessed, and used to train a hidden Markov model (HMM). Different experimental conditions were analyzed, comparing training using only motion data versus incorporating force interaction data. The results demonstrate that including interaction forces improves trajectory reconstruction accuracy, achieving a lower root mean squared error (RMSE) of 0.29 mm, compared to 0.44 mm for the model trained solely on motion data. These findings support the proposed method as an effective strategy for encoding, simplifying, and learning robotic trajectories in surgical applications. Full article
(This article belongs to the Special Issue Sensor Technology in Robotic Surgery)
Show Figures

Figure 1

14 pages, 974 KiB  
Article
N-Dimensional Reduction Algorithm for Learning from Demonstration Path Planning
by Juliana Manrique-Cordoba, Miguel Ángel de la Casa-Lillo and José María Sabater-Navarro
Sensors 2025, 25(7), 2145; https://doi.org/10.3390/s25072145 - 28 Mar 2025
Cited by 1 | Viewed by 379
Abstract
This paper presents an n-dimensional reduction algorithm for Learning from Demonstration (LfD) for robotic path planning, addressing the complexity of high-dimensional data. The method extends the Douglas–Peucker algorithm by incorporating velocity and orientation alongside position, enabling more precise trajectory simplification. A magnitude-based [...] Read more.
This paper presents an n-dimensional reduction algorithm for Learning from Demonstration (LfD) for robotic path planning, addressing the complexity of high-dimensional data. The method extends the Douglas–Peucker algorithm by incorporating velocity and orientation alongside position, enabling more precise trajectory simplification. A magnitude-based normalization process preserves proportional relationships across dimensions, and the reduced dataset is used to train Hidden Markov Models (HMMs), where continuous trajectories are discretized into identifier sequences. The algorithm is evaluated in 2D and 3D environments with datasets combining position and velocity. The results show that incorporating additional dimensions significantly enhances trajectory simplification while preserving key information. Additionally, the study highlights the importance of selecting appropriate encoding parameters to achieve optimal resolution. The HMM-based models generated new trajectories that retained the patterns of the original demonstrations, demonstrating the algorithm’s capacity to generalize learned behaviors for trajectory learning in high-dimensional spaces. Full article
(This article belongs to the Special Issue Advances in Sensing, Control and Path Planning for Robotic Systems)
Show Figures

Figure 1

29 pages, 9718 KiB  
Article
Segment, Compare, and Learn: Creating Movement Libraries of Complex Task for Learning from Demonstration
by Adrian Prados, Gonzalo Espinoza, Luis Moreno and Ramon Barber
Biomimetics 2025, 10(1), 64; https://doi.org/10.3390/biomimetics10010064 - 17 Jan 2025
Viewed by 1447
Abstract
Motion primitives are a highly useful and widely employed tool in the field of Learning from Demonstration (LfD). However, obtaining a large number of motion primitives can be a tedious process, as they typically need to be generated individually for each task to [...] Read more.
Motion primitives are a highly useful and widely employed tool in the field of Learning from Demonstration (LfD). However, obtaining a large number of motion primitives can be a tedious process, as they typically need to be generated individually for each task to be learned. To address this challenge, this work presents an algorithm for acquiring robotic skills through automatic and unsupervised segmentation. The algorithm divides tasks into simpler subtasks and generates motion primitive libraries that group common subtasks for use in subsequent learning processes. Our algorithm is based on an initial segmentation step using a heuristic method, followed by probabilistic clustering with Gaussian Mixture Models. Once the segments are obtained, they are grouped using Gaussian Optimal Transport on the Gaussian Processes (GPs) of each segment group, comparing their similarities through the energy cost of transforming one GP into another. This process requires no prior knowledge, it is entirely autonomous, and supports multimodal information. The algorithm enables generating trajectories suitable for robotic tasks, establishing simple primitives that encapsulate the structure of the movements to be performed. Its effectiveness has been validated in manipulation tasks with a real robot, as well as through comparisons with state-of-the-art algorithms. Full article
(This article belongs to the Special Issue Bio-Inspired and Biomimetic Intelligence in Robotics: 2nd Edition)
Show Figures

Figure 1

14 pages, 13034 KiB  
Article
Learning Underwater Intervention Skills Based on Dynamic Movement Primitives
by Xuejiao Yang, Yunxiu Zhang, Rongrong Li, Xinhui Zheng and Qifeng Zhang
Electronics 2024, 13(19), 3860; https://doi.org/10.3390/electronics13193860 - 29 Sep 2024
Viewed by 980
Abstract
Improving the autonomy of underwater interventions by remotely operated vehicles (ROVs) can help mitigate the impact of communication delays on operational efficiency. Currently, underwater interventions for ROVs usually rely on real-time teleoperation or preprogramming by operators, which is not only time-consuming and increases [...] Read more.
Improving the autonomy of underwater interventions by remotely operated vehicles (ROVs) can help mitigate the impact of communication delays on operational efficiency. Currently, underwater interventions for ROVs usually rely on real-time teleoperation or preprogramming by operators, which is not only time-consuming and increases the cognitive burden on operators but also requires extensive specialized programming. Instead, this paper uses the intuitive learning from demonstrations (LfD) approach that uses operator demonstrations as inputs and models the trajectory characteristics of the task through the dynamic movement primitive (DMP) approach for task reproduction as well as the generalization of knowledge to new environments. Unlike existing applications of DMP-based robot trajectory learning methods, we propose the underwater DMP (UDMP) method to address the problem that the complexity and stochasticity of underwater operational environments (e.g., current perturbations and floating operations) diminish the representativeness of the demonstrated trajectories. First, the Gaussian mixture model (GMM) and Gaussian mixture regression (GMR) are used for feature extraction of multiple demonstration trajectories to obtain typical trajectories as inputs to the DMP method. The UDMP method is more suitable for the LfD of underwater interventions than the method that directly learns the nonlinear terms of the DMP. In addition, we improve the commonly used homomorphic-based teleoperation mode to heteromorphic mode, which allows the operator to focus more on the end-operation task. Finally, the effectiveness of the developed method is verified by simulation experiments. Full article
Show Figures

Figure 1

26 pages, 4238 KiB  
Article
PRF: A Program Reuse Framework for Automated Programming by Learning from Existing Robot Programs
by Tyler Toner, Dawn M. Tilbury and Kira Barton
Robotics 2024, 13(8), 118; https://doi.org/10.3390/robotics13080118 - 6 Aug 2024
Viewed by 1418
Abstract
This paper explores the problem of automated robot program generation from limited historical data when neither accurate geometric environmental models nor online vision feedback are available. The Program Reuse Framework (PRF) is developed, which uses expert-defined motion classes, a novel data structure [...] Read more.
This paper explores the problem of automated robot program generation from limited historical data when neither accurate geometric environmental models nor online vision feedback are available. The Program Reuse Framework (PRF) is developed, which uses expert-defined motion classes, a novel data structure introduced in this work, to learn affordances, workspaces, and skills from historical data. Historical data comprise raw robot joint trajectories and descriptions of the robot task being completed. Given new tasks, motion classes are then used again to formulate an optimization problem capable of generating new open-loop, skill-based programs to complete the tasks. To cope with a lack of geometric models, a technique to learn safe workspaces from demonstrations is developed, allowing the risk of new programs to be estimated before execution. A new learnable motion primitive for redundant manipulators is introduced, called a redundancy dynamical movement primitive, which enables new end-effector goals to be reached while mimicking the whole-arm behavior of a demonstration. A mobile manipulator part transportation task is used throughout to illustrate each step of the framework. Full article
(This article belongs to the Section Industrial Robots and Automation)
Show Figures

Figure 1

33 pages, 3278 KiB  
Review
A Practical Roadmap to Learning from Demonstration for Robotic Manipulators in Manufacturing
by Alireza Barekatain, Hamed Habibi and Holger Voos
Robotics 2024, 13(7), 100; https://doi.org/10.3390/robotics13070100 - 10 Jul 2024
Cited by 2 | Viewed by 4112
Abstract
This paper provides a structured and practical roadmap for practitioners to integrate learning from demonstration (LfD) into manufacturing tasks, with a specific focus on industrial manipulators. Motivated by the paradigm shift from mass production to mass customization, it is crucial to have an [...] Read more.
This paper provides a structured and practical roadmap for practitioners to integrate learning from demonstration (LfD) into manufacturing tasks, with a specific focus on industrial manipulators. Motivated by the paradigm shift from mass production to mass customization, it is crucial to have an easy-to-follow roadmap for practitioners with moderate expertise, to transform existing robotic processes to customizable LfD-based solutions. To realize this transformation, we devise the key questions of “What to Demonstrate”, “How to Demonstrate”, “How to Learn”, and “How to Refine”. To follow through these questions, our comprehensive guide offers a questionnaire-style approach, highlighting key steps from problem definition to solution refinement. This paper equips both researchers and industry professionals with actionable insights to deploy LfD-based solutions effectively. By tailoring the refinement criteria to manufacturing settings, this paper addresses related challenges and strategies for enhancing LfD performance in manufacturing contexts. Full article
(This article belongs to the Special Issue Integrating Robotics into High-Accuracy Industrial Operations)
Show Figures

Figure 1

23 pages, 11021 KiB  
Article
A Trajectory Optimisation-Based Incremental Learning Strategy for Learning from Demonstration
by Yuqi Wang, Weidong Li and Yuchen Liang
Appl. Sci. 2024, 14(11), 4943; https://doi.org/10.3390/app14114943 - 6 Jun 2024
Cited by 2 | Viewed by 1532
Abstract
The insufficient generalisation capability of the conventional learning from demonstration (LfD) model necessitates redemonstrations. In addition, retraining the model can overwrite existing knowledge, making it impossible to perform previously acquired skills in new application scenarios. These are not economical and efficient. To address [...] Read more.
The insufficient generalisation capability of the conventional learning from demonstration (LfD) model necessitates redemonstrations. In addition, retraining the model can overwrite existing knowledge, making it impossible to perform previously acquired skills in new application scenarios. These are not economical and efficient. To address the issues, in this study, a broad learning system (BLS) and probabilistic roadmap (PRM) are integrated with dynamic movement primitive (DMP)-based LfD. Three key innovations are proposed in this paper: (1) segmentation and extended demonstration: a 1D-based topology trajectory segmentation algorithm (1D-SEG) is designed to divide the original demonstration into several segments. Following the segmentation, a Gaussian probabilistic roadmap (G-PRM) is proposed to generate an extended demonstration that retains the geometric features of the original demonstration. (2) DMP modelling and incremental learning updating: BLS-based incremental learning for DMP (Bi-DMP) is performed based on the constructed DMP and extended demonstration. With this incremental learning approach, the DMP is capable of self-updating in response to task demands, preserving previously acquired skills and updating them without training from scratch. (3) Electric vehicle (EV) battery disassembly case study: this study developed a solution suitable for EV battery disassembly and established a decommissioned battery disassembly experimental platform. Unscrewing nuts and battery cell removal are selected to verify the effectiveness of the proposed algorithms based on the battery disassembly experimental platform. In this study, the effectiveness of the algorithms designed in this paper is measured by the success rate and error of the task execution. In the task of unscrewing nuts, the success rate of the classical DMP is 57.14% and the maximum error is 2.760 mm. After the optimisation of 1D-SEG, G-PRM, and Bi-DMP, the success rate of the task is increased to 100% and the maximum error is reduced to 1.477 mm. Full article
(This article belongs to the Section Robotics and Automation)
Show Figures

Figure 1

15 pages, 4518 KiB  
Article
Robust Learning from Demonstration Based on GANs and Affine Transformation
by Kang An, Zhiyang Wu, Qianqian Shangguan, Yaqing Song and Xiaonong Xu
Appl. Sci. 2024, 14(7), 2902; https://doi.org/10.3390/app14072902 - 29 Mar 2024
Cited by 2 | Viewed by 1504
Abstract
Collaborative robots face barriers to widespread adoption due to the complexity of programming them to achieve human-like movement. Learning from demonstration (LfD) has emerged as a crucial solution, allowing robots to learn tasks directly from expert demonstrations, offering versatility and an intuitive programming [...] Read more.
Collaborative robots face barriers to widespread adoption due to the complexity of programming them to achieve human-like movement. Learning from demonstration (LfD) has emerged as a crucial solution, allowing robots to learn tasks directly from expert demonstrations, offering versatility and an intuitive programming approach. However, many existing LfD methods encounter issues such as convergence failure and lack of generalization ability. In this paper, we propose: (1) a generative adversarial network (GAN)-based model with multilayer perceptron (MLP) architecture, coupled with a novel loss function designed to mitigate convergence issues; (2) an affine transformation-based generalization method aimed at enhancing LfD tasks by improving their generalization performance; (3) a data preprocessing method tailored to facilitate deployment on robotics platforms. We conduct experiments on a UR5 robotic platform tasked with handwritten digit recognition. Our results demonstrate that our proposed method significantly accelerates generation speed, achieving a remarkable processing time of 23 ms, which is five times faster than movement primitives (MPs), while preserving key features from demonstrations. This leads to outstanding convergence and generalization performance. Full article
(This article belongs to the Special Issue AI Technologies for Collaborative and Service Robots)
Show Figures

Figure 1

26 pages, 34712 KiB  
Article
Research on LFD System of Humanoid Dual-Arm Robot
by Ze Cui, Lang Kou, Zenghao Chen, Peng Bao, Donghai Qian, Lang Xie and Yue Tang
Symmetry 2024, 16(4), 396; https://doi.org/10.3390/sym16040396 - 28 Mar 2024
Cited by 3 | Viewed by 2130
Abstract
Although robots have been widely used in a variety of fields, the idea of enabling them to perform multiple tasks in the same way that humans do remains a difficulty. To solve this, we investigate the learning from demonstration (LFD) system with our [...] Read more.
Although robots have been widely used in a variety of fields, the idea of enabling them to perform multiple tasks in the same way that humans do remains a difficulty. To solve this, we investigate the learning from demonstration (LFD) system with our independently designed symmetrical humanoid dual-arm robot. We present a novel action feature matching algorithm. This algorithm accurately transforms human demonstration data into task models that robots can directly execute, considerably improving LFD’s generalization capabilities. In our studies, we used motion capture cameras to capture human demonstration actions, which included combinations of simple actions (the action layer) and a succession of complicated operational tasks (the task layer). For the action layer data, we employed Gaussian mixture models (GMM) for processing and constructing an action primitive library. As for the task layer data, we created a “keyframe” segmentation method to transform this data into a series of action primitives and build another action primitive library. Guided by our algorithm, the robot successfully imitated complex human tasks. Results show its excellent task learning and execution, providing an effective solution for robots to learn from human demonstrations and significantly advancing robot technology. Full article
(This article belongs to the Special Issue Symmetry Applied in Computer Vision, Automation, and Robotics)
Show Figures

Figure 1

18 pages, 4983 KiB  
Article
Nonprehensile Manipulation for Rapid Object Spinning via Multisensory Learning from Demonstration
by Ku Jin Shin and Soo Jeon
Sensors 2024, 24(2), 380; https://doi.org/10.3390/s24020380 - 8 Jan 2024
Cited by 1 | Viewed by 1491
Abstract
Dexterous manipulation concerns the control of a robot hand to manipulate an object in a desired manner. While classical dexterous manipulation strategies are based on stable grasping (or force closure), many human-like manipulation tasks do not maintain grasp stability and often utilize the [...] Read more.
Dexterous manipulation concerns the control of a robot hand to manipulate an object in a desired manner. While classical dexterous manipulation strategies are based on stable grasping (or force closure), many human-like manipulation tasks do not maintain grasp stability and often utilize the dynamics of the object rather than the closed form of kinematic relation between the object and the robotic hand. Such manipulation strategies are referred as nonprehensile or dynamic dexterous manipulation in the literature. Nonprehensile manipulation often involves fast and agile movements such as throwing and flipping. Due to the complexity of such motions and uncertainties associated with them, it has been challenging to realize nonprehensile manipulation tasks in a reliable way. In this paper, we propose a new control strategy to realize practical nonprehensile manipulation. First, we make explicit use of multiple modalities of sensory data for the design of control law. Specifically, force data are employed for feedforward control, while position data are used for feedback control. Secondly, control signals (both feedback and feedforward) are obtained through multisensory learning from demonstration (LfD) experiments designed and performed for specific nonprehensile manipulation tasks of concern. To prove the concept of the proposed control strategy, experimental tests were conducted for a dynamic spinning task using a sensory-rich, two-finger robotic hand. The control performance (i.e., the speed and accuracy of the spinning task) was also compared with that of classical dexterous manipulation based on force closure and finger gaiting. Full article
(This article belongs to the Section Sensors and Robotics)
Show Figures

Figure 1

19 pages, 5174 KiB  
Article
Learning Compliant Box-in-Box Insertion through Haptic-Based Robotic Teleoperation
by Sreekanth Kana, Juhi Gurnani, Vishal Ramanathan, Mohammad Zaidi Ariffin, Sri Harsha Turlapati and Domenico Campolo
Sensors 2023, 23(21), 8721; https://doi.org/10.3390/s23218721 - 25 Oct 2023
Cited by 3 | Viewed by 2179
Abstract
In modern logistics, the box-in-box insertion task is representative of a wide range of packaging applications, and automating compliant object insertion is difficult due to challenges in modelling the object deformation during insertion. Using Learning from Demonstration (LfD) paradigms, which are frequently used [...] Read more.
In modern logistics, the box-in-box insertion task is representative of a wide range of packaging applications, and automating compliant object insertion is difficult due to challenges in modelling the object deformation during insertion. Using Learning from Demonstration (LfD) paradigms, which are frequently used in robotics to facilitate skill transfer from humans to robots, can be one solution for complex tasks that are difficult to mathematically model. In order to automate the box-in-box insertion task for packaging applications, this study makes use of LfD techniques. The proposed framework has three phases. Firstly, a master–slave teleoperated robot system is used in the initial phase to haptically demonstrate the insertion task. Then, the learning phase involves identifying trends in the demonstrated trajectories using probabilistic methods, in this case, Gaussian Mixture Regression. In the third phase, the insertion task is generalised, and the robot adjusts to any object position using barycentric interpolation. This method is novel because it tackles tight insertion by taking advantage of the boxes’ natural compliance, making it possible to complete the task even with a position-controlled robot. To determine whether the strategy is generalisable and repeatable, experimental validation was carried out. Full article
(This article belongs to the Section Industrial Sensors)
Show Figures

Figure 1

28 pages, 22741 KiB  
Article
Process of Learning from Demonstration with Paraconsistent Artificial Neural Cells for Application in Linear Cartesian Robots
by João Inácio Da Silva Filho, Cláudio Luís Magalhães Fernandes, Rodrigo Silvério da Silveira, Paulino Machado Gomes, Sérgio Luiz da Conceição Matos, Leonardo do Espirito Santo, Vander Célio Nunes, Hyghor Miranda Côrtes, William Aparecido Celestino Lopes, Mauricio Conceição Mario, Dorotéa Vilanova Garcia, Cláudio Rodrigo Torres, Jair Minoro Abe and Germano Lambert-Torres
Robotics 2023, 12(3), 69; https://doi.org/10.3390/robotics12030069 - 6 May 2023
Cited by 3 | Viewed by 3266
Abstract
Paraconsistent Annotated Logic (PAL) is a type of non-classical logic based on concepts that allow, under certain conditions, for one to accept contradictions without invalidating conclusions. The Paraconsistent Artificial Neural Cell of Learning (lPANCell) algorithm was created from PAL-based equations. With [...] Read more.
Paraconsistent Annotated Logic (PAL) is a type of non-classical logic based on concepts that allow, under certain conditions, for one to accept contradictions without invalidating conclusions. The Paraconsistent Artificial Neural Cell of Learning (lPANCell) algorithm was created from PAL-based equations. With its procedures for learning discrete patterns being represented by values contained in the closed interval between 0 and 1, the lPANCell algorithm presents responses similar to those of nonlinear dynamical systems. In this work, several tests were carried out to validate the operation of the lPANCell algorithm in a learning from demonstration (LfD) framework applied to a linear Cartesian robot (gantry robot), which was moving rectangular metallic workpieces. For the LfD process used in the teaching of trajectories in the x and y axes of the linear Cartesian robot, a Paraconsistent Artificial Neural Network (lPANnet) was built, which was composed of eight lPANCells. The results showed that lPANnet has dynamic properties with a robustness to disturbances, both in the learning process by demonstration, as well as in the imitation process. Based on this work, paraconsistent artificial neural networks of a greater complexity, which are composed of lPANCells, can be formed. This study will provide a strong contribution to research regarding learning from demonstration frameworks being applied in robotics. Full article
(This article belongs to the Section AI in Robotics)
Show Figures

Figure 1

33 pages, 4334 KiB  
Article
An Improved Search and Rescue Algorithm for Global Optimization and Blood Cell Image Segmentation
by Essam H. Houssein, Gaber M. Mohamed, Nagwan Abdel Samee, Reem Alkanhel, Ibrahim A. Ibrahim and Yaser M. Wazery
Diagnostics 2023, 13(8), 1422; https://doi.org/10.3390/diagnostics13081422 - 15 Apr 2023
Cited by 6 | Viewed by 1988
Abstract
Image segmentation has been one of the most active research areas in the last decade. The traditional multi-level thresholding techniques are effective for bi-level thresholding because of their resilience, simplicity, accuracy, and low convergence time, but these traditional techniques are not effective in [...] Read more.
Image segmentation has been one of the most active research areas in the last decade. The traditional multi-level thresholding techniques are effective for bi-level thresholding because of their resilience, simplicity, accuracy, and low convergence time, but these traditional techniques are not effective in determining the optimal multi-level thresholding for image segmentation. Therefore, an efficient version of the search and rescue optimization algorithm (SAR) based on opposition-based learning (OBL) is proposed in this paper to segment blood-cell images and solve problems of multi-level thresholding. The SAR algorithm is one of the most popular meta-heuristic algorithms (MHs) that mimics humans’ exploration behavior during search and rescue operations. The SAR algorithm, which utilizes the OBL technique to enhance the algorithm’s ability to jump out of the local optimum and enhance its search efficiency, is termed mSAR. A set of experiments is applied to evaluate the performance of mSAR, solve the problem of multi-level thresholding for image segmentation, and demonstrate the impact of combining the OBL technique with the original SAR for improving solution quality and accelerating convergence speed. The effectiveness of the proposed mSAR is evaluated against other competing algorithms, including the L’evy flight distribution (LFD), Harris hawks optimization (HHO), sine cosine algorithm (SCA), equilibrium optimizer (EO), gravitational search algorithm (GSA), arithmetic optimization algorithm (AOA), and the original SAR. Furthermore, a set of experiments for multi-level thresholding image segmentation is performed to prove the superiority of the proposed mSAR using fuzzy entropy and the Otsu method as two objective functions over a set of benchmark images with different numbers of thresholds based on a set of evaluation matrices. Finally, analysis of the experiments’ outcomes indicates that the mSAR algorithm is highly efficient in terms of the quality of the segmented image and feature conservation, compared with the other competing algorithms. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

13 pages, 2112 KiB  
Article
MFSR: Light Field Images Spatial Super Resolution Model Integrated with Multiple Features
by Jianfei Zhou and Hongbing Wang
Electronics 2023, 12(6), 1480; https://doi.org/10.3390/electronics12061480 - 21 Mar 2023
Viewed by 2420
Abstract
Light Field (LF) cameras can capture angular and spatial information simultaneously, making them suitable for a wide range of applications such as refocusing, disparity estimation, and virtual reality. However, the limited spatial resolution of the LF images hinders their applicability. In order to [...] Read more.
Light Field (LF) cameras can capture angular and spatial information simultaneously, making them suitable for a wide range of applications such as refocusing, disparity estimation, and virtual reality. However, the limited spatial resolution of the LF images hinders their applicability. In order to address this issue, we propose an end-to-end learning-based light field super-resolution (LFSR) model called MFSR, which integrates multiple features, including spatial, angular, epipolar plane images (EPI), and global features. These features are extracted separately from the LF image and then fused together to obtain a comprehensive feature using the Feature Extract Block (FE Block) iteratively. Gradient loss is added into the loss function to ensure that the MFSR has good performance for LF images with rich texture. Experimental results on synthetic and real-world datasets demonstrate that the proposed method outperforms other state-of-the-art methods, with a peak signal-to-noise ratio (PSNR) improvement of 0.208 dB and 0.274 dB on average for the 2× and 4× super-resolution tasks, and structural similarity (SSIM) of both improvements of 0.01 on average. Full article
(This article belongs to the Special Issue Deep Learning in Computer Vision and Image Processing)
Show Figures

Figure 1

Back to TopTop