Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (2,135)

Search Parameters:
Keywords = sample augmentation

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
23 pages, 586 KB  
Systematic Review
Application of Augmented Reality Technology as a Dietary Monitoring and Control Measure Among Adults: A Systematic Review
by Gabrielle Victoria Gonzalez, Bingjing Mao, Ruxin Wang, Wen Liu, Chen Wang and Tung Sung Tseng
Nutrients 2025, 17(24), 3893; https://doi.org/10.3390/nu17243893 - 12 Dec 2025
Abstract
Background/Objectives: Traditional dietary monitoring methods such as 24 h recalls rely on self-report, leading to recall bias and underreporting. Similarly, dietary control approaches, including portion control and calorie restriction, depend on user accuracy and consistency. Augmented reality (AR) offers a promising alternative for [...] Read more.
Background/Objectives: Traditional dietary monitoring methods such as 24 h recalls rely on self-report, leading to recall bias and underreporting. Similarly, dietary control approaches, including portion control and calorie restriction, depend on user accuracy and consistency. Augmented reality (AR) offers a promising alternative for improving dietary monitoring and control by enhancing engagement, feedback accuracy, and user learning. This systematic review aimed to examine how AR technologies are implemented to support dietary monitoring and control and to evaluate their usability and effectiveness among adults. Methods: A systematic search of PubMed, CINAHL, and Embase identified studies published between 2000 and 2025 that evaluated augmented reality for dietary monitoring and control among adults. Eligible studies included peer-reviewed and gray literature in English. Data extraction focused on study design, AR system type, usability, and effectiveness outcomes. Risk of bias was assessed using the Cochrane RoB 2 tool for randomized controlled trials and ROBINS-I for non-randomized studies. Results: Thirteen studies met inclusion criteria. Since the evidence based was heterogeneous in design, outcomes, and measurement, findings were synthesized qualitatively rather than pooled. Most studies utilized smartphone-based AR systems for portion size estimation, nutrition education, and behavior modification. Usability and satisfaction varied by study: One study found that 80% of participants (N = 15) were satisfied or extremely satisfied with the AR tool. Another reported that 100% of users (N = 26) rated the app easy to use, and a separate study observed a 72.5% agreement rate on ease of use among participants (N = 40). Several studies also examined portion size estimation, with one reporting a 12.2% improvement in estimation accuracy and another showing −6% estimation, though a 12.7% overestimation in energy intake persisted. Additional outcomes related to behavior, dietary knowledge, and physiological or psychological effects were also identified across the review. Common limitations included difficulty aligning markers, overestimation of amorphous foods, and short intervention durations. Despite these promising findings, the existing evidence is limited by small sample sizes, heterogeneity in intervention and device design, short study durations, and variability in usability and accuracy measures. The limitations of this review warrant cautious interpretation of findings. Conclusions: AR technologies show promise for improving dietary monitoring and control by enhancing accuracy, engagement, and behavior change. Future research should focus on longitudinal designs, diverse populations, and integration with multimodal sensors and artificial intelligence. Full article
(This article belongs to the Section Nutrition Methodology & Assessment)
19 pages, 10689 KB  
Article
Research on Augmentation of Wood Microscopic Image Dataset Based on Generative Adversarial Networks
by Shuo Xu, Hang Su and Lei Zhao
J. Imaging 2025, 11(12), 445; https://doi.org/10.3390/jimaging11120445 - 12 Dec 2025
Abstract
Microscopic wood images are vital in wood analysis and classification research. However, the high cost of acquiring microscopic images and the limitations of experimental conditions have led to a severe problem of insufficient sample data, which significantly restricts the training performance and generalization [...] Read more.
Microscopic wood images are vital in wood analysis and classification research. However, the high cost of acquiring microscopic images and the limitations of experimental conditions have led to a severe problem of insufficient sample data, which significantly restricts the training performance and generalization ability of deep learning models. This study first used basic image processing techniques to perform preliminary augmentation of the original dataset. The augmented data were then input into five GAN models, BGAN, DCGAN, WGAN-GP, LSGAN, and StyleGAN2, for training. The quality and model performance of the generated images were assessed by analyzing the degree of fidelity of cellular structure (e.g., earlywood, latewood, and wood rays), image clarity, and diversity of the images for each model-generated image, as well as by using KID, IS, and SSIM. The results showed that images generated by BGAN and WGAN-GP exhibited high quality, with lower KID values and higher IS values, and the generated images were visually close to real images. In contrast, the DCGAN, LSGAN, and StyleGAN2 models experienced mode collapse during training, resulting in lower image clarity and diversity compared to the other models. Through a comparative analysis of different GAN models, this study demonstrates the feasibility and effectiveness of Generative Adversarial Networks in the domain of small-sample image data augmentation, providing an important reference for further research in the field of wood identification. Full article
(This article belongs to the Section Image and Video Processing)
Show Figures

Figure 1

14 pages, 3441 KB  
Article
Improved Biomethane Potential by Substrate Augmentation in Anaerobic Digestion and Biodigestate Utilization in Meeting Circular Bioeconomy
by Wame Bontsi, Nhlanhla Othusitse, Amare Gessesse and Lesedi Lebogang
Energies 2025, 18(24), 6505; https://doi.org/10.3390/en18246505 - 12 Dec 2025
Abstract
Waste generated from agricultural activities is anticipated to increase in the future, especially in less developed countries, and this could cause environmental health risks if these wastes are not well managed. The anaerobic digestion (AD) by co-digesting organic waste is a technology used [...] Read more.
Waste generated from agricultural activities is anticipated to increase in the future, especially in less developed countries, and this could cause environmental health risks if these wastes are not well managed. The anaerobic digestion (AD) by co-digesting organic waste is a technology used to produce biogas while utilizing biodigestate as a biofertilizer; however, AD requires a lot of water to be efficient, which could pose water challenges to arid areas. This study evaluated biogas production under semi-dry conditions by augmenting the process with a high-water content wild melon and determined the nutrient composition of the resultant biodigestate. Batch studies of AD were performed to evaluate methane potential of the different animal waste using an online and standardized Automatic Methane Potential Test System (AMPTS) II light for approximately 506 h (21 days) at 38 °C. The highest biomethane potential (BMP) determined for mono and co-substrate digestion was 29.5 NmL CH4/g VS (CD) and 63.3 NmL CH4/g VS (CMWM), respectively, which was calculated from AMPTS biomethane yield of 3166.2 NmL (CD) and 1480.6 NmL (CMWM). Water-displacement method was also used to compare biogas yield in wet and semi-dry AD. The results showed high biogas yield of 8480 mL for CM (mono-substrate) and 10,975 mL for CMCC in wet AD. Semi-dry AD was investigated by replacing water with a wild melon (WM), and the highest biogas production was 8000 mL from the CMCC combination augmented with WM. Generally, in wet AD, co-digestion was more effective in biogas production than mono-substrate AD. The biodigestate from different substrate combinations were also evaluated for nutrient composition using X-ray Fluorescence (XRF) analysis, and all the samples contained fair amount of essential nutrients such as calcium (Ca), phosphorus (P), potassium (K) and microelements such as chloride (Cl), magnesium (Mn), iron (Fe), zinc (Zn). This study successfully implemented semi-dry AD from co-digested animal wastes to produce biogas as an energy solution and biofertilizer for crop production, thereby creating a closed-loop system that supports a circular bioeconomy. In addition, the study confirmed that lowering the water content in the AD process is feasible without compromising substantial biogas production. This technology, when optimized and well implemented, could provide sustainable biogas production in areas with water scarcity, therefore making the biogas production process accessible to rural communities. Full article
Show Figures

Figure 1

16 pages, 2076 KB  
Article
Mortality Prediction from Patient’s First Day PAAC Radiograph in Internal Medicine Intensive Care Unit Using Artificial Intelligence Methods
by Orhan Gok, Türker Fedai Cavus, Ahmed Cihad Genc, Selcuk Yaylaci and Lacin Tatli Ayhan
Diagnostics 2025, 15(24), 3138; https://doi.org/10.3390/diagnostics15243138 - 10 Dec 2025
Viewed by 117
Abstract
Introduction: This study aims to predict mortality using chest radiographs obtained on the first day of intensive care admission, thereby contributing to better planning of doctors’ treatment strategies and more efficient use of limited resources through early and accurate predictions. Methods: We retrospectively [...] Read more.
Introduction: This study aims to predict mortality using chest radiographs obtained on the first day of intensive care admission, thereby contributing to better planning of doctors’ treatment strategies and more efficient use of limited resources through early and accurate predictions. Methods: We retrospectively analyzed 510 ICU patients. After data augmentation, a total of 3019 chest radiographs were used for model training and validation, while an independent, non-augmented test set of 100 patients (100 images) was reserved for final evaluation. Seventy-four (74) radiomic features were extracted from the images and analyzed using machine learning algorithms. Model performances were evaluated using the area under the ROC curve (AUC), sensitivity, and specificity metrics. Results: A total of 3019 data samples were included in the study. Through feature selection methods, the initial 74 features were gradually reduced to 10. The Subspace KNN algorithm demonstrated the highest prediction accuracy, achieving AUC 0.88, sensitivity 0.80, and specificity 0.87. Conclusions: Machine learning algorithms such as Subspace KNN and features obtained from PAAC radiographs, such as GLCM Contrast, Kurtosis, Cobb angle, Haralick, Bilateral Infiltrates, Cardiomegaly, Skewness, Unilateral Effusion, Median Intensity, and Intensity Range, are promising tools for mortality prediction in patients hospitalized in the internal medicine intensive care unit. These tools can be integrated into clinical decision support systems to provide benefits in patient management. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

31 pages, 9303 KB  
Article
Automatic Quadrotor Dispatch Missions Based on Air-Writing Gesture Recognition
by Pu-Sheng Tsai, Ter-Feng Wu and Yen-Chun Wang
Processes 2025, 13(12), 3984; https://doi.org/10.3390/pr13123984 - 9 Dec 2025
Viewed by 226
Abstract
This study develops an automatic dispatch system for quadrotor UAVs that integrates air-writing gesture recognition with a graphical user interface (GUI). The DJI RoboMaster quadrotor UAV (DJI, Shenzhen, China) was employed as the experimental platform, combined with an ESP32 microcontroller (Espressif Systems, Shanghai, [...] Read more.
This study develops an automatic dispatch system for quadrotor UAVs that integrates air-writing gesture recognition with a graphical user interface (GUI). The DJI RoboMaster quadrotor UAV (DJI, Shenzhen, China) was employed as the experimental platform, combined with an ESP32 microcontroller (Espressif Systems, Shanghai, China) and the RoboMaster SDK (version 3.0). On the Python (version 3.12.7) platform, a GUI was implemented using Tkinter (version 8.6), allowing users to input addresses or landmarks, which were then automatically converted into geographic coordinates and imported into Google Maps for route planning. The generated flight commands were transmitted to the UAV via a UDP socket, enabling remote autonomous flight. For gesture recognition, a Raspberry Pi integrated with the MediaPipe Hands module was used to capture 16 types of air-written flight commands in real time through a camera. The training samples were categorized into one-dimensional coordinates and two-dimensional images. In the one-dimensional case, X/Y axis coordinates were concatenated after data augmentation, interpolation, and normalization. In the two-dimensional case, three types of images were generated, namely font trajectory plots (T-plots), coordinate-axis plots (XY-plots), and composite plots combining the two (XYT-plots). To evaluate classification performance, several machine learning and deep learning architectures were employed, including a multi-layer perceptron (MLP), support vector machine (SVM), one-dimensional convolutional neural network (1D-CNN), and two-dimensional convolutional neural network (2D-CNN). The results demonstrated effective recognition accuracy across different models and sample formats, verifying the feasibility of the proposed air-writing trajectory framework for non-contact gesture-based UAV control. Furthermore, by combining gesture recognition with a GUI-based map planning interface, the system enhances the intuitiveness and convenience of UAV operation. Future extensions, such as incorporating aerial image object recognition, could extend the framework’s applications to scenarios including forest disaster management, vehicle license plate recognition, and air pollution monitoring. Full article
Show Figures

Figure 1

16 pages, 1310 KB  
Article
Intelligent Monitoring of Lost Circulation Risk Based on Shapelet Transformation and Adaptive Model Updating
by Yanlong Zhang, Chenzhan Zhou, Gensheng Li, Chao Fang, Jiasheng Fu, Detao Zhou, Longlian Cui and Bingshan Liu
Processes 2025, 13(12), 3981; https://doi.org/10.3390/pr13123981 - 9 Dec 2025
Viewed by 149
Abstract
As unconventional hydrocarbon resources gain increasing importance, the risk of lost circulation during drilling operations has also grown significantly. Accurate and reliable risk diagnosis methods are essential to ensure safety and operational efficiency in complex drilling environments. This study proposes a novel lost [...] Read more.
As unconventional hydrocarbon resources gain increasing importance, the risk of lost circulation during drilling operations has also grown significantly. Accurate and reliable risk diagnosis methods are essential to ensure safety and operational efficiency in complex drilling environments. This study proposes a novel lost circulation risk monitoring framework based on time-series shapelet transformation, integrated with Generative Adversarial Network (GAN)-based data augmentation and real-time model updating strategies. GANs are employed to synthesize diverse, high-quality samples, enriching the training dataset and improving the model’s ability to capture rare or latent lost circulation signals. Shapelets are then extracted from the time series using a supervised shapelet transform that searches for discriminative subsequences maximizing the separation between normal and lost-circulation samples. Each time series is subsequently represented by its minimum distances to the learned shapelets, so that critical local temporal patterns indicative of early lost circulation can be explicitly captured. To further enhance adaptability during field applications, a real-time model updating mechanism is incorporated. The system incrementally refines the classifier using newly incoming data, where high-confidence predictions are selectively added for online updating. This strategy enables the model to adjust to evolving operating conditions, improves robustness, and provides earlier and more reliable risk warnings. We implemented and evaluated Support Vector Machine (SVM), k-Nearest Neighbors (kNNs), Logistic Regression, and Artificial Neural Networks (ANNs) on the transformed datasets. Experimental results demonstrate that the proposed method improves prediction accuracy by 6.5%, measured as the accuracy gain of the SVM classifier after applying the shapelet transformation (from 84.7% to 91.2%) compared with using raw, untransformed time-series features. Among all models, SVM achieves the best performance, with an accuracy of 91.2%, recall of 90.5%, and precision of 92.3%. Moreover, the integration of real-time updating further boosts accuracy and responsiveness, confirming the effectiveness of the proposed monitoring framework in dynamic drilling environments. The proposed method offers a practical and scalable solution for intelligent lost circulation monitoring in drilling operations, providing a solid theoretical foundation and technical reference for data-driven safety systems in dynamic environments. Full article
(This article belongs to the Special Issue Development of Advanced Drilling Engineering)
Show Figures

Figure 1

25 pages, 2845 KB  
Article
Power Quality Data Augmentation and Processing Method for Distribution Terminals Considering High-Frequency Sampling
by Ruijiang Zeng, Zhiyong Li, Haodong Liu, Wenxuan Che, Jiamu Yang, Sifeng Li and Zhongwei Sun
Energies 2025, 18(24), 6426; https://doi.org/10.3390/en18246426 - 9 Dec 2025
Viewed by 88
Abstract
The safe and stable operation of distribution networks relies on the real-time monitoring, analysis, and feedback of power quality data. However, with the continuous advancement of distribution network construction, the number of distributed power electronic devices has increased significantly, leading to frequent power [...] Read more.
The safe and stable operation of distribution networks relies on the real-time monitoring, analysis, and feedback of power quality data. However, with the continuous advancement of distribution network construction, the number of distributed power electronic devices has increased significantly, leading to frequent power quality issues such as voltage fluctuations, harmonic pollution, and three-phase unbalance in distribution terminals. Therefore, the augmentation and processing of power quality data have become crucial for ensuring the stable operation of distribution networks. Traditional methods for augmenting and processing power quality data fail to consider the differentiated characteristics of burrs in signal sequences and neglect the comprehensive consideration of both time-domain and frequency-domain features in disturbance identification. This results in the distortion of high-frequency fault information, and insufficient robustness and accuracy in identifying Power Quality Disturbance (PQD) against the complex noise background of distribution networks. In response to these issues, we propose a power quality data augmentation and processing method for distribution terminals considering high-frequency sampling. Firstly, a burr removal method of the sampling waveform based on a high-frequency filter operator is proposed. By comprehensively considering the characteristics of concavity and convexity in both burr and normal waveforms, a high-frequency filtering operator is introduced. Additional constraints and parameters are applied to suppress sequences with burr characteristics, thereby accurately eliminating burrs while preserving the key features of valid information. This approach avoids distortion of high-frequency fault information after filtering, which supports subsequent PQD identification. Secondly, a PQD identification method based on a dual-channel time–frequency feature fusion network is proposed. The PQD signals undergo an S-transform and period reconfiguration to construct matrix image features in the time–frequency domain. Finally, these features are input into a Convolutional Neural Network (CNN) and a Transformer encoder to extract highly coupled global features, which are then fused through a cross-attention mechanism. The identification results of PQD are output through a classification layer, thereby enhancing the robustness and accuracy of disturbance identification against the complex noise background of distribution networks. Simulation results demonstrate that the proposed algorithm achieves optimal burr removal and disturbance identification accuracy. Full article
Show Figures

Figure 1

18 pages, 2443 KB  
Article
Teaching-Based Robotic Arm System with BiLSTM Pattern Recognition for Food Processing Automation
by Youngjin Kim and Sangoh Kim
Appl. Sci. 2025, 15(24), 12936; https://doi.org/10.3390/app152412936 - 8 Dec 2025
Viewed by 126
Abstract
Teaching-based robotic systems offer an accessible alternative to complex inverse kinematics programming for food processing automation. Traditional model-based approaches require precise system identification and analytical solutions that are challenging for custom-built robots with manufacturing tolerances and mechanical uncertainties. This study developed a custom [...] Read more.
Teaching-based robotic systems offer an accessible alternative to complex inverse kinematics programming for food processing automation. Traditional model-based approaches require precise system identification and analytical solutions that are challenging for custom-built robots with manufacturing tolerances and mechanical uncertainties. This study developed a custom six-degree-of-freedom robotic arm using modular brushless motors controlled via Controller Area Network communication and Robot Operating System 2, a teaching mode where users manually demonstrate trajectories that are recorded at 100 Hz. Forty-five demonstration trajectories were collected across three geometric patterns (rectangle, triangle, circle) and augmented to 270 samples. A bidirectional Long Short-Term Memory network with attention mechanism was trained to classify patterns, achieving 83.33% test accuracy and outperforming baseline deep learning models (1D-CNN: 77.78%, TCN: 66.67%, GRU: 44.44%), while being marginally exceeded by Random Forest (86.11%). Rectangle patterns showed strongest recognition (78.57% F1-score), while circle patterns achieved highest performance (91.67% F1-score). However, severe overfitting was observed, with validation accuracy peaking at 85.19% at epoch 24 before degradation, indicating insufficient dataset size despite five-fold augmentation. The results demonstrate proof-of-concept feasibility for pattern recognition from limited teaching demonstrations, providing a pathway for robotic food processing without extensive programming expertise, though larger datasets and robust feedback control strategies are required for production deployment. Full article
Show Figures

Figure 1

28 pages, 4585 KB  
Article
Uncertainty-Aware Adaptive Intrusion Detection Using Hybrid CNN-LSTM with cWGAN-GP Augmentation and Human-in-the-Loop Feedback
by Clinton Manuel de Nascimento and Jin Hou
Safety 2025, 11(4), 120; https://doi.org/10.3390/safety11040120 - 5 Dec 2025
Viewed by 163
Abstract
Intrusion detection systems (IDSs) must operate under severe class imbalance, evolving attack behavior, and the need for calibrated decisions that integrate smoothly with security operations. We propose a human-in-the-loop IDS that combines a convolutional neural network and a long short-term memory network (CNN–LSTM) [...] Read more.
Intrusion detection systems (IDSs) must operate under severe class imbalance, evolving attack behavior, and the need for calibrated decisions that integrate smoothly with security operations. We propose a human-in-the-loop IDS that combines a convolutional neural network and a long short-term memory network (CNN–LSTM) classifier with a variational autoencoder (VAE)-seeded conditional Wasserstein generative adversarial network with gradient penalty (cWGAN-GP) augmentation and entropy-based abstention. Minority classes are reinforced offline via conditional generative adversarial (GAN) sampling, whereas high-entropy predictions are escalated for analysts and are incorporated into a curated retraining set. On CIC-IDS2017, the resulting framework delivered well-calibrated binary performance (ACC = 98.0%, DR = 96.6%, precision = 92.1%, F1 = 94.3%; baseline ECE ≈ 0.04, Brier ≈ 0.11) and substantially improved minority recall (e.g., Infiltration from 0% to >80%, Web Attack–XSS +25 pp, and DoS Slowhttptest +15 pp, for an overall +11 pp macro-recall gain). The deployed model remained lightweight (~42 MB, <10 ms per batch; ≈32 k flows/s on RTX-3050 Ti), and only approximately 1% of the flows were routed for human review. Extensive evaluation, including ROC/PR sweeps, reliability diagrams, cross-domain tests on CIC-IoT2023, and FGSM/PGD adversarial stress, highlights both the strengths and remaining limitations, notably residual errors on rare web attacks and limited IoT transfer. Overall, the framework provides a practical, calibrated, and extensible machine learning (ML) tier for modern IDS deployment and motivates future research on domain alignment and adversarial defense. Full article
Show Figures

Graphical abstract

26 pages, 20055 KB  
Article
Design and Development of a Neural Network-Based End-Effector for Disease Detection in Plants with 7-DOF Robot Integration
by Harol Toro, Hector Moncada, Kristhian Dierik Gonzales, Cristian Moreno, Claudia L. Garzón-Castro and Jose Luis Ordoñez-Avila
Processes 2025, 13(12), 3934; https://doi.org/10.3390/pr13123934 - 5 Dec 2025
Viewed by 283
Abstract
This study presents the design and development of an intelligent end-effector integrated into a custom 7-degree-of-freedom (DOF) robotic arm for monitoring the health status of tomato plants during their growth stages. The robotic system combines five rotational and two prismatic joints, enabling both [...] Read more.
This study presents the design and development of an intelligent end-effector integrated into a custom 7-degree-of-freedom (DOF) robotic arm for monitoring the health status of tomato plants during their growth stages. The robotic system combines five rotational and two prismatic joints, enabling both horizontal reach and vertical adaptability to inspect plants of varying heights without repositioning the robot’s base. The integrated vision module employs a YOLOv5 neural network trained with 7864 images of tomato leaves, including both healthy and diseased samples. Image preprocessing included normalization and data augmentation to enhance robustness under natural lighting conditions. The optimized model achieved a detection accuracy of 90.2% and a mean average precision (mAP) of 92.3%, demonstrating high reliability in real-time disease classification. The end-effector, fabricated using additive manufacturing, incorporates a Raspberry Pi 4 for onboard processing, allowing autonomous operation in agricultural environments. The experimental results validate the feasibility of combining a custom 7-DOF robotic structure with a deep learning-based detector for continuous plant monitoring. This research contributes to the field of agricultural robotics by providing a flexible and precise platform capable of early disease detection in dynamic cultivation conditions, promoting sustainable and data-driven crop management. Full article
Show Figures

Figure 1

10 pages, 496 KB  
Article
Adaptive 3D Augmentation in StyleGAN2-ADA for High-Fidelity Lung Nodule Synthesis from Limited CT Volumes
by Oleksandr Fedoruk, Konrad Klimaszewski and Michał Kruk
Sensors 2025, 25(24), 7404; https://doi.org/10.3390/s25247404 - 5 Dec 2025
Viewed by 285
Abstract
Generative adversarial networks (GANs) typically require large datasets for effective training, which poses challenges for volumetric medical imaging tasks where data are scarce. This study addresses this limitation by extending adaptive discriminator augmentation (ADA) for three-dimensional (3D) StyleGAN2 to improve generative performance on [...] Read more.
Generative adversarial networks (GANs) typically require large datasets for effective training, which poses challenges for volumetric medical imaging tasks where data are scarce. This study addresses this limitation by extending adaptive discriminator augmentation (ADA) for three-dimensional (3D) StyleGAN2 to improve generative performance on limited volumetric data. The proposed 3D StyleGAN2-ADA redefines all 2D operations for volumetric processing and incorporates the full set of original augmentation techniques. Experiments are conducted on the NoduleMNIST3D dataset of lung CT scans containing 590 voxel-based samples across two classes. Two augmentation pipelines are evaluated—one using color-based transformations and another employing a comprehensive set of 3D augmentations including geometric, filtering, and corruption augmentations. Performance is compared against the same network and dataset without any augmentations at all by assessing generation quality with Kernel Inception Distance (KID) and 3D Structural Similarity Index Measure (SSIM). Results show that volumetric ADA substantially improves training stability and reduces the risk of a mode collapse, even under severe data constraints. A strong augmentation strategy improves the realism of generated 3D samples and better preserves anatomical structures relative to those without data augmentation. These findings demonstrate that adaptive 3D augmentations effectively enable high-quality synthetic medical image generation from extremely limited volumetric datasets. The source code and the weights of the networks are available in the GitHub repository. Full article
(This article belongs to the Section Biomedical Sensors)
Show Figures

Figure 1

35 pages, 4295 KB  
Article
Simulation-Driven Deep Transfer Learning Framework for Data-Efficient Prediction of Physical Experiments
by Soo-Young Lim, Han-Bok Seo and Seung-Yop Lee
Mathematics 2025, 13(23), 3884; https://doi.org/10.3390/math13233884 - 4 Dec 2025
Viewed by 142
Abstract
Transfer learning, which utilizes extensive simulation data to overcome the limitations of scarce and expensive experimental data, has emerged as a powerful approach for predictive modeling in various physical domains. This study presents a comprehensive framework to improve the predictive performance of transfer [...] Read more.
Transfer learning, which utilizes extensive simulation data to overcome the limitations of scarce and expensive experimental data, has emerged as a powerful approach for predictive modeling in various physical domains. This study presents a comprehensive framework to improve the predictive performance of transfer learning, focusing on quasi-zero stiffness (QZS) systems with limited experimental datasets. The proposed framework systematically examines the interplay among three critical factors in the target domain: data augmentation, layer-freezing configurations, and neural network architecture. Simulation-driven synthetic data are generated to capture dynamic features not represented in the sparse experimental data. The optimal transfer depth is explored by evaluating different scenarios of selective layer freezing and fine-tuning. Results show that partial transfer strategies outperform both full-transfer and non-transfer approaches, leading to more stable and accurate predictions. To investigate hierarchical transfer, both symmetric and asymmetric network architectures are designed, embedding physically meaningful representations from simulations into the deeper layers of the target model. Furthermore, an attention mechanism is integrated to emphasize material-specific characteristics. Building on these components, the proposed simulation-driven framework predicts the full force–displacement responses of QZS systems using only 12 experimental samples. Through a systematic comparison of three datasets (direct transfer, linear correction, FEM-based correction), three network architectures, and seven layer-freezing scenarios, the framework achieves a best test performance of R2 = 0.978 and MAE = 0.34 Newtons. Full article
(This article belongs to the Special Issue Advances in Neural Networks and Their Applications)
Show Figures

Figure 1

22 pages, 1145 KB  
Article
TSMTFN: Two-Stream Temporal Shift Module Network for Efficient Egocentric Gesture Recognition in Virtual Reality
by Muhammad Abrar Hussain, Chanjun Chun and SeongKi Kim
Virtual Worlds 2025, 4(4), 58; https://doi.org/10.3390/virtualworlds4040058 - 4 Dec 2025
Viewed by 132
Abstract
Egocentric hand gesture recognition is vital for natural human–computer interaction in augmented and virtual reality (AR/VR) systems. However, most deep learning models struggle to balance accuracy and efficiency, limiting real-time use on wearable devices. This paper introduces a Two-Stream Temporal Shift Module Transformer [...] Read more.
Egocentric hand gesture recognition is vital for natural human–computer interaction in augmented and virtual reality (AR/VR) systems. However, most deep learning models struggle to balance accuracy and efficiency, limiting real-time use on wearable devices. This paper introduces a Two-Stream Temporal Shift Module Transformer Fusion Network (TSMTFN) that achieves high recognition accuracy with low computational cost. The model integrates Temporal Shift Modules (TSMs) for efficient motion modeling and a Transformer-based fusion mechanism for long-range temporal understanding, operating on dual RGB-D streams to capture complementary visual and depth cues. Training stability and generalization are enhanced through full-layer training from epoch 1 and MixUp/CutMix augmentations. Evaluated on the EgoGesture dataset, TSMTFN attained 96.18% top-1 accuracy and 99.61% top-5 accuracy on the independent test set with only 16 GFLOPs and 21.3M parameters, offering a 2.4–4.7× reduction in computation compared to recent state-of-the-art methods. The model runs at 15.10 samples/s, achieving real-time performance. The results demonstrate robust recognition across over 95% of gesture classes and minimal inter-class confusion, establishing TSMTFN as an efficient, accurate, and deployable solution for next-generation wearable AR/VR gesture interfaces. Full article
Show Figures

Figure 1

23 pages, 1197 KB  
Article
Signal Surface Augmentation for Artificial Intelligence-Based Automatic Modulation Classification
by Alexander Gros, Véronique Moeyaert and Patrice Mégret
Electronics 2025, 14(23), 4760; https://doi.org/10.3390/electronics14234760 - 3 Dec 2025
Viewed by 215
Abstract
Automatic modulation recognition has regained attention as a critical application for cognitive radio, combining artificial intelligence with physical layer monitoring of wireless transmissions. This paper formalizes signal surface augmentation (SSA), a process that decomposes signals into informative subcomponents to enhance AI-based analysis. We [...] Read more.
Automatic modulation recognition has regained attention as a critical application for cognitive radio, combining artificial intelligence with physical layer monitoring of wireless transmissions. This paper formalizes signal surface augmentation (SSA), a process that decomposes signals into informative subcomponents to enhance AI-based analysis. We employ Bivariate Empirical Mode Decomposition (BEMD) to break signals into intrinsic modes while addressing challenges like adjacent trends in long sample decompositions and introducing the concept of data overdispersion. Using a modern, publicly available dataset of synthetic modulated signals under realistic conditions, we validate that the presentation of BEMD-derived components improves recognition accuracy by 13% compared to raw IQ inputs. For extended signal lengths, gains reach up to 36%. These results demonstrate the value of signal surface augmentation for improving the robustness of modulation recognition, with potential applications in real-world scenarios. Full article
Show Figures

Figure 1

20 pages, 55314 KB  
Article
MSFN-YOLOv11: A Novel Multi-Scale Feature Fusion Recognition Model Based on Improved YOLOv11 for Real-Time Monitoring of Birds in Wetland Ecosystems
by Linqi Wang, Lin Ye, Xinbao Chen and Nan Chu
Animals 2025, 15(23), 3472; https://doi.org/10.3390/ani15233472 - 2 Dec 2025
Viewed by 254
Abstract
Intelligent bird species recognition is vital for biodiversity monitoring and ecological conservation. This study tackles the challenge of declining recognition accuracy caused by occlusions and imaging noise in complex natural environments. Focusing on ten representative bird species from the Dongting Lake Wetland, we [...] Read more.
Intelligent bird species recognition is vital for biodiversity monitoring and ecological conservation. This study tackles the challenge of declining recognition accuracy caused by occlusions and imaging noise in complex natural environments. Focusing on ten representative bird species from the Dongting Lake Wetland, we propose an improved YOLOv11n-based model named MSFN-YOLO11, which incorporates multi-scale feature fusion. After selecting YOLOv11n as the baseline through comparison with the most-stable version of YOLOv8n, we enhance its backbone by introducing an MSFN module. This module strengthens global and local feature extraction via parallel dilated convolution and a channel attention mechanism. Experiments are conducted on a self-built dataset containing 4540 images of ten species with 6824 samples. To simulate real-world conditions, 25% of samples are augmented using random occlusion, Gaussian noise (σ = 0.2, 0.3, 0.4), and Poisson noise. The improved model achieves a mAP@50 of 96.4% and mAP@50-95 of 83.2% on the test set. Although the mAP@50 shows a slight improvement of 0.3% compared to the original YOLOv11, it has contributed to an 18% reduction in training time. Furthermore, it also demonstrates practical efficacy in processing dynamic video, attaining an average 63.1% accuracy at 1920 × 1080@72fps on an NVIDIA_Tesla_V100_SXM2_32_GB. The proposed model provides robust technical support for real-time bird monitoring in wetlands and enhances conservation efforts for endangered species. Full article
(This article belongs to the Special Issue Artificial Intelligence as a Useful Tool in Behavioural Studies)
Show Figures

Figure 1

Back to TopTop