Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (3,630)

Search Parameters:
Keywords = generative adversarial networks

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
32 pages, 4750 KB  
Article
Enabling Reliable Industrial Energy Savings Verification Through Hybrid Factored Conditional Restricted Boltzmann Machine and Generative Adversarial Network
by Suziee Sukarti, Mohamad Fani Sulaima, Norashikin Sahadan, Muhamad Hafizul Shamsor, Siaw Wei Yao and Aida Fazliana Abdul Kadir
Algorithms 2026, 19(5), 338; https://doi.org/10.3390/a19050338 (registering DOI) - 28 Apr 2026
Abstract
Reliable quantification of industrial energy savings requires accurate detection of non-routine events (NREs) that distort post-retrofit baselines. Conventional statistical and rule-based anomaly detection methods often misinterpret operational variability, leading to biased or overstated savings under the International Performance Measurement and Verification Protocol (IPMVP). [...] Read more.
Reliable quantification of industrial energy savings requires accurate detection of non-routine events (NREs) that distort post-retrofit baselines. Conventional statistical and rule-based anomaly detection methods often misinterpret operational variability, leading to biased or overstated savings under the International Performance Measurement and Verification Protocol (IPMVP). This study develops a novel IPMVP-compliant hybrid deep learning framework that integrates a deterministic Deep Neural Network (DNN) for baseline modeling with stochastic architectures, namely the Factored Conditional Restricted Boltzmann Machine (FCRBM) and Generative Adversarial Network (GAN), to capture probabilistic reconstruction patterns. Their outputs are fused using a hybrid thresholding mechanism that balances detection sensitivity and specificity. Using high-resolution data from an industrial glove manufacturing facility, the hybrid DNN–FCRBM model achieved the best trade-off, demonstrating an accuracy of 94.3%, a precision of 91.1%, and a low false positive rate of 5.1%. This model validated 11.32% industrial energy savings (approximately 478,050 kWh), equivalent to 237 tonnes of CO2 avoided. The integration of stochastic generative learning within a deterministic framework strengthens transparency, auditability, and IPMVP compliance, offering a scalable pathway for credible industrial energy savings verification. Full article
(This article belongs to the Section Algorithms for Multidisciplinary Applications)
Show Figures

Figure 1

27 pages, 14460 KB  
Article
Reconstructing High-Resolution Coastal Water Quality Data Based on a Deep Learning Multivariate Downscaling Approach
by Xiaoyu Liu, Xuan Wang, Yicong Tong, Wei Li and Guijun Han
Remote Sens. 2026, 18(9), 1346; https://doi.org/10.3390/rs18091346 (registering DOI) - 28 Apr 2026
Abstract
The availability of high-resolution oceanographic data is critical for evidence-based coastal environmental management and climate resilience planning, yet it remains constrained by observational gaps and the prohibitive computational cost of fine-scale hydrodynamic modeling. While downscaling techniques provide a viable pathway, current data-driven approaches [...] Read more.
The availability of high-resolution oceanographic data is critical for evidence-based coastal environmental management and climate resilience planning, yet it remains constrained by observational gaps and the prohibitive computational cost of fine-scale hydrodynamic modeling. While downscaling techniques provide a viable pathway, current data-driven approaches often lack statistical physical associations, overlook multivariate environmental interactions, and struggle to represent complex coastal topography. To address these limitations, we present MEOFGAN—an environmentally informed downscaling framework that integrates multivariate empirical orthogonal function (MEOF) decomposition with a generative adversarial network (GAN). The model extracts physically interpretable spatial modes of coupled ocean variables, learns their cross-scale transitions through adversarial training, and systematically incorporates high-resolution bathymetry as a static environmental constraint to enhance spatial fidelity. When applied to the Bohai Sea, MEOFGAN successfully downscales sea surface temperature (SST) and sea surface height (SSH) from 1/4° to 1/12°, achieving error reductions of 30–68% compared to benchmark methods while preserving ecologically relevant structural patterns (SSIM > 0.92). The framework demonstrates strong generalization by reconstructing 500 m resolution distributions of chlorophyll-a (Chl-a), dissolved oxygen (DO), and salinity in Bohai Bay, capturing fine-scale environmental gradients during a documented algal bloom event. This work establishes a methodological framework that can be transferred as a paradigm for generating high-resolution coastal datasets. Rather than serving as a universally transferable pre-trained model, the framework requires region-specific training and application. Data generated in this manner can directly support water quality monitoring, eutrophication assessment, habitat mapping, and regionally tailored climate adaptation strategies. Full article
Show Figures

Figure 1

12 pages, 863 KB  
Article
High-Fidelity Synthesis of Temporomandibular Joint Cone-Beam Computed Tomography Images via Latent Diffusion Models
by Qinlanhui Zhang, Yunhao Zheng and Jun Wang
J. Clin. Med. 2026, 15(9), 3344; https://doi.org/10.3390/jcm15093344 (registering DOI) - 28 Apr 2026
Abstract
Background: The development of robust artificial intelligence (AI) models for diagnosing Temporomandibular Disorders (TMDs) is severely constrained by data scarcity and patient privacy regulations. Cone-beam computed tomography (CBCT), the gold standard for assessing osseous changes in the temporomandibular joint (TMJ), inherently contains [...] Read more.
Background: The development of robust artificial intelligence (AI) models for diagnosing Temporomandibular Disorders (TMDs) is severely constrained by data scarcity and patient privacy regulations. Cone-beam computed tomography (CBCT), the gold standard for assessing osseous changes in the temporomandibular joint (TMJ), inherently contains sensitive biometric facial features, making de-identification difficult without losing critical anatomical information. This study aims to develop and evaluate TMJCTGenerator, a specialized latent diffusion model (LDM) framework designed to synthesize high-fidelity, diverse, and anonymous TMJ CBCT images. We hypothesize that this LDM approach can achieve superior anatomical fidelity and diversity compared to traditional generative adversarial network (GAN)- and variational autoencoder (VAE)-based methods, specifically in capturing fine osseous details within sagittal and coronal views of the mandibular condyle. Methods: A training dataset comprising 348 anonymized CBCT volumes was obtained in this retrospective comparative study to extract high-resolution sagittal and coronal regions of interest of the mandibular condyle. An independent test set of 39 anonymized CBCT volumes was further included. We developed a class-conditional LDM that integrates a pre-trained VAE for perceptual compression with a conditional U-Net for iterative denoising in the latent space. Performance was evaluated via qualitative anatomical fidelity assessment, Fréchet Inception Distance (FID), and a blinded Visual Turing test conducted by experienced clinicians to determine the distinguishability of synthetic images from real data. Results: Qualitative analysis revealed that TMJCTGenerator produced images with superior sharpness and anatomical consistency compared to baseline models, successfully reconstructing fine bone structures essential for diagnosing degenerative joint disease. TMJCTGenerator achieved lower FID scores than both VAE and GAN baselines. In the visual Turing test, clinicians were unable to reliably distinguish the generated images from real scans, and non-inferiority analysis confirmed that the synthetic data were statistically non-inferior to real data. Furthermore, TMJCTGenerator demonstrated the capability to generate diverse pathological conditions, ranging from normal anatomy to severe osteoarthritic changes. Conclusions: The proposed LDM framework effectively addresses the data scarcity and privacy bottlenecks in TMJ AI research by generating realistic, fully anonymous medical imaging data. TMJCTGenerator outperforms traditional generative methods in both visual fidelity and diversity, offering a viable solution for training downstream diagnostic algorithms. The source code and pre-trained models of TMJCTGenerator have been made open-source. Full article
(This article belongs to the Section Dentistry, Oral Surgery and Oral Medicine)
Show Figures

Figure 1

34 pages, 3976 KB  
Article
Entropy Guided Benchmarking of Classical and Generative Imputation Methods for High-Dimensional Healthcare Survey Data
by Deepa Fernandes Prabhu, Jaeyoung Park and Varadraj P. Gurupur
Appl. Sci. 2026, 16(9), 4262; https://doi.org/10.3390/app16094262 (registering DOI) - 27 Apr 2026
Abstract
Missing data are a persistent challenge in large healthcare datasets, often undermining both statistical validity and machine learning performance when handled using simplistic assumptions. In this work, we examine how entropy-based diagnostics can guide the selection of imputation strategies for high-dimensional health survey [...] Read more.
Missing data are a persistent challenge in large healthcare datasets, often undermining both statistical validity and machine learning performance when handled using simplistic assumptions. In this work, we examine how entropy-based diagnostics can guide the selection of imputation strategies for high-dimensional health survey data using the National Health and Nutrition Examination Survey (NHANES) 2021–2023. Shannon entropy is used to identify variables with structurally complex missingness, and a range of classical approaches (mean imputation, k-nearest neighbors, and multiple imputation by chained equations) are evaluated alongside deep generative methods, including variational autoencoders, generative adversarial networks (GANs), Wasserstein GANs (WGANs), and diffusion-based models. All methods are compared under a controlled masked-entry evaluation using root mean square error (RMSE) and Kolmogorov–Smirnov (KS) statistics to capture both reconstruction accuracy and distributional fidelity. Results show that diffusion-based models provide the most consistent balance between numerical accuracy and distributional preservation across high-entropy dietary variables, while WGAN demonstrates competitive performance for selected distributions. Structural equation modeling further indicates that entropy is a useful diagnostic signal for identifying variables that are difficult to reconstruct. Overall, this study provides a reproducible framework for aligning imputation strategy with missingness complexity in healthcare data, with implications for improving reliability in downstream analytics. Full article
(This article belongs to the Special Issue New Trends in Decision Support Systems and Their Applications)
19 pages, 16316 KB  
Article
Enhancing Adversarial Transferability via Fourier-Based Input Transformation
by Zilin Tian, Xin Wang, Yunfei Long and Liguo Zhang
Big Data Cogn. Comput. 2026, 10(5), 135; https://doi.org/10.3390/bdcc10050135 - 27 Apr 2026
Abstract
Adversarial transferability makes black-box attacks practical and exposes weaknesses of deep neural networks for computer vision, image recognition, and visual understanding. Among various transferability-enhancing methods, input transformation is one of the most effective strategies. However, existing methods often ignore the decoupling of style [...] Read more.
Adversarial transferability makes black-box attacks practical and exposes weaknesses of deep neural networks for computer vision, image recognition, and visual understanding. Among various transferability-enhancing methods, input transformation is one of the most effective strategies. However, existing methods often ignore the decoupling of style and semantics in the input image, as well as the need for customized transformation strategies, resulting in limited performance gains or suboptimal outcomes. In this paper, we propose a novel Fourier-based perspective for input transformation generalization in the context of vision adversarial attacks. The main observations are that the Fourier amplitude captures stylistic information and the phase encompasses richer semantics which are crucial for visual understanding. Motivated by this, we develop a Fourier-based strategy, which performs a stylistic transform and semantic mixup on the input examples to improve transferability. To avoid inconsistent semantics of augmented images for the surrogate model, we mix the original images with the augmentations to maintain semantic consistency and mitigate imprecise gradients. Extensive experiments on ImageNet-compatible datasets demonstrate that our method consistently outperforms existing input transformation attacks. Full article
(This article belongs to the Section Artificial Intelligence and Multi-Agent Systems)
Show Figures

Figure 1

30 pages, 12666 KB  
Article
Human-Inspired Dexterity-Oriented Perception and Trajectory Optimization for Robotic Surface Inspection
by Menghan Zou, Yuchuang Tong, Tianbo Yang and Zhengtao Zhang
Biomimetics 2026, 11(5), 296; https://doi.org/10.3390/biomimetics11050296 - 24 Apr 2026
Viewed by 189
Abstract
Industrial surface inspection is fundamental to advanced manufacturing, yet reliable robotic image acquisition in complex geometries remains challenging due to severe occlusions and the inherent trade-off between resolution and coverage. Inspired by human visual inspection behaviors and perception–action coordination mechanisms, this paper proposes [...] Read more.
Industrial surface inspection is fundamental to advanced manufacturing, yet reliable robotic image acquisition in complex geometries remains challenging due to severe occlusions and the inherent trade-off between resolution and coverage. Inspired by human visual inspection behaviors and perception–action coordination mechanisms, this paper proposes a hierarchical trajectory optimization framework for robotic image acquisition based on measured point clouds. Specifically, a multi-constraint preprocessing model is developed to emulate human-like active perception strategies, enabling occlusion-aware viewpoint generation over complex concave and convex surfaces with adaptive camera orientation. Building upon this, a multi-objective trajectory optimization method is introduced to coordinate global coverage and local motion efficiency, jointly optimizing viewpoint sequencing, path length, and motion smoothness hierarchically. To further enhance flexibility in constrained environments, a Pose Reachability Augmented Generative Adversarial Network (PRAGAN) is proposed to learn feasible and adaptable imaging postures under kinematic constraints. Experimental results on an industrial robotic platform equipped with 2D and 3D vision systems demonstrate 100% coverage of key surface areas, a 47.0% reduction in path length, and a 37.5% decrease in solution time compared with the baseline in the physical experiments, while ensuring collision-free operation. Both simulation and real-world experiments validate that the proposed framework effectively captures human-inspired perception and motion coordination, providing a practical and scalable solution for complex industrial surface inspection. Full article
(This article belongs to the Section Locomotion and Bioinspired Robotics)
34 pages, 1426 KB  
Article
Bi-Level Optimal Scheduling for Bundled Operation of PSH with WP and PV Under Extreme High-Temperature Weather
by Wanji Ma, Hong Zhang, He Qiao and Dacheng Xing
Energies 2026, 19(9), 2048; https://doi.org/10.3390/en19092048 - 23 Apr 2026
Viewed by 114
Abstract
With the increasing occurrence of extreme high-temperature weather events, the traditional bundled operation of wind power (WP), photovoltaic power (PV), and pumped storage hydropower (PSH) is facing dual challenges, namely intensified renewable energy fluctuations and insufficient flexible regulation capability of PSH. Therefore, this [...] Read more.
With the increasing occurrence of extreme high-temperature weather events, the traditional bundled operation of wind power (WP), photovoltaic power (PV), and pumped storage hydropower (PSH) is facing dual challenges, namely intensified renewable energy fluctuations and insufficient flexible regulation capability of PSH. Therefore, this paper proposes an optimal scheduling strategy for bundled operation based on capacity interval matching of PSH with WP and PV under extreme high-temperature weather. First, typical scenarios are generated based on a Time-series Generative Adversarial Network (TimeGAN), and an interval matching transaction model is established based on the forecast intervals of WP and PV capacity and the corrected intervals of PSH capacity. Second, considering PSH as an independent market entity, a bi-level optimization model is constructed, in which the upper-level objective is to maximize the revenue of PSH, while the lower-level objective is to minimize the total cost of the joint clearing of the energy and ancillary service markets. Finally, simulation case studies verify that under extreme high-temperature weather, the proposed optimal scheduling method increases the bundled operation capacity by 17.9% and improves the revenue of PSH in the reserve ancillary service market by 14.8%, thereby effectively enhancing the economic performance of PSH while ensuring the safe and stable operation of the system. Full article
33 pages, 24046 KB  
Article
CoDA: A Cognitive-Inspired Approach for Domain Adaptation
by Cavide Balkı Gemirter, Emin Erkan Korkmaz and Dionysis Goularas
Appl. Sci. 2026, 16(9), 4115; https://doi.org/10.3390/app16094115 - 23 Apr 2026
Viewed by 329
Abstract
Modern neural networks have achieved remarkable success in visual recognition; however, due to their sensitivity to domain shifts, Unsupervised Domain Adaptation (UDA) remains an open research problem. A key reason for this limitation is that source-trained models rely primarily on texture, lacking the [...] Read more.
Modern neural networks have achieved remarkable success in visual recognition; however, due to their sensitivity to domain shifts, Unsupervised Domain Adaptation (UDA) remains an open research problem. A key reason for this limitation is that source-trained models rely primarily on texture, lacking the explicit geometric information required for object recognition. To overcome this problem, we introduce CoDA, an object-centric learning framework inspired by infant cognitive development, specifically the process of object individuation. By introducing a geometric prior, our approach employs a physically grounded generation pipeline that uses a textureless “Sculpture Mode” and object isolation to complement textural information with 3D geometric features, capturing shape information that is often ignored during training. To enable robust training from scratch, we further integrate two control mechanisms: a Network Stability Scheduler to orchestrate training progression based on convergence stability, and a Dynamic Top-K Pseudo-Labeling strategy that adapts confidence thresholds for each individual class. Extensive evaluations on three real-world target datasets (VegFru, Fruits-262, and Open Images v7) demonstrate that CoDA, trained on a source dataset of just 12,000 synthetic images, achieves comparable results to (and in specific domains surpasses) ImageNet-pretrained models (leveraging 1.2 million images), significantly outperforming state-of-the-art adversarial and semi-supervised domain adaptation methods. Full article
(This article belongs to the Special Issue Advanced Signal and Image Processing for Applied Engineering)
Show Figures

Figure 1

18 pages, 39608 KB  
Article
Denoising Domain Adversarial Network Based on Attention Mechanism for Motor Fault Diagnosis in Real Industrial Environment
by Linjie Jin, Zhengqing Liu, Dawei Gu, Baisong Pan, Qiucheng Wang and Mohammad Fard
Machines 2026, 14(5), 462; https://doi.org/10.3390/machines14050462 - 22 Apr 2026
Viewed by 207
Abstract
Acoustic signal-based fault diagnosis offers a promising non-contact approach for rotating machinery. However, its practical application is usually affected by environmental noise. This paper presented a Denoising Attention Domain Adversarial Network (DDAN) for the robust fault diagnosis of wheel hub motors under severe [...] Read more.
Acoustic signal-based fault diagnosis offers a promising non-contact approach for rotating machinery. However, its practical application is usually affected by environmental noise. This paper presented a Denoising Attention Domain Adversarial Network (DDAN) for the robust fault diagnosis of wheel hub motors under severe noise interference. The proposed framework consists of the following two core modules: a DenseNet-based denoising module that adaptively suppresses background noise while retaining critical fault features, and a Stacked Autoencoder Domain Adversarial Network (SADAN) that integrates channel attention, spatial attention, and multi-head self-attention (MHSA) for refined feature extraction and classification. Such a hierarchical attention mechanism facilitates effective local noise suppression and global dependency capture. Validation on a hub motor fault dataset and publicly available online dataset demonstrates that compared to existing methods, DDAN achieves superior diagnostic accuracy across various noise levels and signal-to-noise ratios, improving SNR from -15.97 dB to 1.24 dB, achieving 82.71% accuracy under low SNR condition, and reaching 84.93% and 83.75% accuracy in cross-domain generalization tests. Furthermore, the comparison of the diagnostic accuracy of audio signals from different acoustic acquisition devices further verifies the practicality and potential of the system in low-cost industrial deployment. Full article
(This article belongs to the Section Electrical Machines and Drives)
Show Figures

Figure 1

16 pages, 2289 KB  
Proceeding Paper
An Efficient Hybrid Framework for Weld Defect Detection Using GAN, CNN and XGBoost
by Kalyanaraman Pattabiraman, Ashish Patil, Yash Gulavani, Ritik Malik and Atharva Gai
Eng. Proc. 2026, 130(1), 9; https://doi.org/10.3390/engproc2026130009 - 22 Apr 2026
Viewed by 202
Abstract
Automated detection of defects in welds are inevitable in the assurance of structural integrity, but this faces serious challenges due to the microscopic characteristics of the discontinuities, low visual contrast and infrequent occurrence of defect samples. Conventional deep learning methods, while accurate, often [...] Read more.
Automated detection of defects in welds are inevitable in the assurance of structural integrity, but this faces serious challenges due to the microscopic characteristics of the discontinuities, low visual contrast and infrequent occurrence of defect samples. Conventional deep learning methods, while accurate, often lack interpretability and exhibit low recall for rare defects. This paper proposes a novel hybrid system combining a Generative Adversarial Network (GAN), a Convolutional Neural Network (CNN), and Extreme Gradient Boosting (XGBoost 2.0.0) to enhance weld defect classification performance and transparency. Firstly, a Deep Convolutional GAN (DCGAN) creates synthetic images of the minority classes; thus, the problem of class imbalance is resolved. Then, a pretrained ResNet50V2 CNN is used to extract features of the deep layers from the original images as well as from the generated ones. After that, these features are fed into an XGBoost classifier, which uses tree-based learning to optimize classification results and make the process more understandable to the user. Furthermore, interpretation is also facilitated by Grad-CAM rendering of the CNN regions of interest and SHAP analysis to measure the involvement of the features in XGBoost. Experiments using the available LoHi-WELD datasets show that the overall accuracy is significantly improved, the per-class recall of the rare defects is also enhanced, and the robustness is also improved. The proposed hybrid method not only achieves better results but also generates visual/explainable output, which is very valuable when the system is implemented in industrial welding inspection systems. This paper serves as a liaison between the latest AI technology and the practical interpretability requirements of the mechanical and welding engineering fields. Full article
(This article belongs to the Proceedings of The 19th Global Congress on Manufacturing and Management (GCMM 2025))
Show Figures

Figure 1

25 pages, 6168 KB  
Article
PerDCGAN: A Perceptual Generative Framework for High-Fidelity Bearing Fault Diagnosis
by Yuantao Li, Ao Li, Xiaoli Wang and Jiancheng Yin
Appl. Sci. 2026, 16(8), 4054; https://doi.org/10.3390/app16084054 - 21 Apr 2026
Viewed by 306
Abstract
Data imbalance significantly hinders the performance of deep learning models in rolling bearing fault diagnosis. While Generative Adversarial Networks (GANs) are widely used for data augmentation, traditional architectures employing pixel-level loss functions often fail to capture complex time-frequency textures, resulting in blurred spectrograms [...] Read more.
Data imbalance significantly hinders the performance of deep learning models in rolling bearing fault diagnosis. While Generative Adversarial Networks (GANs) are widely used for data augmentation, traditional architectures employing pixel-level loss functions often fail to capture complex time-frequency textures, resulting in blurred spectrograms and the loss of transient fault characteristics. To address this, we propose a data augmentation framework based on a Perceptually Optimized Deep Convolutional GAN (PerDCGAN). By integrating a perceptual loss function derived from a pre-trained VGG-16 network, the generator is constrained at the feature level rather than the pixel level, explicitly enforcing the preservation of structural details and high-frequency impact patterns. Extensive experiments on the Case Western Reserve University (CWRU) and Paderborn University (PU) datasets demonstrate that the proposed method effectively mitigates spectral blurring. Ablation studies confirm the synergistic effect of the joint loss function. Furthermore, under extreme 0 dB noise conditions, the classifier augmented by PerDCGAN maintains a robust diagnostic accuracy of 89.65% on the PU dataset, significantly outperforming standard DCGAN and demonstrating strong potential for complex industrial applications. Full article
(This article belongs to the Special Issue Mechanical Fault Diagnosis and Signal Processing)
Show Figures

Figure 1

31 pages, 9766 KB  
Article
Benchmarking Conditional GANs in Industrial Marble Texture Synthesis via a Dual-Evaluation Framework
by António Alves de Campos, Margarida Figueiredo, Carlos M. A. Diogo, Gustavo Paneiro and Pedro Amaral
Appl. Sci. 2026, 16(8), 4028; https://doi.org/10.3390/app16084028 - 21 Apr 2026
Viewed by 144
Abstract
Deploying conditional Generative Adversarial Networks (cGANs) for industrial texture synthesis faces two barriers: the prohibitive cost of manual data annotation and the uncertain alignment between automated evaluation metrics and human perception. This study addresses both challenges for marble texture synthesis using 289 high-resolution [...] Read more.
Deploying conditional Generative Adversarial Networks (cGANs) for industrial texture synthesis faces two barriers: the prohibitive cost of manual data annotation and the uncertain alignment between automated evaluation metrics and human perception. This study addresses both challenges for marble texture synthesis using 289 high-resolution industrial scans. We adapt an unsupervised segmentation pipeline combining Simple Linear Iterative Clustering (SLIC) superpixels, Gaussian Mixture Models (GMMs), and graph cut optimization to extract vein structures without manual annotation. Four cGAN architectures—baseline cGAN, Pix2Pix, BicycleGAN, and GauGAN—are benchmarked using a dual-evaluation protocol contrasting ten automated metrics with structured human-centered assessment. The results reveal a significant metric–perception discrepancy. Pix2Pix achieved the best Fréchet Inception Distance (FID = 85.3) yet received the lowest human ratings due to periodic texture artifacts. GauGAN produced textures statistically indistinguishable from real marble, achieving a Visual Turing Pass Rate (VTPR) of 0.533 and a Mean Opinion Score on Marble Authenticity (MOS-MA) of 2.89, despite an inferior FID (87.3). These findings make three contributions: an annotation-free segmentation pipeline, empirical evidence that automated metrics alone are insufficient for architecture selection, and a dual-evaluation framework that establishes human-in-the-loop assessment as essential for quality-critical industrial deployment. Full article
Show Figures

Figure 1

20 pages, 621 KB  
Review
Conditional Generative AI in Oncology Diagnostics
by Chiara Frascarelli, Alberto Concardi, Elisa Mangione, Mariachiara Negrelli, Francesca Maria Porta, Michela Tulino, Joana Sorino, Antonio Marra, Nicola Fusco, Elena Guerini-Rocco and Konstantinos Venetis
Appl. Sci. 2026, 16(8), 4015; https://doi.org/10.3390/app16084015 - 21 Apr 2026
Viewed by 260
Abstract
The increasing complexity of oncology diagnostics requires advanced Clinical Decision Support Systems (CDSS) capable of integrating multimodal data. Traditional discriminative models often struggle with missing data and cross-modal dependencies. This review provides a novel, systematic analysis of conditional generative artificial intelligence (AI), including [...] Read more.
The increasing complexity of oncology diagnostics requires advanced Clinical Decision Support Systems (CDSS) capable of integrating multimodal data. Traditional discriminative models often struggle with missing data and cross-modal dependencies. This review provides a novel, systematic analysis of conditional generative artificial intelligence (AI), including Variational Autoencoders (VAEs), Generative Adversarial Networks (GANs), diffusion models and Multimodal Large Language Models (MLLMs), specifically tailored for oncological CDSS. We examine how these architectures move beyond simple prediction to learn joint data distributions, enabling robust data imputation, virtual staining, and automated clinical reporting. A central focus of this work is the assessment of translational application, identifying the gaps between experimental proof-of-concepts and clinical deployment. We address critical hurdles such as model hallucinations, domain shift, and demographic bias, providing a roadmap for biological consistency and regulatory compliance. This review highlights the transition from task-specific generators to multimodal reasoning systems. Ultimately, we argue that the integration of generative AI into diagnostic workflows is essential for precision oncology, provided that human-in-the-loop validation and uncertainty-aware inference remain central to their implementation. Full article
Show Figures

Figure 1

28 pages, 8218 KB  
Article
Robust and Adaptive Dual-Defense Framework Against Data Poisoning Attacks in Recommendation Systems
by Xiaocui Dang, Priyadarsi Nanda, Heng Xu, Haiyu Deng and Chunpeng Ge
Electronics 2026, 15(8), 1726; https://doi.org/10.3390/electronics15081726 - 19 Apr 2026
Viewed by 155
Abstract
Deep learning-based recommendation systems are highly vulnerable to data poisoning attacks, where adversaries manipulate user interactions to degrade model integrity. We hypothesize that combining an active robust loss with a passive GAN-based detection will significantly reduce poisoning impact in recommendation systems without sacrificing [...] Read more.
Deep learning-based recommendation systems are highly vulnerable to data poisoning attacks, where adversaries manipulate user interactions to degrade model integrity. We hypothesize that combining an active robust loss with a passive GAN-based detection will significantly reduce poisoning impact in recommendation systems without sacrificing utility. We propose a robust and adaptive dual-defense framework: the active defense integrates a crafted loss function to mitigate poisoning effects while maintaining model performance. The passive defense employs a Generative Adversarial Network (GAN)-based detection model to identify and filter poisoned data, enhancing detection accuracy and system security. The framework supports classical matrix factorization (MF) model and large language model (LLM)-based pipelines and scales to large datasets. Extensive experiments across multiple real-world datasets at varying poison rates show that our method outperforms representative defenses, consistently reducing attack success without sacrificing recommendation quality. The framework also admits a federated instantiation, where robust training and GAN-based detection run on clients and only privacy-preserving summaries are aggregated. The proposed method significantly improves the robustness and adaptability of recommendation systems under data poisoning attacks. Full article
Show Figures

Figure 1

25 pages, 8039 KB  
Article
Enhancing the Transferability of Generative Targeted Adversarial Attacks via Cosine-Based Logit Alignment
by Tengfei Shi, Shihai Wang and Bin Liu
Mathematics 2026, 14(8), 1370; https://doi.org/10.3390/math14081370 - 19 Apr 2026
Viewed by 142
Abstract
Adversarial examples reveal critical vulnerabilities in deep neural networks, posing significant risks in real-world deployment. In black-box settings, transferable targeted attacks rely on surrogate models but often suffer from low success rates. We argue that this limitation arises not only from surrogate-boundary overfitting [...] Read more.
Adversarial examples reveal critical vulnerabilities in deep neural networks, posing significant risks in real-world deployment. In black-box settings, transferable targeted attacks rely on surrogate models but often suffer from low success rates. We argue that this limitation arises not only from surrogate-boundary overfitting but also from insufficient alignment with the target semantic space, which restricts the ability of adversarial examples to encode target-specific characteristics. To address this issue, we propose Cosine-Based Logit Alignment (CBLA), a unified framework for transferable targeted attacks. CBLA replaces the conventional cross-entropy loss with a cosine similarity objective to enhance directional alignment in logit space and alleviate gradient saturation. In addition, a semantic-invariant transformation strategy is introduced to improve structural consistency and cross-model generalization. Experiments on the ImageNet validation set demonstrate that CBLA consistently improves targeted attack success rates, achieving an average gain of 4.55% over strong baselines across multiple architectures. Full article
Show Figures

Figure 1

Back to TopTop