Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,821)

Search Parameters:
Keywords = artificial image generation

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 1844 KB  
Article
AI-Enhanced Prognostic Model for Predicting Polyp Recurrence and Guiding Post-Polypectomy Surveillance Intervals Using the ERCPMP-V5 Dataset
by Sri Harsha Boppana, Sachin Sravan Kumar Komati, Ritwik Raj, Gautam Maddineni, Raja Chandra Chakinala, Pradeep Yarra, Venkata C. K. Sunkesula and Cyrus David Mintz
J. Clin. Med. 2026, 15(9), 3303; https://doi.org/10.3390/jcm15093303 (registering DOI) - 26 Apr 2026
Abstract
Introduction: Colorectal cancer remains a leading cause of cancer-related morbidity and mortality, with adenomatous polyps representing a common precursor. Post-polypectomy polyp recurrence represents a significant risk of colorectal cancer, driving periodic colonoscopy surveillance and polypectomy as needed. In this study, we explore a [...] Read more.
Introduction: Colorectal cancer remains a leading cause of cancer-related morbidity and mortality, with adenomatous polyps representing a common precursor. Post-polypectomy polyp recurrence represents a significant risk of colorectal cancer, driving periodic colonoscopy surveillance and polypectomy as needed. In this study, we explore a multimodal machine learning approach that integrates endoscopic imaging with clinical and pathology data to improve recurrence risk prediction and support individualized surveillance planning. Methods: We developed and evaluated a multimodal artificial intelligence (AI) model to predict post-polypectomy colorectal polyp recurrence using the ERCPMP-v5 dataset. The cohort included 217 patients with 796 high-resolution endoscopic RGB images and 21 endoscopic videos; video data were converted to still frames at 2 frames per second. Images and frames were resized to 224 × 224 pixels and normalized. Patient-level demographic, morphological (Paris, Kudo Pit, JNET), anatomical, and pathological variables were encoded using standard scaling for continuous features and one-hot encoding for categorical features. Visual representations were extracted using a pretrained Vision Transformer backbone (ViT-Base-Patch16-224) with frozen weights. Structured metadata (79 variables) was encoded using a multilayer perceptron. A late fusion framework used image and metadata representations to generate a recurrence probability via a sigmoid classifier; probabilities were thresholded at 0.5 for binary prediction. Model performance was evaluated on a held-out test set using accuracy, precision, recall, F1-score, and area under the receiver operating characteristic curve (AUC). We additionally compared fusion performance with image-only and metadata-only baselines. Predicted probabilities were translated to surveillance recommendations using risk tiers: low risk (0.00 ≤ p < 0.20), moderate risk (0.20 ≤ p < 0.50), and high risk (p ≥ 0.50). Results: On the test set, the multimodal fusion model achieved 90.4% accuracy, 86.7% precision, 83.1% recall, 84.9% F1-score, and an AUC of 0.920. The image-only model achieved 84.6% accuracy (AUC 0.880), and the metadata-only model achieved 81.9% accuracy (AUC 0.850), indicating improved performance with multimodal fusion. Risk stratification enabled surveillance recommendations of 1–3 years for low risk, 6–12 months for moderate risk, and 3–6 months for high risk. Conclusions: A late-fusion multimodal model integrating endoscopic imaging with structured clinical and pathology variables demonstrated excellent performance for predicting post-polypectomy recurrence and generated actionable risk-based surveillance intervals. This approach may support individualized follow-up planning and more efficient allocation of surveillance resources, while prioritizing timely evaluation for patients at higher predicted risk. Full article
Show Figures

Figure 1

50 pages, 17736 KB  
Article
Swin–YOLOv12: A Hybrid Transformer-Based Deep Learning Approach for Enhanced Real-Time Brain Tumor Detection in MRI Images
by Mubashar Tariq and Kiho Choi
Mathematics 2026, 14(9), 1447; https://doi.org/10.3390/math14091447 (registering DOI) - 25 Apr 2026
Abstract
Brain tumors (BTs) arise from the abnormal growth of cells within brain tissue and may spread rapidly, making them a major cause of mortality worldwide. Early detection of BTs remains highly challenging due to the brain’s complex structure and the heterogeneous nature of [...] Read more.
Brain tumors (BTs) arise from the abnormal growth of cells within brain tissue and may spread rapidly, making them a major cause of mortality worldwide. Early detection of BTs remains highly challenging due to the brain’s complex structure and the heterogeneous nature of tumors. Magnetic Resonance Imaging (MRI) provides detailed information about tumor size, location, and shape, thereby supporting clinical decision-making for treatments such as chemotherapy, radiation therapy, and surgery. Traditional machine learning (ML) approaches mainly rely on manual feature extraction, whereas recent advances in Computer-Aided Diagnosis (CAD) and deep learning (DL) have enabled more accurate detection of small and complex tumor regions. To improve automated tumor detection, we propose a hybrid Swin–YOLO framework that combines the Swin Transformer (ST) with the latest CNN-based YOLOv12 model. In this framework, the Swin Transformer serves as the main backbone for feature extraction, while the Feature Pyramid Network (FPN) and Path Aggregation Network (PANet) are employed in the neck to better capture multi-scale features. For training, we used the publicly available Br35H dataset and applied data augmentation to enhance the model’s robustness and generalization capability. The experimental results show that the proposed framework achieved 99.7% accuracy, 99.4% mAP@50, and 87.2% mAP@50:95. Furthermore, we incorporated Explainable Artificial Intelligence (XAI) techniques, including Grad-CAM and SHAP, to improve the interpretability of the model by visually highlighting the tumor regions that contributed most to the prediction. In addition, we developed NeuroVision AI, a web-based application designed to support faster and more accurate clinical decision-making. Although the proposed model demonstrated strong performance on the dataset, these results should be interpreted within the context of the current experimental setting. Full article
20 pages, 2912 KB  
Article
Leveraging Generative AI for High-Fidelity 360° Spatial Images: Methodological Validation for Use as Experimental Stimuli
by Yoojin Han and Joowon Jeong
Buildings 2026, 16(9), 1679; https://doi.org/10.3390/buildings16091679 - 24 Apr 2026
Abstract
Despite its efficiency, the structural integrity and geometric accuracy of artificial intelligence (AI)-generated imagery used in environmental psychology experiments have not been sufficiently validated. This study investigated the methodological validity and substitutability of generative AI-generated 360° images as experimental stimuli for indoor environmental [...] Read more.
Despite its efficiency, the structural integrity and geometric accuracy of artificial intelligence (AI)-generated imagery used in environmental psychology experiments have not been sufficiently validated. This study investigated the methodological validity and substitutability of generative AI-generated 360° images as experimental stimuli for indoor environmental research. Using a three-stage framework, we generated base panoramas with controlled structural parameters, integrated greenery via AI-based inpainting, and conducted multifaceted validation through objective quality metrics and expert assessments. Quantitative results confirmed high technical integrity, indicating that structural distortions at panoramic stitching points were effectively minimized. Furthermore, the AI-generated stimuli maintained stable visual quality across varying greenery densities. Expert evaluations confirmed that the AI-driven approach significantly outperforms conventional 3D modeling, particularly in terms of presence and realism. By achieving high usability and spatial integrity scores, we established a novel standard for employing generative AI to create high-fidelity virtual environments for architectural and psychological research. Full article
(This article belongs to the Special Issue Artificial Intelligence in Architecture and Interior Design)
Show Figures

Figure 1

20 pages, 4990 KB  
Article
Curvature Radius Measurement Based on Interferogram Analysis and Deep Learning Model
by Yan-Yi Li, Chuen-Lin Tien, Hsi-Fu Shih, Han-Yen Tu and Chih-Cheng Chen
Photonics 2026, 13(5), 416; https://doi.org/10.3390/photonics13050416 - 24 Apr 2026
Abstract
Accurate estimation of curvature radius from interference fringes is critical in optical metrology and precision manufacturing. Conventional interferogram analytical approaches often require manual intervention and are sensitive to fringe variations related to noise and environmental vibrations. To address these limitations, we combine an [...] Read more.
Accurate estimation of curvature radius from interference fringes is critical in optical metrology and precision manufacturing. Conventional interferogram analytical approaches often require manual intervention and are sensitive to fringe variations related to noise and environmental vibrations. To address these limitations, we combine an improved Twyman–Green interferometer with different artificial intelligence (AI) deep learning models and utilize a self-developed MATLAB analysis program to propose a non-destructive and rapid measurement system for optical coating substrates. The proposed AI-assisted Twyman–Green interferometric system differs fundamentally from conventional wavefront sensing techniques in both principle and implementation. This paper utilizes the Twyman–Green interferometer to generate interference fringe datasets on B270 glass and sapphire substrates, and employs convolutional neural network (CNN), ResNet-18, and VGG-16 models for training and evaluation. The proposed method integrates image enhancement, fringe pattern clustering, and analysis and validation based on fast Fourier transform (FFT). Experimental results show that ResNet-18 outperforms other models, with a mean absolute percentage error of 5.44% on sapphire substrates and 3.40% on B270 glass substrates. These findings highlight the effectiveness and robustness of deep learning models, especially residual networks, in automatic ROC prediction for optical measurement applications. Full article
Show Figures

Figure 1

23 pages, 4572 KB  
Article
LLaMA-XR: A Novel Framework for Radiology Report Generation Using LLaMA and QLoRA Fine Tuning
by Md. Zihad Bin Jahangir, Muhammad Ashad Kabir, Sumaiya Akter, Israt Jahan and Minh Chau
Bioengineering 2026, 13(5), 493; https://doi.org/10.3390/bioengineering13050493 - 23 Apr 2026
Viewed by 164
Abstract
Background: The goal of automated radiology report generation is to help radiologists in their task of creating descriptive reports from chest radiographs. However, the process of creating coherent and contextually accurate reports has been challenging, mainly due to the intricacies of medical language [...] Read more.
Background: The goal of automated radiology report generation is to help radiologists in their task of creating descriptive reports from chest radiographs. However, the process of creating coherent and contextually accurate reports has been challenging, mainly due to the intricacies of medical language and the need to correlate visual data with textual descriptions. Methods: This study presents LLaMA-XR, a novel framework that integrates Meta LLaMA 3.1 Large Language Model with DenseNet-121-based image embeddings and Quantized Low-Rank Adaptation (QLoRA) fine-tuning. Results: The experiment conducted on the IU X-ray dataset demonstrates that LLaMA-XR outperforms a range of state-of-the-art methods. It achieves an ROUGE-L score of 0.433 and a METEOR score of 0.336, establishing new performance benchmarks in the domain. Conclusions: These results underscore LLaMA-XR’s potential as an effective artificial intelligence system for automated radiology reporting, offering enhanced performance. Full article
(This article belongs to the Special Issue AI-Driven Imaging and Analysis for Biomedical Applications)
11 pages, 209 KB  
Article
Epistemic Automation and the Deformation of the Human: Artificial Intelligence and the Reconfiguration of Theological Anthropology
by Åke Elden
Religions 2026, 17(5), 515; https://doi.org/10.3390/rel17050515 (registering DOI) - 23 Apr 2026
Viewed by 74
Abstract
This paper argues that the most significant challenge artificial intelligence poses to theological anthropology is not ontological but epistemic. Rather than asking whether machines can think, feel, or bear the image of God, this paper redirects attention to the prior question of what [...] Read more.
This paper argues that the most significant challenge artificial intelligence poses to theological anthropology is not ontological but epistemic. Rather than asking whether machines can think, feel, or bear the image of God, this paper redirects attention to the prior question of what happens to the human when core epistemic capacities, judgment, discernment, interpretive authority, and moral reasoning are progressively delegated to computational systems. Drawing on the concept of epistemic automation, understood as the systematic transfer of knowledge-producing functions from human agents to algorithmic processes, this paper develops a threefold analytical framework. First, it distinguishes epistemic authority from ontological status as the more productive locus for theological anthropological inquiry. Second, it introduces the distinction between fluency and understanding as an anthropological boundary condition that AI renders newly visible. Third, it analyses delegated cognition as a form of agency deformation with theological significance. The paper concludes that theological anthropology must move beyond reactive commentary on AI and instead generate a theory of the human under conditions of epistemic transformation. The argument engages constructively with philosophy of technology, social epistemology, and Christian theological traditions to offer a framework applicable across confessional boundaries. Full article
23 pages, 47800 KB  
Article
AIGC-Driven Short Video Generation Based on the Controllable Multimodal Fusion Architecture
by Yan Zhu, Wei Li, Caixia Fan and Lu Yu
Electronics 2026, 15(9), 1783; https://doi.org/10.3390/electronics15091783 - 22 Apr 2026
Viewed by 220
Abstract
The utilization of Artificial Intelligence-Generated Content (AIGC) has attracted widespread attention in video content creation. To generate high-quality videos, this paper presents a controllable multimodal fusion architecture for AIGC-driven short-video production. This architecture employs hierarchical constraint mechanisms and a multimodal attention fusion mechanism [...] Read more.
The utilization of Artificial Intelligence-Generated Content (AIGC) has attracted widespread attention in video content creation. To generate high-quality videos, this paper presents a controllable multimodal fusion architecture for AIGC-driven short-video production. This architecture employs hierarchical constraint mechanisms and a multimodal attention fusion mechanism to enhance video content coherence and user controllability. Specifically, a scene coherence scheme is first designed to construct graph-based global and transition-level constraints by integrating text descriptions, reference images, and audio features. By leveraging the extracted style vector data, preliminary video clips are then generated through a combination of the cross-modal fusion unit and the spatio-temporal consistency unit. Finally, a fine-grained adjustment mechanism is implemented to ensure logical consistency and stylistic uniformity in the AIGC-generated videos. Experimental results indicate that the proposed architecture improves generation quality, controllability, and cross-segment coherence under the adopted evaluation settings. Full article
Show Figures

Figure 1

25 pages, 37592 KB  
Article
Deep-Learning-Based Mobile Application for Real-Time Recognition of Cultural Artifacts in Museum Environments
by Pablo Minango, Marcelo Zambrano, Carmen Inés Huerta Suarez and Juan Minango
Appl. Sci. 2026, 16(9), 4064; https://doi.org/10.3390/app16094064 - 22 Apr 2026
Viewed by 157
Abstract
Dissemination and conservation of cultural heritage have been challenged by continued accessibility in museums, where traditional information delivery systems are at times ineffective in terms if interaction with visitors. The current paper investigates RumiArt IA, a mobile application, to identify cultural objects in [...] Read more.
Dissemination and conservation of cultural heritage have been challenged by continued accessibility in museums, where traditional information delivery systems are at times ineffective in terms if interaction with visitors. The current paper investigates RumiArt IA, a mobile application, to identify cultural objects in real-time, remaining fully in the scope of this line of research without relying on internet connectivity. The system, which is developed based on the Rumiñahui Museum and Cultural Center, Ecuador, uses transfer learning in the MobileNetV2 architecture with INT8 post-training quantization to identify 21 cultural artifacts spread across six thematic rooms. The experiment involved building a dataset of 36,000 images under diverse lighting conditions, viewing angles, and distances; furthermore, artificial transformations were explicitly crafted to simulate real museum conditions such as glass reflections and non-frontal capture angles. Quantization was used to reduce each model to 775 KB as compared with the 2.4 MB, with accuracy loss not reaching more than 0.5 percent (DKL < 0.05). Assessment of 9450 validation images yielded a general accuracy of 92.2%, with an inference time of 63 ms on current devices with a high throughput and 215 ms on mid-range hardware from 2020. Practical validation involving 50 visitors of the museum showed a success rate of 93.7%, with average user satisfaction at 8.5/10 and 87%, indicating they would recommend the application. An in-depth error study of the most difficult room (88.3% accuracy) indicated that 47% of the errors were due to the angles of the camera, which blocked out distinguishing features, and 22% were caused by display case reflections and the shadows of the visitors. These results indicate that end-to-end machine learning can provide consistent cultural heritage recognition in resource-constrained settings but its efficiency is susceptible to physical capture factors that cannot be resolved by data augmentation. Offline mode and low memory footprint (less than 90 MB when loaded on six models) of the system are especially relevant to application in situations where there is no guarantee of cloud connectivity. Full article
(This article belongs to the Special Issue Intelligent Interaction in Cultural Heritage)
Show Figures

Figure 1

35 pages, 2823 KB  
Article
FedCycle: An Improved Federated Learning Framework for Assessment Across Modalities and Domains
by Betul Dundar, Ebru Akcapinar Sezer, Feyza Yildirim Okay and Suat Ozdemir
Electronics 2026, 15(8), 1752; https://doi.org/10.3390/electronics15081752 - 21 Apr 2026
Viewed by 195
Abstract
Artificial Intelligence (AI) systems based on traditional Deep Learning (DL) are expected to play a leading role in the early detection of various diseases in healthcare applications. However, there are two major drawbacks of these systems: protecting patient privacy and obtaining sufficiently large, [...] Read more.
Artificial Intelligence (AI) systems based on traditional Deep Learning (DL) are expected to play a leading role in the early detection of various diseases in healthcare applications. However, there are two major drawbacks of these systems: protecting patient privacy and obtaining sufficiently large, high-quality datasets to train reliable models. In traditional DL, collecting data from different sources on a single central server increases system complexity and raises serious privacy and security concerns. Federated Learning (FL) makes it possible to train models locally at multiple data locations while collaboratively improving a global model without exposing raw data, making it a promising architectural solution for privacy preservation. Although previous studies have reported that FL can achieve performance comparable to centralized DL approaches, traditional FL approaches often struggle to maintain consistent performance across different settings. This limitation becomes more noticeable when heterogeneous data distributions, modalities, and domains are involved. In these situations, client drift, overfitting, and generalization capability of the global model arise as major challenges. Thus, this study presents FedCycle as an incremental improvement of the FedAvg algorithm. It modifies the aggregation frequency. It aims to overcome these drawbacks and make the global model more stable and efficient. The FedCycle eliminates centralized data collection, enhances data security, and effectively reduces client drift and overfitting by supporting model training across heterogeneous data distributions, modalities, and domains. The performance evaluation involves extensive experiments using various real-world breast cancer image datasets, namely BREAKHIS, ROBOFLOW, RSNA, BUSI, and BCFPP. The presented method is evaluated against both traditional DL and FL approaches using accuracy, precision, recall, F1-score, and AUC. The findings confirm that applying fine-tuning within FedCycle reduces overfitting during training. As a result, FedCycle achieves performance improvements of 7.75% and 4.65% in accuracy and F1-score on the RSNA and BCFPP datasets compared to traditional DL approaches, while also providing an average improvement of approximately 1.5% in accuracy and F1-score across BREAKHIS, ROBOFLOW, and BUSI datasets compared to FedAvg. Full article
(This article belongs to the Special Issue Federated Learning and Its Application)
Show Figures

Figure 1

22 pages, 4789 KB  
Article
DTF-STCANet: A Dual Time–Frequency Swin Transformer and ConvNeXt Attention Network for Heart Sound Classification
by Mehmet Nail Bilen, Fatih Mehmet Çelik, Mehmet Ali Kobat and Fatih Demir
Diagnostics 2026, 16(8), 1234; https://doi.org/10.3390/diagnostics16081234 - 21 Apr 2026
Viewed by 201
Abstract
Background/Objectives: Cardiovascular diseases are the leading cause of death worldwide. Therefore, early diagnosis and treatment of these diseases are of critical importance. Stethoscopes are the easiest and fastest medical devices for the initial diagnosis of cardiovascular diseases. However, interpreting heart sounds requires [...] Read more.
Background/Objectives: Cardiovascular diseases are the leading cause of death worldwide. Therefore, early diagnosis and treatment of these diseases are of critical importance. Stethoscopes are the easiest and fastest medical devices for the initial diagnosis of cardiovascular diseases. However, interpreting heart sounds requires considerable expertise. The use of artificial intelligence in healthcare for decision support has increased and become popular recently. Methods: The popular 2016 PhysioNet/CinC Challenge dataset, consisting of phonocardiogram (PCG) signals, was used to implement the proposed approach. Spectrogram and continuous wavelet transform (CWT) images of the PCG signals were first generated. This increased the distinguishability of the data in terms of both time and frequency components. These two-input images were tested on the developed Dual Time–Frequency Swin Transformer–ConvNeXt Attention Network (DTF-STCANet) model. To further improve classification accuracy, the Weighted KNN algorithm was preferred during the classification phase. Results: With the proposed approach, a 99.29% classification accuracy was achieved. Performance was compared with other state-of-the-art models. Conclusions: The proposed approach, through the integration of PCG signals with artificial intelligence, further strengthens the concept of early diagnosis of heart disease. Full article
(This article belongs to the Special Issue Artificial Intelligence in Biomedical Diagnostics and Analysis 2025)
Show Figures

Figure 1

18 pages, 349 KB  
Review
Autoimmune Hepatitis: Emerging Frontiers in Research and Clinical Management
by Armando Curto, Irene Scami, Giulia Gliottone, Rocco G. Iamello, Erica N. Lynch and Andrea Galli
Gastrointest. Disord. 2026, 8(2), 20; https://doi.org/10.3390/gidisord8020020 - 20 Apr 2026
Viewed by 312
Abstract
Autoimmune hepatitis (AIH) is a chronic immune-mediated liver disorder that, without treatment, can advance to fibrosis and cirrhosis. Although standard regimens with corticosteroids and thiopurines have significantly improved survival, many patients still experience relapses and drug-related toxicity, highlighting the urgent need for alternative [...] Read more.
Autoimmune hepatitis (AIH) is a chronic immune-mediated liver disorder that, without treatment, can advance to fibrosis and cirrhosis. Although standard regimens with corticosteroids and thiopurines have significantly improved survival, many patients still experience relapses and drug-related toxicity, highlighting the urgent need for alternative strategies. Recent studies underscore AIH’s multifactorial nature, revealing intricate interactions among genetic susceptibility, environmental triggers, and dysregulated immune responses. Next-generation diagnostics, ranging from novel biomarkers to high-resolution imaging, are enhancing early detection and more precise disease classification. At the same time, multi-omics analyses and artificial-intelligence-based models are refining predictions of disease trajectory and therapeutic response. On the treatment horizon, investigational options such as targeted immunomodulators, B-cell–depleting therapies, and cell-based interventions aim to achieve durable remission while minimizing adverse effects. This review critically appraises these advances and explores how integrating epidemiological insights with cutting-edge research in pathogenesis, diagnostics, and therapy could pave the way for more personalized and effective management of AIH. Full article
(This article belongs to the Special Issue Feature Papers in Gastrointestinal Disorders in 2025–2026)
21 pages, 3561 KB  
Article
A CLIP-Guided Multi-Objective Optimization Framework for Sustainable Design: Integrating Aesthetic Evaluation, Energy Efficiency, and Life Cycle Environmental Performance
by Hanwen Zhang, Myun Kim, Hao Hu and Yitong Wang
Sustainability 2026, 18(8), 4064; https://doi.org/10.3390/su18084064 - 19 Apr 2026
Viewed by 299
Abstract
Achieving sustainable design requires balancing environmental performance, resource efficiency, functional feasibility, and aesthetic acceptance throughout the product life cycle. However, traditional design approaches often struggle to quantitatively integrate subjective aesthetic evaluation with objective sustainability indicators such as energy consumption, carbon emissions, and material [...] Read more.
Achieving sustainable design requires balancing environmental performance, resource efficiency, functional feasibility, and aesthetic acceptance throughout the product life cycle. However, traditional design approaches often struggle to quantitatively integrate subjective aesthetic evaluation with objective sustainability indicators such as energy consumption, carbon emissions, and material recyclability. To address this challenge, this study proposes a semantic-guided multi-objective optimization framework for sustainable design that integrates cross-modal aesthetic evaluation with life cycle environmental performance assessment. The proposed framework employs a Contrastive Language–Image Pre-training (CLIP)-based semantic evaluation mechanism to translate abstract sustainability and aesthetic concepts into quantifiable design features, enabling consistent assessment across diverse design solutions. These semantic features are further optimized using a multi-objective evolutionary optimization strategy to simultaneously minimize energy consumption and carbon emissions while maximizing material recovery and design quality. Life cycle environmental indicators derived from OpenLCA datasets are incorporated into the optimization process to ensure practical sustainability relevance. The experimental results demonstrate that the proposed framework achieves a superior performance compared with benchmark optimization methods. Specifically, carbon emission equivalents are reduced to as low as 12.3 kg CO2e, material recovery rates exceed 92%, and total computational energy consumption is reduced by more than 40% relative to comparative models. In addition, the framework shows strong stability and convergence efficiency while maintaining a high aesthetic evaluation accuracy in high-quality design ranges. The findings indicate that the proposed approach provides an effective pathway for integrating aesthetic value with environmental responsibility in sustainable design practice. This framework supports low-carbon and resource-efficient product development and offers practical insights for sustainable manufacturing, circular design, and environmentally conscious innovation. Full article
(This article belongs to the Special Issue Artificial Intelligence and Sustainable Development)
Show Figures

Figure 1

25 pages, 20117 KB  
Article
Intelligent Corrosion Diagnosis of High-Strength Bolts Based on Multi-Modal Feature Fusion and APO-XGBoost
by Hanyue Zhang, Yin Wu, Bo Sun, Yanyi Liu and Wenbo Liu
Sensors 2026, 26(8), 2520; https://doi.org/10.3390/s26082520 - 19 Apr 2026
Viewed by 268
Abstract
High-strength bolts are critical structural components that are highly susceptible to corrosion in complex environments, posing significant threats to structural safety and reliability. Although acoustic emission (AE) technology has been widely applied in structural health monitoring, existing studies mainly focus on damage mode [...] Read more.
High-strength bolts are critical structural components that are highly susceptible to corrosion in complex environments, posing significant threats to structural safety and reliability. Although acoustic emission (AE) technology has been widely applied in structural health monitoring, existing studies mainly focus on damage mode identification or source localization, while the identification of corrosion evolution stages based on AE signals remains insufficient. This study develops an intelligent corrosion diagnosis framework for high-strength bolts by integrating multimodal feature fusion and optimized machine learning. AE signals are first collected from the near-end and far-end of bolts using a wireless sensor network and then transformed into time–frequency representations via continuous wavelet transform (CWT). The resulting time–frequency images are fed into a modified ResNet-18 network to extract deep features, while statistical features are simultaneously extracted from the raw signals to preserve global information. These heterogeneous features are subsequently fused to form a comprehensive representation of corrosion characteristics. Furthermore, an artificial protozoa optimizer (APO) is introduced to adaptively optimize the hyperparameters of the XGBoost model. The results demonstrate that AE signals generated by hammering bolts with different corrosion levels can be successfully distinguished. The proposed method achieves high accuracy in corrosion stage classification and outperforms conventional approaches. Even when evaluated on an additional M30 bolt dataset, the proposed method maintains robust performance, demonstrating excellent generalization capability across different bolt sizes. These results demonstrate the practical potential of the proposed method for intelligent bolt corrosion diagnosis. Full article
(This article belongs to the Section Fault Diagnosis & Sensors)
Show Figures

Figure 1

29 pages, 3416 KB  
Article
Enhancing Collaborative AI Learning: A Blockchain-Secured, Edge-Enabled Platform for Multimodal Education in IIoT Environments
by Ahsan Rafiq, Eduard Melnik, Alexey Samoylov, Alexander Kozlovskiy and Irina Safronenkova
Big Data Cogn. Comput. 2026, 10(4), 123; https://doi.org/10.3390/bdcc10040123 - 17 Apr 2026
Viewed by 421
Abstract
As industries deploy more connected devices in factories, warehouses, and smart facilities, the need for artificial intelligence (AI) systems that can operate securely in distributed, data-intensive environments is growing. Traditional centralized learning and online education platforms struggle when students and systems have to [...] Read more.
As industries deploy more connected devices in factories, warehouses, and smart facilities, the need for artificial intelligence (AI) systems that can operate securely in distributed, data-intensive environments is growing. Traditional centralized learning and online education platforms struggle when students and systems have to process real-time streams (sensors, video, text) with strict latency and privacy requirements. To address this challenge, a blockchain-secured, edge-enabled multimodal federated learning framework tailored for Industrial IoT (IIoT) environments is proposed. The model integrates four key layers: (i) a blockchain layer that provides credentialing, transparency, and token-based incentives; (ii) a multimodal community layer that supports group formation, peer consensus, and cross-modal learning across text, images, audio, and sensor data; (iii) an edge computing layer that enables low-latency task offloading and secure training within Intel SGX enclaves; and (iv) a data layer that applies pre-processing, differential privacy, and synthetic augmentation to safeguard sensitive information. Experiments on industrial multimodal datasets demonstrate 42% faster model aggregation, 78.9% multimodal accuracy, and 1.9% accuracy loss under ε = 1.0 differential privacy. This shows a scalable and practical path for decentralized AI training in next-generation IIoT systems, confirming the possibility of technical support for educational processes. However, the conducted research requires a validation of pedagogical effectiveness. Full article
Show Figures

Figure 1

8 pages, 1309 KB  
Proceeding Paper
NEGOTIA: Developing Visual Literacy and Bias Awareness for GenAI
by Giuseppina Debbi and Federico Rodolfo Maiocco
Proceedings 2026, 139(1), 9; https://doi.org/10.3390/proceedings2026139009 - 17 Apr 2026
Viewed by 222
Abstract
Images generated by artificial intelligence recombine visual fragments learned from datasets, producing representations based on criteria of semantic proximity and aesthetic familiarity. These images lie in an intermediate zone between verisimilitude and statistical construction, requiring new interpretative skills to understand their nature and [...] Read more.
Images generated by artificial intelligence recombine visual fragments learned from datasets, producing representations based on criteria of semantic proximity and aesthetic familiarity. These images lie in an intermediate zone between verisimilitude and statistical construction, requiring new interpretative skills to understand their nature and limitations. This paper explores the need to develop visual literacy for generative AI, understood as the critical ability to analyse generation processes, recognise implicit biases, and verify the consistency of the representations produced. Through some case studies, prompting is analysed as a dialogical and reflective practice that highlights recurring patterns in datasets and diffusion models. The cases highlight how automatic composition tends to reproduce dominant cultural patterns related to gender, posture, and professional role. This paper introduces NEGOTIA, a seven-step framework designed to foster critical and operational visual literacy, applicable in educational and design contexts where synthetic images function as tools for representation, communication, and verification. NEGOTIA offers a replicable model for education and design practice. Full article
Show Figures

Figure 1

Back to TopTop