Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (106)

Search Parameters:
Keywords = multi-view prior

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 3666 KB  
Article
Reinforcement Learning Enabled Intelligent Process Monitoring and Control of Wire Arc Additive Manufacturing
by Allen Love, Saeed Behseresht and Young Ho Park
J. Manuf. Mater. Process. 2025, 9(10), 340; https://doi.org/10.3390/jmmp9100340 (registering DOI) - 18 Oct 2025
Abstract
Wire Arc Additive Manufacturing (WAAM) has been recognized as an efficient and cost-effective metal additive manufacturing technique due to its high deposition rate and scalability for large components. However, the quality and repeatability of WAAM parts are highly sensitive to process parameters such [...] Read more.
Wire Arc Additive Manufacturing (WAAM) has been recognized as an efficient and cost-effective metal additive manufacturing technique due to its high deposition rate and scalability for large components. However, the quality and repeatability of WAAM parts are highly sensitive to process parameters such as arc voltage, current, wire feed rate, and torch travel speed, requiring advanced monitoring and adaptive control strategies. In this study, a vision-based monitoring system integrated with a reinforcement learning framework was developed to enable intelligent in situ control of WAAM. A custom optical assembly employing mirrors and a bandpass filter allowed simultaneous top and side views of the melt pool, enabling real-time measurement of layer height and width. These geometric features provide feedback to a tabular Q-learning algorithm, which adaptively adjusts voltage and wire feed rate through direct hardware-level control of stepper motors. Experimental validation across multiple builds with varying initial conditions demonstrated that the RL controller stabilized layer geometry, autonomously recovered from process disturbances, and maintained bounded oscillations around target values. While systematic offsets between digital measurements and physical dimensions highlight calibration challenges inherent to vision-based systems, the controller consistently prevented uncontrolled drift and corrected large deviations in deposition quality. The computational efficiency of tabular Q-learning enabled real-time operation on standard hardware without specialized equipment, demonstrating an accessible approach to intelligent process control. These results establish the feasibility of reinforcement learning as a robust, data-efficient control technique for WAAM, capable of real-time adaptation with minimal prior process knowledge. With improved calibration methods and expanded multi-physics sensing, this framework can advance toward precise geometric accuracy and support broader adoption of machine learning-based process monitoring and control in metal additive manufacturing. Full article
Show Figures

Figure 1

18 pages, 707 KB  
Article
Reading Minds, Sparking Ideas: How Machiavellian Leaders Boost Team Creativity Through Cross-Understanding
by Yihang Yan, Hongzhen Lei, Hui Xiong, Yuanzhe Liu and Xiaoqian Qu
Adm. Sci. 2025, 15(10), 400; https://doi.org/10.3390/admsci15100400 (registering DOI) - 18 Oct 2025
Abstract
This study investigates the impact of Machiavellian leadership on team creativity through the mediating role of cross-understanding and the moderating effect of task interdependence. While prior research has emphasized the negative consequences of Machiavellian tendencies, we argue that in highly interdependent team settings—such [...] Read more.
This study investigates the impact of Machiavellian leadership on team creativity through the mediating role of cross-understanding and the moderating effect of task interdependence. While prior research has emphasized the negative consequences of Machiavellian tendencies, we argue that in highly interdependent team settings—such as project-based groups in technology, manufacturing, and financial enterprises—such leaders may foster constructive processes that enhance innovation. Drawing on social learning and trait activation theories, we conducted a multi-source survey of 86 teams (379 employees) in Chinese organizations. Team members assessed task interdependence and cross-understanding, while leaders reported their own Machiavellian tendencies and rated team creativity. Results show that Machiavellian leadership predicts team creativity indirectly through cross-understanding, with task interdependence strengthening this pathway. Theoretically, this study enriches leadership and creativity research by providing a nuanced view of how dark traits can stimulate team-level creativity through cognitive interaction mechanisms and by identifying task interdependence as a boundary condition. Practically, the findings suggest that organizations should recognize the creative potential of Machiavellian leaders in high-interdependence contexts, channel their ambition toward innovation goals, and design workflows that promote cross-understanding and collaboration. Full article
(This article belongs to the Special Issue The Role of Leadership in Fostering Positive Employee Relationships)
Show Figures

Figure 1

20 pages, 6483 KB  
Article
Loop-MapNet: A Multi-Modal HDMap Perception Framework with SDMap Dynamic Evolution and Priors
by Yuxuan Tang, Jie Hu, Daode Zhang, Wencai Xu, Feiyu Zhao and Xinghao Cheng
Appl. Sci. 2025, 15(20), 11160; https://doi.org/10.3390/app152011160 - 17 Oct 2025
Abstract
High-definition maps (HDMaps) are critical for safe autonomy on structured roads. Yet traditional production—relying on dedicated mapping fleets and manual quality control—is costly and slow, impeding large-scale, frequent updates. Recently, standard-definition maps (SDMaps) derived from remote sensing have been adopted as priors to [...] Read more.
High-definition maps (HDMaps) are critical for safe autonomy on structured roads. Yet traditional production—relying on dedicated mapping fleets and manual quality control—is costly and slow, impeding large-scale, frequent updates. Recently, standard-definition maps (SDMaps) derived from remote sensing have been adopted as priors to support HDMap perception, lowering cost but struggling with subtle urban changes and localization drift. We propose Loop-MapNet, a self-evolving, multimodal, closed-loop mapping framework. Loop-MapNet effectively leverages surround-view images, LiDAR point clouds, and SDMaps; it fuses multi-scale vision via a weighted BiFPN, and couples PointPillars BEV and SDMap topology encoders for cross-modal sensing. A Transformer-based bidirectional adaptive cross-attention aligns SDMap with online perception, enabling robust fusion under heterogeneity. We further introduce a confidence-guided masked autoencoder (CG-MAE) that leverages confidence and probabilistic distillation to both capture implicit SDMap priors and enhance the detailed representation of low-confidence HDMap regions. With spatiotemporal consistency checks, Loop-MapNet incrementally updates SDMaps to form a perception–mapping–update loop, compensating remote-sensing latency and enabling online map optimization. On nuScenes, within 120 m, Loop-MapNet attains 61.05% mIoU, surpassing the best baseline by 0.77%. Under extreme localization errors, it maintains 60.46% mIoU, improving robustness by 2.77%; CG-MAE pre-training raises accuracy in low-confidence regions by 1.72%. These results demonstrate advantages in fusion and robustness, moving beyond one-way prior injection and enabling HDMap–SDMap co-evolution for closed-loop autonomy and rapid SDMap refresh from remote sensing. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

25 pages, 2213 KB  
Article
Multi-Aligned and Multi-Scale Augmentation for Occluded Person Re-Identification
by Xuan Jiang, Xin Yuan and Xiaolan Yang
Sensors 2025, 25(19), 6210; https://doi.org/10.3390/s25196210 - 7 Oct 2025
Viewed by 396
Abstract
Occluded person re-identification (Re-ID) faces significant challenges, mainly due to the interference of occlusion noise and the scarcity of realistic occluded training data. Although data augmentation is a commonly used solution, the current occlusion augmentation methods suffer from the problem of dual inconsistencies: [...] Read more.
Occluded person re-identification (Re-ID) faces significant challenges, mainly due to the interference of occlusion noise and the scarcity of realistic occluded training data. Although data augmentation is a commonly used solution, the current occlusion augmentation methods suffer from the problem of dual inconsistencies: intra-sample inconsistency is caused by misaligned synthetic occluders (an augmentation operation for simulating real occlusion situations); i.e., randomly pasted occluders ignore spatial prior information and style differences, resulting in unrealistic artifacts that mislead feature learning; inter-sample inconsistency stems from information loss during random cropping (an augmentation operation for simulating occlusion-induced information loss); i.e., single-scale cropping strategies discard discriminative regions, weakening the robustness of the model. To address the aforementioned dual inconsistencies, this study proposes the unified Multi-Aligned and Multi-Scale Augmentation (MA–MSA) framework based on the core principle of ”synthetic data should resemble real-world data”. First, the Frequency–Style–Position Data Augmentation (FSPDA) module is designed: it ensures consistency in three aspects (frequency, style, and position) by constructing an occluder library that conforms to real-world distribution, achieving style alignment via adaptive instance normalization and optimizing the placement of occluders using hierarchical position rules. Second, the Multi-Scale Crop Data Augmentation (MSCDA) strategy is proposed. It eliminates the problem of information loss through multi-scale cropping with non-overlapping ratios and dynamic view fusion. In addition, different from the traditional serial augmentation method, MA–MSA integrates FSPDA and MSCDA in a parallel manner to achieve the collaborative resolution of dual inconsistencies. Extensive experiments on Occluded-Duke and Occluded-REID show that MA–MSA achieves state-leading performance of 73.3% Rank-1 (+1.5%) and 62.9% mAP on Occluded-Duke, and 87.3% Rank-1 (+2.0%) and 82.1% mAP on Occluded-REID, demonstrating superior robustness without auxiliary models. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

21 pages, 4397 KB  
Article
Splatting the Cat: Efficient Free-Viewpoint 3D Virtual Try-On via View-Decomposed LoRA and Gaussian Splatting
by Chong-Wei Wang, Hung-Kai Huang, Tzu-Yang Lin, Hsiao-Wei Hu and Chi-Hung Chuang
Electronics 2025, 14(19), 3884; https://doi.org/10.3390/electronics14193884 - 30 Sep 2025
Viewed by 413
Abstract
As Virtual Try-On (VTON) technology matures, 2D VTON methods based on diffusion models can now rapidly generate diverse and high-quality try-on results. However, with rising user demands for realism and immersion, many applications are shifting towards 3D VTON, which offers superior geometric and [...] Read more.
As Virtual Try-On (VTON) technology matures, 2D VTON methods based on diffusion models can now rapidly generate diverse and high-quality try-on results. However, with rising user demands for realism and immersion, many applications are shifting towards 3D VTON, which offers superior geometric and spatial consistency. Existing 3D VTON approaches commonly face challenges such as barriers to practical deployment, substantial memory requirements, and cross-view inconsistencies. To address these issues, we propose an efficient 3D VTON framework with robust multi-view consistency, whose core design is to decouple the monolithic 3D editing task into a four-stage cascade as follows: (1) We first reconstruct an initial 3D scene using 3D Gaussian Splatting, integrating the SMPL-X model at this stage as a strong geometric prior. By computing a normal-map loss and a geometric consistency loss, we ensure the structural stability of the initial human model across different views. (2) We employ the lightweight CatVTON to generate 2D try-on images, that provide visual guidance for the subsequent personalized fine-tuning tasks. (3) To accurately represent garment details from all angles, we partition the 2D dataset into three subsets—front, side, and back—and train a dedicated LoRA module for each subset on a pre-trained diffusion model. This strategy effectively mitigates the issue of blurred details that can occur when a single model attempts to learn global features. (4) An iterative optimization process then uses the generated 2D VTON images and specialized LoRA modules to edit the 3DGS scene, achieving 360-degree free-viewpoint VTON results. All our experiments were conducted on a single consumer-grade GPU with 24 GB of memory, a significant reduction from the 32 GB or more typically required by previous studies under similar data and parameter settings. Our method balances quality and memory requirement, significantly lowering the adoption barrier for 3D VTON technology. Full article
(This article belongs to the Special Issue 2D/3D Industrial Visual Inspection and Intelligent Image Processing)
Show Figures

Figure 1

35 pages, 70837 KB  
Article
CAM3D: Cross-Domain 3D Adversarial Attacks from a Single-View Image via Mamba-Enhanced Reconstruction
by Ziqi Liu, Wei Luo, Sixu Guo, Jingnan Zhang and Zhipan Wang
Electronics 2025, 14(19), 3868; https://doi.org/10.3390/electronics14193868 - 29 Sep 2025
Viewed by 360
Abstract
With the widespread deployment of deep neural networks in real-world physical environments, assessing their robustness against adversarial attacks has become a central issue in AI safety. However, the existing two-dimensional adversarial methods often lack robustness in the physical world, while three-dimensional adversarial camouflage [...] Read more.
With the widespread deployment of deep neural networks in real-world physical environments, assessing their robustness against adversarial attacks has become a central issue in AI safety. However, the existing two-dimensional adversarial methods often lack robustness in the physical world, while three-dimensional adversarial camouflage generation typically relies on high-fidelity 3D models, limiting practicality. To address these limitations, we propose CAM3D, a cross-domain 3D adversarial camouflage generation framework based on single-view image input. The framework establishes an inverse graphics network based on the Mamba architecture, integrating a hybrid non-causal state-space-duality module and a wavelet-enhanced dual-branch local perception module. This design preserves global dependency modeling while strengthening high-frequency detail representation, enabling high-precision recovery of 3D geometry and texture from a single image and providing a high-quality structural prior for subsequent adversarial camouflage optimization. On this basis, CAM3D employs a progressive three-stage optimization strategy that sequentially performs multi-view pseudo-supervised reconstruction, real-image detail refinement, and cross-domain adversarial camouflage generation, thereby systematically improving the attack effectiveness of adversarial camouflage in both the digital and physical domains. The experimental results demonstrate that CAM3D substantially reduces the detection performance of mainstream object detectors, and comparative as well as ablation studies further confirm its advantages in geometric consistency, texture fidelity, and physical transferability. Overall, CAM3D offers an effective paradigm for adversarial attack research in real-world physical settings, characterized by low data dependency and strong physical generalization. Full article
(This article belongs to the Special Issue Adversarial Attacks and Defenses in AI Safety/Reliability)
Show Figures

Figure 1

13 pages, 787 KB  
Article
Primary Care Clinician Preferences and Perspectives on Multi-Cancer Detection Testing Across an Integrated Healthcare System
by Jessica D. Austin, Ilyse A. Nelson, Jon C. Tilburt, Eric R. Ellinghysen, Claire Yee, Jaxon Quillen, Brian M. Dougan, John R. Presutti, Ryan T. Hurt, Niloy Jewel Samadder, Karthik Ghosh and Steven W. Ressler
J. Pers. Med. 2025, 15(10), 452; https://doi.org/10.3390/jpm15100452 - 28 Sep 2025
Viewed by 317
Abstract
Background/Objectives: Multi-cancer detection (MCD) tests have emerged as a promising tool to redefine the landscape of early cancer detection. Implementation of this novel technology will likely fall to primary care clinicians (PCC). The purpose of this study is to characterize and explore differences [...] Read more.
Background/Objectives: Multi-cancer detection (MCD) tests have emerged as a promising tool to redefine the landscape of early cancer detection. Implementation of this novel technology will likely fall to primary care clinicians (PCC). The purpose of this study is to characterize and explore differences in PCCs perceptions and preferences towards MCD testing. Methods: Between March and May of 2023, this cross-sectional survey was administered to 281 PCCs, including physicians and advanced care providers practicing within an integrated healthcare system spanning five states. The survey collected data on self-reported characteristics, perceptions of MCD testing, and preferences for learning about MCD testing. Analysis was limited to those with no prior experience with MCD testing (N = 181, response rate 22.8%). Descriptive statistics summarized key variables and chi-square tests assessed differences in perceptions and preferences by key characteristics. Results: Most PCCs were interested in MCD testing (66.3%), but limited knowledge/awareness of MCD testing and confidence to manage patients with a positive test were observed, along with concerns around cost (76.7%) and misuse/poor implementation. The primary preferences for learning about MCD testing were online courses or classroom instruction (64.5%). Significant differences in perceptions and preferences for learning were observed by location, degree, and years in practice. Conclusions: PCCs in our study held positive views towards MCD testing, but gaps and variation in knowledge and confidence towards MCD testing and concerns around the cost and misuse/poor implementation were observed. While efforts to train and educate all PCCs on MCD testing is a critical first step, more research is needed to understand how best to support implementation tailored to individual and system-level needs and characteristics. Full article
(This article belongs to the Section Disease Biomarkers)
Show Figures

Figure 1

22 pages, 4736 KB  
Article
Radiometric Cross-Calibration and Validation of KOMPSAT-3/AEISS Using Sentinel-2A/MSI
by Jin-Hyeok Choi, Kyoung-Wook Jin, Dong-Hwan Cha, Kyung-Bae Choi, Yong-Han Jo, Kwang-Nyun Kim, Gwui-Bong Kang, Ho-Yeon Shin, Ji-Yun Lee, Eunyeong Kim, Hojong Chang and Yun Gon Lee
Remote Sens. 2025, 17(19), 3280; https://doi.org/10.3390/rs17193280 - 24 Sep 2025
Viewed by 433
Abstract
The successful launch of Korea Multipurpose Satellite-3/Advanced Earth Imaging Sensor System (KOMPSAT-3/AEISS) on 18 May 2012 allowed the Republic of Korea to meet the growing demand for high-resolution satellite imagery. However, like all satellite sensors, KOMPSAT-3/AEISS experienced temporal changes post-launch and thus requires [...] Read more.
The successful launch of Korea Multipurpose Satellite-3/Advanced Earth Imaging Sensor System (KOMPSAT-3/AEISS) on 18 May 2012 allowed the Republic of Korea to meet the growing demand for high-resolution satellite imagery. However, like all satellite sensors, KOMPSAT-3/AEISS experienced temporal changes post-launch and thus requires ongoing evaluation and calibration. Although more than a decade has passed since launch, the KOMPSAT-3/AEISS mission and its multi-year data archive remain widely used. This study focused on the cross-calibration of KOMPSAT-3/AEISS with Sentinel-2A/Multispectral Instrument (MSI) by comparing the radiometric responses of the two satellite sensors under similar observation conditions, leveraging the linear relationship between Digital Numbers (DN) and top-of-atmosphere (TOA) radiance. Cross-calibration was performed using near-simultaneous satellite images of the same region, and the Spectral Band Adjustment Factor (SBAF) was calculated and applied to account for differences in spectral response functions (SRF). Additionally, Bidirectional Reflectance Distribution Function (BRDF) correction was applied using MODIS-based kernel models to minimize angular reflectance effects caused by differences in viewing and illumination geometry. This study aims to evaluate the radiometric consistency of KOMPSAT-3/AEISS relative to Sentinel-2A/MSI over Baotou scenes acquired in 2022–2023, derive band-specific calibration coefficients and compare them with prior results, and conduct a side-by-side comparison of cross-calibration and vicarious calibration. Furthermore, the cross-calibration yielded band-specific gains of 0.0196 (Blue), 0.0237 (Green), 0.0214 (Red), and 0.0136 (NIR). These findings offer valuable implications for Earth observation, environmental monitoring, and the planning and execution of future satellite missions. Full article
Show Figures

Graphical abstract

16 pages, 254 KB  
Article
Different Sides of University Life: An Exploratory Study Investigating How Multiple Visits to a Campus Nurture a Rounded View of the Setting and Strengthen Intentions Towards Higher Education Progression
by Cherry Canovan, Hibah Sohail and Anna Graham
Trends High. Educ. 2025, 4(3), 55; https://doi.org/10.3390/higheredu4030055 - 19 Sep 2025
Viewed by 365
Abstract
The evidence base supporting practices to widen participation in higher education, such as campus visits and multi-intervention programs for younger students, remains limited. In order to address this gap, this exploratory study examines the impact of repeated university campus exposure on primary-aged children [...] Read more.
The evidence base supporting practices to widen participation in higher education, such as campus visits and multi-intervention programs for younger students, remains limited. In order to address this gap, this exploratory study examines the impact of repeated university campus exposure on primary-aged children in the UK. We studied the influence of a campus tour on the views of a group of 78 primary school children who had visited the setting on a previous occasion. Our cohort (32M, 45F, aged 10–11) was drawn from schools with high populations of pupils from low-socioeconomic status backgrounds. Using a pre- and post-visit survey design, we assessed changes in perceptions following a second campus tour, building on a prior visit. We found that while one visit was enough to establish basic perceptions—for example, a university is big not small—a second visit allowed participants to see a different side of the university experience, adding nuance, expanding university-related vocabulary, and increasing comfort with the campus environment. Notably, repeat visits strengthened intentions to pursue higher education. We conclude that multiple campus visits benefit low-participation groups by fostering familiarity and exposing younger pupils to different motivations for university attendance. While this study provides a useful foundation from which to explore this area, further work is needed to address limitations such as the small sample size and the UK-specific context. Full article
23 pages, 5510 KB  
Article
Research on Intelligent Generation of Line Drawings from Point Clouds for Ancient Architectural Heritage
by Shuzhuang Dong, Dan Wu, Weiliang Kong, Wenhu Liu and Na Xia
Buildings 2025, 15(18), 3341; https://doi.org/10.3390/buildings15183341 - 15 Sep 2025
Viewed by 374
Abstract
Addressing the inefficiency, subjective errors, and limited adaptability of existing methods for surveying complex ancient structures, this study presents an intelligent hierarchical algorithm for generating line drawings guided by structured architectural features. Leveraging point cloud data, our approach integrates prior semantic and structural [...] Read more.
Addressing the inefficiency, subjective errors, and limited adaptability of existing methods for surveying complex ancient structures, this study presents an intelligent hierarchical algorithm for generating line drawings guided by structured architectural features. Leveraging point cloud data, our approach integrates prior semantic and structural knowledge of ancient buildings to establish a multi-granularity feature extraction framework encompassing local geometric features (normal vectors, curvature, Simplified Point Feature Histograms-SPFH), component-level semantic features (utilizing enhanced PointNet++ segmentation and geometric graph matching for specialized elements), and structural relationships (adjacency analysis, hierarchical support inference). This framework autonomously achieves intelligent layer assignment, line type/width selection based on component semantics, vectorization optimization via orthogonal and hierarchical topological constraints, and the intelligent generation of sectional views and symbolic annotations. We implemented an algorithmic toolchain using the AutoCAD Python API (pyautocad version 0.5.0) within the AutoCAD 2023 environment. Validation on point cloud datasets from two representative ancient structures—Guanchang No. 11 (Luoyuan County, Fujian) and Li Tianda’s Residence (Langxi County, Anhui)—demonstrates the method’s effectiveness in accurately identifying key components (e.g., columns, beams, Dougong brackets), generating engineering-standard line drawings with significantly enhanced efficiency over traditional approaches, and robustly handling complex architectural geometries. This research delivers an efficient, reliable, and intelligent solution for digital preservation, restoration design, and information archiving of ancient architectural heritage. Full article
Show Figures

Figure 1

23 pages, 5635 KB  
Article
Attention-Based Transfer Enhancement Network for Cross-Corpus EEG Emotion Recognition
by Zongni Li, Kin-Yeung Wong and Chan-Tong Lam
Sensors 2025, 25(18), 5718; https://doi.org/10.3390/s25185718 - 13 Sep 2025
Viewed by 584
Abstract
A critical challenge in EEG-based emotion recognition is the poor generalization of models across different datasets due to significant domain shifts. Traditional methods struggle because they either overfit to source-domain characteristics or fail to bridge large discrepancies between datasets. To address this, we [...] Read more.
A critical challenge in EEG-based emotion recognition is the poor generalization of models across different datasets due to significant domain shifts. Traditional methods struggle because they either overfit to source-domain characteristics or fail to bridge large discrepancies between datasets. To address this, we propose the Cross-corpus Attention-based Transfer Enhancement network (CATE), a novel two-stage framework. The core novelty of CATE lies in its dual-view self-supervised pre-training strategy, which learns robust, domain-invariant representations by approaching the problem from two complementary perspectives. Unlike single-view models that capture an incomplete picture, our framework synergistically combines: (1) Noise-Enhanced Representation Modeling (NERM), which builds resilience to domain-specific artifacts and noise, and (2) Wavelet Transform Representation Modeling (WTRM), which captures the essential, multi-scale spectral patterns fundamental to emotion. This dual approach moves beyond the brittle assumptions of traditional domain adaptation, which often fails when domains are too dissimilar. In the second stage, a supervised fine-tuning process adapts these powerful features for classification using attention-based mechanisms. Extensive experiments on six transfer tasks across the SEED, SEED-IV, and SEED-V datasets demonstrate that CATE establishes a new state-of-the-art, achieving accuracies from 68.01% to 81.65% and outperforming prior methods by up to 15.65 percentage points. By effectively learning transferable features from these distinct, synergistic views, CATE provides a robust framework that significantly advances the practical applicability of cross-corpus EEG emotion recognition. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

17 pages, 8440 KB  
Article
Three-Dimensional Gaussian Style Transfer Method Based on Two-Dimensional Priors and Iterative Optimization
by Weijing Zhang, Xinyu Wang, Haolin Yin, Wei Xing, Huaizhong Lin, Lixia Chen and Lei Zhao
Appl. Sci. 2025, 15(17), 9678; https://doi.org/10.3390/app15179678 - 3 Sep 2025
Viewed by 634
Abstract
To address the limitations of existing optimization-based 3D style transfer methods in terms of visual quality, 3D consistency, and real-time rendering performance, we propose a novel 3D Gaussian scene style transfer method based on 2D priors and iterative optimization. Our approach introduces a [...] Read more.
To address the limitations of existing optimization-based 3D style transfer methods in terms of visual quality, 3D consistency, and real-time rendering performance, we propose a novel 3D Gaussian scene style transfer method based on 2D priors and iterative optimization. Our approach introduces a progressive training pipeline that alternates between fine-tuning the 3D Gaussian field and updating a set of supervised stylized images. By gradually injecting style information into the 3D scene through iterative refinement, the method effectively preserves the geometric structure and spatial coherence across viewpoints. Furthermore, we incorporated a pre-trained stable diffusion model as a 2D prior to guide the style adaptation of the 3D Gaussian representation. The combination of diffusion priors and differentiable 3D Gaussian rendering enables high-fidelity style transfer while maintaining real-time rendering capability. Extensive experiments demonstrate that our method significantly improves the visual quality and multi-view consistency of 3D stylized scenes, offering an effective and efficient solution for real-time 3D scene stylization. Full article
Show Figures

Figure 1

26 pages, 62819 KB  
Article
Low-Light Image Dehazing and Enhancement via Multi-Feature Domain Fusion
by Jiaxin Wu, Han Ai, Ping Zhou, Hao Wang, Haifeng Zhang, Gaopeng Zhang and Weining Chen
Remote Sens. 2025, 17(17), 2944; https://doi.org/10.3390/rs17172944 - 25 Aug 2025
Viewed by 900
Abstract
The acquisition of nighttime remote-sensing visible-light images is often accompanied by low-illumination effects and haze interference, resulting in significant image quality degradation and greatly affecting subsequent applications. Existing low-light enhancement and dehazing algorithms can handle each problem individually, but their simple cascade cannot [...] Read more.
The acquisition of nighttime remote-sensing visible-light images is often accompanied by low-illumination effects and haze interference, resulting in significant image quality degradation and greatly affecting subsequent applications. Existing low-light enhancement and dehazing algorithms can handle each problem individually, but their simple cascade cannot effectively address unknown real-world degradations. Therefore, we design a joint processing framework, WFDiff, which fully exploits the advantages of Fourier–wavelet dual-domain features and innovatively integrates the inverse diffusion process through differentiable operators to construct a multi-scale degradation collaborative correction system. Specifically, in the reverse diffusion process, a dual-domain feature interaction module is designed, and the joint probability distribution of the generated image and real data is constrained through differentiable operators: on the one hand, a global frequency-domain prior is established by jointly constraining Fourier amplitude and phase, effectively maintaining the radiometric consistency of the image; on the other hand, wavelets are used to capture high-frequency details and edge structures in the spatial domain to improve the prediction process. On this basis, a cross-overlapping-block adaptive smoothing estimation algorithm is proposed, which achieves dynamic fusion of multi-scale features through a differentiable weighting strategy, effectively solving the problem of restoring images of different sizes and avoiding local inconsistencies. In view of the current lack of remote-sensing data for low-light haze scenarios, we constructed the Hazy-Dark dataset. Physical experiments and ablation experiments show that the proposed method outperforms existing single-task or simple cascade methods in terms of image fidelity, detail recovery capability, and visual naturalness, providing a new paradigm for remote-sensing image processing under coupled degradations. Full article
(This article belongs to the Section AI Remote Sensing)
Show Figures

Figure 1

25 pages, 928 KB  
Article
Digital Trust in Transition: Student Perceptions of AI-Enhanced Learning for Sustainable Educational Futures
by Aikumis Omirali, Kanat Kozhakhmet and Rakhima Zhumaliyeva
Sustainability 2025, 17(17), 7567; https://doi.org/10.3390/su17177567 - 22 Aug 2025
Viewed by 1478
Abstract
In the context of the rapid digitalization of higher education, proactive artificial intelligence (AI) agents embedded within multi-agent systems (MAS) offer new opportunities for personalized learning, improved quality of education, and alignment with sustainable development goals. This study aims to analyze how such [...] Read more.
In the context of the rapid digitalization of higher education, proactive artificial intelligence (AI) agents embedded within multi-agent systems (MAS) offer new opportunities for personalized learning, improved quality of education, and alignment with sustainable development goals. This study aims to analyze how such AI solutions are perceived by students at Narxoz University (Kazakhstan) prior to their practical implementation. The research focuses on four key aspects: the level of student trust in AI agents, perceived educational value, concerns related to privacy and autonomy, and individual readiness to use MAS tools. The article also explores how these solutions align with the Sustainable Development Goals—specifically SDG 4 (“Quality Education”) and SDG 8 (“Decent Work and Economic Growth”)—through the development of digital competencies and more equitable access to education. Methodologically, the study combines a bibliometric literature analysis, a theoretical review of pedagogical and technological MAS concepts, and a quantitative survey (n = 150) of students. The results reveal a high level of student interest in AI agents and a general readiness to use them, although this is tempered by moderate trust and significant ethical concerns. The findings suggest that the successful integration of AI into educational environments requires a strategic approach from university leadership, including change management, trust-building, and staff development. Thus, MAS technologies are viewed not only as technical innovations but also as managerial advancements that contribute to the creation of a sustainable, human-centered digital pedagogy. Full article
(This article belongs to the Special Issue Sustainable Management for the Future of Education Systems)
Show Figures

Figure 1

21 pages, 12997 KB  
Article
Aerial-Ground Cross-View Vehicle Re-Identification: A Benchmark Dataset and Baseline
by Linzhi Shang, Chen Min, Juan Wang, Liang Xiao, Dawei Zhao and Yiming Nie
Remote Sens. 2025, 17(15), 2653; https://doi.org/10.3390/rs17152653 - 31 Jul 2025
Viewed by 1074
Abstract
Vehicle re-identification (Re-ID) is a critical computer vision task that aims to match the same vehicle across spatially distributed cameras, especially in the context of remote sensing imagery. While prior research has primarily focused on Re-ID using remote sensing images captured from similar, [...] Read more.
Vehicle re-identification (Re-ID) is a critical computer vision task that aims to match the same vehicle across spatially distributed cameras, especially in the context of remote sensing imagery. While prior research has primarily focused on Re-ID using remote sensing images captured from similar, typically elevated viewpoints, these settings do not fully reflect complex aerial-ground collaborative remote sensing scenarios. In this work, we introduce a novel and challenging task: aerial-ground cross-view vehicle Re-ID, which involves retrieving vehicles in ground-view image galleries using query images captured from aerial (top-down) perspectives. This task is increasingly relevant due to the integration of drone-based surveillance and ground-level monitoring in multi-source remote sensing systems, yet it poses substantial challenges due to significant appearance variations between aerial and ground views. To support this task, we present AGID (Aerial-Ground Vehicle Re-Identification), the first benchmark dataset specifically designed for aerial-ground cross-view vehicle Re-ID. AGID comprises 20,785 remote sensing images of 834 vehicle identities, collected using drones and fixed ground cameras. We further propose a novel method, Enhanced Self-Correlation Feature Computation (ESFC), which enhances spatial relationships between semantically similar regions and incorporates shape information to improve feature discrimination. Extensive experiments on the AGID dataset and three widely used vehicle Re-ID benchmarks validate the effectiveness of our method, which achieves a Rank-1 accuracy of 69.0% on AGID, surpassing state-of-the-art approaches by 2.1%. Full article
Show Figures

Figure 1

Back to TopTop