Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (3,033)

Search Parameters:
Keywords = semantic analysis

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
23 pages, 8017 KB  
Article
Individual-Aware Gradient Boosting Regression for Visual Saliency Prediction of Damaged Regions in Ancient Murals
by Naiyu Xie, Yingchun Cao and Bowen Zhang
Appl. Sci. 2026, 16(4), 2055; https://doi.org/10.3390/app16042055 - 19 Feb 2026
Abstract
Murals are vital cultural heritage assets, yet many are increasingly threatened by long-term natural weathering, environmental erosion, and human-induced damage. Given limited conservation resources, an objective method for prioritizing restoration is urgently needed. This study proposes an Individual-Aware Gradient Boosting Regression (IA-GBR) approach [...] Read more.
Murals are vital cultural heritage assets, yet many are increasingly threatened by long-term natural weathering, environmental erosion, and human-induced damage. Given limited conservation resources, an objective method for prioritizing restoration is urgently needed. This study proposes an Individual-Aware Gradient Boosting Regression (IA-GBR) approach to predict the visual saliency of damaged mural regions by integrating physical damage characteristics, spatial location, and observer identity. We construct an eye-tracking dataset containing complete fixation records from multiple participants viewing diverse mural damage types. IA-GBR employs a two-level feature fusion strategy that combines damage, spatial, and individual features within a gradient boosting residual learning framework. The experimental results demonstrate that IA-GBR consistently outperforms baseline methods, including linear and ridge regression, SVR, decision trees, random forests, AdaBoost, and multilayer perceptrons. Feature importance analysis further reveals the relative contributions of individual differences, damage size, spatial position, and semantic factors to saliency formation. The proposed framework provides data-driven support for restoration prioritization and advances perception-aware saliency modeling in cultural heritage conservation. Full article
Show Figures

Figure 1

29 pages, 19866 KB  
Article
GCF-Net: A Geometric Context and Frequency Domain Fusion Network for Landslide Segmentation in Remote Sensing Imagery
by Chunlong Du, Shaoqun Qi, Luhe Wan, Yin Chen, Zhiwei Lin, Ling Zhu and Xiaona Yu
Remote Sens. 2026, 18(4), 635; https://doi.org/10.3390/rs18040635 - 18 Feb 2026
Viewed by 49
Abstract
Remote sensing-based landslide segmentation is of great significance for geological hazard assessment and post-disaster rescue. Existing convolutional neural network methods, constrained by the inherent limitations of spatial convolution, tend to lose high-frequency edge details during deep semantic extraction, while frequency-domain analysis, although capable [...] Read more.
Remote sensing-based landslide segmentation is of great significance for geological hazard assessment and post-disaster rescue. Existing convolutional neural network methods, constrained by the inherent limitations of spatial convolution, tend to lose high-frequency edge details during deep semantic extraction, while frequency-domain analysis, although capable of globally preserving high-frequency components, struggles to perceive local multi-scale features. The lack of an effective synergistic mechanism between them makes it difficult for networks to balance regional integrity and boundary precision. To address these issues, this paper proposes the Geometric Context and Frequency Domain Fusion Network (GCF-Net), which achieves explicit edge enhancement through a three-stage progressive framework. First, the Pyramid Lightweight Fusion (PGF) block is proposed to aggregate multi-scale context and provide rich hierarchical features for subsequent stages. Second, the Geometric Context and Frequency Domain Fusion (GCF) module is designed, where the frequency-domain branch generates dynamic high-frequency masks via the Fourier transform to locate boundary positions, while the spatial branch models foreground–background relationships to understand boundary semantics, with both branches fused through an adaptive gating mechanism. Finally, Edge-aware Detail Consistency Improvement (EDCI) module is designed to balance boundary preservation and noise suppression based on edge confidence, achieving adaptive output refinement. Under the joint supervision of Focal loss, Dice loss, and Edge loss, experiments on the mixed dataset and LMHLD dataset demonstrate that GCF-Net achieves OAs of 96.42% and 96.71%, respectively. Ablation experiments and visualization results further validate the effectiveness of each module and the significant improvement in boundary segmentation. Full article
Show Figures

Figure 1

31 pages, 3271 KB  
Article
Limits of Transferring User-Defined Quantity Takeoff Rules in 2D CAD and 3D BIM Using Semantic Vertices
by Jaeho Cho
Appl. Sci. 2026, 16(4), 2019; https://doi.org/10.3390/app16042019 - 18 Feb 2026
Viewed by 44
Abstract
In construction projects, dimensioning automation is now implemented with relatively high precision in both 2D CAD and 3D BIM environments when conditions are clearly defined. For identical or similar objects with the same attributes, dimension-based quantity takeoff formulas can be automated, and such [...] Read more.
In construction projects, dimensioning automation is now implemented with relatively high precision in both 2D CAD and 3D BIM environments when conditions are clearly defined. For identical or similar objects with the same attributes, dimension-based quantity takeoff formulas can be automated, and such automated formulas can be repeatedly applied as long as the geometric form and attributes remain unchanged. However, this automation is feasible only within limited environments and under pre-defined rules. Once the geometry is slightly altered or geometric identity is disrupted, directly applying the existing automated mechanism becomes structurally constrained. This is because (1) quantity takeoff formulas are difficult to standardize universally, (2) object orientation is often difficult to determine consistently, and (3) even minor geometric changes can alter the meaning of dimensions, making automatic interpretation problematic. Accordingly, this study aims to systematically and experimentally analyze the practical limits of transferring semantic-vertex-based quantity takeoff formulas. To this end, a single isolated footing is adopted as a common reference object, and the limits of user-defined dimension-based automation in 2D CAD and 3D BIM environments are evaluated through three core comparisons: 1. 2D–3D reliability comparison: consistency of final quantity results within an acceptable tolerance range for the isolated footing; 2. User dependency assessment (3D-focused): the extent to which quantity takeoff formulas in 3D BIM depend on user judgment; 3. User-defined dimension transfer limits: the practical limits of transferring user-defined dimensions between geometrically identical isolated footings in both 2D and 3D environments. Through this analysis, this study empirically confirms that geometric variation is a key factor structurally constraining the transferability of user-defined quantity takeoff formulas. Furthermore, as future work, it proposes directions for linking automated dimension generation with quantity takeoff formulations. Full article
Show Figures

Figure 1

21 pages, 1805 KB  
Article
Introducing LEAF: LLM Edge Assessment Framework for Generative AI on the Edge
by Mustafa Abdulkadhim and Sandor R. Repas
Mach. Learn. Knowl. Extr. 2026, 8(2), 48; https://doi.org/10.3390/make8020048 - 18 Feb 2026
Viewed by 60
Abstract
The transition of Large Language Models (LLMs) from centralized clouds to edge environments is critical for addressing privacy concerns, latency bottlenecks, and operational costs. However, existing edge benchmarking frameworks remain tailored to discriminative Deep Learning tasks (e.g., object detection), failing to capture the [...] Read more.
The transition of Large Language Models (LLMs) from centralized clouds to edge environments is critical for addressing privacy concerns, latency bottlenecks, and operational costs. However, existing edge benchmarking frameworks remain tailored to discriminative Deep Learning tasks (e.g., object detection), failing to capture the multidimensional challenges of generative AI, specifically the trade-offs between token generation speed, semantic accuracy, and hardware sustainability. To address this gap, we introduce LEAF (LLM Edge Assessment Framework), a novel evaluation methodology that integrates Circular Economy principles directly into performance metrics. LEAF assesses edge deployments across five synergistic pillars: Circular Economy Score, Energy Efficiency (Joules/Token), Performance Speed (Tokens/Second), semantic accuracy (BERTScore), and End-to-End Latency. We validate LEAF through an extensive experimental analysis of five distinct hardware classes, ranging from embedded IoT devices (Raspberry Pi 4 and 5, NVIDIA Jetson Nano) to professional edge servers (NVIDIA T400) and repurposed legacy workstations (NVIDIA GTX 1050 Ti). Utilizing 4-bit quantized models via the Ollama runtime, our results reveal a counterintuitive insight: repurposed consumer hardware significantly outperforms modern purpose-built edge SoCs. The legacy GTX 1050 Ti achieved a 20× speedup over the Raspberry Pi 4 and maintained superior energy-per-task efficiency compared to low-power ARM architectures by minimizing active runtime. These findings challenge the prevailing narrative that newer silicon is essential for Edge AI, demonstrating that sustainable, high-performance inference can be achieved by extending the lifecycle of existing hardware. LEAF thus provides a blueprint for a “Green Edge” ecosystem that balances computational capability with environmental responsibility. Full article
(This article belongs to the Section Data)
Show Figures

Figure 1

28 pages, 9652 KB  
Article
A Heritage Information System Based on Point-Clouds: Research and Intervention Analyses Made Accessible
by Paula Redweik, Manuel Sánchez-Fernández, María José Marín-Miranda and José Juan Sanjosé-Blasco
Heritage 2026, 9(2), 77; https://doi.org/10.3390/heritage9020077 - 17 Feb 2026
Viewed by 83
Abstract
Heritage buildings can now be surveyed in great detail using geospatial techniques such as photogrammetry and TLS to produce dense point-clouds. For the purposes of research and building analyses, data about interventions and other relevant semantic data from the building are available from [...] Read more.
Heritage buildings can now be surveyed in great detail using geospatial techniques such as photogrammetry and TLS to produce dense point-clouds. For the purposes of research and building analyses, data about interventions and other relevant semantic data from the building are available from many sources, though not always in a well-organized way. Allying semantic data to point-clouds requires the elaboration of an ontology and the segmentation and classification of the point-clouds in accordance with that ontology. The present paper deals with an approach to make semantic classified point-clouds accessible to researchers, heritage managers and members of the public who wish to explore the 3D point-cloud data with ease and without the need for geospatial expertise. The app presented here, ‘HISTERIA’ (Heritage Information System Tool to Enable Research and Intervention Analysis), was developed with MATLAB 2023 App Designer, an object-oriented programming software module. HISTERIA has an interface in which the user can choose which parts of the heritage building s/he wishes to analyze according to several criteria presented in pre-defined queries. The result of most queries is shown in a point-cloud viewer window inside the app. A point can also be selected in the viewer, and all the values attached to it can be accessed in the different classes. HISTERIA is intended to give to the exploration of semantic heritage data in 3D added value in a simplified way. Full article
Show Figures

Figure 1

30 pages, 1068 KB  
Article
Ontological Foundations for Deterministic Assurance Context Construction and Governed AI Reasoning
by Shao-Fang Wen
Appl. Sci. 2026, 16(4), 1984; https://doi.org/10.3390/app16041984 - 17 Feb 2026
Viewed by 83
Abstract
Security assurance aims to provide justified confidence that a system satisfies its security requirements under defined contextual conditions. In practice, assurance context is often handled implicitly, with assumptions, scope limitations, and boundary conditions embedded in documentation or expert judgment. This limits auditability, reproducibility, [...] Read more.
Security assurance aims to provide justified confidence that a system satisfies its security requirements under defined contextual conditions. In practice, assurance context is often handled implicitly, with assumptions, scope limitations, and boundary conditions embedded in documentation or expert judgment. This limits auditability, reproducibility, and governance, particularly in continuous assurance settings and workflows that rely on automation and AI-assisted reasoning. When reasoning operates over incomplete or underspecified context, implicit assumption formation can alter the basis of assurance conclusions. This paper introduces the Security Assurance Context Ontology (SACO), which reframes assurance context construction and evolution as explicit semantic and governance problems. SACO represents assurance-relevant context elements, their relationships, provenance, and epistemic status as authoritative semantic structures. Missing but required information is preserved as explicit semantic gaps that delimit when assurance claims may be authoritatively accepted. A strict separation between authoritative assurance context and advisory reasoning outputs constrains how automated or AI-assisted analysis may influence the assurance basis. The paper further presents a deterministic realization model for assurance context construction and evolution, where determinism applies to reconstructing authoritative context states from governed inputs. Full article
(This article belongs to the Special Issue Innovative Applications of Ontology and the Semantic Web)
Show Figures

Figure 1

29 pages, 4367 KB  
Article
Contrastive Masked Feature Modeling for Self-Supervised Representation Learning of High-Resolution Remote Sensing Images
by Shiyan Pang, Jianwu Xiang, Zhiqi Zuo, Hanchun Hu and Huiwei Jiang
Remote Sens. 2026, 18(4), 626; https://doi.org/10.3390/rs18040626 - 17 Feb 2026
Viewed by 106
Abstract
As an emerging learning paradigm, self-supervised learning (SSL) has attracted extensive attention due to its ability to mine features with effective representation from massive unlabeled data. In particular, SSL, driven by contrastive learning and masked modeling, shows great potential in general visual tasks. [...] Read more.
As an emerging learning paradigm, self-supervised learning (SSL) has attracted extensive attention due to its ability to mine features with effective representation from massive unlabeled data. In particular, SSL, driven by contrastive learning and masked modeling, shows great potential in general visual tasks. However, because of the diversity of ground target types, the complexity of spectral radiation characteristics, and changes in environmental conditions, existing SSL frameworks exhibit limited feature extraction accuracy and generalization ability when applied to complex remote sensing scenarios. To address this issue, we propose a hybrid SSL framework that integrates the advantages of contrastive learning and masked modeling to extract more robust and reliable features from remote sensing images. The proposed framework includes two parallel branches: one branch uses a contrastive learning strategy to strengthen global feature representation and capture image structural information by constructing positive and negative sample pairs; the other branch adopts a masked modeling strategy, focusing on the fine analysis of local details and predicting the features of masked areas, thereby establishing connections between global and local features. Additionally, to better integrate local and global features, we adopt a hybrid CNN+Transformer architecture, which is particularly suitable for intensive downstream tasks such as semantic segmentation. Extensive experimental results demonstrate that the proposed framework not only exhibits superior feature extraction ability and higher accuracy in small-sample scenarios but also outperforms state-of-the-art mainstream SSL frameworks on large-scale datasets. Full article
Show Figures

Figure 1

8 pages, 6865 KB  
Proceeding Paper
Evaluating Semantic Segmentation Performance Using DeepLabv3+ with Pretrained ResNet Backbones and Multi-Class Annotations
by Matej Spajić, Marija Habijan, Danijel Marinčić and Irena Galić
Eng. Proc. 2026, 125(1), 23; https://doi.org/10.3390/engproc2026125023 - 16 Feb 2026
Viewed by 108
Abstract
Semantic segmentation is a critical task in computer vision, enabling dense classification of image regions. This work investigates the effectiveness of the DeepLabv3+ architecture for binary semantic segmentation using annotated image data. A pretrained ResNet-101 backbone is employed to extract deep features, while [...] Read more.
Semantic segmentation is a critical task in computer vision, enabling dense classification of image regions. This work investigates the effectiveness of the DeepLabv3+ architecture for binary semantic segmentation using annotated image data. A pretrained ResNet-101 backbone is employed to extract deep features, while Atrous Spatial Pyramid Pooling (ASPP) and a decoder module refine the segmentation outputs. The dataset provides per-image annotations indicating class presence, which are leveraged to approximate segmentation masks for training purposes. Various data augmentation techniques and training strategies were applied to support effective learning and reduce overfitting. Experimental results on the MHIST dataset show that the proposed pipeline achieves strong performance despite the lack of pixel-level annotations, with a mean Intersection-over-Union (mIoU) of 0.76 and a mean Dice coefficient of 0.84. These confirm the potential of weakly supervised segmentation using class-aware CAMs and deep pretrained encoders for structured pixel-level prediction tasks in medical imaging. Full article
Show Figures

Figure 1

30 pages, 2117 KB  
Article
Automated Structuring and Analysis of Unstructured Equipment Maintenance Text Data in Manufacturing Using Generative AI Models: A Comparative Study of Pre-Trained Language Models
by Yongju Cho
Appl. Sci. 2026, 16(4), 1969; https://doi.org/10.3390/app16041969 - 16 Feb 2026
Viewed by 173
Abstract
Manufacturing companies face significant challenges in leveraging artificial intelligence for equipment management due to high infrastructure costs and limited availability of labeled data for failures. While most manufacturing AI applications focus on structured sensor data, vast amounts of unstructured textual information containing valuable [...] Read more.
Manufacturing companies face significant challenges in leveraging artificial intelligence for equipment management due to high infrastructure costs and limited availability of labeled data for failures. While most manufacturing AI applications focus on structured sensor data, vast amounts of unstructured textual information containing valuable maintenance knowledge remain underutilized. This study presents a practical generative AI-based framework for structured information extraction that automatically converts unstructured equipment maintenance texts into predefined semantic fields to support predictive maintenance in manufacturing environments. We adopted and evaluated three representative generative models—Bidirectional and Auto-Regressive Transformers (BART) with KoBART, Text-to-Text Transfer Transformer (T5) with pko-t5-base, and the large language model Qwen—to generate structured outputs by extracting three predefined fields: failed components, failure types, and corrective actions. The framework enables the structuring of equipment management text data from Manufacturing Execution Systems (MES) to build predictive maintenance support systems. We validated the approach using a large-scale MES dataset consisting of 29,736 equipment maintenance records from a major automotive parts manufacturer, from which curated subsets were used for model training and evaluation. Our methodology employs Generative Pre-trained Transformer 4 (GPT-4) for initial dataset construction, followed by domain expert validation to ensure data quality. The trained models achieved promising performance when evaluated using extraction-aligned metrics, including exact match (EM) and token-level precision, recall, and F1-score, which directly assess field-level extraction correctness. ROUGE scores are additionally reported as a supplementary indicator of lexical overlap. Among the evaluated models, Qwen consistently outperformed BART and T5 across all extracted fields. The structured outputs are further processed through domain-specific dictionaries and regular expressions to create a comprehensive analytical database supporting predictive maintenance strategies. We implemented a web-based analytics platform enabling time-series analysis, correlation analysis, frequency analysis, and anomaly detection for equipment maintenance optimization. The proposed system converts tacit knowledge embedded in maintenance texts into explicit, actionable insights without requiring additional sensor installations or infrastructure investments. This research contributes to the manufacturing AI field by demonstrating a comprehensive application of generative language models to equipment maintenance text analysis, providing a cost-effective approach for digital transformation in manufacturing environments. The framework’s scalability and cloud-based deployment model present significant opportunities for widespread adoption in the manufacturing sector, supporting the transition from reactive to predictive maintenance strategies. Full article
Show Figures

Figure 1

31 pages, 5533 KB  
Article
Comparative Evaluation of Fusion Strategies Using Multi-Pretrained Deep Learning Fusion-Based (MPDLF) Model for Histopathology Image Classification
by Fatma Alshohoumi and Abdullah Al-Hamdani
Appl. Sci. 2026, 16(4), 1964; https://doi.org/10.3390/app16041964 - 16 Feb 2026
Viewed by 112
Abstract
Histopathological image analysis remains the cornerstone of cancer diagnosis; however, manual assessment is challenged by stain variability, differences in imaging magnification, and complex morphological patterns. The proposed multi-pretrained deep learning fusion (MPDLF) approach combines two widely used CNN architectures: ResNet50, which captures deeper [...] Read more.
Histopathological image analysis remains the cornerstone of cancer diagnosis; however, manual assessment is challenged by stain variability, differences in imaging magnification, and complex morphological patterns. The proposed multi-pretrained deep learning fusion (MPDLF) approach combines two widely used CNN architectures: ResNet50, which captures deeper semantic representations, and VGG16, which extracts fine-grained details. This work differs from previous fusion studies by providing a controlled evaluation of early, intermediate, and late fusion for integrating two pretrained CNN backbones (ResNet50 and VGG16) under single-modality histopathology constraints. To isolate the fusion effect, identical training settings are used across three public H&E datasets. Early fusion achieved the best test performance for the two primary tasks reported here: breast cancer binary classification (accuracy = 0.9070, 95% CI: 0.8742–0.9404; AUC = 0.9707, 95% CI: 0.9541–0.9844) and renal clear cell carcinoma (RCCC) five-class grading (accuracy = 0.8792, 95% CI: 0.8529–0.9041; AUC (OvR, macro) = 0.9895, 95% CI: 0.9859–0.9927). Future work will extend these experiments to additional magnification levels (100×, 200×, and 400×) for breast cancer histopathology images and explore advanced hybrid fusion strategies across different histopathology datasets. Full article
(This article belongs to the Special Issue AI for Medical Systems: Algorithms, Applications, and Challenges)
Show Figures

Figure 1

28 pages, 2555 KB  
Article
Deep Learning-Based Video Watermarking: A Robust Framework for Spatial–Temporal Embedding and Retrieval
by Antonio Cedillo-Hernandez, Lydia Velazquez-Garcia, Francisco Javier Garcia-Ugalde and Manuel Cedillo-Hernandez
Future Internet 2026, 18(2), 104; https://doi.org/10.3390/fi18020104 - 16 Feb 2026
Viewed by 112
Abstract
This paper introduces a deep learning-based framework for video watermarking that achieves robust, imperceptible, and fast embedding under a wide range of visual and temporal conditions. The proposed method is organized into seven modules that collaboratively perform frame encoding, semantic region analysis, block [...] Read more.
This paper introduces a deep learning-based framework for video watermarking that achieves robust, imperceptible, and fast embedding under a wide range of visual and temporal conditions. The proposed method is organized into seven modules that collaboratively perform frame encoding, semantic region analysis, block selection, watermark transformation, and spatiotemporal injection, followed by decoding and multi-objective optimization. A key component of the framework is its ability to learn a visual importance map, which guides a saliency-based block selection strategy. This allows the model to embed the watermark in perceptually redundant regions while minimizing distortion. To enhance resilience, the watermark is distributed across multiple frames, leveraging temporal redundancy to improve recovery under frame loss, insertion, and reordering. Experimental evaluations conducted on a large-scale video dataset demonstrate that the proposed method achieves high fidelity, while preserving low decoding error rates under compression, noise, and temporal distortions. The proposed method operates processing 38 video frames per second on a standard GPU. Additional ablation studies confirm the contribution of each module to the system’s robustness. This framework offers a promising solution for watermarking in streaming, surveillance, and content verification applications. Full article
(This article belongs to the Section Big Data and Augmented Intelligence)
Show Figures

Figure 1

25 pages, 2523 KB  
Article
Link Prediction in Heterogeneous Information Networks: Improved Hypergraph Convolution with Adaptive Soft Voting
by Sheng Zhang, Yuyuan Huang, Ziqiang Luo, Jiangnan Zhou, Bing Wu, Ka Sun and Hongmei Mao
Entropy 2026, 28(2), 230; https://doi.org/10.3390/e28020230 - 16 Feb 2026
Viewed by 181
Abstract
Complex real-world systems are often modeled as heterogeneous information networks with diverse node and relation types, bringing new opportunities and challenges to link prediction. Traditional methods based on similarity or meta-paths fail to fully capture high-order structures and semantics, while existing hypergraph-based models [...] Read more.
Complex real-world systems are often modeled as heterogeneous information networks with diverse node and relation types, bringing new opportunities and challenges to link prediction. Traditional methods based on similarity or meta-paths fail to fully capture high-order structures and semantics, while existing hypergraph-based models homogenize all high-order information without considering their importance differences, diluting core associations with redundant noise and limiting prediction accuracy. Given these issues, we propose the VE-HGCN, a link prediction model for HINs that fuses hypergraph convolution with soft-voting ensemble strategy. The model first constructs multiple heterogeneous hypergraphs from HINs via network frequent subgraph pattern extraction, then leverages hypergraph convolution for node representation learning, and finally employs a soft-voting ensemble strategy to fuse multi-model prediction results. Extensive experiments on four public HIN datasets show that the VE-HGCN outperforms seven mainstream baseline models, thereby validating the effectiveness of the proposed method. This study offers a new perspective for link prediction in HINs and exhibits good generality and practicality, providing a feasible reference for addressing high-order information utilization issues in complex heterogeneous network analysis. Full article
Show Figures

Figure 1

29 pages, 6157 KB  
Article
3D Road Defect Mapping via Differentiable Neural Rendering and Multi-Frame Semantic Fusion in Bird’s-Eye-View Space
by Hongjia Xing and Feng Yang
J. Imaging 2026, 12(2), 83; https://doi.org/10.3390/jimaging12020083 - 15 Feb 2026
Viewed by 100
Abstract
Road defect detection is essential for traffic safety and infrastructure maintenance. Excising automated methods based on 2D image analysis lack spatial context and cannot provide accurate 3D localization required for maintenance planning. We propose a novel framework for road defect mapping from monocular [...] Read more.
Road defect detection is essential for traffic safety and infrastructure maintenance. Excising automated methods based on 2D image analysis lack spatial context and cannot provide accurate 3D localization required for maintenance planning. We propose a novel framework for road defect mapping from monocular video sequences by integrating differentiable Bird’s-Eye-View (BEV) mesh representation, semantic filtering, and multi-frame temporal fusion. Our differentiable mesh-based BEV representation enables efficient scene reconstruction from sparse observations through MLP-based optimization. The semantic filtering strategy leverages road surface segmentation to eliminate off-road false positives, reducing detection errors by 33.7%. Multi-frame fusion with ray-casting projection and exponential moving average update accumulates defect observations across frames while maintaining 3D geometric consistency. Experimental results demonstrate that our framework produces geometrically consistent BEV defect maps with superior accuracy compared to single-frame 2D methods, effectively handling occlusions, motion blur, and varying illumination conditions. Full article
30 pages, 2061 KB  
Article
Target-Aware Bilingual Stance Detection in Social Media Using Transformer Architecture
by Abdul Rahaman Wahab Sait and Yazeed Alkhurayyif
Electronics 2026, 15(4), 830; https://doi.org/10.3390/electronics15040830 - 14 Feb 2026
Viewed by 82
Abstract
Stance detection has emerged as an essential tool in natural language processing for understanding how individuals express agreement, disagreement, or neutrality toward specific targets in social and online discourse. It plays a crucial role in bilingual and multilingual environments, including English-Arabic social media [...] Read more.
Stance detection has emerged as an essential tool in natural language processing for understanding how individuals express agreement, disagreement, or neutrality toward specific targets in social and online discourse. It plays a crucial role in bilingual and multilingual environments, including English-Arabic social media ecosystems, where differences in language structure, discourse style, and data availability pose significant challenges for reliable stance modelling. Existing approaches often struggle with target awareness, cross-lingual generalization, robustness to noisy user-generated text, and the interpretability of model decisions. This study aims to build a reliable, explainable target-aware bilingual stance-detection framework that generalizes across heterogeneous stance formats and languages without retraining on a dataset specific to the target language. Thus, a unified dual-encoder architecture based on mDeBERTa-v3 is proposed. Cross-language contrastive learning offers an auxiliary training objective to align English and Arabic stance representations in a common semantic space. Robustness-oriented regularization is used to mitigate the effects of informal language, vocabulary variation, and adversarial noise. To promote transparency and trustworthiness, the framework incorporates token-level rationale extraction, enables fine-grained interpretability, and supports analysis of hallucination. The proposed model is tested on a combined bilingual test set and two structurally distinct zero-shot benchmarks: MT-CSD and AraStance. Experimental results show consistent performance, with accuracies of 85.0% and 86.8% and F1-scores of 84.7% and 86.8% on the zero-shot benchmarks, confirming stable performance and realistic generalization. Ultimately, these findings reveal that effective bilingual stance detection can be achieved via explicit target conditioning, cross-lingual alignment, and explainability-driven design. Full article
22 pages, 7987 KB  
Article
RioCC: Efficient and Accurate Class-Level Code Recommendation Based on Deep Code Clone Detection
by Hongcan Gao, Chenkai Guo and Hui Yang
Entropy 2026, 28(2), 223; https://doi.org/10.3390/e28020223 - 14 Feb 2026
Viewed by 123
Abstract
Context: Code recommendation plays an important role in improving programming efficiency and software quality. Existing approaches mainly focus on method- or API-level recommendations, which limits their effectiveness to local code contexts. From a multi-stage recommendation perspective, class-level code recommendation aims to efficiently narrow [...] Read more.
Context: Code recommendation plays an important role in improving programming efficiency and software quality. Existing approaches mainly focus on method- or API-level recommendations, which limits their effectiveness to local code contexts. From a multi-stage recommendation perspective, class-level code recommendation aims to efficiently narrow a large candidate code space while preserving essential structural information. Objective: This paper proposes RioCC, a class-level code recommendation framework that leverages deep forest-based code clone detection to progressively reduce the candidate space and improve recommendation efficiency in large-scale code spaces. Method: RioCC models the recommendation process as a coarse-to-fine candidate reduction procedure. In the coarse-grained stage, a quick search-based filtering module performs rapid candidate screening and initial similarity estimation, effectively pruning irrelevant candidates and narrowing the search space. In the fine-grained stage, a deep forest-based analysis with cascade learning and multi-grained scanning captures context- and structure-aware representations of class-level code fragments, enabling accurate similarity assessment and recommendation. This two-stage design explicitly separates coarse candidate filtering from detailed semantic matching to balance efficiency and accuracy. Results: Experiments on a large-scale dataset containing 192,000 clone pairs from BigCloneBench and a collected code pool show that RioCC consistently outperforms state-of-the-art methods, including CCLearner, Oreo, and RSharer, across four types of code clones, while significantly accelerating the recommendation process with comparable detection accuracy. Conclusions: By explicitly formulating class-level code recommendation as a staged retrieval and refinement problem, RioCC provides an efficient and scalable solution for large-scale code recommendation and demonstrates the practical value of integrating lightweight filtering with deep forest-based learning. Full article
(This article belongs to the Section Multidisciplinary Applications)
Back to TopTop