Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (833)

Search Parameters:
Keywords = word embedding

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
21 pages, 922 KB  
Article
Research on Agricultural Meteorological Disaster Event Extraction Method Based on Character–Word Fusion
by Minghui Qiu, Lihua Jiang, Nengfu Xie, Huanping Wu, Ying Chen and Yonglei Li
Agronomy 2025, 15(9), 2135; https://doi.org/10.3390/agronomy15092135 - 5 Sep 2025
Abstract
Meteorological disasters are significant factors that impact agricultural production. Given the vast volume of online agrometeorological disaster information, accurately and automatically extracting essential details—such as the time, location, and extent of damage—is crucial for understanding disaster mechanisms and evolution, as well as for [...] Read more.
Meteorological disasters are significant factors that impact agricultural production. Given the vast volume of online agrometeorological disaster information, accurately and automatically extracting essential details—such as the time, location, and extent of damage—is crucial for understanding disaster mechanisms and evolution, as well as for enhancing disaster prevention capabilities. This paper constructs a comprehensive dataset of agrometeorological disasters in China, providing a robust data foundation and strong support for event extraction tasks. Additionally, we propose a novel model named character and word embedding fusion-based GCN network (CWEF-GCN). This integration of character- and word-level information enhances the model’s ability to better understand and represent text, effectively addressing the challenges of multi-events and argument overlaps in the event extraction process. The experimental results on the agrometeorological disaster dataset indicate that the F1 score of the proposed model is 81.66% for trigger classification and 63.31% for argument classification. Following the extraction of batch agricultural meteorological disaster events, this study analyzes the triggering mechanisms, damage patterns, and disaster response strategies across various disaster types using the extracted event. The findings offer actionable decision-making support for research on agricultural disaster prevention and mitigation. Full article
Show Figures

Figure 1

26 pages, 740 KB  
Article
Enhancement of the Generation Quality of Generative Linguistic Steganographic Texts by a Character-Based Diffusion Embedding Algorithm (CDEA)
by Yingquan Chen, Qianmu Li, Aniruddha Bhattacharjya, Xiaocong Wu, Huifeng Li, Qing Chang, Le Zhu and Yan Xiao
Appl. Sci. 2025, 15(17), 9663; https://doi.org/10.3390/app15179663 - 2 Sep 2025
Viewed by 158
Abstract
Generative linguistic steganography aims to produce texts that remain both perceptually and statistically imperceptible. The existing embedding algorithms often suffer from imbalanced candidate selection, where high-probability words are overlooked and low-probability words dominate, leading to reduced coherence and fluency. We introduce a character-based [...] Read more.
Generative linguistic steganography aims to produce texts that remain both perceptually and statistically imperceptible. The existing embedding algorithms often suffer from imbalanced candidate selection, where high-probability words are overlooked and low-probability words dominate, leading to reduced coherence and fluency. We introduce a character-based diffusion embedding algorithm (CDEA) that uniquely leverages character-level statistics and a power-law-inspired grouping strategy to better balance candidate word selection. Unlike prior methods, the proposed CDEA explicitly prioritizes high-probability candidates, thereby improving both semantic consistency and text naturalness. When combined with XLNet, it effectively generates longer sensitive sequences while preserving quality. The experimental results showed that CDEA not only produces steganographic texts with higher imperceptibility and fluency but also achieves stronger resistance to steganalysis compared with the existing approaches. Future work will be to enhance statistical imperceptibility, integrate CDEA with larger language models such as GPT-5, and extend applications to cross-lingual, multimodal, and practical IoT or blockchain communication scenarios. Full article
(This article belongs to the Special Issue Cyber Security and Software Engineering)
19 pages, 805 KB  
Article
A Multi-Level Feature Fusion Network Integrating BERT and TextCNN
by Yangwu Zhang, Mingxiao Xu and Guohe Li
Electronics 2025, 14(17), 3508; https://doi.org/10.3390/electronics14173508 - 2 Sep 2025
Viewed by 189
Abstract
With the rapid growth of job-related crimes in developing economies, there is an urgent need for intelligent judicial systems to standardize sentencing practices. This study proposes a Multi-Level Feature Fusion Network (MLFFN) to enhance the accuracy and interpretability of sentencing predictions in job-related [...] Read more.
With the rapid growth of job-related crimes in developing economies, there is an urgent need for intelligent judicial systems to standardize sentencing practices. This study proposes a Multi-Level Feature Fusion Network (MLFFN) to enhance the accuracy and interpretability of sentencing predictions in job-related crime cases. The model integrates hierarchical legal feature representation, beginning with benchmark judgments (including starting-point penalties and additional penalties) as the foundational input. The frontend of MLFFN employs an attention mechanism to dynamically fuse word-level, segment-level, and position-level embeddings, generating a global feature encoding that captures critical legal relationships. The backend utilizes sliding-window convolutional kernels to extract localized features from the global feature map, preserving nuanced contextual factors that influence sentencing ranges. Trained on a dataset of job-related crime cases, MLFFN achieves a 6%+ performance improvement over the baseline models (BERT-base-Chinese, TextCNN, and ERNIE) in sentencing prediction accuracy. Our findings demonstrate that explicit modeling of legal hierarchies and contextual constraints significantly improves judicial AI systems. Full article
Show Figures

Figure 1

16 pages, 1007 KB  
Article
Learning SMILES Semantics: Word2Vec and Transformer Embeddings for Molecular Property Prediction
by Saya Hashemian, Zak Khan, Pulkit Kalhan and Yang Liu
Algorithms 2025, 18(9), 547; https://doi.org/10.3390/a18090547 - 1 Sep 2025
Viewed by 167
Abstract
This paper investigates the effectiveness of Word2Vec-based molecular representation learning on SMILES (Simplified Molecular Input Line Entry System) strings for a downstream prediction task related to the market approvability of chemical compounds. Here, market approvability is treated as a proxy classification label derived [...] Read more.
This paper investigates the effectiveness of Word2Vec-based molecular representation learning on SMILES (Simplified Molecular Input Line Entry System) strings for a downstream prediction task related to the market approvability of chemical compounds. Here, market approvability is treated as a proxy classification label derived from approval status, where only the molecular structure is analyzed. We train character-level embeddings using Continuous Bag of Words (CBOW) and Skip-Gram with Negative Sampling architectures and apply the resulting embeddings in a downstream classification task using a multi-layer perceptron (MLP). To evaluate the utility of these lightweight embedding techniques, we conduct experiments on a curated SMILES dataset labeled by approval status under both imbalanced and SMOTE-balanced training conditions. In addition to our Word2Vec-based models, we include a ChemBERTa-based baseline using the pretrained ChemBERTa-77M model. Our findings show that while ChemBERTa achieves a higher performance, the Word2Vec-based models offer a favorable trade-off between accuracy and computational efficiency. This efficiency is especially relevant in large-scale compound screening, where rapid exploration of the chemical space can support early-stage cheminformatics workflows. These results suggest that traditional embedding models can serve as viable alternatives for scalable and interpretable cheminformatics pipelines, particularly in resource-constrained environments. Full article
Show Figures

Figure 1

31 pages, 1503 KB  
Article
From Games to Understanding: Semantrix as a Testbed for Advancing Semantics in Human–Computer Interaction with Transformers
by Javier Sevilla-Salcedo, José Carlos Castillo Montoya, Álvaro Castro-González and Miguel A. Salichs
Electronics 2025, 14(17), 3480; https://doi.org/10.3390/electronics14173480 - 31 Aug 2025
Viewed by 244
Abstract
Despite rapid progress in natural language processing, current interactive AI systems continue to struggle with interpreting ambiguous, idiomatic, and contextually rich human language, a barrier to natural human–computer interaction. Many deployed applications, such as language games or educational tools, showcase surface-level adaptation but [...] Read more.
Despite rapid progress in natural language processing, current interactive AI systems continue to struggle with interpreting ambiguous, idiomatic, and contextually rich human language, a barrier to natural human–computer interaction. Many deployed applications, such as language games or educational tools, showcase surface-level adaptation but do not systematically probe or advance the deeper semantic understanding of user intent in open-ended, creative settings. In this paper, we present Semantrix, a web-based semantic word-guessing platform, not merely as a game but as a living testbed for evaluating and extending the semantic capabilities of state-of-the-art Transformer models in human-facing contexts. Semantrix challenges models to both assess the nuanced meaning of user guesses and generate dynamic, context-sensitive hints in real time, exposing the system to the diversity, ambiguity, and unpredictability of genuine human interaction. To empirically investigate how advanced semantic representations and adaptive language feedback affect user experience, we conducted a preregistered 2 × 2 factorial study (N = 42), independently manipulating embedding depth (Transformers vs. Word2Vec) and feedback adaptivity (dynamic hints vs. minimal feedback). Our findings revealed that only the combination of Transformer-based semantic modelling and adaptive hint generation sustained user engagement, motivation, and enjoyment; conditions lacking either component led to pronounced attrition, highlighting the limitations of shallow or static approaches. Beyond benchmarking game performance, we argue that the methodologies applied in platforms like Semantrix are helpful for improving machine understanding of natural language, paving the way for more robust, intuitive, and human-aligned AI approaches. Full article
(This article belongs to the Special Issue Feature Papers in Artificial Intelligence)
Show Figures

Figure 1

11 pages, 275 KB  
Opinion
Making Historical Consciousness Come Alive: Abstract Concepts, Artificial Intelligence, and Implicit Game-Based Learning
by Julie Madelen Madshaven, Christian Walter Peter Omlin and Apostolos Spanos
Educ. Sci. 2025, 15(9), 1128; https://doi.org/10.3390/educsci15091128 - 30 Aug 2025
Viewed by 331
Abstract
As new technologies shape education, helping students develop historical consciousness remains a challenge. Building on Nordic curricula that emphasize students as both “history-made” and “history-making” citizens, this paper proposes an approach that integrates artificial intelligence (AI) with implicit digital game-based learning (DGBL) to [...] Read more.
As new technologies shape education, helping students develop historical consciousness remains a challenge. Building on Nordic curricula that emphasize students as both “history-made” and “history-making” citizens, this paper proposes an approach that integrates artificial intelligence (AI) with implicit digital game-based learning (DGBL) to learn and develop historical consciousness in education. We outline how traditional, lecture-driven history teaching often fails to convey the abstract principles of historicity (the idea that individual identity, social institutions, values, and ways of thinking are historically conditioned) and the interpretation of the past, understanding of the present, and perspective on the future. Building on Jeismann’s definition of historical consciousness, we identify a gap between the theory-rich notions of historical consciousness and classroom practice, where many educators either do not recognize it or interpret it intuitively from the curriculum’s limited wording, leaving the concept generally absent from the classroom. We then examine three theory-based methods of enriching teaching and learning. Game-based learning provides an interactive environment in which students assume roles, make decisions, and observe consequences, experiencing historical consciousness instead of only reading about it. AI contributes personalized, adaptive content: branching narratives evolve based on individual choices, non-player characters respond dynamically, and analytics guide scaffolding. Implicit learning theory suggests that embedding core principles directly into gameplay allows students to internalize complex ideas without interrupting immersion; they learn by doing, not by explicit instruction. Finally, we propose a model in which these elements combine: (1) game mechanics and narrative embed principles of historical consciousness; (2) AI dynamically adjusts challenges, generates novel scenarios, and delivers feedback; (3) key concepts are embedded into the game narrative so that students absorb them implicitly; and (4) follow-up reflection activities transform tacit understanding into explicit knowledge. We conclude by outlining a research agenda that includes prototyping interactive environments, conducting longitudinal studies to assess students’ learning outcomes, and exploring transferability to other abstract concepts. By situating students within scenarios that explore historicity and temporal interplay, this approach seeks to transform history education into an immersive, reflective practice where students see themselves as history-made and history-making and view the world through a historical lens. Full article
(This article belongs to the Special Issue Unleashing the Potential of E-learning in Higher Education)
Show Figures

Figure 1

18 pages, 2884 KB  
Article
Research on Multi-Path Feature Fusion Manchu Recognition Based on Swin Transformer
by Yu Zhou, Mingyan Li, Hang Yu, Jinchi Yu, Mingchen Sun and Dadong Wang
Symmetry 2025, 17(9), 1408; https://doi.org/10.3390/sym17091408 - 29 Aug 2025
Viewed by 220
Abstract
Recognizing Manchu words can be challenging due to their complex character variations, subtle differences between similar characters, and homographic polysemy. Most studies rely on character segmentation techniques for character recognition or use convolutional neural networks (CNNs) to encode word images for word recognition. [...] Read more.
Recognizing Manchu words can be challenging due to their complex character variations, subtle differences between similar characters, and homographic polysemy. Most studies rely on character segmentation techniques for character recognition or use convolutional neural networks (CNNs) to encode word images for word recognition. However, these methods can lead to segmentation errors or a loss of semantic information, which reduces the accuracy of word recognition. To address the limitations in the long-range dependency modeling of CNNs and enhance semantic coherence, we propose a hybrid architecture to fuse the spatial features of original images and spectral features. Specifically, we first leverage the Short-Time Fourier Transform (STFT) to preprocess the raw input images and thereby obtain their multi-view spectral features. Then, we leverage a primary CNN block and a pair of symmetric CNN blocks to construct a symmetric spectral enhancement module, which is used to encode the raw input features and the multi-view spectral features. Subsequently, we design a feature fusion module via Swin Transformer to fuse multi-view spectral embedding and thereby concat it with the raw input embedding. Finally, we leverage a Transformer decoder to obtain the target output. We conducted extensive experiments on Manchu words benchmark datasets to evaluate the effectiveness of our proposed framework. The experimental results demonstrated that our framework performs robustly in word recognition tasks and exhibits excellent generalization capabilities. Additionally, our model outperformed other baseline methods in multiple writing-style font-recognition tasks. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

19 pages, 1303 KB  
Article
Effect of Deep Recurrent Architectures on Code Vulnerability Detection: Performance Evaluation for SQL Injection in Python
by Asta Slotkienė, Adomas Poška, Pavel Stefanovič and Simona Ramanauskaitė
Electronics 2025, 14(17), 3436; https://doi.org/10.3390/electronics14173436 - 28 Aug 2025
Viewed by 277
Abstract
Security defects in software code can lead to situations that compromise web-based systems, data security, service availability, and the reliability of functionality. Therefore, it is crucial to detect code vulnerabilities as early as possible. During the research, the architectures of the deep learning [...] Read more.
Security defects in software code can lead to situations that compromise web-based systems, data security, service availability, and the reliability of functionality. Therefore, it is crucial to detect code vulnerabilities as early as possible. During the research, the architectures of the deep learning models, peephole LSTM, GRU-Z, and GRU-LN, their element regularizations, and their hyperparameter settings were analysed to achieve the highest performance in detecting SQL injection vulnerabilities in Python code. The results of the research showed that after investigating the effect of hyperparameters on Word2Vector embeddings and applying the most efficient one, the peephole LSTM, delivered the highest performance (F1 = 0.90)—surpassing GRU-Z (0.88) and GRU-LN (0.878)—thereby confirming that the access of the peephole connections to the cell state produces the highest performance score in the architecture of the peephole LSTM model. Comparison of the results with other research indicates that the use of the selected deep learning models and the suggested research methodology allows for improving the performance in detecting SQL injection vulnerabilities in Python-based web applications, with an F1 score reaching 0.90, which is approximately 10% higher than achieved by other researchers. Full article
Show Figures

Figure 1

22 pages, 1926 KB  
Review
Biological Sequence Representation Methods and Recent Advances: A Review
by Hongwei Zhang, Yan Shi, Yapeng Wang, Xu Yang, Kefeng Li, Sio-Kei Im and Yu Han
Biology 2025, 14(9), 1137; https://doi.org/10.3390/biology14091137 - 27 Aug 2025
Viewed by 472
Abstract
Biological-sequence representation methods are pivotal for advancing machine learning in computational biology, transforming nucleotide and protein sequences into formats that enhance predictive modeling and downstream task performance. This review categorizes these methods into three developmental stages: computational-based, word embedding-based, and large language model [...] Read more.
Biological-sequence representation methods are pivotal for advancing machine learning in computational biology, transforming nucleotide and protein sequences into formats that enhance predictive modeling and downstream task performance. This review categorizes these methods into three developmental stages: computational-based, word embedding-based, and large language model (LLM)-based, detailing their principles, applications, and limitations. Computational-based methods, such as k-mer counting and position-specific scoring matrices (PSSM), extract statistical and evolutionary patterns to support tasks like motif discovery and protein–protein interaction prediction. Word embedding-based approaches, including Word2Vec and GloVe, capture contextual relationships, enabling robust sequence classification and regulatory element identification. Advanced LLM-based methods, leveraging Transformer architectures like ESM3 and RNAErnie, model long-range dependencies for RNA structure prediction and cross-modal analysis, achieving superior accuracy. However, challenges persist, including computational complexity, sensitivity to data quality, and limited interpretability of high-dimensional embeddings. Future directions prioritize integrating multimodal data (e.g., sequences, structures, and functional annotations), employing sparse attention mechanisms to enhance efficiency, and leveraging explainable AI to bridge embeddings with biological insights. These advancements promise transformative applications in drug discovery, disease prediction, and genomics, empowering computational biology with robust, interpretable tools. Full article
(This article belongs to the Special Issue Machine Learning Applications in Biology—2nd Edition)
Show Figures

Figure 1

19 pages, 1190 KB  
Article
A Lightweight AI System to Generate Headline Messages for Inventory Status Summarization
by Bongjun Ji, Yukwan Hwang, Donghun Kim, Jungmin Park, Minhyeok Ryu and Yongkyu Cho
Systems 2025, 13(9), 741; https://doi.org/10.3390/systems13090741 - 26 Aug 2025
Viewed by 383
Abstract
In the manufacturing supply chain, management reports often begin with concise messages that summarize key inventory insights. Traditionally, human analysts manually crafted these summary messages by sifting through complex data—a process that is both time-consuming and prone to inconsistency. In this research study, [...] Read more.
In the manufacturing supply chain, management reports often begin with concise messages that summarize key inventory insights. Traditionally, human analysts manually crafted these summary messages by sifting through complex data—a process that is both time-consuming and prone to inconsistency. In this research study, we present an AI-based system that automatically generates high-quality inventory insight summaries, referred to as “headline messages,” using real-world inventory data. The proposed system leverages lightweight natural language processing (NLP) and machine learning models to achieve accurate and efficient performance. Historical messages are first clustered using a sentence-translation MiniLM model that provides fast semantic embedding. This is used to derive key message categories and define structured input features for this purpose. Then, an explainable and low-complexity classifier trained to predict appropriate headline messages based on current inventory metrics using minimal computational resources. Through empirical experiments with real enterprise data, we demonstrate that this approach can reproduce expert-written headline messages with high accuracy while reducing report generation time from hours to minutes. This study makes three contributions. First, it introduces a lightweight approach that transforms inventory data into concise messages. Second, the proposed approach mitigates confusion by maintaining interpretability and fact-based control, and aligns wording with domain-specific terminology. Furthermore, it reports an industrial validation and deployment case study, demonstrating that the system can be integrated with enterprise data pipelines to generate large-scale weekly reports. These results demonstrate the application and technological innovation of combining small-scale language models with interpretable machine learning to provide insights. Full article
(This article belongs to the Special Issue Data-Driven Analysis of Industrial Systems Using AI)
Show Figures

Figure 1

28 pages, 3746 KB  
Article
BERNN: A Transformer-BiLSTM Hybrid Model for Cross-Domain Short Text Classification in Agricultural Expert Systems
by Xueyong Li, Menghao Zhang, Xiaojuan Guo, Jiaxin Zhang, Jiaxia Sun, Xianqin Yun, Liyuan Zheng, Wenyue Zhao, Lican Li and Haohao Zhang
Symmetry 2025, 17(9), 1374; https://doi.org/10.3390/sym17091374 - 22 Aug 2025
Viewed by 449
Abstract
With the advancement of artificial intelligence, Agricultural Expert Systems (AESs) show great potential in enhancing agricultural management efficiency and resource utilization. Accurate extraction of semantic features from agricultural short texts is fundamental to enabling key functions such as intelligent question answering, semantic retrieval, [...] Read more.
With the advancement of artificial intelligence, Agricultural Expert Systems (AESs) show great potential in enhancing agricultural management efficiency and resource utilization. Accurate extraction of semantic features from agricultural short texts is fundamental to enabling key functions such as intelligent question answering, semantic retrieval, and decision support. However, existing single-structure deep neural networks struggle to capture the hierarchical linguistic patterns and contextual dependencies inherent in domain-specific texts. To address this limitation, we propose a hybrid deep learning model—Bidirectional Encoder Recurrent Neural Network (BERNN)—which combines a domain-specific pre-trained Transformer encoder (AgQsBERT) with a Bidirectional Long Short-Term Memory (BiLSTM) network. AgQsBERT generates contextualized word embeddings by leveraging domain-specific pretraining, effectively capturing the semantics of agricultural terminology. These embeddings are then passed to the BiLSTM, which models sequential dependencies in both directions, enhancing the model’s understanding of contextual flow and word disambiguation. Importantly, the bidirectional nature of the BiLSTM introduces a form of architectural symmetry, allowing the model to process input in both forward and backward directions. This symmetric design enables balanced context modeling, which improves the understanding of fragmented and ambiguous phrases frequently encountered in agricultural texts. The synergy between semantic abstraction from AgQsBERT and symmetric contextual modeling from BiLSTM significantly enhances the expressiveness and generalizability of the model. Evaluated on a self-constructed agricultural question dataset with 110,647 annotated samples, BERNN achieved a classification accuracy of 97.19%, surpassing the baseline by 3.2%. Cross-domain validation on the Tsinghua News dataset further demonstrates its robust generalization capability. This architecture provides a powerful foundation for intelligent agricultural question-answering systems, semantic retrieval, and decision support within smart agriculture applications. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

23 pages, 5310 KB  
Article
Greek Sign Language Detection with Artificial Intelligence
by Ioannis Panopoulos, Evangelos Topalis, Nikos Petrellis and Loukas Hadellis
Electronics 2025, 14(16), 3241; https://doi.org/10.3390/electronics14163241 - 15 Aug 2025
Viewed by 627
Abstract
Sign language serves as a vital way to communicate with individuals with hearing loss, deafness, or a speech disorder, yet accessibility remains limited, requiring technological advances to bridge the gap. This study presents the first real-time Greek Sign Language recognition system utilizing deep [...] Read more.
Sign language serves as a vital way to communicate with individuals with hearing loss, deafness, or a speech disorder, yet accessibility remains limited, requiring technological advances to bridge the gap. This study presents the first real-time Greek Sign Language recognition system utilizing deep learning and embedded computers. The recognition system is implemented using You Only Look Once (YOLO11X-seg), an advanced object detection model, which is embedded in a Python-based framework. The model is trained to recognize Greek Sign Language letters and an expandable set of specific words, i.e., the model is capable of distinguishing between static hand shapes (letters) and dynamic gestures (words). The most important advantage of the proposed system is its mobility and scalable processing power. The data are recorded using a mobile IP camera (based on Raspberry Pi 4) via a Motion-Joint Photographic Experts Group (MJPEG) Stream. The image is transmitted over a private ZeroTier network to a remote powerful computer capable of quickly processing large sign language models, employing Moonlight streaming technology. Smaller models can run on an embedded computer. The experimental evaluation shows excellent 99.07% recognition accuracy, while real-time operation is supported, with the image frames processed in 42.7 ms (23.4 frames/s), offering remote accessibility without requiring a direct connection to the processing unit. Full article
(This article belongs to the Special Issue Methods for Object Orientation and Tracking)
Show Figures

Figure 1

39 pages, 3230 KB  
Article
Decoding Wine Narratives with Hierarchical Attention: Classification, Visual Prompts, and Emerging E-Commerce Possibilities
by Vlad Diaconita, Anda Belciu, Alexandra Maria Ioana Corbea and Iuliana Simonca
J. Theor. Appl. Electron. Commer. Res. 2025, 20(3), 212; https://doi.org/10.3390/jtaer20030212 - 14 Aug 2025
Viewed by 513
Abstract
Wine reviews can connect words to flavours; they entwine sensory experiences into vivid stories. This research explores the intersection of artificial intelligence and oenology by using state-of-the-art neural networks to decipher the nuances in wine reviews. For more accurate wine classification and to [...] Read more.
Wine reviews can connect words to flavours; they entwine sensory experiences into vivid stories. This research explores the intersection of artificial intelligence and oenology by using state-of-the-art neural networks to decipher the nuances in wine reviews. For more accurate wine classification and to capture the essence of what matters most to aficionados, we use Hierarchical Attention Networks enhanced with pre-trained embeddings. We also propose an approach to create captivating marketing images using advanced text-to-image generation models, mining a large review corpus for the most important descriptive terms and thus linking textual tasting notes to automatically generated imagery. Compared to more conventional models, our results show that hierarchical attention processes fused with rich linguistic embeddings better reflect the complexities of wine language. In addition to improving the accuracy of wine classification, this method provides consumers with immersive experiences by turning sensory descriptors into striking visual stories. Ultimately, our research helps modernise wine marketing and consumer engagement by merging deep learning with sensory analytics, proving how technology-driven solutions can amplify storytelling and shopping experiences in the digital marketplace. Full article
(This article belongs to the Topic Data Science and Intelligent Management)
Show Figures

Figure 1

21 pages, 1344 KB  
Article
Research on Intelligent Extraction Method of Influencing Factors of Loess Landslide Geological Disasters Based on Soft-Lexicon and GloVe
by Lutong Huang, Yueqin Zhu, Yingfei Li, Tianxiao Yan, Yu Xiao, Dongqi Wei, Ziyao Xing and Jian Li
Appl. Sci. 2025, 15(16), 8879; https://doi.org/10.3390/app15168879 - 12 Aug 2025
Viewed by 233
Abstract
Loess landslide disasters are influenced by a multitude of factors, including slope conditions, triggering mechanisms, and spatial attributes. Extracting these factors from unstructured geological texts is challenging due to nested entities, semantic ambiguity, and rare domain-specific terms. This study proposes a joint extraction [...] Read more.
Loess landslide disasters are influenced by a multitude of factors, including slope conditions, triggering mechanisms, and spatial attributes. Extracting these factors from unstructured geological texts is challenging due to nested entities, semantic ambiguity, and rare domain-specific terms. This study proposes a joint extraction framework guided by a domain ontology that categorizes six types of loess landslide influencing factors, including spatial relationships. The ontology facilitates conceptual classification and semi-automatic nested entity annotation, enabling the construction of a high-quality corpus with eight tag types. The model integrates a Soft-Lexicon mechanism that enhances character-level GloVe embeddings with explicit lexical features, including domain terms, part-of-speech tags, and word boundary indicators derived from a domain-specific lexicon. The resulting hybrid character-level representations are then fed into a BiLSTM-CRF architecture to jointly extract entities, attributes, and multi-level spatial and causal relationships. Extracted results are structured using a content-knowledge model to build a spatially enriched knowledge graph, supporting semantic queries and intelligent reasoning. Experimental results demonstrate improved performance over baseline methods, showcasing the framework’s effectiveness in geohazard information extraction and disaster risk analysis. Full article
(This article belongs to the Special Issue Applications of Big Data and Artificial Intelligence in Geoscience)
Show Figures

Figure 1

23 pages, 6919 KB  
Article
Addressing the Information Asymmetry of Fake News Detection Using Large Language Models and Emotion Embeddings
by Kirishnni Prabagar, Kogul Srikandabala, Nilaan Loganathan, Shalinka Jayatilleke, Gihan Gamage and Daswin De Silva
Symmetry 2025, 17(8), 1290; https://doi.org/10.3390/sym17081290 - 11 Aug 2025
Viewed by 493
Abstract
Fake news generation and propagation occurs in large volumes, at high speed, in diverse formats, while also being short-lived to evade detection and counteraction. Despite its role as an enabler, Artificial Intelligence (AI) has been effective at fake news detection and prediction through [...] Read more.
Fake news generation and propagation occurs in large volumes, at high speed, in diverse formats, while also being short-lived to evade detection and counteraction. Despite its role as an enabler, Artificial Intelligence (AI) has been effective at fake news detection and prediction through diverse techniques of both supervised and unsupervised machine learning. In this article, we propose a novel Artificial Intelligence (AI) approach that addresses the underexplored attribution of information asymmetry in fake news detection. This approach demonstrates how fine-tuned language models and emotion embeddings can be used to detect information asymmetry in intent, emotional framing, and linguistic complexity between content creators and content consumers. The intensity and temperature of emotion, selection of words, and the structure and relationship between words contribute to detecting this asymmetry. An empirical evaluation conducted on five benchmark datasets demonstrates the generalizability and real-time detection capabilities of the proposed AI approach. Full article
Show Figures

Figure 1

Back to TopTop