Topic Editors

Department of Informatics, Ionian University, 491 32 Corfu, Greece
Department of Informatics, Ionian University, 491 32 Corfu, Greece
Department of Informatics, Ionian University, 491 32 Corfu, Greece

Artificial Intelligence Models, Tools and Applications

Abstract submission deadline
31 May 2025
Manuscript submission deadline
31 August 2025
Viewed by
208919

Topic Information

Dear Colleagues,

During the difficult years since the start of the ongoing COVID-19 pandemic, the need for efficient artificial intelligence models, tools, and applications has been more evident than ever. Machine learning and data science, not to mention the huge amount of data they produce, form a clear new source of valuable information. New and innovative approaches are required to tackle the new research challenges faced in this area. In this framework, artificial intelligence is crucial and thus may be described as one of the most important research areas of our time. Since this view is applicable to the research community, it also faces huge challenges from the perspective of data management and involves emerging disciplines in information processing and related tools and applications.

This Topic aims to bring together interdisciplinary approaches focusing on innovative applications and existing artificial intelligence methodologies. Since the typical notion of data is usually focused on heterogeneity and is rather dynamic in nature, computer science researchers are encouraged to develop new or adapt existing suitable artificial intelligence models, tools, and applications to effectively solve these problems. Therefore, this Topic is open to anyone who wants to submit a relevant research manuscript.

Prof. Dr. Phivos Mylonas
Dr. Katia Lida Kermanidis
Prof. Dr. Manolis Maragoudakis
Topic Editors

Keywords

  • artificial intelligence
  • machine learning
  • smart tools and applications
  • computational logic
  • multi-agent systems
  • cross-disciplinary AI applications

Participating Journals

Journal Name Impact Factor CiteScore Launched Year First Decision (median) APC
Applied Sciences
applsci
2.5 5.3 2011 17.8 Days CHF 2400 Submit
Computers
computers
2.6 5.4 2012 17.2 Days CHF 1800 Submit
Digital
digital
- 3.1 2021 23.6 Days CHF 1000 Submit
Electronics
electronics
2.6 5.3 2012 16.8 Days CHF 2400 Submit
Smart Cities
smartcities
7.0 11.2 2018 25.8 Days CHF 2000 Submit

Preprints.org is a multidiscipline platform providing preprint service that is dedicated to sharing your research from the start and empowering your research journey.

MDPI Topics is cooperating with Preprints.org and has built a direct connection between MDPI journals and Preprints.org. Authors are encouraged to enjoy the benefits by posting a preprint at Preprints.org prior to publication:

  1. Immediately share your ideas ahead of publication and establish your research priority;
  2. Protect your idea from being stolen with this time-stamped preprint article;
  3. Enhance the exposure and impact of your research;
  4. Receive feedback from your peers in advance;
  5. Have it indexed in Web of Science (Preprint Citation Index), Google Scholar, Crossref, SHARE, PrePubMed, Scilit and Europe PMC.

Published Papers (92 papers)

Order results
Result details
Journals
Select all
Export citation of selected articles as:
19 pages, 1699 KiB  
Article
Deep Speech Synthesis and Its Implications for News Verification: Lessons Learned in the RTVE-UGR Chair
by Daniel Calderón-González, Nieves Ábalos, Blanca Bayo, Pedro Cánovas, David Griol, Carlos Muñoz-Romero, Carmen Pérez, Pere Vila and Zoraida Callejas
Appl. Sci. 2024, 14(21), 9916; https://doi.org/10.3390/app14219916 - 30 Oct 2024
Viewed by 403
Abstract
This paper presents the multidisciplinary work carried out in the RTVE-UGR Chair within the IVERES project, whose main objective is the development of a tool for journalists to verify the veracity of the audios that reach the newsrooms. In the current context, voice [...] Read more.
This paper presents the multidisciplinary work carried out in the RTVE-UGR Chair within the IVERES project, whose main objective is the development of a tool for journalists to verify the veracity of the audios that reach the newsrooms. In the current context, voice synthesis has both beneficial and detrimental applications, with audio deepfakes being a significant concern in the world of journalism due to their ability to mislead and misinform. This is a multifaceted problem that can only be tackled adopting a multidisciplinary perspective. In this article, we describe the approach we adopted within the RTVE-UGR Chair to successfully address the challenges derived from audio deepfakes involving a team with different backgrounds and a specific methodology of iterative co-creation. As a result, we present several outcomes including the compilation and generation of audio datasets, the development and deployment of several audio fake detection models, and the development of a web audio verification tool addressed to journalists. As a conclusion, we highlight the importance of this systematic collaborative work in the fight against misinformation and the future potential of audio verification technologies in various applications. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

18 pages, 5751 KiB  
Article
An Enhanced Particle Filtering Method Leveraging Particle Swarm Optimization for Simultaneous Localization and Mapping in Mobile Robots Navigating Unknown Environments
by Xu Bian, Wanqiu Zhao, Ling Tang, Hong Zhao and Xuesong Mei
Appl. Sci. 2024, 14(20), 9426; https://doi.org/10.3390/app14209426 - 16 Oct 2024
Viewed by 494
Abstract
With the rapid advancement of mobile robotics technology, Simultaneous Localization and Mapping (SLAM) has become indispensable for enabling robots to autonomously navigate and construct maps of unknown environments in real time. Traditional SLAM algorithms, such as the Extended Kalman Filter (EKF) and FastSLAM, [...] Read more.
With the rapid advancement of mobile robotics technology, Simultaneous Localization and Mapping (SLAM) has become indispensable for enabling robots to autonomously navigate and construct maps of unknown environments in real time. Traditional SLAM algorithms, such as the Extended Kalman Filter (EKF) and FastSLAM, have shown commendable performance in certain applications. However, they encounter significant limitations when dealing with nonlinear systems and non-Gaussian noise distributions, especially in dynamic and complex environments coupled with high computational complexity. To address these challenges, this study proposes an enhanced particle filtering method leveraging particle swarm optimization (PSO) to improve the accuracy of pose estimation and the efficacy of map construction in SLAM algorithms. We begin by elucidating the foundational principles of FastSLAM and its critical role in empowering robots with the ability to autonomously explore and map unknown territories. Subsequently, we delve into the innovative integration of PSO with FastSLAM, highlighting our novel approach of designing a bespoke fitness function tailored to enhance the distribution of particles. This innovation is pivotal in mitigating the degradation issues associated with particle filtering, thereby significantly improving the estimation accuracy and robustness of the SLAM solution in various operational scenarios. A series of simulation experiments and tests were conducted to substantiate the efficacy of the proposed method across diverse environments. The experimental outcomes demonstrate that, compared to the standard particle filtering algorithm, the PSO-enhanced particle filtering effectively mitigates the issue of particle degeneration, ensuring reliable and accurate SLAM performance even in challenging, unknown environments. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

20 pages, 1936 KiB  
Review
Physics Guided Neural Networks with Knowledge Graph
by Kishor Datta Gupta, Sunzida Siddique, Roy George, Marufa Kamal, Rakib Hossain Rifat and Mohd Ariful Haque
Digital 2024, 4(4), 846-865; https://doi.org/10.3390/digital4040042 - 10 Oct 2024
Viewed by 894
Abstract
Over the past few decades, machine learning (ML) has demonstrated significant advancements in all areas of human existence. Machine learning and deep learning models rely heavily on data. Typically, basic machine learning (ML) and deep learning (DL) models receive input data and its [...] Read more.
Over the past few decades, machine learning (ML) has demonstrated significant advancements in all areas of human existence. Machine learning and deep learning models rely heavily on data. Typically, basic machine learning (ML) and deep learning (DL) models receive input data and its matching output. Within the model, these models generate rules. In a physics-guided model, input and output rules are provided to optimize the model’s learning, hence enhancing the model’s loss optimization. The concept of the physics-guided neural network (PGNN) is becoming increasingly popular among researchers and industry professionals. It has been applied in numerous fields such as healthcare, medicine, environmental science, and control systems. This review was conducted using four specific research questions. We obtained papers from six different sources and reviewed a total of 81 papers, based on the selected keywords. In addition, we have specifically addressed the difficulties and potential advantages of the PGNN. Our intention is for this review to provide guidance for aspiring researchers seeking to obtain a deeper understanding of the PGNN. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

22 pages, 56552 KiB  
Article
Towards Urban Accessibility: Modeling Trip Distribution to Assess the Provision of Social Facilities
by Margarita Mishina, Sergey Mityagin, Alexander Belyi, Alexander Khrulkov and Stanislav Sobolevsky
Smart Cities 2024, 7(5), 2741-2762; https://doi.org/10.3390/smartcities7050106 - 18 Sep 2024
Viewed by 862
Abstract
Assessing the accessibility and provision of social facilities in urban areas presents a significant challenge, particularly when direct data on facility utilization are unavailable or incomplete. To address this challenge, our study investigates the potential of trip distribution models in estimating facility utilization [...] Read more.
Assessing the accessibility and provision of social facilities in urban areas presents a significant challenge, particularly when direct data on facility utilization are unavailable or incomplete. To address this challenge, our study investigates the potential of trip distribution models in estimating facility utilization based on the spatial distributions of population demand and facilities’ capacities within a city. We first examine the extent to which traditional gravity-based and optimization-focused models can capture population–facilities interactions and provide a reasonable perspective on facility accessibility and provision. We then explore whether advanced deep learning techniques can produce more robust estimates of facility utilization when data are partially observed (e.g., when some of the district administrations collect and share these data). Our findings suggest that, while traditional models offer valuable insights into facility utilization, especially in the absence of direct data, their effectiveness depends on accurate assumptions about distance-related commute patterns. This limitation is addressed by our proposed novel deep learning model, incorporating supply–demand constraints, which demonstrates the ability to uncover hidden interaction patterns from partly observed data, resulting in accurate estimates of facility utilization and, thereby, more reliable provision assessments. We illustrate these findings through a case study on kindergarten accessibility in Saint Petersburg, Russia, offering urban planners a strategic toolkit for evaluating facility provision in data-limited contexts. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

30 pages, 3456 KiB  
Article
Towards Next-Generation Urban Decision Support Systems through AI-Powered Construction of Scientific Ontology Using Large Language Models—A Case in Optimizing Intermodal Freight Transportation
by Jose Tupayachi, Haowen Xu, Olufemi A. Omitaomu, Mustafa Can Camur, Aliza Sharmin and Xueping Li
Smart Cities 2024, 7(5), 2392-2421; https://doi.org/10.3390/smartcities7050094 - 31 Aug 2024
Cited by 1 | Viewed by 1308
Abstract
The incorporation of Artificial Intelligence (AI) models into various optimization systems is on the rise. However, addressing complex urban and environmental management challenges often demands deep expertise in domain science and informatics. This expertise is essential for deriving data and simulation-driven insights that [...] Read more.
The incorporation of Artificial Intelligence (AI) models into various optimization systems is on the rise. However, addressing complex urban and environmental management challenges often demands deep expertise in domain science and informatics. This expertise is essential for deriving data and simulation-driven insights that support informed decision-making. In this context, we investigate the potential of leveraging the pre-trained Large Language Models (LLMs) to create knowledge representations for supporting operations research. By adopting ChatGPT-4 API as the reasoning core, we outline an applied workflow that encompasses natural language processing, Methontology-based prompt tuning, and Generative Pre-trained Transformer (GPT), to automate the construction of scenario-based ontologies using existing research articles and technical manuals of urban datasets and simulations. From these ontologies, knowledge graphs can be derived using widely adopted formats and protocols, guiding various tasks towards data-informed decision support. The performance of our methodology is evaluated through a comparative analysis that contrasts our AI-generated ontology with the widely recognized pizza ontology, commonly used in tutorials for popular ontology software. We conclude with a real-world case study on optimizing the complex system of multi-modal freight transportation. Our approach advances urban decision support systems by enhancing data and metadata modeling, improving data integration and simulation coupling, and guiding the development of decision support strategies and essential software components. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

18 pages, 5060 KiB  
Article
Generative Adversarial Network-Based Voltage Fault Diagnosis for Electric Vehicles under Unbalanced Data
by Weidong Fang, Yihan Guo and Ji Zhang
Electronics 2024, 13(16), 3131; https://doi.org/10.3390/electronics13163131 - 7 Aug 2024
Cited by 1 | Viewed by 1029
Abstract
The research of electric vehicle power battery fault diagnosis technology is turning to machine learning methods. However, during operation, the time of occurrence of faults is much smaller than the normal driving time, resulting in too small a proportion of fault data as [...] Read more.
The research of electric vehicle power battery fault diagnosis technology is turning to machine learning methods. However, during operation, the time of occurrence of faults is much smaller than the normal driving time, resulting in too small a proportion of fault data as well as a single fault characteristic in the collected data. This has hindered the research progress in this field. To address this problem, this paper proposes a data enhancement method using Least Squares Generative Adversarial Networks (LSGAN). The method consists of training the original power battery fault dataset using LSGAN models to generate diverse sample data representing various fault states. The augmented dataset is then used to develop a fault diagnosis framework called LSGAN-RF-GWO, which combines a random forest (RF) model with a Gray Wolf Optimization (GWO) model for effective fault diagnosis. The performance of the framework is evaluated on the original and enhanced datasets and compared with other commonly used models such as Support Vector Machine (SVM), Gradient Boosting Machine (GBM), and Naïve Bayes (NB). The results show that the proposed fault diagnosis scheme improves the evaluation metrics and accuracy level, proving that the LSGAN-RF-GWO framework can utilize limited data resources to effectively diagnose power battery faults. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

28 pages, 16382 KiB  
Review
Overview of Pest Detection and Recognition Algorithms
by Boyu Guo, Jianji Wang, Minghui Guo, Miao Chen, Yanan Chen and Yisheng Miao
Electronics 2024, 13(15), 3008; https://doi.org/10.3390/electronics13153008 - 30 Jul 2024
Cited by 1 | Viewed by 1425
Abstract
Detecting and recognizing pests are paramount for ensuring the healthy growth of crops, maintaining ecological balance, and enhancing food production. With the advancement of artificial intelligence technologies, traditional pest detection and recognition algorithms based on manually selected pest features have gradually been substituted [...] Read more.
Detecting and recognizing pests are paramount for ensuring the healthy growth of crops, maintaining ecological balance, and enhancing food production. With the advancement of artificial intelligence technologies, traditional pest detection and recognition algorithms based on manually selected pest features have gradually been substituted by deep learning-based algorithms. In this review paper, we first introduce the primary neural network architectures and evaluation metrics in the field of pest detection and pest recognition. Subsequently, we summarize widely used public datasets for pest detection and recognition. Following this, we present various pest detection and recognition algorithms proposed in recent years, providing detailed descriptions of each algorithm and their respective performance metrics. Finally, we outline the challenges that current deep learning-based pest detection and recognition algorithms encounter and propose future research directions for related algorithms. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

15 pages, 1314 KiB  
Article
Joint Extraction Method for Hydraulic Engineering Entity Relations Based on Multi-Features
by Yang Liu, Xingzhi Wang, Xuemei Liu, Zehong Ren, Yize Wang and Qianqian Cai
Electronics 2024, 13(15), 2979; https://doi.org/10.3390/electronics13152979 - 28 Jul 2024
Viewed by 835
Abstract
During the joint extraction of entity and relationship from the operational management data of hydraulic engineering, complex sentences containing multiple triplets and overlapping entity relations often arise. However, traditional joint extraction models suffer from a single-feature representation approach, which hampers the effectiveness of [...] Read more.
During the joint extraction of entity and relationship from the operational management data of hydraulic engineering, complex sentences containing multiple triplets and overlapping entity relations often arise. However, traditional joint extraction models suffer from a single-feature representation approach, which hampers the effectiveness of entity relation extraction in complex sentences within hydraulic engineering datasets. To address this issue, this study proposes a multi-feature joint entity relation extraction method based on global context mechanism and graph convolutional neural networks. This method builds upon the Bidirectional Encoder Representations from Transformers (BERT) pre-trained model and utilizes a bidirectional gated recurrent unit (BiGRU) and global context mechanism (GCM) to supplement the contextual and global features of sentences. Subsequently, a graph convolutional network (GCN) based on syntactic dependencies is employed to learn inter-word dependency features, enhancing the model’s knowledge representation capabilities for complex sentences. Experimental results demonstrate the effectiveness of the proposed model in the joint extraction task on hydraulic engineering datasets. The precision, recall, and F1-score are 86.5%, 84.1%, and 85.3%, respectively, all outperforming the baseline model. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

28 pages, 3159 KiB  
Article
VEPO-S2S: A VEssel Portrait Oriented Trajectory Prediction Model Based on S2S Framework
by Xinyi Yang, Zhonghe Han, Yuanben Zhang, Hu Liu, Siye Liu, Wanzheng Ai and Junyi Liu
Appl. Sci. 2024, 14(14), 6344; https://doi.org/10.3390/app14146344 - 20 Jul 2024
Viewed by 876
Abstract
The prediction of vessel trajectories plays a crucial role in ensuring maritime safety and reducing maritime accidents. Substantial progress has been made in trajectory prediction tasks by adopting sequence modeling methods, containing recurrent neural networks (RNNs) and sequence-to-sequence networks (Seq2Seq). However, (1) most [...] Read more.
The prediction of vessel trajectories plays a crucial role in ensuring maritime safety and reducing maritime accidents. Substantial progress has been made in trajectory prediction tasks by adopting sequence modeling methods, containing recurrent neural networks (RNNs) and sequence-to-sequence networks (Seq2Seq). However, (1) most of these studies focus on the application of trajectory information, such as the longitude, latitude, course, and speed, while neglecting the impact of differing vessel features and behavioral preferences on the trajectories. (2) Challenges remain in acquiring these features and preferences, as well as enabling the model to sensibly integrate and efficiently express them. To address the issue, we introduce a novel deep framework VEPO-S2S, consisting of a Multi-level Vessel Trajectory Representation Module (Multi-Rep) and a Feature Fusion and Decoding Module (FFDM). Apart from the trajectory information, we first defined the Multi-level Vessel Characteristics in Multi-Rep, encompassing Shallow-level Attributes (vessel length, width, draft, etc.) and Deep-level Features (Sailing Location Preference, Voyage Time Preference, etc.). Subsequently, Multi-Rep was designed to obtain trajectory information and Multi-level Vessel Characteristics, applying distinct encoders for encoding. Next, the FFDM selected and integrated the above features from Multi-Rep for prediction by employing both a priori and a posteriori mechanisms, a Feature Fusion Component, and an enhanced decoder. This allows the model to efficiently leverage them and enhance overall performance. Finally, we conducted comparative experiments with several baseline models. The experimental results demonstrate that VEPO-S2S is both quantitatively and qualitatively superior to the models. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

27 pages, 11740 KiB  
Article
Site-Specific Deterministic Temperature and Dew Point Forecasts with Explainable and Reliable Machine Learning
by Mengmeng Han, Tennessee Leeuwenburg and Brad Murphy
Appl. Sci. 2024, 14(14), 6314; https://doi.org/10.3390/app14146314 - 19 Jul 2024
Viewed by 882
Abstract
Site-specific weather forecasts are essential for accurate prediction of power demand and are consequently of great interest to energy operators. However, weather forecasts from current numerical weather prediction (NWP) models lack the fine-scale detail to capture all important characteristics of localised real-world sites. [...] Read more.
Site-specific weather forecasts are essential for accurate prediction of power demand and are consequently of great interest to energy operators. However, weather forecasts from current numerical weather prediction (NWP) models lack the fine-scale detail to capture all important characteristics of localised real-world sites. Instead, they provide weather information representing a rectangular gridbox (usually kilometres in size). Even after post-processing and bias correction, area-averaged information is usually not optimal for specific sites. Prior work on site-optimised forecasts has focused on linear methods, weighted consensus averaging, and time-series methods, among others. Recent developments in machine learning (ML) have prompted increasing interest in applying ML as a novel approach towards this problem. In this study, we investigate the feasibility of optimising forecasts at sites by adopting the popular machine learning model “gradient boosted decision tree”, supported by the XGBoost package (v.1.7.3) in the Python language. Regression trees have been trained with historical NWP and site observations as training data, aimed at predicting temperature and dew point at multiple site locations across Australia. We developed a working ML framework, named “Multi-SiteBoost”, and initial test results show a significant improvement compared with gridded values from bias-corrected NWP models. The improvement from XGBoost (0.1–0.6 °C, 4–27% improvement in temperature) is found to be comparable with non-ML methods reported in the literature. With the insights provided by SHapley Additive exPlanations (SHAP), this study also tests various approaches to understand the ML predictions and increase the reliability of the forecasts generated by ML. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

21 pages, 22426 KiB  
Article
Intelligent Surveillance of Airport Apron: Detection and Location of Abnormal Behavior in Typical Non-Cooperative Human Objects
by Jun Li and Xiangqing Dong
Appl. Sci. 2024, 14(14), 6182; https://doi.org/10.3390/app14146182 - 16 Jul 2024
Viewed by 740
Abstract
Most airport surface surveillance systems focus on monitoring and commanding cooperative objects (vehicles) while neglecting the location and detection of non-cooperative objects (humans). Abnormal behavior by non-cooperative objects poses a potential threat to airport security. This study collects surveillance video data from civil [...] Read more.
Most airport surface surveillance systems focus on monitoring and commanding cooperative objects (vehicles) while neglecting the location and detection of non-cooperative objects (humans). Abnormal behavior by non-cooperative objects poses a potential threat to airport security. This study collects surveillance video data from civil aviation airports in several regions of China, and a non-cooperative abnormal behavior localization and detection framework (NC-ABLD) is established. As the focus of this paper, the proposed framework seamlessly integrates a multi-scale non-cooperative object localization module, a human keypoint detection module, and a behavioral classification module. The framework uses a serial structure, with multiple modules working in concert to achieve precise position, human keypoints, and behavioral classification of non-cooperative objects in the airport field. In addition, since there is no publicly available rich dataset of airport aprons, we propose a dataset called IIAR-30, which consists of 1736 images of airport surfaces and 506 video clips in six frequently occurring behavioral categories. The results of experiments conducted on the IIAR-30 dataset show that the framework performs well compared to mainstream behavior recognition methods and achieves fine-grained localization and refined class detection of typical non-cooperative human abnormal behavior on airport apron surfaces. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

18 pages, 3930 KiB  
Article
Implementation of an Automatic Meeting Minute Generation System Using YAMNet with Speaker Identification and Keyword Prompts
by Ching-Ta Lu and Liang-Yu Wang
Appl. Sci. 2024, 14(13), 5718; https://doi.org/10.3390/app14135718 - 29 Jun 2024
Viewed by 1104
Abstract
Producing conference/meeting minutes requires a person to simultaneously identify a speaker and the speaking content during the course of the meeting. This recording process is a heavy task. Reducing the workload for meeting minutes is an essential task for most people. In addition, [...] Read more.
Producing conference/meeting minutes requires a person to simultaneously identify a speaker and the speaking content during the course of the meeting. This recording process is a heavy task. Reducing the workload for meeting minutes is an essential task for most people. In addition, providing conference/meeting highlights in real time is helpful to the meeting process. In this study, we aim to implement an automatic meeting minutes generation system (AMMGS) for recording conference/meeting minutes. A speech recognizer transforms speech signals to obtain the conference/meeting text. Accordingly, the proposed AMMGS can reduce the effort in recording the minutes. All meeting members can concentrate on the meeting; taking minutes is unnecessary. The AMMGS includes speaker identification for Mandarin Chinese speakers, keyword spotting, and speech recognition. Transferring learning on YAMNet lets the network identify specified speakers. So, the proposed AMMGS can automatically generate conference/meeting minutes with labeled speakers. Furthermore, the AMMGS applies the Jieba segmentation tool for keyword spotting. The system detects the frequency of words’ occurrence. Keywords are determined from the highly segmented words. These keywords help an attendant to stay with the agenda. The experimental results reveal that the proposed AMMGS can accurately identify speakers and recognize speech. Accordingly, the AMMGS can generate conference/meeting minutes while the keywords are spotted effectively. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

50 pages, 3271 KiB  
Review
Unlocking Artificial Intelligence Adoption in Local Governments: Best Practice Lessons from Real-World Implementations
by Tan Yigitcanlar, Anne David, Wenda Li, Clinton Fookes, Simon Elias Bibri and Xinyue Ye
Smart Cities 2024, 7(4), 1576-1625; https://doi.org/10.3390/smartcities7040064 - 28 Jun 2024
Viewed by 3800
Abstract
In an era marked by rapid technological progress, the pivotal role of Artificial Intelligence (AI) is increasingly evident across various sectors, including local governments. These governmental bodies are progressively leveraging AI technologies to enhance service delivery to their communities, ranging from simple task [...] Read more.
In an era marked by rapid technological progress, the pivotal role of Artificial Intelligence (AI) is increasingly evident across various sectors, including local governments. These governmental bodies are progressively leveraging AI technologies to enhance service delivery to their communities, ranging from simple task automation to more complex engineering endeavours. As more local governments adopt AI, it is imperative to understand the functions, implications, and consequences of these advanced technologies. Despite the growing importance of this domain, a significant gap persists within the scholarly discourse. This study aims to bridge this void by exploring the applications of AI technologies within the context of local government service provision. Through this inquiry, it seeks to generate best practice lessons for local government and smart city initiatives. By conducting a comprehensive review of grey literature, we analysed 262 real-world AI implementations across 170 local governments worldwide. The findings underscore several key points: (a) there has been a consistent upward trajectory in the adoption of AI by local governments over the last decade; (b) local governments from China, the US, and the UK are at the forefront of AI adoption; (c) among local government AI technologies, natural language processing and robotic process automation emerge as the most prevalent ones; (d) local governments primarily deploy AI across 28 distinct services; and (e) information management, back-office work, and transportation and traffic management are leading domains in terms of AI adoption. This study enriches the existing body of knowledge by providing an overview of current AI applications within the sphere of local governance. It offers valuable insights for local government and smart city policymakers and decision-makers considering the adoption, expansion, or refinement of AI technologies in urban service provision. Additionally, it highlights the importance of using these insights to guide the successful integration and optimisation of AI in future local government and smart city projects, ensuring they meet the evolving needs of communities. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

24 pages, 21619 KiB  
Article
A Vehicle Velocity Prediction Method with Kinematic Segment Recognition
by Benxiang Lin, Chao Wei and Fuyong Feng
Appl. Sci. 2024, 14(12), 5030; https://doi.org/10.3390/app14125030 - 9 Jun 2024
Viewed by 1153
Abstract
Accurate vehicle velocity prediction is of great significance in vehicle energy distribution and road traffic management. In light of the high time variability of vehicle velocity itself and the limitation of single model prediction, a velocity prediction method based on K-means-QPSO-LSTM with kinematic [...] Read more.
Accurate vehicle velocity prediction is of great significance in vehicle energy distribution and road traffic management. In light of the high time variability of vehicle velocity itself and the limitation of single model prediction, a velocity prediction method based on K-means-QPSO-LSTM with kinematic segment recognition is proposed in this paper. Firstly, the K-means algorithm was used to cluster samples with similar characteristics together, extract kinematic fragment samples in typical driving conditions, calculate their feature parameters, and carry out principal component analysis on the feature parameters to achieve dimensionality reduction transformation of information. Then, the vehicle velocity prediction sub-neural network models based on long short-term memory (LSTM) with the QPSO algorithm optimized were trained under different driving condition datasets. Furthermore, the kinematic segment recognition and traditional vehicle velocity prediction were integrated to form an adaptive vehicle velocity prediction method based on driving condition identification. Finally, the current driving condition type was identified and updated in real-time during vehicle velocity prediction, and then the corresponding sub-LSTM model was used for vehicle velocity prediction. The simulation experiment demonstrated a significant enhancement in both the velocity and accuracy of prediction through the proposed method. The proposed hybrid method has the potential to improve the accuracy and reliability of vehicle velocity prediction, making it applicable in various fields such as autonomous driving, traffic management, and energy management strategies for hybrid electric vehicles. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

16 pages, 4629 KiB  
Article
Characterizing Smart Cities Based on Artificial Intelligence
by Laaziza Hammoumi, Mehdi Maanan and Hassan Rhinane
Smart Cities 2024, 7(3), 1330-1345; https://doi.org/10.3390/smartcities7030056 - 7 Jun 2024
Cited by 5 | Viewed by 2140
Abstract
Cities worldwide are attempting to be labelled as smart, but truly classifying as such remains a great challenge. This study aims to use artificial intelligence (AI) to classify the performance of smart cities and identify the factors linked to their smartness. Based on [...] Read more.
Cities worldwide are attempting to be labelled as smart, but truly classifying as such remains a great challenge. This study aims to use artificial intelligence (AI) to classify the performance of smart cities and identify the factors linked to their smartness. Based on residents’ perceptions of urban structures and technological applications, this study included 200 cities globally. For 147 cities, we gathered the perceptions of 120 residents per city through a survey of 39 questions covering two main pillars: ‘Structures’, referring to the existing infrastructure of the city, and the ‘Technology’ pillar that describes the technological provisions and services available to the inhabitants. These pillars were evaluated across five key areas: health and safety, mobility, activities, opportunities, and governance. For the remaining 53 cities, scores were derived by analyzing pertinent data collected from various online resources. Multiple machine learning algorithms, including Random Forest, Artificial Neural Network, Support Vector Machine, and Gradient Boost, were tested and compared in order to select the best one. The results showed that Random Forest and the Artificial Neural Network are the best trained models that achieved the highest levels of accuracy. This study provides a robust framework for using machine learning to identify and assess smart cities, offering valuable insights for future research and urban planning. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

26 pages, 6942 KiB  
Article
Effectiveness of the Fuzzy Logic Control to Manage the Microclimate Inside a Smart Insulated Greenhouse
by Jamel Riahi, Hamza Nasri, Abdelkader Mami and Silvano Vergura
Smart Cities 2024, 7(3), 1304-1329; https://doi.org/10.3390/smartcities7030055 - 6 Jun 2024
Cited by 1 | Viewed by 1019
Abstract
Agricultural greenhouses incorporate intricate systems to regulate the internal climate. Among the crucial climatic variables, indoor temperature and humidity take precedence in establishing an optimal environment for plant production and growth. The present research emphasizes the efficacy of employing intelligent control systems in [...] Read more.
Agricultural greenhouses incorporate intricate systems to regulate the internal climate. Among the crucial climatic variables, indoor temperature and humidity take precedence in establishing an optimal environment for plant production and growth. The present research emphasizes the efficacy of employing intelligent control systems in the automation of the indoor climate for smart insulated greenhouses (SIGs), utilizing a fuzzy logic controller (FLC). This paper proposes the use of an FLC to reduce the energy consumption of a greenhouse. In the first step, a thermodynamic model is presented and experimentally validated based on thermal heat exchanges between the indoor and outdoor climatic variables. The outcomes show the effectiveness of the proposed model in controlling indoor air temperature and relative humidity with a low error percentage. Secondly, several fuzzy logic control models have been developed to regulate the indoor temperature and humidity for cold and hot periods. The results show the good performance of the proposed FLC model as highlighted by the statistical analysis. In fact, the root mean squared error (RMSE) is very small and equal to 0.69% for temperature and 0.23% for humidity, whereas the efficiency factor (EF) of the fuzzy logic control is equal to 99.35% for temperature control and 99.86% for humidity control. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

17 pages, 2534 KiB  
Article
LeakPred: An Approach for Identifying Components with Resource Leaks in Android Mobile Applications
by Josias Gomes Lima, Rafael Giusti and Arilo Claudio Dias-Neto
Computers 2024, 13(6), 140; https://doi.org/10.3390/computers13060140 - 3 Jun 2024
Viewed by 621
Abstract
Context: Mobile devices contain some resources, for example, the camera, battery, and memory, that are allocated, used, and then deallocated by mobile applications. Whenever a resource is allocated and not correctly released, a defect called a resource leak occurs, which can cause [...] Read more.
Context: Mobile devices contain some resources, for example, the camera, battery, and memory, that are allocated, used, and then deallocated by mobile applications. Whenever a resource is allocated and not correctly released, a defect called a resource leak occurs, which can cause crashes and slowdowns. Objective: In this study, we intended to demonstrate the usefulness of the LeakPred approach in terms of the number of components with resource leak problems identified in applications. Method: We compared the approach’s effectiveness with three state-of-the-art methods in identifying leaks in 15 Android applications. Result: LeakPred obtained the best median (85.37%) of components with identified leaks, the best coverage (96.15%) of the classes of leaks that could be identified in the applications, and an accuracy of 81.25%. The Android Lint method achieved the second best median (76.92%) and the highest accuracy (100%), but only covered 1.92% of the leak classes. Conclusions: LeakPred is effective in identifying leaky components in applications. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

13 pages, 432 KiB  
Article
Deep Pre-Training Transformers for Scientific Paper Representation
by Jihong Wang, Zhiguang Yang and Zhanglin Cheng
Electronics 2024, 13(11), 2123; https://doi.org/10.3390/electronics13112123 - 29 May 2024
Cited by 2 | Viewed by 1390
Abstract
In the age of scholarly big data, efficiently navigating and analyzing the vast corpus of scientific literature is a significant challenge. This paper introduces a specialized pre-trained BERT-based language model, termed SPBERT, which enhances natural language processing tasks specifically tailored to the domain [...] Read more.
In the age of scholarly big data, efficiently navigating and analyzing the vast corpus of scientific literature is a significant challenge. This paper introduces a specialized pre-trained BERT-based language model, termed SPBERT, which enhances natural language processing tasks specifically tailored to the domain of scientific paper analysis. Our method employs a novel neural network embedding technique that leverages textual components, such as keywords, titles, abstracts, and full texts, to represent papers in a vector space. By integrating recent advancements in text representation and unsupervised feature aggregation, SPBERT offers a sophisticated approach to encode essential information implicitly, thereby enhancing paper classification and literature retrieval tasks. We applied our method to several real-world academic datasets, demonstrating notable improvements over existing methods. The findings suggest that SPBERT not only provides a more effective representation of scientific papers but also facilitates a deeper understanding of large-scale academic data, paving the way for more informed and accurate scholarly analysis. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

17 pages, 3557 KiB  
Article
EDUNet++: An Enhanced Denoising Unet++ for Ice-Covered Transmission Line Images
by Yu Zhang, Yinke Dou, Liangliang Zhao, Yangyang Jiao and Dongliang Guo
Electronics 2024, 13(11), 2085; https://doi.org/10.3390/electronics13112085 - 27 May 2024
Cited by 1 | Viewed by 679
Abstract
New technology has made it possible to monitor and analyze the condition of ice-covered transmission lines based on images. However, the collected images are frequently accompanied by noise, which results in inaccurate monitoring. Therefore, this paper proposes an enhanced denoising Unet++ for ice-covered [...] Read more.
New technology has made it possible to monitor and analyze the condition of ice-covered transmission lines based on images. However, the collected images are frequently accompanied by noise, which results in inaccurate monitoring. Therefore, this paper proposes an enhanced denoising Unet++ for ice-covered transmission line images (EDUNet++). This algorithm mainly comprises three modules: a feature encoding and decoding module (FEADM), a shared source feature fusion module (SSFFM), and an error correction module (ECM). In the FEADM, a residual attention module (RAM) and a multilevel feature attention module (MFAM) are proposed. The RAM incorporates the cascaded residual structure and hybrid attention mechanism, that effectively preserve the mapping of feature information. The MFAM uses dilated convolution to obtain features at different levels, and then uses feature attention for weighting. This module effectively combines local and global features, which can better capture the details and texture information in the image. In the SSFFM, the source features are fused to preserve low-frequency information like texture and edges in the image, hence enhancing the realism and clarity of the image. The ECM utilizes the discrepancy between the generated image and the original image to effectively capture all the potential information in the image, hence enhancing the realism of the generated image. We employ a novel piecewise joint loss. On the dataset of ice-covered transmission lines, PSNR (peak signal to noise ratio) and SSIM (structural similarity) achieved values of 29.765 dB and 0.968, respectively. Additionally, the visual effects exhibited more distinct detailed features. The proposed method exhibits superior noise suppression capabilities and robustness compared to alternative approaches. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

18 pages, 7828 KiB  
Article
A Few-Shot Object Detection Method for Endangered Species
by Hongmei Yan, Xiaoman Ruan, Daixian Zhu, Haoran Kong and Peixuan Liu
Appl. Sci. 2024, 14(11), 4443; https://doi.org/10.3390/app14114443 - 23 May 2024
Cited by 1 | Viewed by 757
Abstract
Endangered species detection plays an important role in biodiversity conservation and is significant in maintaining ecological balance. Existing deep learning-based object detection methods are overly dependent on a large number of supervised samples, and building such endangered species datasets is usually costly. Aiming [...] Read more.
Endangered species detection plays an important role in biodiversity conservation and is significant in maintaining ecological balance. Existing deep learning-based object detection methods are overly dependent on a large number of supervised samples, and building such endangered species datasets is usually costly. Aiming at the problems faced by endangered species detection, such as low accuracy and easy loss of location information, an efficient endangered species detection method with fewer samples is proposed to extend the few-shot object detection technique to the field of endangered species detection, which requires only a small number of training samples to obtain excellent detection results. First, SE-Res2Net is proposed to optimize the feature extraction capability. Secondly, an RPN network with multiple attention mechanism is proposed. Finally, for the classification confusion problem, a weighted prototype-based comparison branch is introduced to construct weighted category prototype vectors, which effectively improves the performance of the original classifier. Under the setting of 30 samples in the endangered species dataset, the average detection accuracy value of the method, mAP50, reaches 76.54%, which is 7.98% higher than that of the pre-improved FSCE method. This paper also compares the algorithm on the PASCOL VOC dataset, which is optimal and has good generalization ability compared to the other five algorithms. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

14 pages, 1252 KiB  
Article
An Artificial Intelligence Approach for Estimating the Turbidity of Artisanal Wine and Dosage of Clarifying Agents
by Erika Mishell De La Cruz Rojas, Jimmy Nuñez-Pérez, Marco Lara-Fiallos, José-Manuel Pais-Chanfrau, Rosario Espín-Valladares and Juan Carlos DelaVega-Quintero
Appl. Sci. 2024, 14(11), 4416; https://doi.org/10.3390/app14114416 - 23 May 2024
Viewed by 1029
Abstract
Red wine is a beverage consumed worldwide and contains suspended solids that cause turbidity. The study’s purpose was to mathematically model estimated turbidity in artisanal wines concerning the dosage and types of fining agents based on previous studies presenting positive results. Burgundy grape [...] Read more.
Red wine is a beverage consumed worldwide and contains suspended solids that cause turbidity. The study’s purpose was to mathematically model estimated turbidity in artisanal wines concerning the dosage and types of fining agents based on previous studies presenting positive results. Burgundy grape wine (Vitis lambrusca) was made and clarified with ‘yausabara’ (Pavonia sepium) and bentonite at different concentrations. The system was modelled using several machine learning models, including MATLAB’s Neural Net Fitting and Regression Learner applications. The results showed that the validation of the neural network trained with the Levenberg–Marquardt algorithm obtained significant statistical indicators, such as the coefficient of determination (R2) of 0.985, mean square error (MSE) of 0.004, normalized root mean square error (NRSME) of 6.01 and Akaike information criterion (AIC) of −160.12, selecting it as the representative model of the system. It presents an objective and simple alternative for measuring wine turbidity that is useful for artisanal winemakers who can improve quality and consistency. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

15 pages, 4056 KiB  
Article
Advanced Swine Management: Infrared Imaging for Precise Localization of Reproductive Organs in Livestock Monitoring
by Iyad Almadani, Brandon Ramos, Mohammed Abuhussein and Aaron L. Robinson
Digital 2024, 4(2), 446-460; https://doi.org/10.3390/digital4020022 - 2 May 2024
Viewed by 1305
Abstract
Traditional methods for predicting sow reproductive cycles are not only costly but also demand a larger workforce, exposing workers to respiratory toxins, repetitive stress injuries, and chronic pain. This occupational hazard can even lead to mental health issues due to repeated exposure to [...] Read more.
Traditional methods for predicting sow reproductive cycles are not only costly but also demand a larger workforce, exposing workers to respiratory toxins, repetitive stress injuries, and chronic pain. This occupational hazard can even lead to mental health issues due to repeated exposure to violence. Managing health and welfare issues becomes pivotal in group-housed animal settings, where individual care is challenging on large farms with limited staff. The necessity for computer vision systems to analyze sow behavior and detect deviations indicative of health problems is apparent. Beyond observing changes in behavior and physical traits, computer vision can accurately detect estrus based on vulva characteristics and analyze thermal imagery for temperature changes, which are crucial indicators of estrus. By automating estrus detection, farms can significantly enhance breeding efficiency, ensuring optimal timing for insemination. These systems work continuously, promptly alerting staff to anomalies for early intervention. In this research, we propose part of the solution by utilizing an image segmentation model to localize the vulva. We created our technique to identify vulvae on pig farms using infrared imagery. To accomplish this, we initially isolate the vulva region by enclosing it within a red rectangle and then generate vulva masks by applying a threshold to the red area. The system is trained using U-Net semantic segmentation, where the input for the system consists of grayscale images and their corresponding masks. We utilize U-Net semantic segmentation to find the vulva in the input image, making it lightweight, simple, and robust enough to be tested on many images. To evaluate the performance of our model, we employ the intersection over union (IOU) metric, which is a suitable indicator for determining the model’s robustness. For the segmentation model, a prediction is generally considered ‘good’ when the intersection over union score surpasses 0.5. Our model achieved this criterion with a score of 0.58, surpassing the scores of alternative methods such as the SVM with Gabor (0.515) and YOLOv3 (0.52). Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

22 pages, 13917 KiB  
Article
PDED-ConvLSTM: Pyramid Dilated Deeper Encoder–Decoder Convolutional LSTM for Arctic Sea Ice Concentration Prediction
by Deyu Zhang, Changying Wang, Baoxiang Huang, Jing Ren, Junli Zhao and Guojia Hou
Appl. Sci. 2024, 14(8), 3278; https://doi.org/10.3390/app14083278 - 13 Apr 2024
Viewed by 810
Abstract
Arctic sea ice concentration plays a key role in the global ecosystem. However, accurate prediction of Arctic sea ice concentration remains a challenging task due to its inherent nonlinearity and complex spatiotemporal correlations. To address these challenges, we propose an innovative encoder–decoder pyramid [...] Read more.
Arctic sea ice concentration plays a key role in the global ecosystem. However, accurate prediction of Arctic sea ice concentration remains a challenging task due to its inherent nonlinearity and complex spatiotemporal correlations. To address these challenges, we propose an innovative encoder–decoder pyramid dilated convolutional long short-term memory network (DED-ConvLSTM). The model is constructed based on the convolutional long short-term memory network (ConvLSTM) and, for the first time, integrates the encoder–decoder architecture of ConvLSTM (ED-ConvLSTM) with a pyramidal dilated convolution strategy. This approach aims to efficiently capture the spatiotemporal properties of the sea ice concentration and to enhance the identification of its nonlinear relationships. By applying convolutional layers with different dilation rates, the PDED-ConvLSTM model can capture spatial features at multiple scales and increase the receptive field without losing resolution. Further, the integration of the pyramid convolution module significantly enhances the model’s ability to understand complex spatiotemporal relationships, resulting in notable improvements in prediction accuracy and generalization ability. The experimental results show that the sea ice concentration distribution predicted by the PDED-ConvLSTM model is in high agreement with ground-based observations, with the residuals between the predictions and observations maintained within a range from −20% to 20%. PDED-ConvLSTM outperforms other models in terms of prediction performance, reducing the RMSE by 3.6% compared to the traditional ConvLSTM model and also performing well over a five-month prediction period. These experiments demonstrate the potential of PDED-ConvLSTM in predicting Arctic sea ice concentrations, making it a viable tool to meet the requirements for accurate prediction and provide technical support for safe and efficient operations in the Arctic region. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

20 pages, 6477 KiB  
Article
Integrating Multi-Criteria Decision Models in Smart Urban Planning: A Case Study of Architectural and Urban Design Competitions
by Tomaž Berčič, Marko Bohanec and Lucija Ažman Momirski
Smart Cities 2024, 7(2), 786-805; https://doi.org/10.3390/smartcities7020033 - 18 Mar 2024
Cited by 2 | Viewed by 1473
Abstract
The focus of this study is to integrate the DEX (Decision EXpert) decision-modeling method in architectural and urban design (A & UD) competitions. This study aims to assess the effectiveness of integrating the DEX (Decision EXpert) decision-modeling method into the evaluation process of [...] Read more.
The focus of this study is to integrate the DEX (Decision EXpert) decision-modeling method in architectural and urban design (A & UD) competitions. This study aims to assess the effectiveness of integrating the DEX (Decision EXpert) decision-modeling method into the evaluation process of A & UD competitions to enhance decision-making transparency, objectivity, and efficiency. By using symbolic values in decision models, the approach offers a more user-friendly alternative to the conventional jury decision-making process. The practical application of the DEX method is demonstrated in the Rhinoceros 3D environment to show its effectiveness in evaluating A & UD competition project solutions related to the development of the smart city. The results indicate that the DEX method, with its hierarchical and symbolic values, significantly improves the simplicity of the evaluation process in A & UD competitions, aligning it with the objectives of the smart cities. This method provides an efficient, accessible, and viable alternative to other multi-criteria decision-making approaches. This study importantly contributes to the field of architectural decision making by merging qualitative multi-criteria decision models into the CAD environment, thus supporting more informed, objective, and transparent decision-making processes in the planning and development of smart cities. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

18 pages, 1100 KiB  
Article
Leveraging Large Language Models for Sensor Data Retrieval
by Alberto Berenguer, Adriana Morejón, David Tomás and Jose-Norberto Mazón
Appl. Sci. 2024, 14(6), 2506; https://doi.org/10.3390/app14062506 - 15 Mar 2024
Cited by 1 | Viewed by 2145
Abstract
The growing significance of sensor data in the development of information technology services finds obstacles due to disparate data presentations and non-adherence to FAIR principles. This paper introduces a novel approach for sensor data gathering and retrieval. The proposal leverages large language models [...] Read more.
The growing significance of sensor data in the development of information technology services finds obstacles due to disparate data presentations and non-adherence to FAIR principles. This paper introduces a novel approach for sensor data gathering and retrieval. The proposal leverages large language models to convert sensor data into FAIR-compliant formats and to provide word embedding representations of tabular data for subsequent exploration, enabling semantic comparison. The proposed system comprises two primary components. The first focuses on gathering data from sensors and converting it into a reusable structured format, while the second component aims to identify the most relevant sensor data to augment a given user-provided dataset. The evaluation of the proposed approach involved comparing the performance of various large language models in generating representative word embeddings for each table to retrieve related sensor data. The results show promising performance in terms of precision and MRR (0.90 and 0.94 for the best-performing model, respectively), indicating the system’s ability to retrieve pertinent sensor data that fulfil user requirements. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

18 pages, 1061 KiB  
Article
Automatic Spell-Checking System for Spanish Based on the Ar2p Neural Network Model
by Eduard Puerto, Jose Aguilar and Angel Pinto
Computers 2024, 13(3), 76; https://doi.org/10.3390/computers13030076 - 12 Mar 2024
Viewed by 1544
Abstract
Currently, approaches to correcting misspelled words have problems when the words are complex or massive. This is even more serious in the case of Spanish, where there are very few studies in this regard. So, proposing new approaches to word recognition and correction [...] Read more.
Currently, approaches to correcting misspelled words have problems when the words are complex or massive. This is even more serious in the case of Spanish, where there are very few studies in this regard. So, proposing new approaches to word recognition and correction remains a research topic of interest. In particular, an interesting approach is to computationally simulate the brain process for recognizing misspelled words and their automatic correction. Thus, this article presents an automatic recognition and correction system of misspelled words in Spanish texts, for the detection of misspelled words, and their automatic amendments, based on the systematic theory of pattern recognition of the mind (PRTM). The main innovation of the research is the use of the PRTM theory in this context. Particularly, a corrective system of misspelled words in Spanish based on this theory, called Ar2p-Text, was designed and built. Ar2p-Text carries out a recursive process of analysis of words by a disaggregation/integration mechanism, using specialized hierarchical recognition modules that define formal strategies to determine if a word is well or poorly written. A comparative evaluation shows that the precision and coverage of our Ar2p-Text model are competitive with other spell-checkers. In the experiments, the system achieves better performance than the three other systems. In general, Ar2p-Text obtains an F-measure of 83%, above the 73% achieved by the other spell-checkers. Our hierarchical approach reuses a lot of information, allowing for the improvement of the text analysis processes in both quality and efficiency. Preliminary results show that the above will allow for future developments of technologies for the correction of words inspired by this hierarchical approach. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

13 pages, 3968 KiB  
Article
Electrocardiogram Signals Classification Using Deep-Learning-Based Incorporated Convolutional Neural Network and Long Short-Term Memory Framework
by Alaa Eleyan and Ebrahim Alboghbaish
Computers 2024, 13(2), 55; https://doi.org/10.3390/computers13020055 - 18 Feb 2024
Cited by 5 | Viewed by 3329
Abstract
Cardiovascular diseases (CVDs) like arrhythmia and heart failure remain the world’s leading cause of death. These conditions can be triggered by high blood pressure, diabetes, and simply the passage of time. The early detection of these heart issues, despite substantial advancements in artificial [...] Read more.
Cardiovascular diseases (CVDs) like arrhythmia and heart failure remain the world’s leading cause of death. These conditions can be triggered by high blood pressure, diabetes, and simply the passage of time. The early detection of these heart issues, despite substantial advancements in artificial intelligence (AI) and technology, is still a significant challenge. This research addresses this hurdle by developing a deep-learning-based system that is capable of predicting arrhythmias and heart failure from abnormalities in electrocardiogram (ECG) signals. The system leverages a model that combines long short-term memory (LSTM) networks with convolutional neural networks (CNNs). Extensive experiments were conducted using ECG data from both the MIT-BIH and BIDMC databases under two scenarios. The first scenario employed data from five distinct ECG classes, while the second focused on classifying data from three classes. The results from both scenarios demonstrated that the proposed deep-learning-based classification approach outperformed existing methods. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

16 pages, 3778 KiB  
Article
Multi-Layer Fusion 3D Object Detection via Lidar Point Cloud and Camera Image
by Yuhao Guo and Hui Hu
Appl. Sci. 2024, 14(4), 1348; https://doi.org/10.3390/app14041348 - 6 Feb 2024
Cited by 1 | Viewed by 1615
Abstract
Object detection is a key task in automatic driving, and the poor performance of small object detection is a challenge that needs to be overcome. Previously, object detection networks could detect large-scale objects in ideal environments, but detecting small objects was very difficult. [...] Read more.
Object detection is a key task in automatic driving, and the poor performance of small object detection is a challenge that needs to be overcome. Previously, object detection networks could detect large-scale objects in ideal environments, but detecting small objects was very difficult. To address this problem, we propose a multi-layer fusion 3D object detection network. First, a dense fusion (D-fusion) method is proposed, which is different from the traditional fusion method. By fusing the feature maps of each layer, more semantic information of the fusion network can be preserved. Secondly, in order to preserve small objects at the feature map level, we designed a feature extractor with an adaptive fusion module (AFM), which reduces the impact of the background on small objects by weighting and fusing different feature layers. Finally, an attention mechanism was added to the feature extractor to accelerate the training efficiency and convergence speed of the network by suppressing information that is irrelevant to the task. The experimental results show that our proposed approach greatly improves the baseline and outperforms most state-of-the-art methods on KITTI object detection benchmarks. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

22 pages, 3852 KiB  
Article
Local-Global Spatial-Temporal Graph Convolutional Network for Traffic Flow Forecasting
by Xinlu Zong, Zhen Chen, Fan Yu and Siwei Wei
Electronics 2024, 13(3), 636; https://doi.org/10.3390/electronics13030636 - 2 Feb 2024
Cited by 1 | Viewed by 1206
Abstract
Traffic forecasting’s key challenge is to extract dynamic spatial-temporal features within intricate traffic systems. This paper introduces a novel framework for traffic prediction, named Local-Global Spatial-Temporal Graph Convolutional Network (LGSTGCN). The framework consists of three core components. Firstly, a graph attention residual network [...] Read more.
Traffic forecasting’s key challenge is to extract dynamic spatial-temporal features within intricate traffic systems. This paper introduces a novel framework for traffic prediction, named Local-Global Spatial-Temporal Graph Convolutional Network (LGSTGCN). The framework consists of three core components. Firstly, a graph attention residual network layer is proposed to capture global spatial dependencies by evaluating traffic mode correlations between different nodes. The context information added in the residual connection can improve the generalization ability of the model. Secondly, a T-GCN module, combining a Graph Convolution Network (GCN) with a Gated Recurrent Unit (GRU), is introduced to capture real-time local spatial-temporal dependencies. Finally, a transformer layer is designed to extract long-term temporal dependence and to identify the sequence characteristics of traffic data through positional encoding. Experiments conducted on four real traffic datasets validate the forecasting performance of the LGSTGCN model. The results demonstrate that LGSTGCN can achieve good performance and be applicable to traffic forecasting tasks. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

14 pages, 5369 KiB  
Article
Identification Method for XRF Spectral Analysis Based on an AGA-BP-Attention Neural Network
by Zeyuan Chang, Qi Zhang, Yuanfeng Li, Xiangjun Xin, Ran Gao, Yun Teng, Lan Rao and Meng Sun
Electronics 2024, 13(3), 507; https://doi.org/10.3390/electronics13030507 - 25 Jan 2024
Cited by 2 | Viewed by 1342
Abstract
X-ray fluorescence (XRF) spectroscopy is a non-destructive differential measurement technique widely utilized in elemental analysis. However, due to its inherent high non-linearity and noise issues, it is challenging for XRF spectral analysis to achieve high levels of accuracy. In response to these challenges, [...] Read more.
X-ray fluorescence (XRF) spectroscopy is a non-destructive differential measurement technique widely utilized in elemental analysis. However, due to its inherent high non-linearity and noise issues, it is challenging for XRF spectral analysis to achieve high levels of accuracy. In response to these challenges, this paper proposes a method for XRF spectral analysis that integrates an adaptive genetic algorithm with a backpropagation neural network, enhanced by an attention mechanism, termed as the AGA-BP-Attention method. By leveraging the robust feature extraction capabilities of the neural network and the ability of the attention mechanism to focus on significant features, spectral features are extracted for elemental identification. The adaptive genetic algorithm is subsequently employed to optimize the parameters of the BP neural network, such as weights and thresholds, which enhances the model’s accuracy and stability. The experimental results demonstrate that, compared to traditional BP neural networks, the AGA-BP-Attention network can more effectively address the non-linearity and noise issues of XRF spectral signals. In XRF spectral analysis of air pollutant samples, it achieved superior prediction accuracy, effectively suppressing the impact of background noise on spectral element recognition. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

25 pages, 574 KiB  
Article
SCALE-BOSS-MR: Scalable Time Series Classification Using Multiple Symbolic Representations
by Apostolos Glenis and George A. Vouros
Appl. Sci. 2024, 14(2), 689; https://doi.org/10.3390/app14020689 - 13 Jan 2024
Cited by 1 | Viewed by 951
Abstract
Time-Series-Classification (TSC) is an important machine learning task for many branches of science. Symbolic representations of time series, especially Symbolic Fourier Approximation (SFA), have been proven very effective for this task, given their abilities to reduce noise. In this paper, we improve upon [...] Read more.
Time-Series-Classification (TSC) is an important machine learning task for many branches of science. Symbolic representations of time series, especially Symbolic Fourier Approximation (SFA), have been proven very effective for this task, given their abilities to reduce noise. In this paper, we improve upon SCALE-BOSS using multiple symbolic representations of time series. More specifically, the proposed SCALE-BOSS-MR incorporates into the process a variety of window sizes combined with multiple dilation parameters applied to the original and to first-order differences’ time series, with the latter modeling trend information. SCALE-BOSS-MR has been evaluated using the eight datasets with the largest training size of the UCR time series repository. The results indicate that SCALE-BOSS-MR can be instantiated to classifiers that are able to achieve state-of-the-art accuracy and can be tuned for scalability. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

21 pages, 5704 KiB  
Article
Deep Convolutional Neural Network for Indoor Regional Crowd Flow Prediction
by Qiaoshuang Teng, Shangyu Sun, Weidong Song, Jinzhong Bei and Chongchang Wang
Electronics 2024, 13(1), 172; https://doi.org/10.3390/electronics13010172 - 30 Dec 2023
Viewed by 1050
Abstract
Crowd flow prediction plays a vital role in modern city management and public safety prewarning. However, the existing approaches related to this topic mostly focus on single sites or road segments, and indoor regional crowd flow prediction has yet to receive sufficient academic [...] Read more.
Crowd flow prediction plays a vital role in modern city management and public safety prewarning. However, the existing approaches related to this topic mostly focus on single sites or road segments, and indoor regional crowd flow prediction has yet to receive sufficient academic attention. Therefore, this paper proposes a novel prediction model, named the spatial–temporal attention-based crowd flow prediction network (STA-CFPNet), to forecast the indoor regional crowd flow volume. The model has four branches of temporal closeness, periodicity, tendency and external factors. Each branch of this model takes a convolutional neural network (CNN) as its principal component, which computes spatial correlations from near to distant areas by stacking multiple CNN layers. By incorporating the output of the four branches into the model’s fusion layer, it is possible to utilize ensemble learning to mine the temporal dependence implicit within the data. In order to improve both the convergence speed and prediction performance of the model, a building block based on spatial–temporal attention mechanisms was designed. Furthermore, a fully convolutional structure was applied to the external factors branch to provide globally shared external factors contexts for the research area. The empirical study demonstrates that STA-CFPNet outperforms other well-known crowd flow prediction methods in processing the experimental datasets. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

12 pages, 451 KiB  
Article
Interpretable Single-dimension Outlier Detection (ISOD): An Unsupervised Outlier Detection Method Based on Quantiles and Skewness Coefficients
by Yuehua Huang, Wenfen Liu, Song Li, Ying Guo and Wen Chen
Appl. Sci. 2024, 14(1), 136; https://doi.org/10.3390/app14010136 - 22 Dec 2023
Cited by 2 | Viewed by 1182
Abstract
A crucial area of study in data mining is outlier detection, particularly in the areas of network security, credit card fraud detection, industrial flaw detection, etc. Existing outlier detection algorithms, which can be divided into supervised methods, semi-supervised methods, and unsupervised methods, suffer [...] Read more.
A crucial area of study in data mining is outlier detection, particularly in the areas of network security, credit card fraud detection, industrial flaw detection, etc. Existing outlier detection algorithms, which can be divided into supervised methods, semi-supervised methods, and unsupervised methods, suffer from missing labeled data, the curse of dimensionality, low interpretability, etc. To address these issues, in this paper, we present an unsupervised outlier detection method based on quantiles and skewness coefficients called ISOD (Interpretable Single dimension Outlier Detection). ISOD first fulfils the empirical cumulative distribution function before computing the quantile and skewness coefficients of each dimension. Finally, it outputs the outlier score. This paper’s contributions are as follows: (1) we propose an unsupervised outlier detection algorithm called ISOD, which has high interpretability and scalability; (2) massive experiments on benchmark datasets demonstrated the superior performance of the ISOD algorithm compared with state-of-the-art baselines in terms of ROC and AP. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

16 pages, 1720 KiB  
Article
Multi-Label Diagnosis of Arrhythmias Based on a Modified Two-Category Cross-Entropy Loss Function
by Junjiang Zhu, Cheng Ma, Yihui Zhang, Hao Huang, Dongdong Kong and Wangjin Ni
Electronics 2023, 12(24), 4976; https://doi.org/10.3390/electronics12244976 - 12 Dec 2023
Cited by 3 | Viewed by 1222
Abstract
The 12-lead resting electrocardiogram (ECG) is commonly used in hospitals to assess heart health. The ECG can reflect a variety of cardiac abnormalities, requiring multi-label classification. However, the diagnosis results in previous studies have been imprecise. For example, in some previous studies, some [...] Read more.
The 12-lead resting electrocardiogram (ECG) is commonly used in hospitals to assess heart health. The ECG can reflect a variety of cardiac abnormalities, requiring multi-label classification. However, the diagnosis results in previous studies have been imprecise. For example, in some previous studies, some cardiac abnormalities that cannot coexist often appeared in the diagnostic results. In this work, we explore how to realize the effective multi-label diagnosis of ECG signals and prevent the prediction of cardiac arrhythmias that cannot coexist. In this work, a multi-label classification method based on a convolutional neural network (CNN), long short-term memory (LSTM), and an attention mechanism is presented for the multi-label diagnosis of cardiac arrhythmia using resting ECGs. In addition, this work proposes a modified two-category cross-entropy loss function by introducing a regularization term to avoid the existence of arrhythmias that cannot coexist. The effectiveness of the modified cross-entropy loss function is validated using a 12-lead resting ECG database collected by our team. Using traditional and modified cross-entropy loss functions, three deep learning methods are employed to classify six types of ECG signals. Experimental results show the modified cross-entropy loss function greatly reduces the number of non-coexisting label pairs while maintaining prediction accuracy. Deep learning methods are effective in the multi-label diagnosis of ECG signals, and diagnostic efficiency can be improved by using the modified cross-entropy loss function. In addition, the modified cross-entropy loss function helps prevent diagnostic models from outputting two arrhythmias that cannot coexist, further reducing the false positive rate of non-coexisting arrhythmic diseases, thereby demonstrating the potential value of the modified loss function in clinical applications. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

15 pages, 2577 KiB  
Article
Multi-Task Learning and Temporal-Fusion-Transformer-Based Forecasting of Building Power Consumption
by Wenxian Ji, Zeyu Cao and Xiaorun Li
Electronics 2023, 12(22), 4656; https://doi.org/10.3390/electronics12224656 - 15 Nov 2023
Cited by 1 | Viewed by 2474
Abstract
Improving the accuracy of the forecasting of building power consumption is helpful in reducing commercial expenses and carbon emissions. However, challenges such as the shortage of training data and the absence of efficient models are the main obstacles in this field. To address [...] Read more.
Improving the accuracy of the forecasting of building power consumption is helpful in reducing commercial expenses and carbon emissions. However, challenges such as the shortage of training data and the absence of efficient models are the main obstacles in this field. To address these issues, this work introduces a model named MTLTFT, combining multi-task learning (MTL) with the temporal fusion transformer (TFT). The MTL approach is utilized to maximize the effectiveness of the limited data by introducing multiple related forecasting tasks. This method enhances the learning process by enabling the model to learn shared representations across different tasks, although the physical number of data remains unchanged. The TFT component, which is optimized for feature learning, is integrated to further improve the model’s performance. Based on a dataset from a large exposition building in Hangzhou, we conducted several forecasting experiments. The results demonstrate that MTLTFT outperforms most baseline methods (such as LSTM, GRU, N-HiTS) in terms of Root Mean Squared Error (RMSE) and Mean Absolute Percentage Error (MAPE), suggesting that MTLTFT is a promising approach for the forecasting of building power consumption and other similar tasks. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

22 pages, 11823 KiB  
Article
Adaptive Smart eHealth Framework for Personalized Asthma Attack Prediction and Safe Route Recommendation
by Eman Alharbi, Asma Cherif and Farrukh Nadeem
Smart Cities 2023, 6(5), 2910-2931; https://doi.org/10.3390/smartcities6050130 - 20 Oct 2023
Cited by 3 | Viewed by 1848
Abstract
Recently, there has been growing interest in using smart eHealth systems to manage asthma. However, limitations still exist in providing smart services and accurate predictions tailored to individual patients’ needs. This study aims to develop an adaptive ubiquitous computing framework that leverages different [...] Read more.
Recently, there has been growing interest in using smart eHealth systems to manage asthma. However, limitations still exist in providing smart services and accurate predictions tailored to individual patients’ needs. This study aims to develop an adaptive ubiquitous computing framework that leverages different bio-signals and spatial data to provide personalized asthma attack prediction and safe route recommendations. We proposed a smart eHealth framework consisting of multiple layers that employ telemonitoring application, environmental sensors, and advanced machine-learning algorithms to deliver smart services to the user. The proposed smart eHealth system predicts asthma attacks and uses spatial data to provide a safe route that drives the patient away from any asthma trigger. Additionally, the framework incorporates an adaptation layer that continuously updates the system based on real-time environmental data and daily bio-signals reported by the user. The developed telemonitoring application collected a dataset containing 665 records used to train the prediction models. The testing result demonstrates a remarkable 98% accuracy in predicting asthma attacks with a recall of 96%. The eHealth system was tested online by ten asthma patients, and its accuracy achieved 94% of accuracy and a recall of 95.2% in generating safe routes for asthma patients, ensuring a safer and asthma-trigger-free experience. The test shows that 89% of patients were satisfied with the safer recommended route than their usual one. This research contributes to enhancing the capabilities of smart healthcare systems in managing asthma and improving patient outcomes. The adaptive feature of the proposed eHealth system ensures that the predictions and recommendations remain relevant and personalized to the current conditions and needs of the individual. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

13 pages, 703 KiB  
Article
Double Consistency Regularization for Transformer Networks
by Yuxian Wan, Wenlin Zhang and Zhen Li
Electronics 2023, 12(20), 4357; https://doi.org/10.3390/electronics12204357 - 20 Oct 2023
Cited by 1 | Viewed by 1282
Abstract
The large-scale and deep-layer deep neural network based on the Transformer model is very powerful in sequence tasks, but it is prone to overfitting for small-scale training data. Moreover, the prediction result of the model with a small disturbance input is significantly lower [...] Read more.
The large-scale and deep-layer deep neural network based on the Transformer model is very powerful in sequence tasks, but it is prone to overfitting for small-scale training data. Moreover, the prediction result of the model with a small disturbance input is significantly lower than that without disturbance. In this work, we propose a double consistency regularization (DOCR) method for the end-to-end model structure, which separately constrains the output of the encoder and decoder during the training process to alleviate the above problems. Specifically, on the basis of the cross-entropy loss function, we build the mean model by integrating the model parameters of the previous rounds and measure the consistency between the models by calculating the KL divergence between the features of the encoder output and the probability distribution of the decoder output of the mean model and the base model so as to impose regularization constraints on the solution space of the model. We conducted extensive experiments on machine translation tasks, and the results show that the BLEU score increased by 2.60 on average, demonstrating the effectiveness of DOCR in improving model performance and its complementary impacts with other regularization techniques. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

18 pages, 2497 KiB  
Article
A Design and Its Application of Multi-Granular Fuzzy Model with Hierarchical Tree Structures
by Chan-Uk Yeom and Keun-Chang Kwak
Appl. Sci. 2023, 13(20), 11175; https://doi.org/10.3390/app132011175 - 11 Oct 2023
Cited by 1 | Viewed by 934
Abstract
This paper is concerned with the design of a context-based fuzzy C-means (CFCM)-based multi-granular fuzzy model (MGFM) with hierarchical tree structures. For this purpose, we propose three types of hierarchical tree structures (incremental, aggregated, and cascaded types) in the design of MGFM. In [...] Read more.
This paper is concerned with the design of a context-based fuzzy C-means (CFCM)-based multi-granular fuzzy model (MGFM) with hierarchical tree structures. For this purpose, we propose three types of hierarchical tree structures (incremental, aggregated, and cascaded types) in the design of MGFM. In general, the conventional fuzzy inference system (FIS) has problems, such as time consumption and an exponential increase in the number of if–then rules when processing large-scale multivariate data. Meanwhile, the existing granular fuzzy model (GFM) reduces the number of rules that increase exponentially. However, the GFM not only has overlapping rules as the cluster centers become closer but also has problems that are difficult to interpret due to many input variables. To solve these problems, the CFCM-based MGFM can be designed as a smaller tree of interconnected GFMs. Here, the inputs of the high-level GFMs are taken from the output to the low-level GFMs. The hierarchical tree structure is more computationally efficient and easier to understand than a single GFM. Furthermore, since the output of the CFCM-based MGFM is a triangular fuzzy number, it is evaluated based on a performance measurement method suitable for the GFM. The prediction performance is analyzed from the automobile fuel consumption and Boston housing database to present the validity of the proposed approach. The experimental results demonstrate that the proposed CFCM-based MGFM based on the hierarchical tree structure creates a small number of meaningful rules and solves prediction-related problems by making them explainable. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

12 pages, 756 KiB  
Article
Unsupervised Vehicle Re-Identification Method Based on Source-Free Knowledge Transfer
by Zhigang Song, Daisong Li, Zhongyou Chen and Wenqin Yang
Appl. Sci. 2023, 13(19), 11013; https://doi.org/10.3390/app131911013 - 6 Oct 2023
Cited by 2 | Viewed by 1111
Abstract
The unsupervised domain-adaptive vehicle re-identification approach aims to transfer knowledge from a labeled source domain to an unlabeled target domain; however, there are knowledge differences between the target domain and the source domain. To mitigate domain discrepancies, existing unsupervised domain-adaptive re-identification methods typically [...] Read more.
The unsupervised domain-adaptive vehicle re-identification approach aims to transfer knowledge from a labeled source domain to an unlabeled target domain; however, there are knowledge differences between the target domain and the source domain. To mitigate domain discrepancies, existing unsupervised domain-adaptive re-identification methods typically require access to source domain data to assist in retraining the target domain model. However, for security reasons, such as data privacy, data exchange between different domains is often infeasible in many scenarios. To this end, this paper proposes an unsupervised domain-adaptive vehicle re-identification method based on source-free knowledge transfer. First, by constructing a source-free domain knowledge migration module, the target domain is consistent with the source domain model output to train a generator to generate the “source-like samples”. Then, it can effectively reduce the model knowledge difference and improve the model’s generalization performance. In the experiment, two mainstream public datasets in this field, VeRi776 and VehicleID, are tested experimentally, and the obtained rank-k (the cumulative matching features) and mAP (the mean Average Precision) indicators are both improved, which are suitable for object re-identification tasks when data between domains cannot be interoperated. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

17 pages, 2402 KiB  
Article
Recommendation Method of Power Knowledge Retrieval Based on Graph Neural Network
by Rongxu Hou, Yiying Zhang, Qinghai Ou, Siwei Li, Yeshen He, Hongjiang Wang and Zhenliu Zhou
Electronics 2023, 12(18), 3922; https://doi.org/10.3390/electronics12183922 - 18 Sep 2023
Cited by 2 | Viewed by 1439
Abstract
With the development of the digital and intelligent transformation of the power grid, the structure and operation and maintenance technology of the power grid are constantly updated, which leads to problems such as difficulties in information acquisition and screening. Therefore, we propose a [...] Read more.
With the development of the digital and intelligent transformation of the power grid, the structure and operation and maintenance technology of the power grid are constantly updated, which leads to problems such as difficulties in information acquisition and screening. Therefore, we propose a recommendation method for power knowledge retrieval based on a graph neural network (RPKR-GNN). The method first uses a graph neural network to learn the network structure information of the power fault knowledge graph and realize the deep semantic embedding of power entities and relations. After this, it fuses the power knowledge graph paths to mine the potential power entity relationships and completes the power fault knowledge graph through knowledge inference. At the same time, we combine the user retrieval behavior features for knowledge aggregation to form a personal subgraph, and we analyze the user retrieval subgraph by matching the similarity of retrieval keyword features. Finally, we form a fusion subgraph based on the subgraph topology and reorder the entities of the subgraph to generate a recommendation list for the target users for the prediction of user retrieval intention. Through experimental comparison with various classical models, the results show that the models have a certain generalization ability in knowledge inference. The method performs well in terms of the MR and Hit@10 indexes on each dataset, and the F1 value can reach 87.3 in the retrieval recommendation effect, which effectively enhances the automated operation and maintenance capability of the power system. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

13 pages, 737 KiB  
Article
Web-Based Malware Detection System Using Convolutional Neural Network
by Ali Alqahtani, Sumayya Azzony, Leen Alsharafi and Maha Alaseri
Digital 2023, 3(3), 273-285; https://doi.org/10.3390/digital3030017 - 12 Sep 2023
Cited by 2 | Viewed by 4692
Abstract
In this article, we introduce a web-based malware detection system that leverages a deep-learning approach. Our primary objective is the development of a robust deep-learning model designed for classifying malware in executable files. In contrast to conventional malware detection systems, our approach relies [...] Read more.
In this article, we introduce a web-based malware detection system that leverages a deep-learning approach. Our primary objective is the development of a robust deep-learning model designed for classifying malware in executable files. In contrast to conventional malware detection systems, our approach relies on static detection techniques to unveil the true nature of files as either malicious or benign. Our method makes use of a one-dimensional convolutional neural network 1D-CNN due to the nature of the portable executable file. Significantly, static analysis aligns perfectly with our objectives, allowing us to uncover static features within the portable executable header. This choice holds particular significance given the potential risks associated with dynamic detection, often necessitating the setup of controlled environments, such as virtual machines, to mitigate dangers. Moreover, we seamlessly integrate this effective deep-learning method into a web-based system, rendering it accessible and user-friendly via a web interface. Empirical evidence showcases the efficiency of our proposed methods, as demonstrated in extensive comparisons with state-of-the-art models across three diverse datasets. Our results undeniably affirm the superiority of our approach, delivering a practical, dependable, and rapid mechanism for identifying malware within executable files. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

19 pages, 3595 KiB  
Article
Joint Location–Allocation Model for Multi-Level Maintenance Service Network in Agriculture
by Jinliang Li, Weibo Ren and Xibin Wang
Appl. Sci. 2023, 13(18), 10167; https://doi.org/10.3390/app131810167 - 9 Sep 2023
Cited by 1 | Viewed by 1018
Abstract
The maintenance service network is always designed as a multi-level service network to provide timely maintenance service for failed machinery, and is rarely studied in agriculture. Thus, this paper focuses on a three-level maintenance service network location–allocation problem in agriculture, which contains several [...] Read more.
The maintenance service network is always designed as a multi-level service network to provide timely maintenance service for failed machinery, and is rarely studied in agriculture. Thus, this paper focuses on a three-level maintenance service network location–allocation problem in agriculture, which contains several spare part centres, service stations, and service units. This research aims to obtain the optimal location of spare part centres and service stations while determining service vehicle allocation results for service stations, and the problem can be called a multi-level facility location and allocation problem (MLFLAP). Considering contiguity constraints and hierarchical relationships, the proposed MLFLAP is formulated as a mixed-integer linear programming (MILP) model integrating with P-region and set covering location problems to minimize total service costs, including spare part centre construction costs, service vehicle usage costs, and service mileage costs of service stations. The Benders decomposition-based solution method with several improvements is then applied to decompose the original MLFLAP into master problem and subproblems to find the optimal solutions effectively. Finally, a real-world case in China is proposed to evaluate the performance of the model and algorithm in agriculture, and sensitivity analysis is also conducted to demonstrate the impact of several parameters. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

16 pages, 5133 KiB  
Article
A New Linear Model for the Calculation of Routing Metrics in 802.11s Using ns-3 and RStudio
by Juan Ochoa-Aldeán and Carlos Silva-Cárdenas
Computers 2023, 12(9), 172; https://doi.org/10.3390/computers12090172 - 28 Aug 2023
Viewed by 1161
Abstract
Wireless mesh networks (WMNs) offer a pragmatic solution with a cost-effective ratio when provisioning ubiquitous broadband internet access and diverse telecommunication systems. The conceptual underpinning of mesh networks finds application not only in IEEE networks, but also in 3GPP networks like LTE and [...] Read more.
Wireless mesh networks (WMNs) offer a pragmatic solution with a cost-effective ratio when provisioning ubiquitous broadband internet access and diverse telecommunication systems. The conceptual underpinning of mesh networks finds application not only in IEEE networks, but also in 3GPP networks like LTE and the low-power wide area network (LPWAN) tailored for the burgeoning Internet of Things (IoT) landscape. IEEE 802.11s is well known for its facto standard for WMN, which defines the hybrid wireless mesh protocol (HWMP) as a layer-2 routing protocol and airtime link (ALM) as a metric. In this intricate landscape, artificial intelligence (AI) plays a prominent role in the industry, particularly within the technology and telecommunication realms. This study presents a novel methodology for the computation of routing metrics, specifically the ALM. This methodology implements the network simulator ns-3 and the RStudio as a statistical computing environment for data analysis. The former has enabled for the creation of scripts that elicit a variety of scenarios for WMN where information is gathered and stored in databases. The latter (RStudio) takes this information, and at this point, two linear predictions are supported. The first uses linear models (lm) and the second employs general linear models (glm). To conclude this process, statistical tests are applied to the original model, as well as to the new suggested ones. This work substantially contributes in two ways: first, through the methodological tool for the metric calculation of the HWMP protocol that belongs to the IEEE 802.11s standard, using lm and glm for the selection and validation of the model regressors. At this stage the ANOVA and STEPWIZE tools of RStudio are used. The second contribution is a linear predictor that improves the WMN’s performance as a priori mechanism before the use of the ns-3 simulator. The ANCOVA tool of RStudio is employed in the latter. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

27 pages, 9370 KiB  
Article
Efficient On-Chip Learning of Multi-Layer Perceptron Based on Neuron Multiplexing Method
by Zhenyu Zhang, Guangsen Wang, Kang Wang, Bo Gan and Guoyong Chen
Electronics 2023, 12(17), 3607; https://doi.org/10.3390/electronics12173607 - 26 Aug 2023
Viewed by 981
Abstract
An efficient on-chip learning method based on neuron multiplexing is proposed in this paper to address the limitations of traditional on-chip learning methods, including low resource utilization and non-tunable parallelism. The proposed method utilizes a configurable neuron calculation unit (NCU) to calculate neural [...] Read more.
An efficient on-chip learning method based on neuron multiplexing is proposed in this paper to address the limitations of traditional on-chip learning methods, including low resource utilization and non-tunable parallelism. The proposed method utilizes a configurable neuron calculation unit (NCU) to calculate neural networks in different degrees of parallelism through multiplexing NCUs at different levels, and resource utilization can be increased by reducing the number of NCUs since the resource consumption is predominantly determined by the number of NCUs and the data bit-width, which are decoupled from the specific topology. To better support the proposed method and minimize RAM block usage, a weight segmentation and recombination method is introduced, accompanied by a detailed explanation of the access order. Moreover, a performance model is developed to facilitate parameter selection process. Experimental results conducted on an FPGA development board demonstrate that the proposed method has lower resource consumption, higher resource utilization, and greater generality compared to other methods. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

19 pages, 5924 KiB  
Article
Pm2.5 Time Series Imputation with Deep Learning and Interpolation
by Anibal Flores, Hugo Tito-Chura, Deymor Centty-Villafuerte and Alejandro Ecos-Espino
Computers 2023, 12(8), 165; https://doi.org/10.3390/computers12080165 - 16 Aug 2023
Cited by 4 | Viewed by 2601
Abstract
Commonly, regression for time series imputation has been implemented directly through regression models, statistical, machine learning, and deep learning techniques. In this work, a novel approach is proposed based on a classification model that determines the NA value class, and from this, two [...] Read more.
Commonly, regression for time series imputation has been implemented directly through regression models, statistical, machine learning, and deep learning techniques. In this work, a novel approach is proposed based on a classification model that determines the NA value class, and from this, two types of interpolations are implemented: polynomial or flipped polynomial. An hourly pm2.5 time series from Ilo City in southern Peru was chosen as a study case. The results obtained show that for gaps of one NA value, the proposal in most cases presents superior results to techniques such as ARIMA, LSTM, BiLSTM, GRU, and BiGRU; thus, on average, in terms of R2, the proposal exceeds implemented benchmark models by between 2.4341% and 19.96%. Finally, supported by the results, it can be stated that the proposal constitutes a good alternative for short-gaps imputation in pm2.5 time series. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

25 pages, 4877 KiB  
Article
A Mobile Solution for Enhancing Tourist Safety in Warm and Humid Destinations
by Sairoong Dinkoksung, Rapeepan Pitakaso, Chawis Boonmee, Thanatkit Srichok, Surajet Khonjun, Ganokgarn Jirasirilerd, Ponglert Songkaphet and Natthapong Nanthasamroeng
Appl. Sci. 2023, 13(15), 9027; https://doi.org/10.3390/app13159027 - 7 Aug 2023
Cited by 1 | Viewed by 3410
Abstract
This research introduces a mobile application specifically designed to enhance tourist safety in warm and humid destinations. The proposed solution integrates advanced functionalities, including a comprehensive warning system, health recommendations, and a life rescue system. The study showcases the exceptional effectiveness of the [...] Read more.
This research introduces a mobile application specifically designed to enhance tourist safety in warm and humid destinations. The proposed solution integrates advanced functionalities, including a comprehensive warning system, health recommendations, and a life rescue system. The study showcases the exceptional effectiveness of the implemented system, consistently providing tourists with precise and timely weather and safety information. Notably, the system achieves an impressive average accuracy rate of 100%, coupled with an astonishingly rapid response time of just 0.001 s. Furthermore, the research explores the correlation between the System Usability Scale (SUS) score and tourist engagement and loyalty. The findings reveal a positive relationship between the SUS score and the level of tourist engagement and loyalty. The proposed mobile solution holds significant potential for enhancing the safety and comfort of tourists in hot and humid climates, thereby making a noteworthy contribution to the advancement of the tourism business in smart cities. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

27 pages, 2495 KiB  
Article
Toward Improved Machine Learning-Based Intrusion Detection for Internet of Things Traffic
by Sarah Alkadi, Saad Al-Ahmadi and Mohamed Maher Ben Ismail
Computers 2023, 12(8), 148; https://doi.org/10.3390/computers12080148 - 27 Jul 2023
Cited by 10 | Viewed by 2439
Abstract
The rapid development of Internet of Things (IoT) networks has revealed multiple security issues. On the other hand, machine learning (ML) has proven its efficiency in building intrusion detection systems (IDSs) intended to reinforce the security of IoT networks. In fact, the successful [...] Read more.
The rapid development of Internet of Things (IoT) networks has revealed multiple security issues. On the other hand, machine learning (ML) has proven its efficiency in building intrusion detection systems (IDSs) intended to reinforce the security of IoT networks. In fact, the successful design and implementation of such techniques require the use of effective methods in terms of data and model quality. This paper encloses an empirical impact analysis for the latter in the context of a multi-class classification scenario. A series of experiments were conducted using six ML models, along with four benchmarking datasets, including UNSW-NB15, BOT-IoT, ToN-IoT, and Edge-IIoT. The proposed framework investigates the marginal benefit of employing data pre-processing and model configurations considering IoT limitations. In fact, the empirical findings indicate that the accuracy of ML-based IDS detection rapidly increases when methods that use quality data and models are deployed. Specifically, data cleaning, transformation, normalization, and dimensionality reduction, along with model parameter tuning, exhibit significant potential to minimize computational complexity and yield better performance. In addition, MLP- and clustering-based algorithms outperformed the remaining models, and the obtained accuracy reached up to 99.97%. One should note that the performance of the challenger models was assessed using similar test sets, and this was compared to the results achieved using the relevant pieces of research. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

28 pages, 1744 KiB  
Article
An Incident Detection Model Using Random Forest Classifier
by Osama ElSahly and Akmal Abdelfatah
Smart Cities 2023, 6(4), 1786-1813; https://doi.org/10.3390/smartcities6040083 - 17 Jul 2023
Cited by 3 | Viewed by 1991
Abstract
Traffic incidents have adverse effects on traffic operations, safety, and the economy. Efficient Automatic Incident Detection (AID) systems are crucial for timely and accurate incident detection. This paper develops a realistic AID model using the Random Forest (RF), which is a machine learning [...] Read more.
Traffic incidents have adverse effects on traffic operations, safety, and the economy. Efficient Automatic Incident Detection (AID) systems are crucial for timely and accurate incident detection. This paper develops a realistic AID model using the Random Forest (RF), which is a machine learning technique. The model is trained and tested on simulated data from VISSIM traffic simulation software. The model considers the variations in four critical factors: congestion levels, incident severity, incident location, and detector distance. Comparative evaluation with existing AID models, in the literature, demonstrates the superiority of the developed model, exhibiting higher Detection Rate (DR), lower Mean Time to Detect (MTTD), and lower False Alarm Rate (FAR). During training, the RF model achieved a DR of 96.97%, MTTD of 1.05 min, and FAR of 0.62%. During testing, it achieved a DR of 100%, MTTD of 1.17 min, and FAR of 0.862%. Findings indicate that detecting minor incidents during low traffic volumes is challenging. FAR decreases with the increase in Demand to Capacity ratio (D/C), while MTTD increases with D/C. Higher incident severity leads to lower MTTD values, while greater distance between an incident and upstream detector has the opposite effect. The FAR is inversely proportional to the incident’s location from the upstream detector, while being directly proportional to the distance between detectors. Larger detector spacings result in longer detection times. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

9 pages, 243 KiB  
Perspective
We Are Also Metabolites: Towards Understanding the Composition of Sweat on Fingertips via Hyperspectral Imaging
by Emanuela Marasco, Karl Ricanek and Huy Le
Digital 2023, 3(2), 137-145; https://doi.org/10.3390/digital3020010 - 19 Jun 2023
Cited by 1 | Viewed by 2075
Abstract
AI-empowered sweat metabolite analysis is an emerging and open research area with great potential to add a third category to biometrics: chemical. Current biometrics use two types of information to identify humans: physical (e.g., face, eyes) and behavioral (i.e., gait, typing). Sweat offers [...] Read more.
AI-empowered sweat metabolite analysis is an emerging and open research area with great potential to add a third category to biometrics: chemical. Current biometrics use two types of information to identify humans: physical (e.g., face, eyes) and behavioral (i.e., gait, typing). Sweat offers a promising solution for enriching human identity with more discerning characteristics to overcome the limitations of current technologies (e.g., demographic differential and vulnerability to spoof attacks). The analysis of a biometric trait’s chemical properties holds potential for providing a meticulous perspective on an individual. This not only changes the taxonomy for biometrics, but also lays a foundation for more accurate and secure next-generation biometric systems. This paper discusses existing evidence about the potential held by sweat components in representing the identity of a person. We also highlight emerging methodologies and applications pertaining to sweat analysis and guide the scientific community towards transformative future research directions to design AI-empowered systems of the next generation. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
23 pages, 4738 KiB  
Article
A Nonintrusive Load Identification Method Based on Improved Gramian Angular Field and ResNet18
by Jingqin Wang, Yufeng Wu and Liang Shu
Electronics 2023, 12(11), 2540; https://doi.org/10.3390/electronics12112540 - 5 Jun 2023
Cited by 2 | Viewed by 1735
Abstract
Image classification methods based on deep learning have been widely used in the study of nonintrusive load identification. However, in the process of encoding the load electrical signals into images, how to fully retain features of the raw data and thus increase the [...] Read more.
Image classification methods based on deep learning have been widely used in the study of nonintrusive load identification. However, in the process of encoding the load electrical signals into images, how to fully retain features of the raw data and thus increase the recognizability of loads carried with very similar current signals are still challenging, and the loss of load features will cause the overall accuracy of load identification to decrease. To deal with this problem, this paper proposes a nonintrusive load identification method based on the improved Gramian angular field (iGAF) and ResNet18. In the proposed method, fast Fourier transform is used to calculate the amplitude spectrum and the phase spectrum to reconstruct the pixel matrices of the B channel, G channel, and R channel of generated GAF images so that the color image fused by the three channels contains more information. This improvement to the GAF method enables generated images to retain the amplitude feature and phase feature of the raw data that are usually missed in the general GAF image. ResNet18 is trained with iGAF images for nonintrusive load identification. Experiments are conducted on two private datasets, ESEAD and EMCAD, and two public datasets, PLAID and WHITED. Experimental results suggest that the proposed method performs well on both private and public datasets, achieving overall identification accuracies of 99.545%, 99.375%, 98.964%, and 100% on the four datasets, respectively. In particular, the method demonstrates significant identification effects for loads with similar current waveforms in private datasets. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

18 pages, 4357 KiB  
Article
Assisting Heart Valve Diseases Diagnosis via Transformer-Based Classification of Heart Sound Signals
by Dongru Yang, Yi Lin, Jianwen Wei, Xiongwei Lin, Xiaobo Zhao, Yingbang Yao, Tao Tao, Bo Liang and Sheng-Guo Lu
Electronics 2023, 12(10), 2221; https://doi.org/10.3390/electronics12102221 - 13 May 2023
Cited by 6 | Viewed by 1792
Abstract
Background: In computer-aided medical diagnosis or prognosis, the automatic classification of heart valve diseases based on heart sound signals is of great importance since the heart sound signal contains a wealth of information that can reflect the heart status. Traditional binary classification algorithms [...] Read more.
Background: In computer-aided medical diagnosis or prognosis, the automatic classification of heart valve diseases based on heart sound signals is of great importance since the heart sound signal contains a wealth of information that can reflect the heart status. Traditional binary classification algorithms (normal and abnormal) currently cannot comprehensively assess the heart valve diseases based on analyzing various heart sounds. The differences between heart sound signals are relatively subtle, but the reflected heart conditions differ significantly. Consequently, from a clinical point of view, it is of utmost importance to assist in the diagnosis of heart valve disease through the multiple classification of heart sound signals. Methods: We utilized a Transformer model for the multi-classification of heart sound signals. It has achieved results from four abnormal heart sound signals and the typical type. Results: According to 5-fold cross-validation strategy as well as 10-fold cross-validation strategy, e.g., in 5-fold cross-validation, the proposed method achieved a highest accuracy of 98.74% and a mean AUC of 0.99. Furthermore, the classification accuracy for Aortic Stenosis, Mitral Regurgitation, Mitral Stenosis, Mitral Valve Prolapse, and standard heart sound signals is 98.72%, 98.50%, 98.30%, 98.56%, and 99.61%, respectively. In 10-fold cross-validation, our model obtained the highest accuracy, sensitivity, specificity, precision, and F1 score all at 100%. Conclusion: The results indicate that the framework can precisely classify five classes of heart sound signals. Our method provides an effective tool for the ancillary detection of heart valve diseases in the clinical setting. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

28 pages, 5922 KiB  
Review
A Review of Plant Disease Detection Systems for Farming Applications
by Mbulelo S. P. Ngongoma, Musasa Kabeya and Katleho Moloi
Appl. Sci. 2023, 13(10), 5982; https://doi.org/10.3390/app13105982 - 12 May 2023
Cited by 10 | Viewed by 4931
Abstract
The globe and more particularly the economically developed regions of the world are currently in the era of the Fourth Industrial Revolution (4IR). Conversely, the economically developing regions in the world (and more particularly the African continent) have not yet even fully passed [...] Read more.
The globe and more particularly the economically developed regions of the world are currently in the era of the Fourth Industrial Revolution (4IR). Conversely, the economically developing regions in the world (and more particularly the African continent) have not yet even fully passed through the Third Industrial Revolution (3IR) wave, and Africa’s economy is still heavily dependent on the agricultural field. On the other hand, the state of global food insecurity is worsening on an annual basis thanks to the exponential growth in the global human population, which continuously heightens the food demand in both quantity and quality. This justifies the significance of the focus on digitizing agricultural practices to improve the farm yield to meet the steep food demand and stabilize the economies of the African continent and countries such as India that are dependent on the agricultural sector to some extent. Technological advances in precision agriculture are already improving farm yields, although several opportunities for further improvement still exist. This study evaluated plant disease detection models (in particular, those over the past two decades) while aiming to gauge the status of the research in this area and identify the opportunities for further research. This study realized that little literature has discussed the real-time monitoring of the onset signs of diseases before they spread throughout the whole plant. There was also substantially less focus on real-time mitigation measures such as actuation operations, spraying pesticides, spraying fertilizers, etc., once a disease was identified. Very little research has focused on the combination of monitoring and phenotyping functions into one model capable of multiple tasks. Hence, this study highlighted a few opportunities for further focus. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

13 pages, 1772 KiB  
Article
Chinese News Text Classification Method via Key Feature Enhancement
by Bin Ge, Chunhui He, Hao Xu, Jibing Wu and Jiuyang Tang
Appl. Sci. 2023, 13(9), 5399; https://doi.org/10.3390/app13095399 - 26 Apr 2023
Cited by 3 | Viewed by 1680
Abstract
(1) Background: Chinese news text is a popular form of media communication, which can be seen everywhere in China. Chinese news text classification is an important direction in natural language processing (NLP). How to use high-quality text classification technology to help humans to [...] Read more.
(1) Background: Chinese news text is a popular form of media communication, which can be seen everywhere in China. Chinese news text classification is an important direction in natural language processing (NLP). How to use high-quality text classification technology to help humans to efficiently organize and manage the massive amount of web news is an urgent problem to be solved. It is noted that the existing deep learning methods rely on a large-scale tagged corpus for news text classification tasks and this model is poorly interpretable because the size is large. (2) Methods: To solve the above problems, this paper proposes a Chinese news text classification method based on key feature enhancement named KFE-CNN. It can effectively expand the semantic information of key features to enhance sample data and then combine the zero–one binary vector representation to transform text features into binary vectors and input them into CNN model for training and implementation, thus improving the interpretability of the model and effectively compressing the size of the model. (3) Results: The experimental results show that our method can significantly improve the overall performance of the model and the average accuracy and F1-score of the THUCNews subset of the public dataset reached 97.84% and 98%. (4) Conclusions: this fully proved the effectiveness of the KFE-CNN method for the Chinese news text classification task and it also fully demonstrates that key feature enhancement can improve classification performance. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

12 pages, 2577 KiB  
Article
Health Status Evaluation of Welding Robots Based on the Evidential Reasoning Rule
by Bang-Cheng Zhang, Ji-Dong Wang, Shuo Gao, Xiao-Jing Yin and Zhi Gao
Electronics 2023, 12(8), 1755; https://doi.org/10.3390/electronics12081755 - 7 Apr 2023
Cited by 1 | Viewed by 1347
Abstract
It is extremely important to monitor the health status of welding robots for the safe and stable operation of a body-in-white (BIW) welding production line. In the actual production process, the robot degradation rate is slow and the effective data are poor, which [...] Read more.
It is extremely important to monitor the health status of welding robots for the safe and stable operation of a body-in-white (BIW) welding production line. In the actual production process, the robot degradation rate is slow and the effective data are poor, which can reflect a degradation state in the large amount of obtained monitoring data, which causes difficulties in health status evaluation. In order to realize the accurate evaluation of the health status of welding robots, this paper proposes a health status evaluation method based on the evidential reasoning (ER) rule, which reflects the health status of welding robots by using the running state data monitored in actual engineering and through the qualitative knowledge of experts, which makes up for the lack of effective data. In the ER rule evaluation model, the covariance matrix adaptive evolutionary strategy (CMA-ES) algorithm is used to optimize the initial parameters of the evaluation model, which improved the accuracy of health status evaluations. Finally, a BIW welding robot was taken as an example for verification. The results show that the proposed model is able to accurately estimate the health status of the welding robot by using the monitored degradation data. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

19 pages, 331 KiB  
Article
Clustering of Monolingual Embedding Spaces
by Kowshik Bhowmik and Anca Ralescu
Digital 2023, 3(1), 48-66; https://doi.org/10.3390/digital3010004 - 23 Feb 2023
Viewed by 1794
Abstract
Suboptimal performance of cross-lingual word embeddings for distant and low-resource languages calls into question the isomorphic assumption integral to the mapping-based methods of obtaining such embeddings. This paper investigates the comparative impact of typological relationship and corpus size on the isomorphism between monolingual [...] Read more.
Suboptimal performance of cross-lingual word embeddings for distant and low-resource languages calls into question the isomorphic assumption integral to the mapping-based methods of obtaining such embeddings. This paper investigates the comparative impact of typological relationship and corpus size on the isomorphism between monolingual embedding spaces. To that end, two clustering algorithms were applied to three sets of pairwise degrees of isomorphisms. It is also the goal of the paper to determine the combination of the isomorphism measure and clustering algorithm that best captures the typological relationship among the chosen set of languages. Of the three measures investigated, Relational Similarity seemed to capture best the typological information of the languages encoded in their respective embedding spaces. These language clusters can help us identify, without any pre-existing knowledge about the real-world linguistic relationships shared among a group of languages, the related higher-resource languages of low-resource languages. The presence of such languages in the cross-lingual embedding space can help improve the performance of low-resource languages in a cross-lingual embedding space. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

22 pages, 10803 KiB  
Article
Learning and Compressing: Low-Rank Matrix Factorization for Deep Neural Network Compression
by Gaoyuan Cai, Juhu Li, Xuanxin Liu, Zhibo Chen and Haiyan Zhang
Appl. Sci. 2023, 13(4), 2704; https://doi.org/10.3390/app13042704 - 20 Feb 2023
Cited by 11 | Viewed by 4481
Abstract
Recently, the deep neural network (DNN) has become one of the most advanced and powerful methods used in classification tasks. However, the cost of DNN models is sometimes considerable due to the huge sets of parameters. Therefore, it is necessary to compress these [...] Read more.
Recently, the deep neural network (DNN) has become one of the most advanced and powerful methods used in classification tasks. However, the cost of DNN models is sometimes considerable due to the huge sets of parameters. Therefore, it is necessary to compress these models in order to reduce the parameters in weight matrices and decrease computational consumption, while maintaining the same level of accuracy. In this paper, in order to deal with the compression problem, we first combine the loss function and the compression cost function into a joint function, and optimize it as an optimization framework. Then we combine the CUR decomposition method with this joint optimization framework to obtain the low-rank approximation matrices. Finally, we narrow the gap between the weight matrices and the low-rank approximations to compress the DNN models on the image classification task. In this algorithm, we not only solve the optimal ranks by enumeration, but also obtain the compression result with low-rank characteristics iteratively. Experiments were carried out on three public datasets under classification tasks. Comparisons with baselines and current state-of-the-art results can conclude that our proposed low-rank joint optimization compression algorithm can achieve higher accuracy and compression ratios. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

14 pages, 2655 KiB  
Article
A Multi-Channel Contrastive Learning Network Based Intrusion Detection Method
by Jian Luo, Yiying Zhang, Yannian Wu, Yao Xu, Xiaoyan Guo and Boxiang Shang
Electronics 2023, 12(4), 949; https://doi.org/10.3390/electronics12040949 - 14 Feb 2023
Cited by 6 | Viewed by 2007
Abstract
Network intrusion data are characterized by high feature dimensionality, extreme category imbalance, and complex nonlinear relationships between features and categories. The actual detection accuracy of existing supervised intrusion-detection models performs poorly. To address this problem, this paper proposes a multi-channel contrastive learning network-based [...] Read more.
Network intrusion data are characterized by high feature dimensionality, extreme category imbalance, and complex nonlinear relationships between features and categories. The actual detection accuracy of existing supervised intrusion-detection models performs poorly. To address this problem, this paper proposes a multi-channel contrastive learning network-based intrusion-detection method (MCLDM), which combines feature learning in the multi-channel supervised contrastive learning stage and feature extraction in the multi-channel unsupervised contrastive learning stage to train an effective intrusion-detection model. The objective is to research whether feature enrichment and the use of contrastive learning for specific classes of network intrusion data can improve the accuracy of the model. The model is based on an autoencoder to achieve feature reconstruction with supervised contrastive learning and for implementing multi-channel data reconstruction. In the next stage of unsupervised contrastive learning, the extraction of features is implemented using triplet convolutional neural networks (TCNN) to achieve the classification of intrusion data. Through experimental analysis, the multichannel contrastive learning network-based intrusion-detection method achieves 98.43% accuracy in dataset CICIDS17 and 93.94% accuracy in dataset KDDCUP99. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

18 pages, 2231 KiB  
Article
Symbiotic Combination of a Bayesian Network and Fuzzy Logic to Quantify the QoS in a VANET: Application in Logistic 4.0
by Hafida Khalfaoui, Abdellah Azmani, Abderrazak Farchane and Said Safi
Computers 2023, 12(2), 40; https://doi.org/10.3390/computers12020040 - 14 Feb 2023
Cited by 3 | Viewed by 2015
Abstract
Intelligent transportation systems use new technologies to improve road safety. In them, vehicles have been equipped with wireless communication systems called on-board units (OBUs) to be able to communicate with each other. This type of wireless network refers to vehicular ad hoc networks [...] Read more.
Intelligent transportation systems use new technologies to improve road safety. In them, vehicles have been equipped with wireless communication systems called on-board units (OBUs) to be able to communicate with each other. This type of wireless network refers to vehicular ad hoc networks (VANET). The primary problem in a VANET is the quality of service (QoS) because a small problem in the services can extremely damage both human lives and the economy. From this perspective, this article makes a contribution within the framework of a new conceptual project called the Smart Digital Logistic Services Provider (Smart DLSP). This is intended to give freight vehicles more intelligence in the service of logistics on a global scale. This article proposes a model that combines two approaches—a Bayesian network and fuzzy logic for calculating the QoS in a VANET as a function of multiple criteria—and provides a database that helps determine the originality of the risk of degrading the QoS in the network. The outcome of this approach was employed in an event tree analysis to assess the impact of the system’s security mechanisms. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

16 pages, 10359 KiB  
Article
An Improved Algorithm for Insulator and Defect Detection Based on YOLOv4
by Gujing Han, Qiwei Yuan, Feng Zhao, Ruijie Wang, Liu Zhao, Saidian Li, Min He, Shiqi Yang and Liang Qin
Electronics 2023, 12(4), 933; https://doi.org/10.3390/electronics12040933 - 13 Feb 2023
Cited by 10 | Viewed by 2136
Abstract
To further improve the accuracy and speed of UAV inspection of transmission line insulator defects, this paper proposes an insulator detection and defect identification algorithm based on YOLOv4, which is called DSMH-YOLOv4. In the feature extraction network of the YOLOv4 model, the improved [...] Read more.
To further improve the accuracy and speed of UAV inspection of transmission line insulator defects, this paper proposes an insulator detection and defect identification algorithm based on YOLOv4, which is called DSMH-YOLOv4. In the feature extraction network of the YOLOv4 model, the improved algorithm improves the residual edges of the residual structure based on feature reuse and designs the backbone network D-CSPDarknet53, which greatly reduces the number of parameters and computation of the model. The SA-Net (Shuffle Attention Neural Networks) attention model is embedded in the feature fusion network to strengthen the attention of target features and improve the weight of the target. Multi-head output is added to the output layer to improve the ability of the model to recognize the small target of insulator damage. The experimental results show that the number of parameters of the improved algorithm model is only 25.98% of that of the original model, and the mAP (mean Average Precision) of the insulator and defect is increased from 92.44% to 96.14%, which provides an effective way for the implementation of edge end algorithm deployment. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

18 pages, 5512 KiB  
Article
Single Image Reflection Removal Based on Residual Attention Mechanism
by Yubin Guo, Wanzhou Lu, Ximing Li and Qiong Huang
Appl. Sci. 2023, 13(3), 1618; https://doi.org/10.3390/app13031618 - 27 Jan 2023
Cited by 2 | Viewed by 2878
Abstract
Affected by shooting angle and light intensity, shooting through transparent media may cause light reflections in an image and influence picture quality, which has a negative effect on the research of computer vision tasks. In this paper, we propose a Residual Attention Based [...] Read more.
Affected by shooting angle and light intensity, shooting through transparent media may cause light reflections in an image and influence picture quality, which has a negative effect on the research of computer vision tasks. In this paper, we propose a Residual Attention Based Reflection Removal Network (RABRRN) to tackle the issue of single image reflection removal. We hold that reflection removal is essentially an image separation problem sensitive to both spatial and channel features. Therefore, we integrate spatial attention and channel attention into the model to enhance spatial and channel feature representation. For a more feasible solution to solve the problem of gradient disappearance in the iterative training of deep neural networks, the attention module is combined with a residual network to design a residual attention module so that the performance of reflection removal can be ameliorated. In addition, we establish a reflection image dataset named the SCAU Reflection Image Dataset (SCAU-RID), providing sufficient real training data. The experimental results show that the proposed method achieves a PSNR of 23.787 dB and an SSIM value of 0.885 from four benchmark datasets. Compared with the other most advanced methods, our method has only 18.524M parameters, but it obtains the best results from test datasets. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1