Sign in to use this feature.

Years

Between: -

Article Types

Countries / Regions

Search Results (48)

Search Parameters:
Journal = MAKE
Section = Network

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 1926 KB  
Article
A Novel Approach to Company Bankruptcy Prediction Using Convolutional Neural Networks and Generative Adversarial Networks
by Alessia D’Ercole and Gianluigi Me
Mach. Learn. Knowl. Extr. 2025, 7(3), 63; https://doi.org/10.3390/make7030063 - 7 Jul 2025
Cited by 1 | Viewed by 2155
Abstract
Predicting company bankruptcy is a critical task in financial risk assessment. This study introduces a novel approach using Convolutional Neural Networks (CNNs) and Generative Adversarial Networks (GANs) to enhance bankruptcy prediction accuracy. By transforming financial statements into grayscale images and leveraging synthetic data [...] Read more.
Predicting company bankruptcy is a critical task in financial risk assessment. This study introduces a novel approach using Convolutional Neural Networks (CNNs) and Generative Adversarial Networks (GANs) to enhance bankruptcy prediction accuracy. By transforming financial statements into grayscale images and leveraging synthetic data generation, we analyze a dataset of 6249 companies, including 3256 active and 2993 bankrupt firms. Our methodology innovates by addressing dataset limitations through GAN-based data augmentation. CNNs are employed to take advantage of their ability to extract hierarchical patterns from financial statement images, providing a new approach to financial analysis, while GANs help mitigate dataset imbalance by generating realistic synthetic data for training. We generate synthetic financial data that closely mimics real-world patterns, expanding the training dataset and potentially improving classifier performance. The CNN model is trained on a combination of real and synthetic data, with strict separation between training/validation and testing. Full article
(This article belongs to the Section Network)
Show Figures

Graphical abstract

21 pages, 3561 KB  
Article
Artificial Intelligence Meets Bioequivalence: Using Generative Adversarial Networks for Smarter, Smaller Trials
by Anastasios Nikolopoulos and Vangelis D. Karalis
Mach. Learn. Knowl. Extr. 2025, 7(2), 47; https://doi.org/10.3390/make7020047 - 23 May 2025
Viewed by 2100
Abstract
This study introduces artificial intelligence as a powerful tool to transform bioequivalence (BE) trials. We apply advanced generative models, specifically Wasserstein Generative Adversarial Networks (WGANs), to create virtual subjects and reduce the need for real human participants in generic drug assessment. Although BE [...] Read more.
This study introduces artificial intelligence as a powerful tool to transform bioequivalence (BE) trials. We apply advanced generative models, specifically Wasserstein Generative Adversarial Networks (WGANs), to create virtual subjects and reduce the need for real human participants in generic drug assessment. Although BE studies typically involve small sample sizes (usually 24 subjects), which may limit the use of AI-generated populations, our findings show that these models can successfully overcome this challenge. To show the utility of generative AI algorithms in BE testing, this study applied Monte Carlo simulations of 2 × 2 crossover BE trials, combined with WGANs. After training of the WGAN model, several scenarios were explored, including sample size, the proportion of subjects used for the synthesis of virtual subjects, and variabilities. The performance of the AI-synthesized populations was tested in two ways: (a) first, by assessing the similarity of the performance with the actual population, and (b) second, by evaluating the statistical power achieved, which aimed to be as high as that of the entire original population. The results demonstrated that WGANs could generate virtual populations with BE acceptance percentages and similarity levels that matched or exceeded those of the original population. This approach proved effective across various scenarios, enhancing BE study sample sizes, reducing costs, and accelerating trial durations. This study highlights the potential of WGANs to improve data augmentation and optimize subject recruitment in BE studies. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

36 pages, 11592 KB  
Article
A Novel Approach Based on Hypergraph Convolutional Neural Networks for Cartilage Shape Description and Longitudinal Prediction of Knee Osteoarthritis Progression
by John B. Theocharis, Christos G. Chadoulos and Andreas L. Symeonidis
Mach. Learn. Knowl. Extr. 2025, 7(2), 40; https://doi.org/10.3390/make7020040 - 26 Apr 2025
Cited by 2 | Viewed by 1378
Abstract
Knee osteoarthritis (KOA) is a highly prevalent muscoloskeletal joint disorder affecting a significant portion of the population worldwide. Accurate predictions of KOA progression can assist clinicians in drawing preventive strategies for patients. In this paper, we present an integrated approach based [...] Read more.
Knee osteoarthritis (KOA) is a highly prevalent muscoloskeletal joint disorder affecting a significant portion of the population worldwide. Accurate predictions of KOA progression can assist clinicians in drawing preventive strategies for patients. In this paper, we present an integrated approach based on hypergraph convolutional networks (HGCNs) for longitudinal predictions of KOA grades and progressions from MRI images. We propose two novel models, namely, the C_Shape.Net and the predictor network. The C_Shape.Net operates on a hypergraph of volumetric nodes, especially designed to represent the surface and volumetric features of the cartilage. It encompasses deep HGCN convolutions, graph pooling, and readout operations in a hierarchy of layers, providing, at the output, expressive 3D shape descriptors of the cartilage volume. The predictor is a spatio-temporal HGCN network (ST_HGCN), following the sequence-to-sequence learning scheme. Concretely, it transforms sequences of knee representations at the historical stage into sequences of KOA predictions at the prediction stage. The predictor includes spatial HGCN convolutions, attention-based temporal fusion of feature embeddings at multiple layers, and a transformer module that generates longitudinal predictions at follow-up times. We present comprehensive experiments on the Osteoarthritis Initiative (OAI) cohort to evaluate the performance of our methodology for various tasks, including node classification, longitudinal KL grading, and progression. The basic finding of the experiments is that the larger the depth of the historical stage, the higher the accuracy of the obtained predictions in all tasks. For the maximum historic depth of four years, our method yielded an average balanced accuracy (BA) of 85.94% in KOA grading, and accuracies of 91.89% (+1), 88.11% (+2), 84.35% (+3), and 79.41% (+4) for the four consecutive follow-up visits. Under the same setting, we also achieved an average value of Area Under Curve (AUC) of 0.94 for the prediction of progression incidence, and follow-up AUC values of 0.81 (+1), 0.77 (+2), 0.73 (+3), and 0.68 (+4), respectively. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

22 pages, 17414 KB  
Article
Advancing Particle Tracking: Self-Organizing Map Hyperparameter Study and Long Short-Term Memory-Based Outlier Detection
by Max Klein, Niklas Dormagen, Lukas Wimmer, Markus H. Thoma and Mike Schwarz
Mach. Learn. Knowl. Extr. 2025, 7(2), 37; https://doi.org/10.3390/make7020037 - 17 Apr 2025
Cited by 1 | Viewed by 1642
Abstract
Particle tracking velocimetry (PTV) forms the basis for many fluid dynamic experiments, in which individual particles are tracked across multiple successive images. However, when the experimental setup involves high-speed, high-density particles that are indistinguishable and follow complex or unknown flow fields, matching particles [...] Read more.
Particle tracking velocimetry (PTV) forms the basis for many fluid dynamic experiments, in which individual particles are tracked across multiple successive images. However, when the experimental setup involves high-speed, high-density particles that are indistinguishable and follow complex or unknown flow fields, matching particles between images becomes significantly more challenging. Reliable PTV algorithms are crucial in such scenarios. Previous work has demonstrated that the Self-Organizing Map (SOM) machine learning approach offers superior outcomes on complex-plasma data compared with traditional methods, though its performance is sensitive to hyperparameter calibration, which requires optimization for specific flow scenarios. In this article, we describe how the dependence of the various hyperparameters on different flow scenarios was studied and the optimal settings for diverse flow conditions were identified. Based on these results, automatic hyperparameter calibration was implemented in the PTV framework. Furthermore, the SOM’s performance was directly compared with that of the preceding conventional PTV method, Trackpy, for complex plasmas using synthetic data. Finally, as a new approach to identifying incorrectly matched particle traces, a Long Short-Term Memory (LSTM) neural network was developed to sort out all inaccuracies to further improve the outcome. Combined with automatic hyperparameter calibration, outlier detection and additional computational speed optimization, this work delivers a robust, versatile and efficient framework for PTV analysis. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

21 pages, 2466 KB  
Article
Enhancing Performance of Credit Card Model by Utilizing LSTM Networks and XGBoost Algorithms
by Kianeh Kandi and Antonio García-Dopico
Mach. Learn. Knowl. Extr. 2025, 7(1), 20; https://doi.org/10.3390/make7010020 - 21 Feb 2025
Cited by 15 | Viewed by 5736
Abstract
This research paper presents novel approaches for detecting credit card risk through the utilization of Long Short-Term Memory (LSTM) networks and XGBoost algorithms. Facing the challenge of securing credit card transactions, this study explores the potential of LSTM networks for their ability to [...] Read more.
This research paper presents novel approaches for detecting credit card risk through the utilization of Long Short-Term Memory (LSTM) networks and XGBoost algorithms. Facing the challenge of securing credit card transactions, this study explores the potential of LSTM networks for their ability to understand sequential dependencies in transaction data. This research sheds light on which model is more effective in addressing the challenges posed by imbalanced datasets in credit risk assessment. The methodology utilized for imbalanced datasets includes the use of the Synthetic Minority Oversampling Technique (SMOTE) to address any imbalance in class distribution. This paper conducts an extensive literature review, comparing various machine learning methods, and proposes an innovative framework that compares LSTM with XGBoost to improve fraud detection accuracy. LSTM, a recurrent neural network renowned for its ability to capture temporal dependencies within sequences of transactions, is compared with XGBoost, a formidable ensemble learning algorithm that enhances feature-based classification. By meticulously carrying out preprocessing tasks, constructing competent training models, and implementing ensemble techniques, our proposed framework demonstrates unwavering performance in accurately identifying fraudulent transactions. The comparison of LSTM and XGBoost shows that LSTM is more effective for our imbalanced dataset. Compared with XGBOOST’s 97% accuracy, LSTM’s accuracy is 99%. The final result emphasizes how crucial it is to select the optimal algorithm based on particular criteria within financial concerns, which will ultimately result in more reliable and knowledgeable credit score decisions. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

12 pages, 1257 KB  
Article
ExShall-CNN: An Explainable Shallow Convolutional Neural Network for Medical Image Segmentation
by Vahid Khalkhali, Sayed Mehedi Azim and Iman Dehzangi
Mach. Learn. Knowl. Extr. 2025, 7(1), 19; https://doi.org/10.3390/make7010019 - 15 Feb 2025
Cited by 2 | Viewed by 2741
Abstract
Explainability is essential for AI models, especially in clinical settings where understanding the model’s decisions is crucial. Despite their impressive performance, black-box AI models are unsuitable for clinical use if their operations cannot be explained to clinicians. While deep neural networks (DNNs) represent [...] Read more.
Explainability is essential for AI models, especially in clinical settings where understanding the model’s decisions is crucial. Despite their impressive performance, black-box AI models are unsuitable for clinical use if their operations cannot be explained to clinicians. While deep neural networks (DNNs) represent the forefront of model performance, their explanations are often not easily interpreted by humans. On the other hand, hand-crafted features extracted to represent different aspects of the input data and traditional machine learning models are generally more understandable. However, they often lack the effectiveness of advanced models due to human limitations in feature design. To address this, we propose ExShall-CNN, a novel explainable shallow convolutional neural network for medical image processing. This model improves upon hand-crafted features to maintain human interpretability, ensuring that its decisions are transparent and understandable. We introduce the explainable shallow convolutional neural network (ExShall-CNN), which combines the interpretability of hand-crafted features with the performance of advanced deep convolutional networks like U-Net for medical image segmentation. Built on recent advancements in machine learning, ExShall-CNN incorporates widely used kernels while ensuring transparency, making its decisions visually interpretable by physicians and clinicians. This balanced approach offers both the accuracy of deep learning models and the explainability needed for clinical applications. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

23 pages, 1352 KB  
Article
A Hybrid Gradient Boosting and Neural Network Model for Predicting Urban Happiness: Integrating Ensemble Learning with Deep Representation for Enhanced Accuracy
by Gregorius Airlangga and Alan Liu
Mach. Learn. Knowl. Extr. 2025, 7(1), 4; https://doi.org/10.3390/make7010004 - 7 Jan 2025
Cited by 14 | Viewed by 6172
Abstract
Urban happiness prediction presents a complex challenge, due to the nonlinear and multifaceted relationships among socio-economic, environmental, and infrastructural factors. This study introduces an advanced hybrid model combining a gradient boosting machine (GBM) and neural network (NN) to address these complexities. Unlike traditional [...] Read more.
Urban happiness prediction presents a complex challenge, due to the nonlinear and multifaceted relationships among socio-economic, environmental, and infrastructural factors. This study introduces an advanced hybrid model combining a gradient boosting machine (GBM) and neural network (NN) to address these complexities. Unlike traditional approaches, this hybrid leverages a GBM to handle structured data features and an NN to extract deeper nonlinear relationships. The model was evaluated against various baseline machine learning and deep learning models, including a random forest, CNN, LSTM, CatBoost, and TabNet, using metrics such as RMSE, MAE, R2, and MAPE. The GBM + NN hybrid achieved superior performance, with the lowest RMSE of 0.3332, an R2 of 0.9673, and an MAPE of 7.0082%. The model also revealed significant insights into urban indicators, such as a 10% improvement in air quality correlating to a 5% increase in happiness. These findings underscore the potential of hybrid models in urban analytics, offering both predictive accuracy and actionable insights for urban planners. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

20 pages, 1278 KB  
Article
Application of Bayesian Neural Networks in Healthcare: Three Case Studies
by Lebede Ngartera, Mahamat Ali Issaka and Saralees Nadarajah
Mach. Learn. Knowl. Extr. 2024, 6(4), 2639-2658; https://doi.org/10.3390/make6040127 - 16 Nov 2024
Cited by 12 | Viewed by 7246
Abstract
This study aims to explore the efficacy of Bayesian Neural Networks (BNNs) in enhancing predictive modeling for healthcare applications. Advancements in artificial intelligence have significantly improved predictive modeling capabilities, with BNNs offering a probabilistic framework that addresses the inherent uncertainty and variability in [...] Read more.
This study aims to explore the efficacy of Bayesian Neural Networks (BNNs) in enhancing predictive modeling for healthcare applications. Advancements in artificial intelligence have significantly improved predictive modeling capabilities, with BNNs offering a probabilistic framework that addresses the inherent uncertainty and variability in healthcare data. This study demonstrates the real-world applicability of BNNs through three key case studies: personalized diabetes treatment, early Alzheimer’s disease detection, and predictive modeling for HbA1c levels. By leveraging the Bayesian approach, these models provide not only enhanced predictive accuracy but also uncertainty quantification, a critical factor in clinical decision making. While the findings are promising, future research should focus on optimizing scalability and integration for real-world applications. This work lays a foundation for future studies, including the development of rating scales based on BNN predictions to improve clinical outcomes. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

47 pages, 1749 KB  
Article
Automatic Extraction and Visualization of Interaction Networks for German Fairy Tales
by David Schmidt and Frank Puppe
Mach. Learn. Knowl. Extr. 2024, 6(4), 2447-2493; https://doi.org/10.3390/make6040121 - 1 Nov 2024
Viewed by 2223
Abstract
Interaction networks are a method of displaying the significant characters in a narrative text and their interactions. We automatically construct interaction networks from dialogues in German fairy tales by the Brothers Grimm and subsequently visualize these networks. This requires the combination of algorithms [...] Read more.
Interaction networks are a method of displaying the significant characters in a narrative text and their interactions. We automatically construct interaction networks from dialogues in German fairy tales by the Brothers Grimm and subsequently visualize these networks. This requires the combination of algorithms for several tasks: coreference resolution for the identification of characters and their appearances, as well as speaker/addressee detection and the detection of dialogue boundaries for the identification of interactions. After an evaluation of the individual algorithms, the predicted networks are evaluated against benchmarks established by networks based on manually annotated coreference and speaker/addressee information. The evaluation focuses on specific components of the predicted networks, such as the nodes, as well as the overall network, employing a newly devised score. This is followed by an analysis of various types of errors that the algorithms can make, like a coreference resolution algorithm not realizing that the frog has transformed into a prince, and their impact on the created networks. We find that the quality of many predicted networks is satisfactory for use cases in which the reliability of edges and character types are not of critical importance. However, there is considerable room for improvement. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

18 pages, 893 KB  
Article
Temporal Relational Graph Convolutional Network Approach to Financial Performance Prediction
by Brindha Priyadarshini Jeyaraman, Bing Tian Dai and Yuan Fang
Mach. Learn. Knowl. Extr. 2024, 6(4), 2303-2320; https://doi.org/10.3390/make6040113 - 10 Oct 2024
Cited by 5 | Viewed by 4582
Abstract
Accurately predicting financial entity performance remains a challenge due to the dynamic nature of financial markets and vast unstructured textual data. Financial knowledge graphs (FKGs) offer a structured representation for tackling this problem by representing complex financial relationships and concepts. However, constructing a [...] Read more.
Accurately predicting financial entity performance remains a challenge due to the dynamic nature of financial markets and vast unstructured textual data. Financial knowledge graphs (FKGs) offer a structured representation for tackling this problem by representing complex financial relationships and concepts. However, constructing a comprehensive and accurate financial knowledge graph that captures the temporal dynamics of financial entities is non-trivial. We introduce FintechKG, a comprehensive financial knowledge graph developed through a three-dimensional information extraction process that incorporates commercial entities and temporal dimensions and uses a financial concept taxonomy that ensures financial domain entity and relationship extraction. We propose a temporal and relational graph convolutional network (RGCN)-based representation for FintechKG data across multiple timesteps, which captures temporal dependencies. This representation is then combined with FinBERT embeddings through a projection layer, enabling a richer feature space. To demonstrate the efficacy of FintechKG, we evaluate its performance using the example task of financial performance prediction. A logistic regression model uses these combined features and social media embeddings for performance prediction. We classify whether the revenue will increase or decrease. This approach demonstrates the effectiveness of FintechKG combined with textual information for accurate financial forecasting. Our work contributes a systematic FKG construction method and a framework that utilizes both relational and textual embeddings for improved financial performance prediction. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

16 pages, 2094 KB  
Article
Graph Convolutional Networks for Predicting Cancer Outcomes and Stage: A Focus on cGAS-STING Pathway Activation
by Mateo Sokač, Borna Skračić, Danijel Kučak and Leo Mršić
Mach. Learn. Knowl. Extr. 2024, 6(3), 2033-2048; https://doi.org/10.3390/make6030100 - 11 Sep 2024
Cited by 1 | Viewed by 2969
Abstract
The study presented in this paper evaluated gene expression profiles from The Cancer Genome Atlas (TCGA). To reduce complexity, we focused on genes in the cGAS–STING pathway, crucial for cytosolic DNA detection and immune response. The study analyzes three clinical variables: disease-specific survival [...] Read more.
The study presented in this paper evaluated gene expression profiles from The Cancer Genome Atlas (TCGA). To reduce complexity, we focused on genes in the cGAS–STING pathway, crucial for cytosolic DNA detection and immune response. The study analyzes three clinical variables: disease-specific survival (DSS), overall survival (OS), and tumor stage. To effectively utilize the high-dimensional gene expression data, we needed to find a way to project these data meaningfully. Since gene pathways can be represented as graphs, a novel method of presenting genomics data using graph data structure was employed, rather than the conventional tabular format. To leverage the gene expression data represented as graphs, we utilized a graph convolutional network (GCN) machine learning model in conjunction with the genetic algorithm optimization technique. This allowed for obtaining an optimal graph representation topology and capturing important activations within the pathway for each use case, enabling a more insightful analysis of the cGAS–STING pathway and its activations across different cancer types and clinical variables. To tackle the problem of unexplainable AI, graph visualization alongside the integrated gradients method was employed to explain the GCN model’s decision-making process, identifying key nodes (genes) in the cGAS–STING pathway. This approach revealed distinct molecular mechanisms, enhancing interpretability. This study demonstrates the potential of GCNs combined with explainable AI to analyze gene expression, providing insights into cancer progression. Further research with more data is needed to validate these findings. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

17 pages, 2683 KB  
Article
Forecasting the Right Crop Nutrients for Specific Crops Based on Collected Data Using an Artificial Neural Network (ANN)
by Sairoel Amertet and Girma Gebresenbet
Mach. Learn. Knowl. Extr. 2024, 6(3), 1936-1952; https://doi.org/10.3390/make6030095 - 26 Aug 2024
Cited by 2 | Viewed by 1952
Abstract
In farming technologies, it is difficult to properly provide the accurate crop nutrients for respective crops. For this reason, farmers are experiencing enormous problems. Although various types of machine learning (deep learning and convolutional neural networks) have been used to identify crop diseases, [...] Read more.
In farming technologies, it is difficult to properly provide the accurate crop nutrients for respective crops. For this reason, farmers are experiencing enormous problems. Although various types of machine learning (deep learning and convolutional neural networks) have been used to identify crop diseases, as has crop classification-based image processing, they have failed to forecast accurate crop nutrients for various crops, as crop nutrients are numerical instead of visual. Neural networks represent an opportunity for the precision agriculture sector to more accurately forecast crop nutrition. Recent technological advancements in neural networks have begun to provide greater precision, with an array of opportunities in pattern recognition. Neural networks represent an opportunity to effectively solve numerical data problems. The aim of the current study is to estimate the right crop nutrients for the right crops based on the data collected using an artificial neural network. The crop data were collected from the MNIST dataset. To forecast the precise nutrients for the crops, ANN models were developed. The entire system was simulated in a MATLAB environment. The obtained results for forecasting accurate nutrients were 99.997%, 99.996%, and 99.997% for validation, training, and testing, respectively. Therefore, the proposed algorithm is suitable for forecasting accurate crop nutrients for the crops. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

14 pages, 7188 KB  
Article
Accuracy Improvement of Debonding Damage Detection Technology in Composite Blade Joints for 20 kW Class Wind Turbine
by Hakgeun Kim, Hyeongjin Kim and Kiweon Kang
Mach. Learn. Knowl. Extr. 2024, 6(3), 1857-1870; https://doi.org/10.3390/make6030091 - 7 Aug 2024
Viewed by 1691
Abstract
Securing the structural safety of blades has become crucial, owing to the increasing size and weight of blades resulting from the recent development of large wind turbines. Composites are primarily used for blade manufacturing because of their high specific strength and specific stiffness. [...] Read more.
Securing the structural safety of blades has become crucial, owing to the increasing size and weight of blades resulting from the recent development of large wind turbines. Composites are primarily used for blade manufacturing because of their high specific strength and specific stiffness. However, in composite blades, joints may experience fractures from the loads generated during wind turbine operation, leading to deformation caused by changes in structural stiffness. In this study, 7132 debonding damage data, classified by damage type, position, and size, were selected to predict debonding damage based on natural frequency. The change in the natural frequency caused by debonding damage was acquired through finite element (FE) modeling and modal analysis. Synchronization between the FE analysis model and manufactured blades was achieved through modal testing and data analysis. Finally, the relationship between debonding damage and the change in natural frequency was examined using artificial neural network techniques. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

20 pages, 4689 KB  
Article
Extending Multi-Output Methods for Long-Term Aboveground Biomass Time Series Forecasting Using Convolutional Neural Networks
by Efrain Noa-Yarasca, Javier M. Osorio Leyton and Jay P. Angerer
Mach. Learn. Knowl. Extr. 2024, 6(3), 1633-1652; https://doi.org/10.3390/make6030079 - 17 Jul 2024
Cited by 6 | Viewed by 3325
Abstract
Accurate aboveground vegetation biomass forecasting is essential for livestock management, climate impact assessments, and ecosystem health. While artificial intelligence (AI) techniques have advanced time series forecasting, a research gap in predicting aboveground biomass time series beyond single values persists. This study introduces RECMO [...] Read more.
Accurate aboveground vegetation biomass forecasting is essential for livestock management, climate impact assessments, and ecosystem health. While artificial intelligence (AI) techniques have advanced time series forecasting, a research gap in predicting aboveground biomass time series beyond single values persists. This study introduces RECMO and DirRecMO, two multi-output methods for forecasting aboveground vegetation biomass. Using convolutional neural networks, their efficacy is evaluated across short-, medium-, and long-term horizons on six Kenyan grassland biomass datasets, and compared with that of existing single-output methods (Recursive, Direct, and DirRec) and multi-output methods (MIMO and DIRMO). The results indicate that single-output methods are superior for short-term predictions, while both single-output and multi-output methods exhibit a comparable effectiveness in long-term forecasts. RECMO and DirRecMO outperform established multi-output methods, demonstrating a promising potential for biomass forecasting. This study underscores the significant impact of multi-output size on forecast accuracy, highlighting the need for optimal size adjustments and showcasing the proposed methods’ flexibility in long-term forecasts. Short-term predictions show less significant differences among methods, complicating the identification of the best performer. However, clear distinctions emerge in medium- and long-term forecasts, underscoring the greater importance of method choice for long-term predictions. Moreover, as the forecast horizon extends, errors escalate across all methods, reflecting the challenges of predicting distant future periods. This study suggests advancing hybrid models (e.g., RECMO and DirRecMO) to improve extended horizon forecasting. Future research should enhance adaptability, investigate multi-output impacts, and conduct comparative studies across diverse domains, datasets, and AI algorithms for robust insights. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

10 pages, 10910 KB  
Article
Enhancing Computation-Efficiency of Deep Neural Network Processing on Edge Devices through Serial/Parallel Systolic Computing
by Iraj Moghaddasi and Byeong-Gyu Nam
Mach. Learn. Knowl. Extr. 2024, 6(3), 1484-1493; https://doi.org/10.3390/make6030070 - 1 Jul 2024
Cited by 4 | Viewed by 2433
Abstract
In recent years, deep neural networks (DNNs) have addressed new applications with intelligent autonomy, often achieving higher accuracy than human experts. This capability comes at the expense of the ever-increasing complexity of emerging DNNs, causing enormous challenges while deploying on resource-limited edge devices. [...] Read more.
In recent years, deep neural networks (DNNs) have addressed new applications with intelligent autonomy, often achieving higher accuracy than human experts. This capability comes at the expense of the ever-increasing complexity of emerging DNNs, causing enormous challenges while deploying on resource-limited edge devices. Improving the efficiency of DNN hardware accelerators by compression has been explored previously. Existing state-of-the-art studies applied approximate computing to enhance energy efficiency even at the expense of a little accuracy loss. In contrast, bit-serial processing has been used for improving the computational efficiency of neural processing without accuracy loss, exploiting a simple design, dynamic precision adjustment, and computation pruning. This research presents Serial/Parallel Systolic Array (SPSA) and Octet Serial/Parallel Systolic Array (OSPSA) processing elements for edge DNN acceleration, which exploit bit-serial processing on systolic array architecture for improving computational efficiency. For evaluation, all designs were described at the RTL level and synthesized in 28 nm technology. Post-synthesis cycle-accurate simulations of image classification over DNNs illustrated that, on average, a sample 16 × 16 systolic array indicated remarkable improvements of 17.6% and 50.6% in energy efficiency compared to the baseline, with no loss of accuracy. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

Back to TopTop