Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (144)

Search Parameters:
Keywords = stochastic gradient descent (SGD)

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
32 pages, 999 KB  
Article
A Robust Hybrid Metaheuristic Framework for Training Support Vector Machines
by Khalid Nejjar, Khalid Jebari and Siham Rekiek
Algorithms 2026, 19(1), 70; https://doi.org/10.3390/a19010070 - 13 Jan 2026
Viewed by 92
Abstract
Support Vector Machines (SVMs) are widely used in critical decision-making applications, such as precision agriculture, due to their strong theoretical foundations and their ability to construct an optimal separating hyperplane in high-dimensional spaces. However, the effectiveness of SVMs is highly dependent on the [...] Read more.
Support Vector Machines (SVMs) are widely used in critical decision-making applications, such as precision agriculture, due to their strong theoretical foundations and their ability to construct an optimal separating hyperplane in high-dimensional spaces. However, the effectiveness of SVMs is highly dependent on the efficiency of the optimization algorithm used to solve their underlying dual problem, which is often complex and constrained. Classical solvers, such as Sequential Minimal Optimization (SMO) and Stochastic Gradient Descent (SGD), present inherent limitations: SMO ensures numerical stability but lacks scalability and is sensitive to heuristics, while SGD scales well but suffers from unstable convergence and limited suitability for nonlinear kernels. To address these challenges, this study proposes a novel hybrid optimization framework based on Open Competency Optimization and Particle Swarm Optimization (OCO–PSO) to enhance the training of SVMs. The proposed approach combines the global exploration capability of PSO with the adaptive competency-based learning mechanism of OCO, enabling efficient exploration of the solution space, avoidance of local minima, and strict enforcement of dual constraints on the Lagrange multipliers. Across multiple datasets spanning medical (diabetes), agricultural yield, signal processing (sonar and ionosphere), and imbalanced synthetic data, the proposed OCO-PSO–SVM consistently outperforms classical SVM solvers (SMO and SGD) as well as widely used classifiers, including decision trees and random forests, in terms of accuracy, macro-F1-score, Matthews correlation coefficient (MCC), and ROC-AUC. On the Ionosphere dataset, OCO-PSO achieves an accuracy of 95.71%, an F1-score of 0.954, and an MCC of 0.908, matching the accuracy of random forest while offering superior interpretability through its kernel-based structure. In addition, the proposed method yields a sparser model with only 66 support vectors compared to 71 for standard SVC (a reduction of approximately 7%), while strictly satisfying the dual constraints with a near-zero violation of 1.3×103. Notably, the optimal hyperparameters identified by OCO-PSO (C=2, γ0.062) differ substantially from those obtained via Bayesian optimization for SVC (C=10, γ0.012), indicating that the proposed approach explores alternative yet equally effective regions of the hypothesis space. The statistical significance and robustness of these improvements are confirmed through extensive validation using 1000 bootstrap replications, paired Student’s t-tests, Wilcoxon signed-rank tests, and Holm–Bonferroni correction. These results demonstrate that the proposed metaheuristic hybrid optimization framework constitutes a reliable, interpretable, and scalable alternative for training SVMs in complex and high-dimensional classification tasks. Full article
Show Figures

Figure 1

18 pages, 4726 KB  
Article
Advancing Prostate Cancer Assessment: A Biparametric MRI (T2WI and DWI/ADC)-Based Radiomic Approach to Predict Tumor–Stroma Ratio
by Jiangqin Ma, Xiling Gu, Zhonglin Zhang, Jun Chen, Yunfan Liu, Yang Qiu, Guangyong Ai, Xuxiang Jia, Zhenghao Li, Bo Xiang and Xiaojing He
Diagnostics 2025, 15(21), 2722; https://doi.org/10.3390/diagnostics15212722 - 27 Oct 2025
Viewed by 831
Abstract
Objectives: This study aimed to develop and validate a biparametric MRI (bpMRI)-based radiomics model for the noninvasive prediction of tumor–stroma ratio (TSR) in prostate cancer (PCa). Additionally, we sought to explore lesion distribution patterns in the peripheral zone (PZ) and transition zone (TZ) [...] Read more.
Objectives: This study aimed to develop and validate a biparametric MRI (bpMRI)-based radiomics model for the noninvasive prediction of tumor–stroma ratio (TSR) in prostate cancer (PCa). Additionally, we sought to explore lesion distribution patterns in the peripheral zone (PZ) and transition zone (TZ) to provide deeper insights into the biological behavior of PCa. Methods: This multicenter retrospective study included 223 pathologically confirmed PCa patients, with 146 for training and 39 for internal validation at Center 1, and 38 for external testing at Center 2. All patients underwent preoperative bpMRI (T2WI, DWI acquired with a b-value of 1400 s/mm2, and ADC maps), with TSR histopathologically quantified. Regions of interest (ROIs) were manually segmented on bpMRI images using ITK-SNAP software (version 4.0.1), followed by high-throughput radiomic feature extraction. Redundant features were eliminated via Spearman correlation analysis and least absolute shrinkage and selection operator (LASSO) regression. Five machine learning (ML) classifiers—Logistic Regression (LR), Support Vector Machine (SVM), BernoulliNBBayes, Ridge, and Stochastic Gradient Descent (SGD)—were trained and optimized. Model performance was rigorously evaluated using receiver operating characteristic (ROC) curves and decision curve analysis (DCA). Results: The Ridge demonstrated superior diagnostic performance, achieving AUCs of 0.846, 0.789, and 0.745 in the training, validation, and test cohorts, respectively. Lesion distribution analysis revealed no significant differences between High-TSR and Low-TSR groups (p = 0.867), suggesting that TSR may not be strongly associated with zonal localization. Conclusions: This exploratory study suggests that a bpMRI-based radiomic model holds promise for noninvasive TSR estimation in prostate cancer and may provide complementary insights into tumor aggressiveness beyond conventional pathology. Full article
(This article belongs to the Special Issue Innovations in Medical Imaging for Precision Diagnostics)
Show Figures

Figure 1

23 pages, 3153 KB  
Article
Domain-Specific Acceleration of Gravity Forward Modeling via Hardware–Software Co-Design
by Yong Yang, Daying Sun, Zhiyuan Ma and Wenhua Gu
Micromachines 2025, 16(11), 1215; https://doi.org/10.3390/mi16111215 - 25 Oct 2025
Viewed by 1033
Abstract
The gravity forward modeling algorithm is a compute-intensive method and is widely used in scientific computing, particularly in geophysics, to predict the impact of subsurface structures on surface gravity fields. Traditional implementations rely on CPUs, where performance gains are mainly achieved through algorithmic [...] Read more.
The gravity forward modeling algorithm is a compute-intensive method and is widely used in scientific computing, particularly in geophysics, to predict the impact of subsurface structures on surface gravity fields. Traditional implementations rely on CPUs, where performance gains are mainly achieved through algorithmic optimization. With the rise of domain-specific architectures, FPGA offers a promising platform for acceleration, but faces challenges such as limited programmability and the high cost of nonlinear function implementation. This work proposes an FPGA-based co-processor to accelerate gravity forward modeling. A RISC-V core is integrated with a custom instruction set targeting key computation steps. Tasks are dynamically scheduled and executed on eight fully pipeline processing units, achieving high parallelism while retaining programmability. To address nonlinear operations, we introduce a piecewise linear approximation method optimized via stochastic gradient descent (SGD), significantly reducing resource usage and latency. The design is implemented on the AMD UltraScale+ ZCU102 FPGA (Advanced Micro Devices, Inc. (AMD), Santa Clara, CA, USA) and evaluated across several forward modeling scenarios. At 250 MHz, the system achieves up to 179× speedup over an Intel Xeon 5218R CPU (Intel Corporation, Santa Clara, CA, USA) and improves energy efficiency by 2040×. To the best of our knowledge, this is the first FPGA-based gravity forward modeling accelerate design. Full article
(This article belongs to the Special Issue Advances in Field-Programmable Gate Arrays (FPGAs))
Show Figures

Figure 1

13 pages, 1587 KB  
Article
Glioma Grading by Integrating Radiomic Features from Peritumoral Edema in Fused MRI Images and Automated Machine Learning
by Amir Khorasani
J. Imaging 2025, 11(10), 336; https://doi.org/10.3390/jimaging11100336 - 27 Sep 2025
Cited by 1 | Viewed by 1065
Abstract
We aimed to investigate the utility of peritumoral edema-derived radiomic features from magnetic resonance imaging (MRI) image weights and fused MRI sequences for enhancing the performance of machine learning-based glioma grading. The present study utilized the Multimodal Brain Tumor Segmentation Challenge 2023 (BraTS [...] Read more.
We aimed to investigate the utility of peritumoral edema-derived radiomic features from magnetic resonance imaging (MRI) image weights and fused MRI sequences for enhancing the performance of machine learning-based glioma grading. The present study utilized the Multimodal Brain Tumor Segmentation Challenge 2023 (BraTS 2023) dataset. Laplacian Re-decomposition (LRD) was employed to fuse multimodal MRI sequences. The fused image quality was evaluated using the Entropy, standard deviation (STD), peak signal-to-noise ratio (PSNR), and structural similarity index measure (SSIM) metrics. A comprehensive set of radiomic features was subsequently extracted from peritumoral edema regions using PyRadiomics. The Boruta algorithm was applied for feature selection, and an optimized classification pipeline was developed using the Tree-based Pipeline Optimization Tool (TPOT). Model performance for glioma grade classification was evaluated based on accuracy, precision, recall, F1-score, and area under the curve (AUC) parameters. Analysis of fused image quality metrics confirmed that the LRD method produces high-quality fused images. From 851 radiomic features extracted from peritumoral edema regions, the Boruta algorithm selected different sets of informative features in both standard MRI and fused images. Subsequent TPOT automated machine learning optimization analysis identified a fine-tuned Stochastic Gradient Descent (SGD) classifier, trained on features from T1Gd+FLAIR fused images, as the top-performing model. This model achieved superior performance in glioma grade classification (Accuracy = 0.96, Precision = 1.0, Recall = 0.94, F1-Score = 0.96, AUC = 1.0). Radiomic features derived from peritumoral edema in fused MRI images using the LRD method demonstrated distinct, grade-specific patterns and can be utilized as a non-invasive, accurate, and rapid glioma grade classification method. Full article
(This article belongs to the Topic Machine Learning and Deep Learning in Medical Imaging)
Show Figures

Figure 1

26 pages, 423 KB  
Article
Enhancing Privacy-Preserving Network Trace Synthesis Through Latent Diffusion Models
by Jin-Xi Yu, Yi-Han Xu, Min Hua, Gang Yu and Wen Zhou
Information 2025, 16(8), 686; https://doi.org/10.3390/info16080686 - 12 Aug 2025
Cited by 1 | Viewed by 1578
Abstract
Network trace is a comprehensive record of data packets traversing a computer network, serving as a critical resource for analyzing network behavior. However, in practice, the limited availability of high-quality network traces, coupled with the presence of sensitive information such as IP addresses [...] Read more.
Network trace is a comprehensive record of data packets traversing a computer network, serving as a critical resource for analyzing network behavior. However, in practice, the limited availability of high-quality network traces, coupled with the presence of sensitive information such as IP addresses and MAC addresses, poses significant challenges to advancing network trace analysis. To address these issues, this paper focuses on network trace synthesis in two practical scenarios: (1) data expansion, where users create synthetic traces internally to diversify and enhance existing network trace utility; (2) data release, where synthesized network traces are shared externally. Inspired by the powerful generative capabilities of latent diffusion models (LDMs), this paper introduces NetSynDM, which leverages LDM to address the challenges of network trace synthesis in data expansion scenarios. To address the challenges in the data release scenario, we integrate differential privacy (DP) mechanisms into NetSynDM, introducing DPNetSynDM, which leverages DP Stochastic Gradient Descent (DP-SGD) to update NetSynDM, incorporating privacy-preserving noise throughout the training process. Experiments on five widely used network trace datasets show that our methods outperform prior works. NetSynDM achieves an average 166.1% better performance in fidelity compared to baselines. DPNetSynDM strikes an improved balance between privacy and fidelity, surpassing previous state-of-the-art network trace synthesis method fidelity scores of 18.4% on UGR16 while reducing privacy risk scores by approximately 9.79%. Full article
Show Figures

Figure 1

26 pages, 7587 KB  
Article
PAC–Bayes Guarantees for Data-Adaptive Pairwise Learning
by Sijia Zhou, Yunwen Lei and Ata Kabán
Entropy 2025, 27(8), 845; https://doi.org/10.3390/e27080845 - 8 Aug 2025
Viewed by 1478
Abstract
We study the generalization properties of stochastic optimization methods under adaptive data sampling schemes, focusing on the setting of pairwise learning, which is central to tasks like ranking, metric learning, and AUC maximization. Unlike pointwise learning, pairwise methods must address statistical dependencies between [...] Read more.
We study the generalization properties of stochastic optimization methods under adaptive data sampling schemes, focusing on the setting of pairwise learning, which is central to tasks like ranking, metric learning, and AUC maximization. Unlike pointwise learning, pairwise methods must address statistical dependencies between input pairs—a challenge that existing analyses do not adequately handle when sampling is adaptive. In this work, we extend a general framework that integrates two algorithm-dependent approaches—algorithmic stability and PAC–Bayes analysis for this purpose. Specifically, we examine (1) Pairwise Stochastic Gradient Descent (Pairwise SGD), widely used across machine learning applications, and (2) Pairwise Stochastic Gradient Descent Ascent (Pairwise SGDA), common in adversarial training. Our analysis avoids artificial randomization and leverages the inherent stochasticity of gradient updates instead. Our results yield generalization guarantees of order n1/2 under non-uniform adaptive sampling strategies, covering both smooth and non-smooth convex settings. We believe these findings address a significant gap in the theory of pairwise learning with adaptive sampling. Full article
(This article belongs to the Section Information Theory, Probability and Statistics)
Show Figures

Figure 1

14 pages, 6691 KB  
Article
Remote Sensing Extraction of Damaged Buildings in the Shigatse Earthquake, 2025: A Hybrid YOLO-E and SAM2 Approach
by Zhimin Wu, Chenyao Qu, Wei Wang, Zelang Miao and Huihui Feng
Sensors 2025, 25(14), 4375; https://doi.org/10.3390/s25144375 - 12 Jul 2025
Viewed by 1694
Abstract
In January 2025, a magnitude 6.8 earthquake struck Dingri County, Shigatse, Tibet, causing severe damage. Rapid and precise extraction of damaged buildings is essential for emergency relief and rebuilding efforts. This study proposes an approach integrating YOLO-E (Real-Time Seeing Anything) and the Segment [...] Read more.
In January 2025, a magnitude 6.8 earthquake struck Dingri County, Shigatse, Tibet, causing severe damage. Rapid and precise extraction of damaged buildings is essential for emergency relief and rebuilding efforts. This study proposes an approach integrating YOLO-E (Real-Time Seeing Anything) and the Segment Anything Model 2 (SAM2) to extract damaged buildings with multi-source remote sensing images, including post-earthquake Gaofen-7 imagery (0.80 m), Beijing-3 imagery (0.30 m), and pre-earthquake Google satellite imagery (0.15 m), over the affected region. In this hybrid approach, YOLO-E functions as the preliminary segmentation module for initial segmentation. It leverages its real-time detection and segmentation capability to locate potential damaged building regions and generate coarse segmentation masks rapidly. Subsequently, SAM2 follows as a refinement step, incorporating shapefile information from pre-disaster sources to apply precise, pixel-level segmentation. The dataset used for training contained labeled examples of damaged buildings, and the model optimization was carried out using stochastic gradient descent (SGD), with cross-entropy and mean squared error as the selected loss functions. Upon evaluation, the model reached a precision of 0.840, a recall of 0.855, an F1-score of 0.847, and an IoU of 0.735. It successfully extracted 492 suspected damaged building patches within a radius of 20 km from the earthquake epicenter, clearly showing the distribution characteristics of damaged buildings concentrated in the earthquake fault zone. In summary, this hybrid YOLO-E and SAM2 approach, leveraging multi-source remote sensing imagery, delivers precise and rapid extraction of damaged buildings with a precision of 0.840, recall of 0.855, and IoU of 0.735, effectively supporting targeted earthquake rescue and post-disaster reconstruction efforts in the Dingri County fault zone. Full article
Show Figures

Figure 1

19 pages, 1442 KB  
Article
Hyperspectral Imaging for Enhanced Skin Cancer Classification Using Machine Learning
by Teng-Li Lin, Arvind Mukundan, Riya Karmakar, Praveen Avala, Wen-Yen Chang and Hsiang-Chen Wang
Bioengineering 2025, 12(7), 755; https://doi.org/10.3390/bioengineering12070755 - 11 Jul 2025
Cited by 9 | Viewed by 2063
Abstract
Objective: The classification of skin cancer is very helpful in its early diagnosis and treatment, considering the complexity involved in differentiating AK from BCC and SK. These conditions are generally not easily detectable due to their comparable clinical presentations. Method: This paper presents [...] Read more.
Objective: The classification of skin cancer is very helpful in its early diagnosis and treatment, considering the complexity involved in differentiating AK from BCC and SK. These conditions are generally not easily detectable due to their comparable clinical presentations. Method: This paper presents a new approach to hyperspectral imaging for enhancing the visualization of skin lesions called the Spectrum-Aided Vision Enhancer (SAVE), which has the ability to convert any RGB image into a narrow-band image (NBI) by combining hyperspectral imaging (HSI) to increase the contrast of the area of the cancerous lesions when compared with the normal tissue, thereby increasing the accuracy of classification. The current study investigates the use of ten different machine learning algorithms for the purpose of classification of AK, BCC, and SK, including convolutional neural network (CNN), random forest (RF), you only look once (YOLO) version 8, support vector machine (SVM), ResNet50, MobileNetV2, Logistic Regression, SVM with stochastic gradient descent (SGD) Classifier, SVM with logarithmic (LOG) Classifier and SVM- Polynomial Classifier, in assessing the capability of the system to differentiate AK from BCC and SK with heightened accuracy. Results: The results demonstrated that SAVE enhanced classification performance and increased its accuracy, sensitivity, and specificity compared to a traditional RGB imaging approach. Conclusions: This advanced method offers dermatologists a tool for early and accurate diagnosis, reducing the likelihood of misclassification and improving patient outcomes. Full article
Show Figures

Figure 1

16 pages, 1400 KB  
Article
An RMSprop-Incorporated Latent Factorization of Tensor Model for Random Missing Data Imputation in Structural Health Monitoring
by Jingjing Yang
Algorithms 2025, 18(6), 351; https://doi.org/10.3390/a18060351 - 6 Jun 2025
Viewed by 1344
Abstract
In structural health monitoring (SHM), ensuring data completeness is critical for enhancing the accuracy and reliability of structural condition assessments. SHM data are prone to random missing values due to signal interference or connectivity issues, making precise data imputation essential. A latent factorization [...] Read more.
In structural health monitoring (SHM), ensuring data completeness is critical for enhancing the accuracy and reliability of structural condition assessments. SHM data are prone to random missing values due to signal interference or connectivity issues, making precise data imputation essential. A latent factorization of tensor (LFT)-based method has proven effective for such problems, with optimization typically achieved via stochastic gradient descent (SGD). However, SGD-based LFT models and other imputation methods exhibit significant sensitivity to learning rates and slow tail-end convergence. To address these limitations, this study proposes an RMSprop-incorporated latent factorization of tensor (RLFT) model, which integrates an adaptive learning rate mechanism to dynamically adjust step sizes based on gradient magnitudes. Experimental validation on a scaled bridge accelerometer dataset demonstrates that RLFT achieves faster convergence and higher imputation accuracy compared to state-of-the-art models including SGD-based LFT and the long short-term memory (LSTM) network, with improvements of at least 10% in both imputation accuracy and convergence rate, offering a more efficient and reliable solution for missing data handling in SHM. Full article
Show Figures

Figure 1

21 pages, 7991 KB  
Article
Machine Learning–Based Calibration and Performance Evaluation of Low-Cost Internet of Things Air Quality Sensors
by Mehmet Taştan
Sensors 2025, 25(10), 3183; https://doi.org/10.3390/s25103183 - 19 May 2025
Cited by 5 | Viewed by 4411
Abstract
Low-cost air quality sensors (LCSs) are increasingly being used in environmental monitoring due to their affordability and portability. However, their sensitivity to environmental factors can lead to measurement inaccuracies, necessitating effective calibration methods to enhance their reliability. In this study, an Internet of [...] Read more.
Low-cost air quality sensors (LCSs) are increasingly being used in environmental monitoring due to their affordability and portability. However, their sensitivity to environmental factors can lead to measurement inaccuracies, necessitating effective calibration methods to enhance their reliability. In this study, an Internet of Things (IoT)-based air quality monitoring system was developed and tested using the most commonly preferred sensor types for air quality measurement: fine particulate matter (PM2.5), carbon dioxide (CO2), temperature, and humidity sensors. To improve sensor accuracy, eight different machine learning (ML) algorithms were applied: Decision Tree (DT), Linear Regression (LR), Random Forest (RF), k-Nearest Neighbors (kNN), AdaBoost (AB), Gradient Boosting (GB), Support Vector Machines (SVM), and Stochastic Gradient Descent (SGD). Sensor performance was evaluated by comparing measurements with a reference device, and the best-performing ML model was determined for each sensor. The results indicate that GB and kNN achieved the highest accuracy. For CO2 sensor calibration, GB achieved R2 = 0.970, RMSE = 0.442, and MAE = 0.282, providing the lowest error rates. For the PM2.5 sensor, kNN delivered the most successful results, with R2 = 0.970, RMSE = 2.123, and MAE = 0.842. Additionally, for temperature and humidity sensors, GB demonstrated the highest accuracy with the lowest error values (R2 = 0.976, RMSE = 2.284). These findings demonstrate that, by identifying suitable ML methods, ML-based calibration techniques can significantly enhance the accuracy of LCSs. Consequently, they offer a viable and cost-effective alternative to traditional high-cost air quality monitoring systems. Future studies should focus on long-term data collection, testing under diverse environmental conditions, and integrating additional sensor types to further advance this field. Full article
(This article belongs to the Special Issue Intelligent Sensor Calibration: Techniques, Devices and Methodologies)
Show Figures

Figure 1

21 pages, 16775 KB  
Article
Non-Iterative Phase-Only Hologram Generation via Stochastic Gradient Descent Optimization
by Alejandro Velez-Zea and John Fredy Barrera-Ramírez
Photonics 2025, 12(5), 500; https://doi.org/10.3390/photonics12050500 - 16 May 2025
Cited by 1 | Viewed by 1571
Abstract
In this work, we explored, for the first time, to the best of our knowledge, the potential of stochastic gradient descent (SGD) to optimize random phase functions for application in non-iterative phase-only hologram generation. We defined and evaluated four loss functions based on [...] Read more.
In this work, we explored, for the first time, to the best of our knowledge, the potential of stochastic gradient descent (SGD) to optimize random phase functions for application in non-iterative phase-only hologram generation. We defined and evaluated four loss functions based on common image quality metrics and compared the performance of SGD-optimized random phases with those generated using Gerchberg–Saxton (GS) optimization. The quality of the reconstructed holograms was assessed through numerical simulations, considering both accuracy and computational efficiency. Our results demonstrate that SGD-based optimization can produce higher-quality phase holograms for low-contrast target scenes and presents nearly identical performance to GS-optimized random phases for high-contrast targets. Experimental validation confirmed the practical feasibility of the proposed method and its potential as a flexible alternative to conventional GS-based optimization. Full article
(This article belongs to the Special Issue Advances in Optical Imaging)
Show Figures

Figure 1

21 pages, 26641 KB  
Article
A CNN-Based Method for Quantitative Assessment of Steel Microstructures in Welded Zones
by Cássio Danelon de Almeida, Thales Tozatto Filgueiras, Moisés Luiz Lagares, Bruno da Silva Macêdo, Camila Martins Saporetti, Matteo Bodini and Leonardo Goliatt
Fibers 2025, 13(5), 66; https://doi.org/10.3390/fib13050066 - 15 May 2025
Cited by 1 | Viewed by 2703
Abstract
The mechanical performance of metallic components is intrinsically linked to their microstructural features. However, the manual quantification of microconstituents in metallographic images remains a time-consuming and subjective task, often requiring over 15 min per image by a trained expert. To address this limitation, [...] Read more.
The mechanical performance of metallic components is intrinsically linked to their microstructural features. However, the manual quantification of microconstituents in metallographic images remains a time-consuming and subjective task, often requiring over 15 min per image by a trained expert. To address this limitation, this study proposes an automated approach for quantifying the microstructural constituents from low-carbon steel welded zone images using convolutional neural networks (CNNs). A dataset of 210 micrographs was expanded to 720 samples through data augmentation to improve model generalization. Two architectures (AlexNet and VGG16) were trained from scratch, while three pre-trained models (VGG19, InceptionV3, and Xception) were fine-tuned. Among these, VGG19 optimized with stochastic gradient descent (SGD) achieved the best predictive performance, with an R2 of 0.838, MAE of 5.01%, and RMSE of 6.88%. The results confirm the effectiveness of CNNs for reliable and efficient microstructure quantification, offering a significant contribution to computational metallography. Full article
Show Figures

Figure 1

10 pages, 332 KB  
Proceeding Paper
Optimizing Brain Tumor Classification: Integrating Deep Learning and Machine Learning with Hyperparameter Tuning
by Vijaya Kumar Velpula, Kamireddy Rasool Reddy, K. Naga Prakash, K. Prasanthi Jasmine and Vadlamudi Jyothi Sri
Eng. Proc. 2025, 87(1), 64; https://doi.org/10.3390/engproc2025087064 - 12 May 2025
Viewed by 2051
Abstract
Brain tumors significantly impact global health and pose serious challenges for accurate diagnosis due to their diverse nature and complex characteristics. Effective diagnosis and classification are essential for selecting the best treatment strategies and forecasting patient outcomes. Currently, histopathological examination of biopsy samples [...] Read more.
Brain tumors significantly impact global health and pose serious challenges for accurate diagnosis due to their diverse nature and complex characteristics. Effective diagnosis and classification are essential for selecting the best treatment strategies and forecasting patient outcomes. Currently, histopathological examination of biopsy samples is the standard method for brain tumor identification and classification. However, this method is invasive, time-consuming, and prone to human error. To address these limitations, a fully automated approach is proposed for brain tumor classification. Recent advancements in deep learning, particularly convolutional neural networks (CNNs), have shown promise in improving the accuracy and efficiency of tumor detection from magnetic resonance imaging (MRI) scans. In response, a model was developed that integrates machine learning (ML) and deep learning (DL) techniques. The process began by splitting the data into training, testing, and validation sets. Images were then resized and cropped to enhance model quality and efficiency. Relevant texture features were extracted using a modified Visual Geometry Group (VGG) architecture. These features were fed into various supervised ML models, including support vector machine (SVM), k-nearest neighbors (KNN), logistic regression (LR), stochastic gradient descent (SGD), random forest (RF), and AdaBoost, with GridSearchCV used for hyperparameter tuning. The model’s performance was evaluated using key metrics such as accuracy, precision, recall, F1-score, and specificity. Experimental results demonstrate that the proposed approach offers a robust and automated solution for brain tumor classification, achieving the highest accuracy of 94.02% with VGG19 and 96.30% with VGG16. This model can significantly assist healthcare professionals in early tumor detection and in improving diagnostic accuracy. Full article
(This article belongs to the Proceedings of The 5th International Electronic Conference on Applied Sciences)
Show Figures

Figure 1

35 pages, 15625 KB  
Article
Influence of Optimization Algorithms and Computational Complexity on Concrete Compressive Strength Prediction Machine Learning Models for Concrete Mix Design
by Patryk Ziolkowski
Materials 2025, 18(6), 1386; https://doi.org/10.3390/ma18061386 - 20 Mar 2025
Cited by 5 | Viewed by 1403
Abstract
The proper design of concrete mixtures is a critical task in concrete technology, where optimal strength, eco-friendliness, and production efficiency are increasingly demanded. While traditional analytical methods, such as the Three Equations Method, offer foundational approaches to mix design, they often fall short [...] Read more.
The proper design of concrete mixtures is a critical task in concrete technology, where optimal strength, eco-friendliness, and production efficiency are increasingly demanded. While traditional analytical methods, such as the Three Equations Method, offer foundational approaches to mix design, they often fall short in handling the complexity of modern concrete technology. Machine learning-based models have demonstrated notable efficacy in predicting concrete compressive strength, addressing the limitations of conventional methods. This study builds on previous research by investigating not only the impact of computational complexity on the predictive performance of machine learning models but also the influence of different optimization algorithms. The study evaluates the effectiveness of three optimization techniques: the Quasi-Newton Method (QNM), the Adaptive Moment Estimation (ADAM) algorithm, and Stochastic Gradient Descent (SGD). A total of forty-five deep neural network models of varying computational complexity were trained and tested using a comprehensive database of concrete mix designs and their corresponding compressive strength test results. The findings reveal a significant interaction between optimization algorithms and model complexity in enhancing prediction accuracy. Models utilizing the QNM algorithm outperformed those using the ADAM and SGD in terms of error reduction (SSE, MSE, RMSE, NSE, and ME) and increased coefficient of determination (R2). These insights contribute to the development of more accurate and efficient AI-driven methods in concrete mix design, promoting the advancement of concrete technology and the potential for future research in this domain. Full article
(This article belongs to the Collection Concrete and Building Materials)
Show Figures

Figure 1

13 pages, 1559 KB  
Article
Evaluation of Convolutional Neural Network-Based Posture Identification Model of Older Adults: From Silhouette of Sagittal Photographs
by Naoki Sugiyama, Yoshihiro Kai, Hitoshi Koda, Toru Morihara and Noriyuki Kida
Geriatrics 2025, 10(2), 49; https://doi.org/10.3390/geriatrics10020049 - 19 Mar 2025
Viewed by 1282
Abstract
Background/Objectives: Posture is a significant indicator of health status in older adults. This study aimed to develop an automatic posture assessment tool based on sagittal photographs by validating recognition models using convolutional neural networks. Methods: A total of 9140 images were [...] Read more.
Background/Objectives: Posture is a significant indicator of health status in older adults. This study aimed to develop an automatic posture assessment tool based on sagittal photographs by validating recognition models using convolutional neural networks. Methods: A total of 9140 images were collected with data augmentation, and each image was labeled as either Ideal or Non-Ideal posture by physical therapists. The hidden and output layers of the models remained unchanged, while the loss function and optimizer were varied to construct four different model configurations: mean squared error and Adam (MSE & Adam), mean squared error and stochastic gradient descent (MSE & SGD), binary cross-entropy and Adam (BCE & Adam), and binary cross-entropy and stochastic gradient descent (BCE & SGD). Results: All four models demonstrated an improved accuracy in both the training and validation phases. However, the two BCE models exhibited divergence in validation loss, suggesting overfitting. Conversely, the two MSE models showed stability during learning. Therefore, we focused on the MSE models and evaluated their reliability using sensitivity, specificity, and Prevalence-Adjusted Bias-Adjusted Kappa (PABAK) based on the model’s output and correct label. Sensitivity and specificity were 85% and 84% for MSE & Adam and 67% and 77% for MSE & SGD, respectively. Moreover, PABAK values for agreement with the correct label were 0.69 and 0.43 for MSE & Adam and MSE & SGD, respectively. Conclusions: Our findings indicate that the MSE & Adam model, in particular, can serve as a useful tool for screening inspections. Full article
Show Figures

Figure 1

Back to TopTop