Next Article in Journal
Digital Serious Games for Undergraduate Nursing Education: A Review of Serious Games Key Design Characteristics and Gamification Elements
Previous Article in Journal
Advances in Facial Micro-Expression Detection and Recognition: A Comprehensive Review
Previous Article in Special Issue
A Novel Approach to the Collatz Conjecture with Petri Nets
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

A Comparison of Different Transformer Models for Time Series Prediction

by
Emek Utku Capoglu
and
Aboozar Taherkhani
*
School of Computer Science and Informatics, De Montfort University, Leicester LE1 9BH, UK
*
Author to whom correspondence should be addressed.
Information 2025, 16(10), 878; https://doi.org/10.3390/info16100878 (registering DOI)
Submission received: 8 August 2025 / Revised: 25 September 2025 / Accepted: 30 September 2025 / Published: 9 October 2025
(This article belongs to the Special Issue Intelligent Information Technology, 2nd Edition)

Abstract

Accurate estimation of the Remaining Useful Life (RUL) of lithium-ion batteries is essential for enhancing the reliability and efficiency of energy storage systems. This study explores custom deep learning models to predict RUL using a dataset from the Hawaii Natural Energy Institute (HNEI). Three approaches are investigated: an Encoder-only Transformer model, its enhancement with SimSiam transfer learning, and a CNN–Encoder hybrid model. These models leverage advanced mechanisms such as multi-head attention, robust feedforward networks, and self-supervised learning to capture complex degradation patterns in the data. Rigorous preprocessing and optimisation ensure optimal performance, reducing key metrics such as mean squared error (MSE) and mean absolute error (MAE). Experimental results demonstrated that Transformer–CNN with Noise Augmentation outperforms other methods, highlighting its potential for battery health monitoring and predictive maintenance.
Keywords: hyperparameter tuning; mask augmentation; noise augmentation; time series; transformers hyperparameter tuning; mask augmentation; noise augmentation; time series; transformers

Share and Cite

MDPI and ACS Style

Capoglu, E.U.; Taherkhani, A. A Comparison of Different Transformer Models for Time Series Prediction. Information 2025, 16, 878. https://doi.org/10.3390/info16100878

AMA Style

Capoglu EU, Taherkhani A. A Comparison of Different Transformer Models for Time Series Prediction. Information. 2025; 16(10):878. https://doi.org/10.3390/info16100878

Chicago/Turabian Style

Capoglu, Emek Utku, and Aboozar Taherkhani. 2025. "A Comparison of Different Transformer Models for Time Series Prediction" Information 16, no. 10: 878. https://doi.org/10.3390/info16100878

APA Style

Capoglu, E. U., & Taherkhani, A. (2025). A Comparison of Different Transformer Models for Time Series Prediction. Information, 16(10), 878. https://doi.org/10.3390/info16100878

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop