1. Introduction
The stock price of PT Indika Energy Tbk (INDY) reflects the dynamic nature of Indonesia’s energy sector. Its movements are highly influenced by global market conditions, national energy policies, and various geopolitical factors [
1]. As such, INDY stock serves as a key indicator in the analysis of the energy market [
1]. The volatility exhibited by INDY’s stock prices highlights the unpredictable and fast-paced nature of the sector [
1]. This volatility underlines the importance of accurate stock price forecasting, which plays a crucial role for companies, investors, and policymakers [
2,
3,
4].
The growing complexity and nonlinearity of financial time series data have prompted researchers to explore more advanced modeling techniques. One promising approach is the development of hybrid models, such as the CNN-BiGRU-AM architecture [
5,
6,
7,
8]. This model integrates the strengths of three powerful components: Convolutional Neural Networks (CNNs), Bidirectional Gated Recurrent Units (BiGRUs), and the Attention Mechanism [
5,
6,
7,
8]. The CNN is particularly effective in extracting local features from historical stock data and in reducing irrelevant noise [
5,
9,
10,
11]. The BiGRU, on the other hand, is capable of capturing long-term dependencies in both the forward and backward directions [
12,
13,
14,
15]. Meanwhile, the Attention Mechanism allows the model to focus more precisely on the most relevant segments of historical data during the prediction process [
7,
16,
17,
18].
To further enhance the modeling of financial time series data, recent advances also explore neurodynamic-based optimization frameworks. For instance, the work by Leung et al. [
19] presents a collaborative neurodynamic approach to tackle minimax and biobjective portfolio selection problems. Their framework optimizes both return and risk under nonlinear and constrained environments—conditions often encountered in stock market forecasting. While our work adopts a hybrid deep learning framework (CNN-BiGRU-AM) rather than a neurodynamic scheme, both approaches share a common objective of improving predictive robustness in high-volatility, high-noise financial domains. This connection affirms the need for further exploration of hybrid and adaptive architectures that account for multi-dimensional objectives and optimization landscapes in finance.
The integration of deep learning and attention-based mechanisms makes the hybrid model more adaptive to shifts in market patterns [
7,
9,
15,
16,
18]. Unlike traditional models that often struggle with nonstationary behavior, the CNN-BiGRU-AM architecture offers increased flexibility in adjusting to sudden market changes [
8,
20,
21]. This adaptability is critical in stock price forecasting, where patterns are not only complex but also subject to rapid external shocks [
22,
23,
24,
25]. By assigning higher weights to the most informative historical features, the model enhances its learning process and improves prediction accuracy [
9,
21,
26]. As a result, the forecast outcomes are more aligned with real-world market conditions [
27,
28,
29].
This research provides practical value by offering a tool that supports strategic decision making. Accurate stock price predictions can reduce the uncertainty faced by stakeholders in the energy and financial sectors [
30,
31,
32]. Companies can plan their financial strategies more effectively, investors can optimize their portfolios, and shareholders can make informed decisions [
30,
31,
32]. Therefore, the proposed model is not only a theoretical contribution but also a practical innovation with high relevance to real-world applications.
Despite the advances in deep learning, conventional models and standard Recurrent Neural Networks (RNNs) still present significant limitations [
6,
33,
34]. One of the most persistent challenges is the vanishing gradient problem, which hinders the model’s ability to retain information over long sequences [
35,
36,
37]. This issue becomes particularly problematic in financial data, where temporal dependencies often span extended periods [
38,
39,
40]. Traditional models, while useful in capturing short-term trends, typically fail to recognize long-term structural patterns [
28,
29,
33]. This limitation calls for more sophisticated and adaptive modeling approaches that can preserve important historical information.
To date, there is still a lack of studies that focus specifically on forecasting INDY’s stock price using complex hybrid architectures. Most existing research tends to concentrate on large-cap stocks or general market indices, leaving mid-cap and sector-specific stocks underexplored. INDY, as a representative of Indonesia’s energy sector, possesses unique characteristics that merit in-depth analysis [
1]. The energy sector is highly sensitive to external influences, including global commodity prices, regulatory changes, and geopolitical events [
41,
42]. This sensitivity further amplifies the need for a focused and customized prediction model.
In recent years, the INDY stock has demonstrated significant instability. Yet, few studies attempt to analyze this behavior using nonlinear modeling approaches. Linear models often fail to capture extreme price fluctuations and regime shifts inherent in stock data. Given that INDY stock exhibits patterns of high variance and nonlinearity, it is more appropriate to approach its forecasting using nonlinear and adaptive methods. This study addresses this gap by employing a hybrid deep learning model tailored to the stock’s unique characteristics
Currently, there are limited machine learning solutions that effectively handle high-volatility stock data, especially in emerging markets like Indonesia. Many models do not incorporate long-term temporal dependencies, which are critical in understanding trends in the energy sector. Furthermore, most forecasting models are not designed to respond dynamically to evolving market conditions. As a result, there is a pressing need for the development of more reliable prediction tools that combine deep learning with temporal and attention-based features.
The demand for accurate forecasting of INDY’s stock price continues to grow. The uncertainty in the energy sector requires models that are capable of adapting quickly to shifting patterns and external influences. Stakeholders need predictive insights that are both timely and accurate to make sound investment and operational decisions. A robust model can help anticipate risks and seize market opportunities, thereby enhancing financial resilience and strategic positioning.
This study aims to construct a hybrid CNN-BiGRU-AM model specifically designed to predict INDY’s stock prices with a high degree of accuracy. The model was trained using historical data ranging from 2019 to 2025, enabling it to learn from various market cycles and anomalies. The ultimate goal is to produce a forecasting system that is not only precise but also efficient in capturing the intricate behaviors of financial time series data. Through this research, we seek to contribute both to academic knowledge and to the practical field of investment analytics.
This study contributes to the literature by presenting a tailored hybrid architecture— CNN-BiGRU-AM—specifically designed to address the forecasting challenges of mid-cap stocks in emerging markets, with INDY stock as a representative case. Unlike conventional models, this approach effectively captures nonlinear dependencies and adapts to regime changes commonly observed in the energy sector. Methodologically, the study introduces a sequence modeling pipeline that combines convolutional feature extraction, bidirectional memory retention, and attention-based weighting. This configuration not only improves predictive accuracy but also enhances the interpretability of model focus on temporally significant data. The proposed framework thus bridges the gap between theoretical advancements in deep learning and their practical applicability in dynamic financial forecasting.
2. Materials and Methods
2.1. Data
This study used secondary data consisting of daily stock prices (on working days) of INDY from 28 August 2019 to 24 February 2025. The data were obtained from the Yahoo Finance website and comprised a total of 1333 observations. The dataset was divided into three subsets, training, validation, and testing data, with a ratio of 60:20:20.
The selection of 28 August 2019 as the starting point was based on the availability of complete, consistent daily trading data from that date onward. This range enabled the model to capture both the pre-pandemic and post-pandemic market conditions, particularly those influenced by global energy demand shocks and economic disruptions. Including the early months preceding COVID-19 ensured that the model learned from baseline price behavior while still reflecting the volatility that emerged during and after the pandemic. This period was also sufficient to meet the data volume requirements for training deep learning architectures effectively.
2.2. Data Pre-Processing
In this study, the dataset underwent a normalization procedure to comply with the modeling standards commonly required in neural-network-based forecasting. Normalization functions as a transformation mechanism designed to reduce substantial differences in the range of feature values [
43] by mapping each data point into a standardized interval between 0 and 1 [
44]. This process facilitates a more efficient and stable training phase for deep learning models.
The normalization approach adopted in this research is the MinMax Normalization method, also widely recognized as the MinMax Scaler. The transformation is mathematically formulated as expressed in Equation (
1):
In this notation, indicates the normalized value of the observation, which is constrained within the interval . The symbol denotes the original, unnormalized observation, while and represent the maximum and minimum values within the corresponding feature, respectively. The use of a breve mark over z highlights that the variable has been transformed from its original state, aiding in distinguishing the normalized values throughout the analysis.
This formulation ensures that each input feature contributes proportionally during the model training process and avoids the domination of features with inherently larger scales. As a result, it enhances both learning stability and convergence speed in deep learning architectures.
2.3. CNN-BiGRU-AM Model Design
The design process of the CNN-BiGRU-AM model involved several layers, beginning with convolutional and pooling operations using a one-dimensional Convolutional Neural Network (1D-CNN), followed by the BiGRU sequence processing layer, the weight allocation using the Attention Mechanism (AM), and final integration through a fully connected layer.
2.3.1. Convolutional Layer (Conv1D)
The
Conv1D layer functions as a feature extractor by performing convolution operations on the sequential input data, resulting in a set of representations known as feature maps that are forwarded to the subsequent layers. Mathematically, the convolutional output at time step
and convolutional filter index
, represented as
, are formulated as shown in Equation (
2) [
45]:
In Equation (
2),
denotes the activation output (feature map) at temporal index
for the
-th convolutional filter. The term
refers to the pre-processed input data (i.e., normalized and/or imputed) at the shifted time index
. The matrix
represents the learnable weights associated with the
-th kernel at the
-th position of the receptive field, while
is the corresponding bias term, typically initialized to zero for simplification in predictive modeling. The index
iterates across the window length
, capturing localized temporal dependencies in the input sequence. The ReLU function is applied to introduce nonlinearity and enhance the model’s ability to represent complex patterns in the data. This convolutional mechanism is particularly effective in isolating important short-term trends and local dynamics from time series input such as stock prices.
2.3.2. Pooling Layer
The pooling layer is designed to extract the most significant and informative characteristics from the feature maps produced by the preceding convolutional process while concurrently reducing their spatial resolution. This operation enhances computational efficiency and reduces the likelihood of overfitting by limiting the number of trainable parameters [
46]. In this study, an average pooling strategy was utilized to preserve the general pattern of localized features, especially in sequential data like stock prices.
Mathematically, the result of the pooling process for the
-th region, denoted by
, is described by Equation (
3):
In this formulation, represents the number of elements within the pooling segment , and corresponds to the activation value at position in that region. This averaging operation results in a downsampled yet information-rich representation of the original feature map, effectively summarizing the critical temporal or spatial signals needed for the next modeling stages, such as the bidirectional GRU or attention mechanism.
2.3.3. BiGRU Layer
The output generated from the pooling layer is subsequently directed to the Bidirectional Gated Recurrent Unit (BiGRU) architecture, which is responsible for extracting temporal dependencies in both chronological and reverse chronological directions. This dual-path mechanism enhances the model’s ability to learn complex sequential behavior in time series data, such as stock price fluctuations.
The BiGRU executes two distinct operations: a forward GRU denoted as
and a backward GRU expressed as
. These are later combined to form the bidirectional hidden representation,
. In each GRU path, the process begins with the computation of the reset gate
, followed by the update gate
, and then the generation of a candidate hidden state
. These steps are mathematically defined in Equations (
4)–(
7):
In these expressions, denotes the input vector at time step , and are learnable weight matrices associated with the input and previous hidden state, respectively, and represents the bias term for each gate. The operator ⊙ denotes element-wise multiplication.
The construction of the bidirectional hidden state continues with the forward and backward computations defined as follows:
In Equation (
10), the symbol ‖ represents the concatenation of the forward and backward hidden states. This final representation
encapsulates context from both temporal directions and is used as input to the subsequent attention mechanism or output layer.
2.3.4. Application of Attention Mechanism
The Attention Mechanism was applied to the latent representations generated by the preceding BiGRU layer, aiming to emphasize information at each time step based on its relative significance in producing the final predictive outcome. In this framework, the Bahdanau attention—also known as additive attention—was adopted due to its efficacy in managing nonlinear sequential relationships. The computation involved is detailed in Equations (
11)–(
13). The context vector, denoted by
, was computed as a weighted aggregation of temporal hidden representations
, where the weights
quantify the attention level assigned to each temporal unit, as shown in Equation (
11):
The attention weight
was determined by normalizing the attention energy score
using a softmax function to ensure probabilistic interpretation over the time steps, which is expressed as
The attention energy
was then computed using a combination of the decoder’s previous hidden representation
and the encoder output
passed through a nonlinear transformation involving a hyperbolic tangent function, as shown in Equation (
13):
In these equations, represents the final context vector capturing the global attention-weighted information; denotes the encoder’s latent vector at time step ; indicates the learned importance of each encoder output; is the decoder’s hidden state from the previous step; and are the weight matrices corresponding to decoder and encoder states, respectively; is the bias term; and is a learned parameter vector that projects the combined transformation into a scalar attention score.
2.3.5. Fully Connected Layer
In the final processing stage of the hybrid CNN-BiGRU-AM architecture, the aggregated context vector derived from the attention mechanism was forwarded as the input to the dense (fully connected) layer. This transformation was designed to project the high-level temporal representations into the target output space. Mathematically, the output of the dense layer at time step
, denoted as
, is defined in Equation (
14):
In this expression, represents the final activated output of the dense layer; is the trainable weight matrix associated with the fully connected transformation; is the context vector summarizing information from all time steps as produced by the attention mechanism; refers to the bias parameter specific to the dense layer; and denotes the activation function—commonly a sigmoid function—applied element-wise to introduce nonlinearity into the final prediction. This operation ensured the model’s ability to map complex, sequential dependencies into the desired predictive output form using the comprehensive information encoded in .
2.4. Model Evaluation Metrics
To evaluate the predictive performance of the proposed hybrid model, a model validation process was carried out using a widely accepted error metric in time series forecasting [
47]. Among various evaluation criteria such as the Root Mean Squared Error (RMSE) and Mean Absolute Error (MAE), the Mean Absolute Percentage Error (MAPE) was selected as the primary indicator of model accuracy in this study, owing to its interpretability in percentage terms and its sensitivity to relative errors. The MAPE was computed based on the following formulation:
In this expression, denotes the total number of time instances observed in the evaluation set; refers to the ground truth value at the j-th point in time, and indicates the corresponding predicted value produced by the model. Notably, the symbol is used to emphasize that the predicted output is derived from the imputed and learned sequences rather than raw observations. The use of the MAPE enabled the assessment of relative prediction errors across all forecasting points, facilitating a comprehensive understanding of the model’s generalization ability in capturing temporal patterns.
3. Results
3.1. Determination of Hyperparameter Tuning Intervals
This study utilized the
Optuna library in Python 3 for conducting the
hyperparameter tuning process. The range of values for each hyperparameter considered during the tuning process is shown in
Table 1.
The selection of the most optimal hyperparameter combination from the specified intervals was based on the minimum value of the loss function. A time step of 5 was used in this study, as stock price movements were highly influenced by short-term factors. After performing the tuning, the combination of hyperparameters that yielded the
lowest training loss (0.0004623) and
testing loss (0.0001786) is presented in
Table 2.
This combination was then used for training the CNN-BiGRU-AM model.
3.2. Model Training
The CNN-BiGRU-AM model was trained to achieve optimal performance in forecasting the stock prices of INDY. The training process was conducted in a sequential manner using a batch size of 62, meaning that a batch of data was processed only after the previous one was completed.
During training, the model was run for a total of 1000 epochs. However, based on the training and validation loss graph shown in
Figure 1, the training stopped early after
257 iterations, when no further improvement was observed. At this point, the
training loss reached
0.00045, while the
validation loss settled at
0.00035.
The close resemblance between the training and validation loss trends indicated that the model generalized well to unseen data and was not overfitting. This suggests that the model has strong potential to perform robust predictions even on data with different characteristics.
3.3. Model Visualization
The results from the CNN-BiGRU-AM model training were visualized to compare the actual and predicted stock prices. These visualizations are presented in
Figure 2,
Figure 3,
Figure 4 and
Figure 5.
The graph for the validation data (
Figure 3) demonstrates that the predicted curve closely followed the actual stock price values, although there were a few deviations during periods of significant change, such as between March and May 2024. Similarly, in the testing data (
Figure 4), the predicted prices successfully captured the overall trend of the actual prices, with visible alignment even during large fluctuations, such as those observed around August to September 2024.
Overall, the CNN-BiGRU-AM model was able to forecast stock prices with high fidelity, even during periods of considerable market volatility. This performance was particularly notable in the context of major external factors such as the COVID-19 pandemic and the Russia–Ukraine conflict, which heavily influenced global coal prices during the 2021–2022 period.
3.4. Model Evaluation
The model was evaluated using the Mean Absolute Percentage Error (MAPE), a common metric that measures the average magnitude of forecasting errors as a percentage of actual values. The evaluation results for training, validation, testing, and overall data are summarized in
Table 3.
As observed in
Table 3, all MAPE values were below 3%, indicating that the model achieved excellent prediction accuracy across all subsets of data. These results suggest that the CNN-BiGRU-AM model is well suited for forecasting stock prices with high precision and reliability.
3.5. Benchmarking with Traditional Models and Related Works
To evaluate the relative performance of the proposed CNN-BiGRU-AM model, a benchmarking comparison was conducted against two widely used classical forecasting approaches: the AutoRegressive Integrated Moving Average (ARIMA) model and Exponential Smoothing. Both models were applied to the same normalized INDY stock price dataset used for our deep learning model. The optimal ARIMA configuration was determined using the Akaike Information Criterion (AIC), resulting in ARIMA(2,1,0). For Exponential Smoothing, the Holt–Winters additive method was employed.
Table 4 summarizes the comparative results in terms of the Mean Absolute Percentage Error (MAPE). The CNN-BiGRU-AM model outperformed the traditional models, achieving a significantly lower MAPE value. This demonstrates its superior capacity in capturing the nonlinear and dynamic patterns present in the INDY time series data.
These results confirm that the hybrid deep learning architecture not only performed better than classical models but also surpassed the results reported in recent studies. The integration of CNN, BiGRU, and Attention Mechanism components contributes significantly to this improvement by capturing short-term features, long-range dependencies, and salient temporal information simultaneously.
3.6. Error Direction Analysis
The forecasted prices for the five-day prediction window, as presented in
Table 5, consistently exceeded the actual prices observed in the market. This pattern may initially suggest a directional bias in the model’s predictions. All five forecasts showed overestimated values, with deviations ranging between 56 and 146 units from the corresponding actual prices. Although this overestimation is evident in the short-term forecast window, it is important to examine whether such a bias persists across the broader test set.
To address this, an extended error direction analysis was performed over the entire test dataset, as presented in the next subsection. The results reveal that approximately 53.2% of the predictions were overestimated, while 46.8% were underestimated. This indicates a mild upward tendency in forecast outputs but does not reflect a dominant systematic bias. This observation highlights the need for continuous recalibration, particularly in volatile or bearish market conditions.
Analysis of Prediction Error Direction
To evaluate whether the model’s predictions exhibit directional bias, we analyzed the sign of prediction errors on the entire test set. Let the prediction error be defined as
. The results show that out of all predictions, 53.2% were overestimations (
), while 46.8% were underestimations (
). This distribution suggests a slight upward bias, but the imbalance is not substantial enough to conclude the presence of a systematic forecasting error. The error direction histogram is presented in
Figure 6, which confirms the near-balanced error spread around zero.
Figure 6 displays the histogram of prediction errors (
) generated by the proposed CNN-BiGRU-AM model across the full test dataset. The distribution illustrates both positive and negative deviations from actual stock prices, allowing for a clearer assessment of model bias. The red dashed vertical line at zero represents a perfect prediction. The relatively symmetric distribution around this zero point suggests that the model produces both overestimations and underestimations in comparable proportions. This indicates an absence of a dominant systematic forecasting bias. The histogram also confirms that the majority of prediction errors fall within a narrow range, further supporting the model’s robustness and generalization capability.
3.7. Forecasting
After completing the training and evaluation stages, the final step involved forecasting the INDY stock price for the next five working days based on the last available data from the testing period. The results of this short-term forecast are shown in
Table 5 and visualized in
Figure 7.
The forecasted prices showed a gradual downward trend, which closely mirrored the actual movement of the stock prices during the same period. The MAPE for this 5-day forecast was calculated at 8.03%, which was still considered acceptable given the highly volatile nature of stock markets.
These forecast results could serve as valuable references for corporate decision making, particularly in the domains of operational planning, investment strategies, and financial policy formulation. Moreover, for investors and shareholders, the ability to anticipate short-term market movements enables more informed buy–sell decisions, potentially leading to profit maximization and risk mitigation.
4. Discussion
The process of determining the hyperparameter tuning interval employed the Optuna library in Python 3. The intervals for each hyperparameter were selected carefully to cover a wide range of possible optimal values, ensuring a thorough exploration of the parameter space. The tuning procedure relied on the minimization of the loss function, with the objective of identifying the best-performing model configuration. The tuning results indicate that the optimal combination includes 181 filters, a kernel size of 2, 200 GRU units, and a dropout rate of 0.0474. In addition, the selection of a time step of 5 supports the model’s ability to capture short-term trends, which is crucial in stock price forecasting, where market behavior is often driven by immediate, short-lived factors.
The training process was conducted sequentially with a batch size of 62 observations per iteration. Although the training was initially set for 1000 epochs, the implementation of early stopping allowed the process to terminate automatically after 257 iterations. This technique helps prevent overfitting and reduces unnecessary computational time. The loss graph illustrates a parallel movement between training and validation loss curves, suggesting that the model maintains a strong generalization capability and does not overfit the training data. The final training loss and validation loss values were 0.00045 and 0.00035, respectively, both of which reflect excellent model performance and stability.
The results of the modeling process have been visualized through graphs comparing the actual and predicted values for the training, validation, and testing datasets. These visualizations provide critical insights into the model’s predictive capabilities. The model accurately followed the movement patterns of stock prices, even though some discrepancies between actual and predicted values occurred, particularly during periods of abrupt price changes. Notably, during March–May 2024 and August–September 2024, there were visible significant fluctuations in the actual stock prices. The model was still able to capture these changes with reasonable accuracy, demonstrating its adaptive response to volatile market conditions, which could be influenced by external events such as political instability or global economic disruptions.
Model evaluation was carried out using the Mean Absolute Percentage Error (MAPE) as the primary metric, which quantifies the average percentage difference between actual and predicted values. This metric is widely used in time series forecasting to assess model accuracy. Based on the evaluation results, the model recorded an MAPE of 2.68% for the training data, 1.88% for the validation data, and 1.99% for the testing data. The overall MAPE of 2.38% indicates a high level of accuracy, supporting the conclusion that the CNN-BiGRU-AM model is both robust and reliable. These values are well within acceptable forecasting error thresholds and affirm that the model can be trusted for real-world implementation in financial decision making.
The stock price forecasting was conducted for the next five working days beyond the testing dataset, providing practical insights into the model’s predictive utility. The forecasting results exhibited a gradual downward trend in stock prices. While some variation exists between predicted and actual values, the forecasting maintained an acceptable MAPE of 8.03%. This level of accuracy remains practical and insightful, especially for short-term forecasting under fluctuating market conditions. The comparison graph between the predicted and actual values further confirms that the model successfully captures the general movement trend, even when minor discrepancies are present.
The implications of this forecasting exercise are substantial for both corporate and investor perspectives. For companies, such predictive capabilities support strategic decision making in areas such as inventory planning, resource allocation, and capital investment. Reliable short-term price forecasts also help firms prepare for market shifts and mitigate risks associated with price volatility. From an investor’s standpoint, the model’s ability to provide advance warnings of potential price movements allows for better-informed decisions regarding stock buying, selling, or holding strategies. Thus, the model not only serves academic and analytical purposes but also holds practical relevance for stakeholders involved in financial markets.
In summary, the use of the hybrid CNN-BiGRU-AM model yields promising results in forecasting stock prices. The careful selection and tuning of hyperparameters, combined with an effective model training strategy and a robust evaluation framework, enable the model to generate predictions that align closely with actual data. The model demonstrates reliability across training, validation, and testing datasets, as well as in unseen data used for real-time forecasting. These results validate the suitability of this model for short-term stock price forecasting and highlight its potential application in broader financial analytics contexts.
5. Conclusions
This study confirms the effectiveness of the CNN-BiGRU-Attention (CNN-BiGRU-AM) model in forecasting INDY stock prices with high accuracy. The model combines three powerful deep learning components: a CNN for extracting essential features and reducing noise, a BiGRU for capturing bidirectional temporal dependencies, and an Attention Mechanism for focusing on the most relevant parts of the data. This integration enables the model to better understand and adapt to complex patterns in volatile stock price movements.
The results show that the model achieved a low forecasting error, with a Mean Absolute Percentage Error (MAPE) consistently below 3% across different data splits. This indicates that the model successfully addresses the nonlinear and highly fluctuating nature of stock market data. It also maintains robust performance during periods affected by external factors such as global coal prices, energy policy shifts, and geopolitical events.
Practically, the model’s accurate forecasts provide valuable support for companies and investors in making more informed decisions. Businesses can utilize these predictions to plan strategic actions, manage financial risk, and respond more effectively to market shifts. Theoretically, this research contributes to the development of advanced predictive models in financial analytics, highlighting the potential of hybrid deep learning architectures in time series forecasting.
In addition to its predictive success, the study offers significant methodological contributions. By integrating three distinct neural components into a unified forecasting model, the CNN-BiGRU-AM architecture provides a replicable framework for time series modeling in volatile and nonlinear financial environments. Its performance on INDY stock demonstrates its applicability to other sector-specific assets with similar volatility profiles. Moreover, the modular design of the model allows for adaptation across various forecasting horizons and domains, paving the way for future applications in energy economics, portfolio optimization, and risk assessment. These findings contribute to the growing body of research on hybrid deep learning methods and support their continued adoption in financial analytics.
Overall, the CNN-BiGRU-AM model not only enhances forecasting performance but also offers practical and theoretical benefits that align with the evolving demands of financial market analysis.