Next Article in Journal
AffectiVR: A Database for Periocular Identification and Valence and Arousal Evaluation in Virtual Reality
Previous Article in Journal
Online Action Detection Incorporating an Additional Action Classifier
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Stock Prediction Method Based on Multidimensional and Multilevel Feature Dynamic Fusion

CAD Research Center, Tongji University, 1239 Siping Road, Shanghai 200092, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(20), 4111; https://doi.org/10.3390/electronics13204111
Submission received: 15 September 2024 / Revised: 15 October 2024 / Accepted: 16 October 2024 / Published: 18 October 2024
(This article belongs to the Section Artificial Intelligence)

Abstract

:
Stock price prediction has long been a topic of interest in academia and the financial industry. Numerous factors influence stock prices, such as a company’s performance, industry development, national policies, and other macroeconomic factors. These factors are challenging to quantify, making predicting stock price movements difficult. This paper presents a novel deep neural network framework that leverages the dynamic fusion of multi-dimensional and multi-level features for stock price prediction, which means we utilize fundamental trading data and technical indicators as multi-dimensional data and local and global multi-level information. Firstly, the model dynamically assigns weights to multi-dimensional features of stocks to capture the impact of each feature on stock prices. Next, it applies the Fourier transform to the global features to capture the long-term trends of the global environment and dynamically fuses these with local and global features of the stocks to capture the overall market environment’s impact on individual stocks. Finally, temporal features are captured using an attention layer and an RNN-based model, which incorporates historical price data to forecast future prices. Experiments on stocks from various industries within the Chinese CSI 300 index reveal that the proposed model outperforms traditional methods and other deep learning approaches in terms of stock price prediction. This paper proposes a method that facilitates the dynamic integration of multi-dimensional and multi-level features in an efficient manner and experimental results show that it improves the accuracy of stock price predictions.

1. Introduction

Stock price prediction is one of the core applications of financial data mining, and the stock market reflects the overall state and expectations of the economy. Studying the stock market gives valuable insights regarding economic trends, industry development, and company performance. By forecasting stock market movements, investors and financial institutions can develop appropriate risk management strategies in advance, reducing potential risks, obtaining more options, and achieving higher efficiency, thus making more informed investment decisions [1]. Consequently, there has been significant interest in stock prediction from both academia and the industry in recent years [2,3,4]. However, the inherent complexity and randomness of the stock market present significant challenges. Historical stock price data are fraught with noise and irregularities, complicating the decision-making process for investors. Moreover, stock prices are influenced by a multitude of factors—including macroeconomic trends, regulatory policies, and investor sentiment—that are often difficult to quantify [5,6]. This vast and intricate dataset poses a formidable barrier to achieving accurate predictions of stock market movements.
The Efficient Market Hypothesis (EMH) [7] suggests that stock prices reflect all available information, and thus, using price information alone can efficiently accomplish prediction tasks [8,9]. In the past, researchers have attempted to use traditional time series analysis methods for stock prediction, such as Vector Autoregression (VAR) [10] and Autoregressive Integrated Moving Average (ARIMA) [11]. However, these methods are predicated on the assumptions of linearity and stationarity, which do not hold true for the inherently nonlinear and non-stationary nature of stock price data. Consequently, these traditional approaches often falter in the face of the noise embedded in historical stock data [12]. Another common approach, fundamental analysis, entails evaluating a company’s financial statements and broader industry conditions. However, this method is resource-intensive and requires a depth of expertise, making it challenging to apply consistently across diverse market conditions. As a result, fundamental analysis can yield suboptimal predictive performance.
In recent years, deep learning technology has rapidly developed and has been widely applied in fields such as medicine, transportation, and finance. Due to the limitations of traditional methods, deep learning techniques have gradually become mainstream in stock prediction [13]. Deep learning can automatically extract useful features from large amounts of historical data without requiring manual feature definition, reducing the workload of feature engineering and improving prediction accuracy. Moreover, deep learning models can adapt to dynamic changes in market conditions through continuous training and updates [14], which significantly enhances their robustness and predictive performance. However, despite these advantages, most existing methods primarily focus on the impact of different features on stock prices in a separate way. This narrow approach often overlooks the complex interactions between multiple features and fails to consider the broader global environment’s influence on stock market dynamics. Such limitations can lead to suboptimal predictions, as stock prices are affected by a multitude of interrelated factors.
Recognizing these challenges, our research aims to develop a stock prediction method that employs the dynamic fusion of multi-dimensional and multi-level features. By integrating data smoothing strategies, we can mitigate the noise often present in global historical stock data, enhancing the reliability of our predictions. Additionally, incorporating attention mechanisms allows us to effectively capture both global and local environmental impacts on stock prices, thereby providing a richer understanding of the market.
To validate the effectiveness of our MDML model, we utilized stock data from the China Securities Index 300 (CSI 300), covering the period from January 2020 to April 2024, as a representation of local market features. Simultaneously, data from the CSI 300 index for the same period were employed to capture global market dynamics. The key contributions of this study are listed below.
  • We provide a stock prediction method based on the dynamic fusion of multidimensional and multilevel data that successfully captures the impact of both global and local factors on stock prices. By combining variables from many dimensions and levels, the model provides a thorough knowledge of the factors that influence stock price movements. This novel approach analyzes not only the influence of individual features but also the interactions between features, hence increasing the model’s expressive power.
  • A dynamic weight allocation method is presented that enables the model to dynamically modify the weights of various characteristics according to their relative importance. This guarantees an accurate representation of each feature’s effect on stock prices. The model can represent the changing importance of information across several time points and market conditions by dynamically allocating weights, which enhances the accuracy and dependability of prediction outcomes.
  • We introduce a Fourier transform method for global features, applying Fourier transform to global features to capture long-term trends in the global environment. This technique helps in understanding the impact of macroeconomic and other broad factors on stock prices over an extended period, providing the model with support from long-term information.
  • We conducted extensive experiments on stocks from different industries within the CSI 300 index in the Chinese market. The results indicate that the proposed model performs exceptionally well in stock price prediction, significantly outperforming traditional methods and other deep learning approaches, thereby demonstrating its substantial potential in practical applications.
The remainder of the paper will be organized as follows: Section 2 will review related work on stock prediction, including both traditional methods and deep learning approaches. Section 3 will provide a detailed description of the proposed stock prediction method based on dynamic fusion of multi-dimensional and multi-level features. Section 4 will present the design and results of the experiments. Section 5 discusses the limitations of this work and future research directions, while Section 6 summarizes the main findings and contributions of the study.

2. Related Work

There are various methods available for stock price prediction, including traditional time series forecasting methods, classic machine learning algorithms, and deep learning approaches. Table 1 provides a summary of the literature review below.

2.1. Traditional Method

The stock prediction problem is often modeled as a time series forecasting problem. Classic models include the Autoregressive Conditional Heteroskedasticity (ARCH) [15] model, the Generalized Autoregressive Conditional Heteroskedasticity (GARCH) [16] model, and the AutoRegressive Integrated Moving Average (ARIMA) [11] model, and so on, which are based on statistical theory and can provide relatively accurate predictions when market trends are clear or volatility is predictable. These models are widely used in the field of time series forecasting. However, they often require manual feature selection and have stringent requirements for data stationarity, which limits their applicability in the domain of stock prediction.
Machine learning methods have also found numerous applications in the field of stock prediction, such as Support Vector Machines (SVM) [17] and Random Forests (RF) [18]. These methods often do not require data to be stationary and can effectively capture the nonlinear relationships between features, making them widely used in stock prediction. For instance, in the paper [19], various machine learning algorithms, including support vector machines, decision trees, and neural networks, are explored to predict stock market trends. Additionally, in [20], the authors compare the performance of different machine learning algorithms, such as linear regression, SVM, and k-nearest neighbors (KNN) [21], in predicting stock prices. However, classic machine learning algorithms typically perform poorly when dealing with high-dimensional and complex nonlinear data. In contrast, deep learning models possess powerful nonlinear modeling capabilities, allowing them to capture intricate patterns and relationships within the data. Furthermore, classic machine learning algorithms often require additional feature extraction and transformation, such as sliding windows and time lags, when handling time series data.
Numerous deep learning-based stock prediction models have been developed in the past, such as Long Short-Term Memory (LSTM) networks [8,22], CNN-LSTM [23,24], ALSTM [25] and so on. They are specifically designed to process time series data, enabling them to better capture temporal dependencies and dynamic changes. There have been numerous applications in this area. For instance, the study [26] examines the use of LSTM networks in this scenario, predicting future stock price trends based on historical prices and technical analysis indicators. Hence, in this paper, we employ LSTM networks to capture temporal features from historical data.

2.2. Predict with Frequency Transformation

In [25], it was found that separating various features before capturing temporal features and then assigning weights based on the cell’s hidden states can enhance prediction accuracy. Therefore, in this paper, we also adopt this two-stage approach to perform multidimensional feature fusion and temporal information capture.
To better utilize global features, it is essential to process these features initially. Frequency transformation has been widely applied in the field of time series forecasting, often combining frequency transformation with deep learning models to enhance prediction accuracy. The significance of frequency transformation lies in its ability to capture the periodicity, trends, and multi-scale patterns of time series by converting them to the frequency domain, thereby providing a more global perspective, as illustrated in the Figure 1. Common frequency transformations include the Discrete Fourier Transform (DFT), Discrete Cosine Transform (DCT), and Discrete Wavelet Transform (DWT) [27]. In the study [28], a frequency domain block is proposed in the model named ATFN to capture the dynamic and complex periodic patterns of time series data, thereby improving prediction accuracy. Additionally, in [29,30], new modules are designed using Fourier transformation to capture patterns in time series. Moreover, ref. [31] combines Fourier analysis with the Transformer architecture, enhancing the Transformer’s ability to capture the global properties of time series.
In the field of stock prediction, many researchers have applied frequency transformation techniques. For instance, in [8], trading patterns are decomposed into various frequency components, with each component modeling the specific frequency of underlying trading patterns that influence stock price fluctuations. In this study, we employ Fourier transformation to extract the top K frequencies from the global features, aiming to capture long-term trends in the overall market to guide the prediction of individual stock prices.

2.3. Predict with Dynamic Fusion

In previous work, there have been few studies that utilize multidimensional features for prediction, while the use of attention mechanisms can effectively facilitate their integration. For example, in [32], the attention patterns in the encoder and decoder are modified to reduce the space and time cost of training while improving the accuracy of long-term predictions. In this paper, we introduce cross-attention and self-attention mechanisms [33] to capture future information contained in both global and local features. There have been precedents for using global and local information for stock prediction. For instance, in [34], a context aggregation module is employed to fuse global and local information, though this process introduces hyperparameters. To better integrate global and local information, we introduce the Mixture-of-Experts (MoE) [35] mechanism, which dynamically adjusts the weights of global and local information. The MoE mechanism typically comprises two parts: expert networks and a gating network, which enhances the scalability of information fusion without adding hyperparameters. This approach has been widely adopted in various fields. Jinguo Zhu et al. (2022) [36] utilized the Mixture-of-Experts (MoE) mechanism to alleviate task interference in general models, thereby enhancing performance across multiple tasks. They proposed Conditional Mixture-of-Experts (CMoEs), which activate sub-networks based on specific conditions to mitigate interference while maintaining versatility. Similarly, Jinhua Liu et al. (2020) [37] introduced Att-MoE, which employs a single gating network in combination with multiple expert networks. These experts collaborate on different sub-tasks, with the gating network guiding their adaptive and cooperative utilization to achieve efficient cell segmentation.
Upon surveying the literature mentioned above, we have identified that there remains significant room for improvement in the accuracy of stock prediction. Moreover, previous research focusing on the fusion of multi-dimensional and multi-level features has been relatively limited. Consequently, we propose a novel model that synergistically combines the capabilities of DFT to capture long-term trends, along with the feature-capturing and fusion prowess of MOE and attention mechanisms.

3. Methodology

3.1. Problem Formulation

Our goal is to predict the future price of a stock given historical data X = { x 1 , x 2 , , x t } , where x i is the feature vector of the stock on the i-th day. The observation window is defined as the time interval of t days, and our objective is to predict the stock price on the t + 1 -th day, denoted as y t + 1 . We divide the stock features into global features X g and local features X l , where the global features represent the overall market trend and are represented by the CSI 300 index data, while the local features are specific to an individual stock. The input to our model consists of the global features X g and local features X l , and the output is the close stock price on the t + 1 -th day y t + 1 .

3.2. Data Process

We utilize historical stock data to train our model, encompassing a variety of data points such as the stock’s opening price, closing price, highest price, lowest price, trading volume, and so forth. To enhance the model training process, we have normalized the data, scaling the input data for the model to a range between 0 and 1. The expression is as follows:
x n o r m = x min ( x ) max ( x ) min ( x )

3.3. Model Architecture

The model proposed in this study is built upon a dynamic fusion mechanism that integrates both multi-dimensional and multi-level features. The term “multi-dimensional” refers to a diverse set of features that encapsulate various aspects of financial data, including fundamental attributes such as opening and closing prices, trading volumes, and technical indicators like the Relative Strength Index (RSI). These features collectively provide a comprehensive representation of the financial market’s micro-level dynamics. In contrast, “multi-level” features pertain to the aggregation of data across different scales, capturing localized characteristics at the level of individual stocks, as well as global attributes that reflect macro-level trends and conditions in the broader stock market. This dual-level structure enables a more holistic analysis by accounting for both granular stock-specific details and overarching market-wide patterns. Taking into account the distinctive characteristics of our model, which synergistically combines these Multi-Dimensional and Multi-Level features, we have designated it as the MDML model.
The MDML model is divided into four components: the Global Feature Encoder, the Local Feature Encoder, the Dynamic Feature Fusion Module, and the Temporal Feature Capture Module. The overall architecture of the model is depicted in Figure 2:

3.3.1. Global Feature Encoder

The Global Feature Encoder aims to extract long-term trends from the global features. We employ the Fourier Transform to identify the top K frequencies within the global features, thereby capturing the long-term trends inherent in these features. The expression for the Fourier Transform is as follows:
X k = n = 0 N 1 x n e j 2 π k n / N
We input the global features X g into the Fourier Transform module to obtain the frequency components of the global features. Subsequently, we select the top K frequencies and perform an inverse transform on the global features. The expression for the inverse transform is as follows:
x n = 1 K k = 0 K 1 X k e j 2 π k n / K
After obtaining the result of the inverse transform, we feed this output into the feature mapping module to derive the latent feature representation of the global long-term trends (a fully connected network followed by a ReLU activation function) as mentioned in [38]:
h g = ReLU ( x n W g + b g )
Here, the global features post-Fourier Transform are denoted as x n R t × d , with W g R d × d p and b g R d p being trainable parameters. The latent feature representation of the global features, h g , will then be inputted alongside the local features into the dynamic feature fusion module.

3.3.2. Local Feature Encoder

The purpose of the local feature encoder is to extract the prospective trends from the features of individual stocks. We utilize an LSTM network in conjunction with an attention mechanism to integrate and encode the local features. Specifically, we input the local features X l into the LSTM network. Unlike the method for capturing temporal features, we treat the local information as a sequence of features, where for the n-th feature, our input is x n = ( x 1 n , x 2 n , , x t n ) R t . Based on the hidden state h t and cell state s t of the LSTM cell, we calculate the weights for various features and then perform a weighted summation of these weights and features, which can be expressed as follows:
e t n = v e tanh ( W e [ h t 1 ; s t 1 ] + U e x n )
α t n = exp ( e t n ) i = 1 d exp ( e t i )
x l = α t 1 x 1 + α t 2 x 2 + + α t d x d
After this, the output passes through a feature mapping module to obtain the latent feature representation of the local features h l R t × d p , which will then be inputted together with the global features into the dynamic feature fusion module.

3.3.3. Dynamic Feature Fusion Module

After obtaining the latent feature representations of both global and local features, we dynamically integrate these features to capture the impact of the global and local environments on stock prices. The self-attention mechanism is capable of capturing dependencies between different time points, aiding the model in comprehending how price fluctuations within a specific past period influence future price trends. On the other hand, the cross-attention mechanism can be utilized to manage the interactions between multiple time series, namely the correlation between stock prices and macroeconomic indicators. Therefore, We employ self-attention mechanisms to capture the historical patterns of the global information h g , which can be specifically represented as:
Q = h g W Q , K = h g W K , V = h g W V
α = softmax Q K T d k
h ˜ g = α V
Here, W Q , W K , W V R d p × d k are trainable parameters, and d k is the dimension of the attention mechanism.
To capture the relationship between global and local information, we introduce a cross-attention mechanism that performs cross-attention between global and local information. Additionally, to preserve the features of individual stocks, we incorporate residual connections, which can be represented as:
Q = h g W Q , K = h l W K , V = h l W V
α = softmax Q K T d k V
h ˜ l = α V + h l
After obtaining h ˜ g and h ˜ l , we need to fuse these two types of information. We introduce the MOE (Mixture of Experts) mechanism, which dynamically adjusts the weights of global and local information without introducing hyperparameters:
h e m b = Concat ( h l , h g )
α i = softmax ( W h e m b + b )
h e m b ^ = α 1 h g + α 2 h l
In this representation, α 1 and α 2 are the weights of the global and local information, respectively, which are dynamically adjusted during the training process. h e m b ^ is the fused result of global and local information, which will then be input into the temporal feature capture module.

3.3.4. Temporal Feature Capture Module

In the temporal feature module, we employ an LSTM network combined with a temporal attention mechanism to capture the temporal features within the historical data. We input h ^ emb into the LSTM network to obtain the hidden state of the LSTM cell, h t . Subsequently, we feed h t into the temporal attention mechanism to acquire the weights for each time step and then perform a weighted summation of these weights with h t to obtain the final prediction result. This can be specifically represented as:
e t = v e tanh ( W e [ h t 1 ; s t 1 ] + U e h t )
α t = exp ( e t ) i = 1 t exp ( e i ) ,
h ^ t + 1 = α t h t
Finally, h t + 1 ^ is input into a fully connected network to yield the ultimate prediction result y t + 1 ^ .

4. Experiments

4.1. Dataset

During the data collection process, taking into account that stocks in the same or related industries may exhibit similar trend variations, we combine the classification results based on the enterprise types from the East Money website with the clustering results based on the close price as shown in Table 2. This approach allows us to segment the stocks into different datasets for separate training.

4.2. Data Process and Model Training

The dataset is divided temporally into training, testing, and validation sets with a ratio of approximately 8:1:1, followed by min–max normalization of the data.
An Adam optimizer is utilized with a learning rate of 0.001 and a batch size of 64, employing the Mean Squared Error (MSE) as the loss function. To investigate the impact of the observation window on the predictive outcome, experiments were conducted with observation windows set to 5, 10, 15, 20, and 25. The results indicated that the model performed optimally with an observation window of 15; thus, this setting was adopted. During the Fourier transformation of global features, the top K frequencies were selected, and it was discovered through experimentation that the model’s predictive performance was best when the top K was set to 7. To mitigate overfitting, an early stopping strategy was employed, halting the training process if the loss value on the validation set did not decrease for five consecutive epochs.
The historical data indicators obtained from the website encompass a total of 20 types, as shown in Table 3, which include: Opening Price, Closing Price, Highest Price, Lowest Price, Trading Volume, Transaction Value, Turnover Rate, Price Change Percentage, Price Change Amount, Amplitude, Pivot and some technical indicators. Among the historical data indicators obtained from the website, the term “Pivot” refers to the average value of the Highest Price, Lowest Price, and Closing Price. In the global information capture module, we utilize the CSI 300 Index as a representative of global information. The CSI 300 is a significant index in China’s securities market, encompassing the stocks of 300 companies and representing the overall trend of China’s securities market.

4.3. Evaluation Parameters

In our quest to forecast stock prices with precision, we evaluated the reliability of the experimental results by incorporating three widely used metrics: Mean Absolute Error (MAE), Mean Absolute Percentage Error (MAPE), and Root Mean Square Error (RMSE). These metrics serve as a multidimensional framework for assessing the predictive efficacy of our model. The Mean Absolute Error (MAE) offers a direct measure of the average prediction error magnitude, while the Mean Absolute Percentage Error (MAPE) translates these errors into a percentage of actual values, highlighting relative inaccuracies especially pertinent in finance. The Root Mean Square Error (RMSE) amplifies the impact of larger errors through squaring, making it an effective tool for gauging average error severity and identifying outliers [25]. The formulas for the calculation of these metrics are as follows:
M A E = 1 n i = 1 n | y i y ^ i |
M A P E = 1 n i = 1 n | y i y ^ i | y i
R M S E = 1 n i = 1 n ( y i y ^ i ) 2
where n is the number of samples, y i is the actual value, y ^ i is the predicted value, and y ¯ is the average value.

4.4. Baseline

Our experiments are benchmarked against a diverse array of analytical techniques, including traditional models like ARIMA and Moving Average (MA) [39,40], which are cornerstones in time series forecasting. We also leverage machine learning algorithms such as SVM and RF for their efficacy with complex data patterns. Additionally, we engage deep learning models: CNN-LSTM, DTML, and LSTM-BN to harness advanced pattern recognition capabilities. This comprehensive comparison with our approach aims to affirm its effectiveness and superiority in predictive performance.
Here is a brief introduction to the baselines:
  • ARIMA [41]: The ARIMA model is a time series forecasting model composed of the Autoregressive (AR) and Moving Average (MA) components. In this approach, we utilize historical closing prices as variables to predict future closing prices. To identify the optimal hyperparameters, we utilize the Autocorrelation Function (ACF) and Partial Autocorrelation Function (PACF) for parameter specification.
  • MA [42]: The MA model is a time series forecasting model that consists solely of the Moving Average component. The predictions of the MA model are influenced by historical data and are suitable for stationary time series data.
  • ES [43]: Exponential Smoothing (ES) effectively captures trends and seasonal variations in data, demonstrating strong adaptability to rapidly respond to market changes and adjust forecasts in a timely manner. This makes it suitable for dynamic financial markets. The relevant code for this part of the experiment can be found at https://github.com/DONGTangYuan/multi-dimension-data (accessed on 1 October 2024) for reference.
  • lightGBM [44]: LightGBM is an efficient gradient-boosting tree algorithm that excels in handling nonlinear relationships and high-dimensional data, enabling it to capture complex market patterns.
  • SVM [45]: Support Vector Machine is a supervised learning algorithm that finds an optimal hyperplane to separate data into two categories. It applies to both classification and regression problems. The SVM model can capture complex nonlinear relationships in historical data and is therefore widely used in stock price forecasting.
  • RF [46]: Random Forest is an ensemble learning algorithm that makes predictions through multiple decision trees. In this stock prediction model, in addition to utilizing fundamental features such as opening price, highest price, lowest price, and trading volume, we have also incorporated technical indicators such as moving averages, Relative Strength Index (RSI), Bollinger Bands, and Moving Average Convergence Divergence (MACD).
  • CNN-LSTM [23]: The CNN-LSTM model is a model that combines Convolutional Neural Networks and Recurrent Neural Networks. It extracts features from the input sequence using convolutional networks and then makes predictions using recurrent networks.
  • DTML [34]: The DTML model is a transformer-based model. It learns the correlations between stocks in an end-to-end manner. DTML captures asymmetric and dynamic correlations by learning the temporal correlations within each stock and generates multi-level context based on the global market context.
  • LSTM-BN [47]: The LSTM-BN model is a model based on Recurrent Neural Networks and Batch Normalization. It accelerates the training process and improves prediction accuracy through batch normalization.

4.5. Experiment Result

To substantiate the efficacy of our model, we conducted experiments on both the baseline models and our proposed model. During the validation process of the model’s performance, we set different random seeds and performed the experiments five times, taking the average results. The following are the outcomes from testing our model and the baseline models on a test set composed of stocks from eight different industries. The results are shown in Table 4, Table 5, Table 6 and Table 7. We have bolded the best metrics for all models and underscored the second-best metrics.
Table 6 and Table 7 demonstrate that the ES model and SVM model exhibit notably strong performance in the Pharm. and Biotech. and Finance sectors, respectively. This indicates that traditional time series forecasting methods and conventional machine learning algorithms can perform well in certain cases. On the other hand, the LSTM-BN and CNN-LSTM algorithms show superior performance in the Chemical Industry and Information Technology datasets, respectively, suggesting that employing alternative deep learning methods for stock price prediction also holds significant promise. Overall, across all eight datasets, our proposed models consistently achieve the lowest or second-lowest values for MAE, RMSE, and MAPE metrics. This signifies that the absolute and relative errors between predicted and actual results are minimal, indicating that our models possess a degree of stability and accuracy. This further validates the effectiveness of our proposed method for dynamic fusion of multi-level and multi-dimensional features.

4.6. Case Study

To substantiate the efficacy of our predictive model, an in-depth case study was conducted focusing on a select group of the most representative stocks within the CSI 300 index. This study was designed to meticulously assess the model’s forecasting capabilities by comparing its predicted values against the actual market outcomes. The selected stocks were chosen based on their market capitalization, liquidity, and influence on the overall index, ensuring a comprehensive reflection of the model’s performance across various sectors and market conditions. The testing period for this case study spanned from 1 April 2024, to 1 June 2024, a timeframe that encapsulates a variety of market conditions, including seasonal fluctuations and potential macroeconomic events that could influence stock prices.
The results of the case study were compelling, with the model demonstrating a high degree of accuracy in its predictions. The predicted values were found to be near the actual values as shown in Figure 3a,b, Figure 4a,b, Figure 5a,b and Figure 6a,b, indicating the model’s robustness and reliability in forecasting stock prices.

4.7. Ablation Study

To verify the effectiveness of our model, we conducted ablation studies. The setup of the ablation studies is as follows:
  • my-model-1: The Discrete Fourier Transform (DFT) is removed, while dynamic weights and feature dynamic fusion are retained, to verify the impact of the DFT on the model.
  • my-model-2: Dynamic weights are removed, while the DFT and feature dynamic fusion are retained, to verify the impact of dynamic weights on the model.
  • my-model-3: Feature dynamic fusion is removed, while the DFT and dynamic weights are retained, to verify the impact of feature dynamic fusion on the model.
  • my-model: The complete model, including the DFT, dynamic weights, and feature dynamic fusion.
In this experiment, we conducted experiments on two datasets from the finance and construction industries within the CSI 300. The results of the experiment are as follows:
The results of the ablation experiments in Table 8 indicate that the removal of the DFT module, dynamic weights module, and feature dynamic fusion all lead to an increase in MAE, RMSE, and MAPE metrics, implying a decline in the model’s predictive accuracy. Thus, it can be concluded that the DFT module, MOE module, and feature fusion module all contribute to the enhancement of the model’s performance. Specifically, the impact of removing the DFT module on the experimental results is minimal, while the removal of the feature fusion module causes a significant increase in the MAE metric. Conversely, the removal of the MOE module results in a more substantial rise in RMSE and MAPE metrics. These experiments demonstrate that adding fusion modules is beneficial for reducing both the relative and absolute discrepancies in predictions, thus validating the effectiveness of multi-dimensional and multi-feature fusion. The minimal impact of the DFT module suggests that there is still room for improvement in capturing the global characteristics of stocks, which could be a focus for future work.

5. Limitations and Future Research

The model discussed in this text has demonstrated impressive performance in stock price forecasting, yet it is not without its limitations. We will examine these limitations and suggest possible avenues for future research. The limitations can be summarized as follows:
  • The model captures long-term trends through Fourier transformation, but balancing the influence of long-term trends and short-term fluctuations in practical applications remains a challenge. Over-reliance on long-term trends may overlook short-term market changes, while an excessive focus on short-term fluctuations could lead to instability in the forecast results.
  • Financial markets are significantly influenced by external factors such as policy changes, economic fluctuations, and unexpected events. These factors are often nonlinear and sudden, making them difficult to predict through historical data. The model may not adapt to these changes promptly, especially in the face of black swan events or economic crises, which could significantly reduce the model’s forecasting power.
Based on the aforementioned limitations, we propose the following directions for future research:
  • Incorporate a dynamic weight adjustment mechanism based on market conditions to flexibly adjust the influence of long-term trends and short-term fluctuations under different market states. For example, using market volatility indicators to dynamically adjust the model’s attention to long-term trends and short-term fluctuations to cope with various market conditions.
  • Develop an event-driven model capable of capturing the impact of unexpected events. Introduce exogenous variables based on events (such as policy changes, news events, etc.) into the time series model to help the model identify and respond to these sudden changes. For instance, using an attention mechanism-based event detection model to automatically identify and quantify the impact of unexpected events on the market.

6. Conclusions

In conclusion, this paper presents a novel deep neural network model for stock price prediction that effectively integrates multi-dimensional and multi-level features. By dynamically assigning weights to various stock features and applying the Fourier transform to capture long-term trends, the model successfully combines global and local information to reflect the overall market environment’s impact on individual stocks. The incorporation of an attention mechanism and RNN-based structure further enhances the model’s ability to capture temporal dynamics, leveraging historical price data to improve prediction accuracy. Experimental results on stocks from different industries within the CSI 300 index demonstrate the model’s superior performance compared to traditional methods and other deep learning approaches, highlighting its potential for more accurate and robust stock price prediction. However, the research presented in this paper still has several limitations. For instance, the experiments were conducted exclusively within the context of the Chinese stock market and relied solely on numerical data, neglecting textual information, which may limit the model’s predictive performance and expressive capability. Therefore, in future work, it would be beneficial to incorporate a broader range of information into the model, including textual data, and to validate the model’s effectiveness across stock markets in other countries.

Author Contributions

Conceptualization, Y.D. and Y.H.; methodology, Y.D.; software, Y.D.; formal analysis, Y.D.; investigation, Y.D.; resources, Y.D.; data curation, Y.D.; writing—original draft preparation, Y.D. and Y.H.; writing—review and editing, Y.D.; visualization, Y.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

In this study, the selected stocks are derived from the stocks included in the CSI 300 Index, all of which are Chinese A-shares. To ensure the accuracy and consistency of the data, we excluded stocks with incomplete data during the experimental time frame. The stock data was obtained using the shared library in Python 3.9, while the classification information for the stocks was sourced from Eastmoney.com. The data and classification results used in this study have been made publicly available on https://github.com/DONGTangYuan/multi-dimension-data (accessed on 1 October 2024), allowing access for other researchers and interested parties.

Acknowledgments

We would like to express our sincere gratitude to the CAD research center, Tongji University, for their invaluable support and assistance with the computational resources utilized in this research.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Tkáč, M.; Verner, R. Artificial neural networks in business: Two decades of research. Appl. Soft Comput. 2016, 38, 788–804. [Google Scholar] [CrossRef]
  2. Xu, Y.; Cohen, S.B. Stock movement prediction from tweets and historical prices. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Melbourne, Australia, 15–20 July 2018; pp. 1970–1979. [Google Scholar]
  3. Albahli, S.; Nazir, T.; Mehmood, A.; Irtaza, A.; Alkhalifah, A.; Albattah, W. AEI-DNET: A novel densenet model with an autoencoder for the stock market predictions using stock technical indicators. Electronics 2022, 11, 611. [Google Scholar] [CrossRef]
  4. Daradkeh, M.K. A hybrid data analytics framework with sentiment convergence and multi-feature fusion for stock trend prediction. Electronics 2022, 11, 250. [Google Scholar] [CrossRef]
  5. Valencia, F.; Gómez-Espinosa, A.; Valdés-Aguirre, B. Price movement prediction of cryptocurrencies using sentiment analysis and machine learning. Entropy 2019, 21, 589. [Google Scholar] [CrossRef] [PubMed]
  6. Jin, Z.; Yang, Y.; Liu, Y. Stock closing price prediction based on sentiment analysis and LSTM. Neural Comput. Appl. 2020, 32, 9713–9729. [Google Scholar] [CrossRef]
  7. Miller, C.N.; Roll, R.; Taylor, W. Efficient capital markets: A review of theory and empirical work. J. Financ. 1970, 25, 383–417. [Google Scholar]
  8. Zhang, L.; Aggarwal, C.; Qi, G.J. Stock price prediction via discovering multi-frequency trading patterns. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Halifax, NS, Canada, 13–17 August 2017; pp. 2141–2149. [Google Scholar]
  9. Stoean, C.; Paja, W.; Stoean, R.; Sandita, A. Deep architectures for long-term stock price prediction with a heuristic-based strategy for trading simulations. PLoS ONE 2019, 14, e0223593. [Google Scholar] [CrossRef]
  10. Jung, C.; Boyd, R. Forecasting UK stock prices. Appl. Financ. Econ. 1996, 6, 279–286. [Google Scholar] [CrossRef]
  11. Ariyo, A.A.; Adewumi, A.O.; Ayo, C.K. Stock price prediction using the ARIMA model. In Proceedings of the 2014 UKSim-AMSS 16th International Conference on Computer Modelling and Simulation, Cambridge, UK, 26–28 March 2014; pp. 106–112. [Google Scholar]
  12. Alotaibi, S.S. Ensemble technique with optimal feature selection for Saudi stock market prediction: A novel hybrid red deer-grey algorithm. IEEE Access 2021, 9, 64929–64944. [Google Scholar] [CrossRef]
  13. Corbet, S.; Eraslan, V.; Lucey, B.; Sensoy, A. The effectiveness of technical trading rules in cryptocurrency markets. Financ. Res. Lett. 2019, 31, 32–37. [Google Scholar] [CrossRef]
  14. Olivas, E.S.; Guerrero, J.D.M.; Martinez-Sober, M.; Magdalena-Benedito, J.R.; Serrano Lopez, A.J. Handbook of Research on Machine Learning Applications and Trends: Algorithms, Methods, and Techniques: Algorithms, Methods, and Techniques; IGI Global: Hershey, PA, USA, 2009. [Google Scholar]
  15. Kirchgässner, G.; Wolters, J.; Hassler, U. Autoregressive Conditional Heteroscedasticity. In Introduction to Modern Time Series Analysis; Springer: Berlin/Heidelberg, Germany, 2013; pp. 281–310. [Google Scholar] [CrossRef]
  16. Bollerslev, T. Generalized autoregressive conditional heteroskedasticity. J. Econom. 1986, 31, 307–327. [Google Scholar] [CrossRef]
  17. Cortes, C.; Vapnik, V. Support-Vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  18. Ma, Y.; Han, R.; Fu, X. Stock prediction based on random forest and LSTM neural network. In Proceedings of the 2019 19th International Conference on Control, Automation and Systems (ICCAS), Jeju, Republic of Korea, 15–18 October 2019; pp. 126–130. [Google Scholar]
  19. Shen, S.; Jiang, H.; Zhang, T. Stock Market Forecasting Using Machine Learning Algorithms; Department of Electrical Engineering, Stanford University: Stanford, CA, USA, 2012; pp. 1–5. [Google Scholar]
  20. Kumar, I.; Dogra, K.; Utreja, C.; Yadav, P. A comparative study of supervised machine learning algorithms for stock market trend prediction. In Proceedings of the 2018 Second International Conference on Inventive Communication and Computational Technologies (ICICCT), Coimbatore, India, 20–21 April 2018; pp. 1003–1007. [Google Scholar]
  21. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 2003, 13, 21–27. [Google Scholar] [CrossRef]
  22. Akita, R.; Yoshihara, A.; Matsubara, T.; Uehara, K. Deep learning for stock prediction using numerical and textual information. In Proceedings of the 2016 IEEE/ACIS 15th International Conference on Computer and Information Science (ICIS), Okayama, Japan, 26–29 June 2016; pp. 1–6. [Google Scholar]
  23. Lu, W.; Li, J.; Li, Y.; Sun, A.; Wang, J. A CNN-LSTM-based model to forecast stock prices. Complexity 2020, 2020, 6622927. [Google Scholar] [CrossRef]
  24. Wang, H.; Wang, J.; Cao, L.; Li, Y.; Sun, Q.; Wang, J. A stock closing price prediction model based on CNN-BiSLSTM. Complexity 2021, 2021, 5360828. [Google Scholar] [CrossRef]
  25. Qin, Y.; Song, D.; Cheng, H.; Cheng, W.; Jiang, G.; Cottrell, G.W. A dual-stage attention-based recurrent neural network for time series prediction. In Proceedings of the 26th International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 2627–2633. [Google Scholar]
  26. Nelson, D.M.; Pereira, A.C.; De Oliveira, R.A. Stock market’s price movement prediction with LSTM neural networks. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 1419–1426. [Google Scholar]
  27. Yi, K.; Zhang, Q.; Cao, L.; Wang, S.; Long, G.; Hu, L.; He, H.; Niu, Z.; Fan, W.; Xiong, H. A Survey on Deep Learning based Time Series Analysis with Frequency Transformation. arXiv 2023, arXiv:2302.02173. [Google Scholar]
  28. Yang, Z.; Yan, W.; Huang, X.; Mei, L. Adaptive temporal-frequency network for time-series forecasting. IEEE Trans. Knowl. Data Eng. 2020, 34, 1576–1587. [Google Scholar] [CrossRef]
  29. Woo, G.; Liu, C.; Sahoo, D.; Kumar, A.; Hoi, S. Etsformer: Exponential smoothing transformers for time-series forecasting. arXiv 2022, arXiv:2202.01381. [Google Scholar]
  30. Woo, G.; Liu, C.; Sahoo, D.; Kumar, A.; Hoi, S. Learning deep time-index models for time series forecasting. In Proceedings of the International Conference on Machine Learning, Seattle, WA, USA, 30 November–1 December 2023; pp. 37217–37237. [Google Scholar]
  31. Zhou, T.; Ma, Z.; Wen, Q.; Wang, X.; Sun, L.; Jin, R. Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting. In Proceedings of the International Conference on Machine Learning, Baltimore, MA, USA, 17–23 July 2022; pp. 27268–27286. [Google Scholar]
  32. Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, online, 2–9 February 2021; Volume 35, pp. 11106–11115. [Google Scholar]
  33. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.u.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  34. Yoo, J.; Soun, Y.; Park, Y.c.; Kang, U. Accurate multivariate stock movement prediction via data-axis transformer with multi-level contexts. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, Virtual Event, 14–18 August 2021; pp. 2037–2045. [Google Scholar]
  35. Jacobs, R.A.; Jordan, M.I.; Nowlan, S.J.; Hinton, G.E. Adaptive Mixtures of Local Experts. Neural Comput. 1991, 3, 79–87. [Google Scholar] [CrossRef]
  36. Zhu, J.; Zhu, X.; Wang, W.; Wang, X.; Li, H.; Wang, X.; Dai, J. Uni-perceiver-moe: Learning sparse generalist models with conditional moes. Adv. Neural Inf. Process. Syst. 2022, 35, 2664–2678. [Google Scholar]
  37. Liu, J.; Desrosiers, C.; Zhou, Y. Att-MoE: Attention-based mixture of experts for nuclear and cytoplasmic segmentation. Neurocomputing 2020, 411, 139–148. [Google Scholar] [CrossRef]
  38. Feng, F.; He, X.; Wang, X.; Luo, C.; Liu, Y.; Chua, T.S. Temporal Relational Ranking for Stock Prediction. ACM Trans. Inf. Syst. 2019, 37, 1–30. [Google Scholar] [CrossRef]
  39. Taslim, D.G.; Murwantara, I.M. A Comparative Study of ARIMA and LSTM in Forecasting Time Series Data. In Proceedings of the 2022 9th International Conference on Information Technology, Computer, and Electrical Engineering (ICITACEE), Semarang, Indonesia, 25–26 August 2022; pp. 231–235. [Google Scholar] [CrossRef]
  40. Bichescu, B.; Polak, G.G. Time series modeling and forecasting by mathematical programming. Comput. Oper. Res. 2023, 151, 106079. [Google Scholar] [CrossRef]
  41. Parzen, E.; Brown, R.G. Smoothing, Forecasting and Prediction of Discrete Time Series. J. Am. Stat. Assoc. 2004, 59, 973. [Google Scholar] [CrossRef]
  42. Billah, M.; Sultana, A.; Bhuiyan, F.; Kaosar, M. Stock price prediction: Comparison of different moving average techniques using deep learning model. Neural Comput. Appl. 2024, 36, 5861–5871. [Google Scholar] [CrossRef]
  43. He, X.J. Crude Oil Prices Forecasting: Time Series Vs. SVR Models. J. Int. Technol. Inf. Manag. 2018, 27, 25–42. [Google Scholar] [CrossRef]
  44. Hartanto, A.D.; Kholik, Y.N.; Pristyanto, Y. Stock Price Time Series Data Forecasting Using the Light Gradient Boosting Machine (lightgbm) Model. Int. J. Inform. Vis. 2023, 7, 4. [Google Scholar]
  45. Hu, Z.; Zhu, J.; Tse, K. Stocks market prediction using Support Vector Machine. In Proceedings of the 2013 6th International Conference on Information Management, Innovation Management and Industrial Engineering, Xi’an, China, 23–24 November 2013; Volume 2, pp. 115–118. [Google Scholar] [CrossRef]
  46. Aliyeva, A. Predicting Stock Prices Using Random Forest and Logistic Regression Algorithms. In Proceedings of the 11th International Conference on Theory and Application of Soft Computing, Computing with Words and Perceptions and Artificial Intelligence—ICSCCW-2021, Antalya, Turkey, 23–24 August 2021; Aliev, R.A., Kacprzyk, J., Pedrycz, W., Jamshidi, M., Babanli, M., Sadikoglu, F.M., Eds.; Springer: Cham, Switzerland, 2022; pp. 95–101. [Google Scholar]
  47. Fang, Z.; Ma, X.; Pan, H.; Yang, G.; Arce, G.R. Movement forecasting of financial time series based on adaptive LSTM-BN network. Expert Syst. Appl. 2023, 213, 119207. [Google Scholar] [CrossRef]
Figure 1. The results of performing the Fourier Transform on the CSI 300 index.
Figure 1. The results of performing the Fourier Transform on the CSI 300 index.
Electronics 13 04111 g001
Figure 2. The architecture of the proposed method.
Figure 2. The architecture of the proposed method.
Electronics 13 04111 g002
Figure 3. Prediction for (a) Midea Group Co., Ltd. and (b) Contemporary Amperex Technology Co., Ltd. (CATL).
Figure 3. Prediction for (a) Midea Group Co., Ltd. and (b) Contemporary Amperex Technology Co., Ltd. (CATL).
Electronics 13 04111 g003
Figure 4. Prediction for (a) China Merchants Bank Co., Ltd. and (b) Kweichow Moutai Co., Ltd.
Figure 4. Prediction for (a) China Merchants Bank Co., Ltd. and (b) Kweichow Moutai Co., Ltd.
Electronics 13 04111 g004
Figure 5. Prediction for (a) China Yangtze Power Co., Ltd. and (b) Bank Co., Ltd.
Figure 5. Prediction for (a) China Yangtze Power Co., Ltd. and (b) Bank Co., Ltd.
Electronics 13 04111 g005
Figure 6. Prediction for (a) Ping An Insurance (Group) Company of China, Ltd. and (b) Zijin Mining Group Co., Ltd.
Figure 6. Prediction for (a) Ping An Insurance (Group) Company of China, Ltd. and (b) Zijin Mining Group Co., Ltd.
Electronics 13 04111 g006
Table 1. Information of literature review.
Table 1. Information of literature review.
ModelInputsStrengthsWeaknesses
ARCH/GARCH/…Single featureProvides relatively accurate predictionsRequires manual feature selection
SVM/RF/KNN/…Multiple featuresEffectively captures nonlinear relationships between featuresStruggles with highly complex nonlinear data
LSTM/CNN-LSTM/ALSTM/…Multiple featuresCaptures temporal dependencies wellCannot utilize multi-level features
ATFN/SFM/…Multiple featuresCaptures periodicity, trends, and multi-scale patternsFails to capture interrelationships between features
CMoEs/Att-MoE/…Multiple featuresCombines results from multiple expert networksNot widely applied in the finance sector
Table 2. Dataset information.
Table 2. Dataset information.
IndustryStocksTraining SetValidation SetTest Set
Finance3831,42638383838
Lifestyle Services3629,77236363636
Pharmaceuticals and Biotechnology3428,11834343434
Electronic Technology3226,46432323232
Chemical Industry3024,81030303030
Manufacturing and Transportation2923,98329292929
Information Technology2218,19422222222
Construction1915,71319191919
Table 3. Features used in the proposed model.
Table 3. Features used in the proposed model.
TypeFeatures
Fundamental trading attributesOpening Price, Closing Price, Highest Price, Lowest Price, Trading Volume, Transaction Value, Turnover Rate, Price Change Percentage, Price Change Amount, Amplitude and Pivot
Technical indicatorsMA5, MA30, EMA5, EMA30, RSI14, MOM14, OBV, WILLR, CMO14
Table 4. Experiment results of chemical industry and electronic technology.
Table 4. Experiment results of chemical industry and electronic technology.
ModelChemical IndustryElectronic Technology
MAERMSEMAPEMAERMSEMAPE
MA0.40790.73081.84671.26872.47552.3378
ARIMA0.41570.74101.87251.30422.51162.4076
ES0.39280.70401.77601.21972.35182.2518
lightGBM0.78171.86092.81791.06742.67822.8231
SVM0.50061.37102.55000.77112.15572.8213
RF0.59281.49522.75480.95902.13072.7948
CNN-LSTM0.48800.87552.82092.52693.61652.7770
DTML0.55780.98441.73091.01001.87432.5719
LSTM-BN0.20880.45171.70281.21742.15482.5328
MDML0.10270.14591.68720.82341.80062.2485
The bold text indicates the model with the best performance on a specific metric, while the underlined text indicates the second-best model.
Table 5. Experimental results of information technology and lifestyle services.
Table 5. Experimental results of information technology and lifestyle services.
ModelInformation TechnologyLifestyle Services
MAERMSEMAPEMAERMSEMAPE
MA1.08782.92882.81321.51864.85751.8150
ARIMA1.12002.98032.81941.50124.95191.8251
ES1.05172.75992.63051.40414.45681.7240
lightGBM0.93102.20633.04211.30213.77004.8085
SVM0.70521.83053.88330.96293.16423.6098
RF0.72440.96162.59480.65432.13952.6010
CNN-LSTM0.46210.83273.66090.70241.39083.4719
DTML0.58230.68543.23090.67041.97332.7539
LSTM-BN0.51650.95172.93280.84421.91482.9648
MDML0.46890.81642.15340.64611.01782.6474
The bold text indicates the model with the best performance on a specific metric, while the underlined text indicates the second-best model.
Table 6. Experimental results of manufacturing and transportation and pharmaceuticals andbiotechnology.
Table 6. Experimental results of manufacturing and transportation and pharmaceuticals andbiotechnology.
ModelMfg. & TransPharm. & Biotech.
MAERMSEMAPEMAERMSEMAPE
MA0.63801.4411.76891.07081.96952.0742
ARIMA0.61461.35781.75511.06981.96882.0773
ES0.59911.34981.66701.01411.90271.9762
lightGBM0.49441.26182.69412.69986.01715.9236
SVM0.45310.60331.56011.50482.45744.7238
RF0.39011.06982.04902.19835.26354.7586
CNN-LSTM0.36741.00102.02342.22385.30724.8086
DTML0.38100.98451.23461.94563.68833.3224
LSTM-BN0.50871.85962.03281.84653.65494.5415
MDML0.30140.57121.79070.55651.25451.7354
The bold text indicates the model with the best performance on a specific metric, while the underlined text indicates the second-best model.
Table 7. Experimental results of finance and construction.
Table 7. Experimental results of finance and construction.
ModelFinanceConstruction
MAERMSEMAPEMAERMSEMAPE
MA0.16230.29361.32870.19890.34651.8554
ARIMA0.16570.29951.34500.20080.34921.8639
ES0.15260.27901.24190.18460.32591.7004
lightGBM0.10570.28021.01620.75421.67497.6478
SVM0.05340.17580.60860.60851.43186.1272
RF0.05400.17660.60610.40590.54432.1129
CNN-LSTM0.12020.22651.65420.38740.54251.5445
DTML0.18230.28143.24550.53610.57332.3559
LSTM-BN0.13440.39991.23570.58780.81092.4365
MDML0.06790.09381.10580.10930.20931.1401
The bold text indicates the model with the best performance on a specific metric, while the underlined text indicates the second-best model.
Table 8. Ablation analysis of MDML model.
Table 8. Ablation analysis of MDML model.
ModelFinanceConstruction
MAERMSEMAPEMAERMSEMAPE
my-model-10.07860.13851.13260.19640.27371.2735
my-model-20.09060.24981.43580.20560.40511.8960
my-model-30.09790.20321.32540.30310.43961.9256
MDML0.06790.09381.10580.10930.20931.1401
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dong, Y.; Hao, Y. A Stock Prediction Method Based on Multidimensional and Multilevel Feature Dynamic Fusion. Electronics 2024, 13, 4111. https://doi.org/10.3390/electronics13204111

AMA Style

Dong Y, Hao Y. A Stock Prediction Method Based on Multidimensional and Multilevel Feature Dynamic Fusion. Electronics. 2024; 13(20):4111. https://doi.org/10.3390/electronics13204111

Chicago/Turabian Style

Dong, Yuxin, and Yongtao Hao. 2024. "A Stock Prediction Method Based on Multidimensional and Multilevel Feature Dynamic Fusion" Electronics 13, no. 20: 4111. https://doi.org/10.3390/electronics13204111

APA Style

Dong, Y., & Hao, Y. (2024). A Stock Prediction Method Based on Multidimensional and Multilevel Feature Dynamic Fusion. Electronics, 13(20), 4111. https://doi.org/10.3390/electronics13204111

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop