Next Article in Journal
Binary Decision Support Using AHP: A Model for Alternative Analysis
Previous Article in Journal
A Review of Key Signal Processing Techniques for Structural Health Monitoring: Highlighting Non-Parametric Time-Frequency Analysis, Adaptive Decomposition, and Deconvolution
Previous Article in Special Issue
Solution of Bin Packing Instances in Falkenauer T Class: Not So Hard
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Weight Assignment-Enhanced Convolutional Neural Network (WACNN) for Freight Volume Prediction of Sea–Rail Intermodal Container Systems

1
Xiangyang Sub-Center of National Engineering Research Center for Water Transport Safety, Hubei University of Arts and Science, Xiangyang 441053, China
2
School of Automotive and Transportation Engineering, Hubei University of Arts and Science, Xiangyang 441053, China
3
Department of Civil and Environmental Engineering, National University of Singapore, Singapore 119077, Singapore
4
Hubei Institute of Logistics Technology, Xiangyang 441100, China
5
Xiaohe Lingang Development Group Co., Ltd., Xiangyang 441401, China
*
Author to whom correspondence should be addressed.
Algorithms 2025, 18(6), 319; https://doi.org/10.3390/a18060319
Submission received: 24 April 2025 / Revised: 17 May 2025 / Accepted: 26 May 2025 / Published: 27 May 2025
(This article belongs to the Special Issue Hybrid Intelligent Algorithms (2nd Edition))

Abstract

:
In order to integrate the use of transportation resources, develop a reasonable sea–rail intermodal container transportation plan, and achieve cost reduction and efficiency improvement of the multimodal transportation system, a method for predicting the daily freight volume of sea–rail intermodal transportation based on a convolutional neural network (CNN) algorithm is proposed and a new feature processing method is used: weight assignment (WA). Firstly, we use qualitative methods to preliminarily select the indicators, and then use multiple interpolation to fill in the missing raw data. Next, Pearson and Spearman quantitative analysis methods are used, and the analysis results are grouped using the k-means, with the high correlation groups assigned high weights. Next, we use quadratic interpolation to obtain the daily data. Finally, a weight assignment-enhanced convolutional neural network (WACNN) model and seven other mainstream models are constructed, using the Yingkou port container throughput prediction as a case study. The research results indicate that the WACNN prediction model has the best performance and strong robustness. The research results can provide a reference basis for the planning of sea–rail intermodal container transportation and the allocation of transportation resources, and achieve the overall efficiency improvement of logistics systems.

1. Introduction

1.1. Background and Motivation

Multimodal transport, through the combination of various modes of transport, makes full use of the advantages of multiple modes of transport to complete door-to-door transport services. Sea–rail intermodal transport, an important subset of intermodal transport combining water and rail transportation, can greatly reduce the transit time of goods, simplify the transport process, and improve the efficiency of transport services [1]. In recent years, traffic and transportation networks have been getting better and better, and the mileage of highways and railroads is increasing. At the same time, the container throughput of railroads and waterways is also showing an upward trend, and the market scale of China’s smart port industry is developing rapidly [2,3,4]. The growth volume of the number of containers and public railroad mileage is shown in Figure 1.
However, due to weather, policy, and other reasons, sea–rail transportation faces inefficiencies, and these problems lead to the accumulation of cargo in ports and the waste of energy [5]. In response to this problem, a reasonable container throughput forecast could enable a port to layout scheduling in advance and improve the overall efficiency of the port, which could effectively solve this problem. Considering the high-dimensional nonlinear coupling effect and spatiotemporal dynamic characteristics among the multidimensional influencing factors affecting the throughput of sea–rail intermodal containers, traditional time series models that rely solely on historical throughput data have low prediction accuracy. Therefore, this article proposes a prediction method based on multi-indicator feature fusion to improve the accuracy of container throughput prediction.

1.2. Multimethodological Approach and Gaps

The forecasting of container throughput is a typical time series forecasting problem. Changes in a time series are not only influenced by their own historical values, but also closely related to external factors [6]. However, most of the existing studies have focused on the unidimensional correlation between a port’s hinterland economy and the container throughput, and lack a systematic consideration of the synergistic mechanism of multiple factors. But, a reasonable selection of the indicator data can effectively improve the predictive effect of the model [7]. Based on this, how to select suitable index data is one of the problems to be solved. And reasonable data processing is conducive to the improvement of the prediction accuracy of the model. Therefore, after obtaining the indicators, data enhancement methods should be used to improve the data quality.
In the past decade or so, many scholars at home and abroad have tried to use traditional statistical models [8,9], machine learning models [10,11,12,13], and combined models [14,15,16,17] for regression prediction and have achieved fruitful research results. However, statistical models only model linear relationships and are insensitive to nonlinear trends. And they rely on smoothness assumptions and require manual adjustment of the difference and seasonal cycle parameters. Therefore, in recent years, researchers have focused their attention on ML and DL models. For example, Temizceri [18] used various machine learning methods to predict the CO2 emissions from intermodal transport and to optimize transport planning. Bassiouni [19] used four deep learning architectures for recognizing the order status in complex supply chains. It was found that deep learning prediction models work better than machine learning prediction models. Gao [20] used a CNN prediction model and two common machine learning prediction models to predict highway–railway-level crossing accidents, and compared their prediction performances. It was found that the CNN prediction model had a better prediction performance. However, in time series forecasting problems, researchers have focused their goals on traditional forecasting models, and less research has been conducted on deep learning models. Based on this, how to apply deep learning-related models to container throughput forecasting problems is what we need to consider.
In summary, the current research gap can be summarized in two parts. The first is data selection and processing, and the second is the selection of the prediction models. Regarding the data selection and processing, single data decomposition is commonly used for regression prediction. On the contrary, this article uses processing methods, such as multi-index data upsampling and weight assignment. In terms of model selection, unlike previous studies that have used statistical models to predict container throughput, this paper applies a CNN model to time series prediction problems through multi-index time series data.

1.3. Contribution and Paper Organization

Regression prediction using deep learning models can effectively improve prediction accuracy, but there are still some problems. The first is the data level, with a variety of data metrics, and unsuitable data can lead to lower prediction accuracy. Secondly, there is the feature engineering aspect. Considering that deep learning models require more data, we need to choose a suitable method for data enhancement. Finally, the problem of container throughput prediction using DL is faced with a complex network construction and parameter selection, which requires the construction of a suitable structure for the specific problem.
These problems make the container throughput prediction problem difficult to solve. In order to solve these problems and improve the accuracy of container throughput prediction, first of all, for the selection of the data indicators, we summarize the selected indicators in existing container throughput prediction problems and obtain a set of division standards by generalization. Next, we group the indicators and assign weights, and at the same time increase the frequency of the data to achieve the goal of improving the quality of the data. And then, in order to select the appropriate network structure as well as the parameter combinations, we construct multiple sets of network structures as well as parameter combinations. Finally, in order to validate the WACNN model’s effect, multiple mainstream DL and ML models are used for validation and comparison. The main contributions of this study are summarized as follows:
  • Firstly, a preliminary screening of the indicators based on qualitative analysis methods, completion of the selection of indicators by formulating scientific division standards, and using missing value analysis to exclude the variables with too high a missing rate; then, using multiple interpolation to fill in the missing data.
  • For the feature engineering, we perform a Pearson as well as a Spearman correlation analysis on the selected indicators. Based on the correlation results obtained, the metrics are categorized into multiple groups using a k-means clustering analysis and assigned appropriate weight sizes based on their combined weights. At the same time, the dataset size is increased horizontally using quadratic interpolation.
  • In order to select the appropriate model structure, we propose a variety of different network structures and select the most appropriate network structure as well as parameter selection through a comparative analysis. At the same time, in order to verify the prediction effect of the WACNN model, this paper uses a variety of mainstream DL as well as ML models to verify that the WACNN model has the best prediction performance.
The rest of the paper is organized as follows: Section 2 briefly reviews the relevant literature on indicator selection, feature engineering, and DL prediction. Section 3 is an introduction to the methodology, including the indicator selection, feature engineering, WACNN model construction, and evaluation index system selection. The data processing; constructing the complete network structure; and finally, conducting the experiments are shown in Section 4. Finally, in Section 5, the research results are summarized.

2. Literature Review

Through the above analysis of the container throughput forecasting problem, we find that it is very important to improve the accuracy of container forecasting, but there are still some unsolved technical problems. Therefore, we analyze the existing studies in depth by reviewing the literature.

2.1. Indicator Selection

The research on the selection of the indicators for container throughput forecasting has experienced an evolution from a single economic driver to the integration of multiple dimensions [21]. Early studies generally focused on the correlation between macroeconomic indicators and socially productive activities, such as the GDP; the volume of import and export trade; investment in fixed assets; and the output value of secondary industries as the key influencing factors [22]. However, these studies were mostly limited to economic variables, ignoring the impact of other port factors on the throughput. In recent years, scholars have begun to try to construct more complex indicator systems to solve this problem, such as using dimensionality reduction techniques, such as principal component analysis (PCA) and dynamic factor analysis (DFA), to screen the economic indicators with high information density [23,24]. This can significantly reduce the data redundancy while ensuring prediction accuracy. Other researchers have gradually incorporated port-specific operational indicators into the model by studying the port-influencing factors [25], providing a new perspective for explaining throughput differences among ports in the context of regional economic homogenization. Although researchers have started using various indicators for regression prediction, there is no standardized system for the selection of indicators. Therefore, it is not comprehensive enough. In summary, we should summarize the classification criteria in order to select the indicators more reasonably.

2.2. Feature Engineering

For time series forecasting research, the existing methods mostly mine the time series data features through decomposition techniques; for example, decomposing the time series data into multiple sets of feature data and then using correlation regression models for forecasting [26]. However, such methods neglect the optimization value of the data density. Studies have shown that enhancing the temporal frequency can improve the prediction accuracy more effectively than traditional data enhancement means [27]. Based on this, this study proposes to convert low-frequency data into high-frequency sequences by means of a quadratic interpolation algorithm, which directly increases the size of the dataset in order to enhance the data. In terms of feature processing, in addition to increasing the dataset, WA has been proven to improve the prediction accuracy of a model [28]. WA strengthens the impact of the key features and weakens noise interference by quantifying the indicator correlations and setting the weighting coefficients in a hierarchical manner, resulting in higher-quality input data. As for data grouping, correlation analyses, such as Pearson, have commonly been used to detect the magnitude of the correlation between indicators [29]. Meanwhile, a k-means clustering analysis is computationally efficient, works better in low-dimensional grouping problems, and the results can be visualized intuitively [30]. Combining these methods, we first use two correlation indicators, Pearson and Spearman, to check the correlation size between each group of indicators and the container throughput, and then use a k-means to group them together and assign different weights to each group in accordance with the combination of weights. Finally, through the synergistic optimization of data upscaling and feature weighting, a high-quality dataset adapted to this time series prediction task is constructed to provide better input data for the deep learning model.

2.3. Prediction Model Selection Strategy

Deep learning models can not only effectively fit complex nonlinear relationships, but also avoid overfitting problems in shallow structures [31]. In recent years, many researchers have applied deep learning models to time series prediction problems in the transportation field and achieved fruitful results. For example, for traffic flow prediction, Gao introduced a spatiotemporal dynamic model (ST) to integrate the spatial correlation and temporal evolution of traffic flow to capture the dynamic interactions of different nodes in a road network [32]; Pan combined basic graph (FD), Markov, and LSTM models to achieve the progressive prediction of traffic flow [33]; and Zhang designed CASAformer for low-speed congestion scenarios for traffic speed prediction, using a sparse attention mechanism to improve the prediction accuracy [34]. These achievements indicate that the success of deep learning in the field of transportation is due to its ability to mine the complex patterns in high-dimensional spatiotemporal data, especially its ability to dynamically model local features.
However, in the field of container throughput prediction, the application of deep learning has not yet formed a scale. Unlike traffic flow forecasting, container throughput is dominated by long-term global factors, such as macroeconomic cycles and international trade policies, and its data often exhibit strong trends, low-frequency, and high-noise characteristics. Traditional statistical models are still widely used due to their ability to capture the explicit seasonality and linear trends [35]. However, these methods do not take into account the complexity of and variability in the data caused by external environmental changes. Fortunately, deep learning can explore the intrinsic patterns in data and analyze their features [36]. In recent years, scholars have attempted to break through this limitation. For example, Cui [37] and Hirata [38] demonstrated, through comparative experiments, that LSTM outperforms SARIMA and Prophet in most scenarios. Cuong et al. [39] combined a discrete wavelet transform (DWT) with LSTM to validate the effectiveness of deep learning for throughput regression prediction. Liu [40] further introduced the BiLSTM model to improve prediction accuracy using bidirectional time series modeling. However, the existing research has mostly focused on RNNs and their variants, which can capture temporal dependencies but lack the ability to extract local features.
To compensate for this deficiency, some scholars have turned to models with stronger feature extraction capabilities, such as CNNs and GNNs. In practice, Zhang proposed a dynamic graph attention multi-attention network (DGAT-MAN), which significantly improved the prediction performance by modeling the spatiotemporal correlations through graph structures [41]. But such methods rely on precise node relationship definitions. In multi-index time series scenarios, the lack of prior knowledge about node relationships can easily lead to a decrease in model performance. Therefore, this article chooses to apply a CNN model to predict the throughput of sea–rail intermodal containers, as its convolutional kernel can adaptively learn the local correlation patterns between multiple indicators. By extracting the features through a multi-layer network structure, high prediction accuracy can be achieved.

2.4. Summary

Based on the above literature review summary, we use the multiple-input single-output method for container throughput prediction; specifically, by summarizing the indicators used by researchers to derive the division criteria, as a way to select the indicators, and then pre-processing the data. To enhance the data, we use Pearson and Spearman analyses to explore the relationship between each set of metrics and the container throughput, and use this for a k-means grouping of the dataset for a more rational WA. Furthermore, to increase the dataset size, we use the quadratic interpolation method. For model selection, we try to use a CNN model for the container throughput prediction. The following Table 1 shows a comparison between the models that have been used for the container throughput prediction problem and that used in this paper.

3. Methodology

3.1. Problem Statement

Accurate prediction of port container throughput is a key link for optimizing intermodal cooperative operation. In order to break through the limitations of traditional forecasting methods, this paper proposes a data-driven forecasting framework to improve the accuracy of throughput prediction by constructing a multidimensional time series regression model, so as to optimize the operational efficiency of ports and promote the efficient collaboration between logistics chains and suppliers. To this end, we design a multidimensional indicator screening framework and innovatively propose a WACNN prediction model. The model takes the screened and data-enhanced multi-source time series indicator data as input, and makes a regression prediction of target data through multiple indicators. The input matrix is shown below:
X 1 , X 2 ,   , X m = x 11 x 12 x 1 m x 21 x 22 x 2 m x n 1 x n 2 x n m
where X 1 , X 2 , , X m represent m sets of indicator data, and x 11 , x 12 , , x n m represent m sets of indicators, each with a total of n days of data.

3.2. Feature Selection Method

  • Step 1. Preliminary selection of indicators
By summarizing the literature, we systematically categorize and establish four core evaluation dimensions—transportation capacity, infrastructure investment, economic development level, and international openness—thereby constructing a multidimensional assessment framework. Considering regional developmental disparities, we implement a stratified indicator pool strategy for flexible metric adaptation. Specifically, the transportation capacity dimension focuses on freight volume and turnover rates across highway, waterway, and railway networks. Infrastructure investment incorporates fixed-asset indicators from key projects of road, rail, and port construction. Economic development metrics integrate GDP aggregates and industrial structure parameters. International openness evaluation encompasses total import–export volume. Based on this framework, we then collect indicator data.
  • Step 2. Data processing
For the collected data, we use a more robust multiple imputation (MI) to complete the data.
The Imputation Phase (IP) is the process of constructing a probability model and generating m complete datasets based on observed data. Assuming the interpolation depends on the fully observed covariate X and partially observed Y , the specific formula is as follows:
Y miss ( k ) P ( Y miss | Y obs , X ) , k = 1 , 2 , , m
where Y obs is the observed data, Y miss is the missing data, m is the number of interpolated datasets, and P ( · ) is the model-based predicted distribution. After applying multiple imputation techniques, complete monthly data for multiple indicators are derived. This is subsequently followed by outlier detection via the interquartile range (IQR), calculated as follows:
IQR = Q 3 Q 1
where Q 1 is the first quartile, which is the 25th percentile; and Q 3 is the third or fourth percentile, which is the 75th percentile. From this, it can be concluded that the boundary of outliers is
Q 1 1.5 × IQR , Q 3 + 1.5 × IQR
  • Step 3. Correlation analysis
To validate the correlation between the collected data and container throughput, we employ Pearson’s correlation coefficient for the primary correlation analysis and Spearman correlation coefficient for supplementary verification. The formula is as follows:
γ = i = 1 n X i X ¯ Y i Y ¯ i = 1 n X i X ¯ 2 i = 1 n Y i Y ¯ 2 = Cov ( X , Y ) σ X σ X
where n represents the sample size; X i , Y i are the observed values of the variables; and X ¯ , Y ¯ are the means of the variables. In the second expression, C ov ( X , Y ) represents the covariance between X and Y , and σ X and σ Y represent the standard deviation of X , Y . Spearman’s rank correlation coefficient is the process of converting raw data into ranks and calculating the Pearson correlation coefficient of ranks. The specific formula is as follows:
ρ = Cov rank X , rank Y σ rank X σ rank Y
where rank X and rank Y are the sorted position numbers.
  • Step 4. K-means
Next, in this section, in order to more rationally divide the data into multiple groups and give appropriate weights, we use k-means cluster analysis to deal with the correlation of these indicators. Firstly, the objective function of k-means is defined, and here the correlation distance is used with the following formula:
J = i = 1 k x C i d x , μ i
where J indicates the total distance within the cluster, which needs to be minimized. k denotes the number of clusters. C i indicates i cluster, x is a normalized data point, and μ i denotes the mean vector within the i cluster. d x , μ i denotes the distance associated with data point x from center of mass μ i , which involves the previously mentioned Pearson correlation, and is formulated as follows:
d x , y = 1 γ
Next, the contour coefficient is chosen as the evaluation criterion, with s ( x ) denoting the contour coefficient of data point x , and the overall contour coefficient is the mean of all data points s ( x ) for assessing the clustering quality. The specific formula is as follows:
s ( x ) = b ( x ) a ( x ) max { a ( x ) , b ( x ) }
where a ( x ) = 1 C i 1 y C i , x y d ( x , y ) denotes the average distance of data point x from other points in the same cluster, and b ( x ) = min C k C i 1 C k y C k d ( x , y ) denotes the average distance from data point x to the nearest other cluster. Using k - means + + initialization, initialized centers of mass are selected through a probability distribution to maximize the center of mass spacing, where the first center of mass is chosen randomly and the probability of subsequent center of mass selection is inversely proportional to the distance squared with the following formula:
P ( x ) = D ( x ) 2 x X D ( x ) 2
where D ( x ) denotes the minimum distance from the current data point to the selected center of mass and X denotes all data points in the data set.
  • Step 5. Upsampling
Interpolation, as a meaningful input transformation, is based on the basic principle of using mathematical methods to reasonably infer the values of unknown points between known data points, thereby transforming the data into a form more suitable for model processing. We use the interpolation method to convert monthly data into daily data. Although the statistical characteristics of interpolated data differ from real daily data, its advantages far outweigh the possible disadvantages. The most important aspect is the adaptability of the data. Increasing the amount of data is beneficial for deep learning models and machine learning models to achieve better learning results, while also alleviating the problem of small samples and reducing the risk of overfitting. On the other hand, linear interpolation not only maintains the overall monthly trend but also introduces small fluctuations to simulate natural changes in actual scenarios.
To enhance data diversity and yield more accurate prediction results, quadratic interpolation is applied to the data, transforming the monthly data into daily data. Assuming the time point of monthly data is the first day of each month, the value is y 0 , y 1 , , y n 1 ; the corresponding time coordinate is x 0 , x 1 , , x n 1 ; and x 0 indicates the first day of the first month, x 1 indicates the first day of the second month, and so forth. Based on all data points ( x 0 , y 0 ) , ( x 1 , y 1 ) , ( x n 1 , y n 1 ) , divided into multiple overlapping three-month windows, with each window containing three consecutive points, the kth window is ( x k , y k ) , ( x k + 1 y k + 1 ) , ( x k + 2 , y k + 2 ) ; and k = 0 , 1 , , n 3 .
Construct a quadratic polynomial— y ( x ) = a x 2 + b x + c —the formula is as follows:
a k x k 2 + b k x k + c k = y k a k x k + 1 2 + b k x k + 1 + c k = y k + 1 a k x k + 2 2 + b k x k + 2 + c k = y k + 2
Solve for the coefficient
a k = ( y k + 2 y k + 1 ) ( x k + 1 x k ) ( y k + 1 y k ) ( x k + 2 x k + 1 ) ( x k + 2 x k + 1 ) ( x k + 1 x k ) ( x k + 2 x k ) b k = ( y k + 1 y k ) a ( x k + 1 2 x k 2 ) x k + 1 x k c k = y k a k x k 2 b k x k
For any target point x , the interpolation result is
y ( x ) = a k x 2 + b k x + c k
After generating daily data using the above formula, calculate the total difference and the corresponding scaling factor for each month to adjust the daily value. Assuming that interval x s t a r t , x e n d corresponds to raw data y t o t a l , the calibration formula is as follows:
y c a l i b r a t i o n x = y t o t a l x = x s t a r t x e n d x s t a r t y ( x ) y ( x )
  • Step 6. Weight assignment
In neural networks, when the dimensional difference between input data is significant, models tend to focus solely on indicators with large values while overlooking some crucial ones with relatively small values. Data normalization is the process of scaling data proportionally to a specific range, which is mainly used to eliminate dimensional differences and thereby enhance a model’s performance. Therefore, data normalization is generally required before feeding the data into the neural network. Here, the min-max normalization method is adopted and the specific formula is as follows:
X n o r m = X X min X max X min
where X n o r m is the normalized value, X min is the minimum value in a set of data, and X max is the maximum value in a set of data. The data are effectively linearly mapped onto the [0, 1] interval through Min-Max normalization.
The weight allocation methodology is implemented through correlation-based analysis and k-means. During data normalization, indicators are ranked and grouped according to their correlation coefficients with container throughput, assigning higher weights to strongly correlated groups. The input matrix obtained after assigning weights is as follows:
X ( 1 , 1 ) , X ( 2 , 2 ) ,   , X ( k , m ) = X 11 X 12 X 1 m X 21 X 22 X 2 m X n 1 X n 2 X n m
where m represents the number of indicators, k represents the number of groups sorted by relevance, X ( k , m ) represents the m indicator and the k group, and n represents the total number of days of data for each indicator. In the formula, X 11 , X 12 , X 21 , , X m n are represented as
X 11 = ω 1 ( X X min ) X max X min X 21 = ω 1 ( X X min ) X max X min X 12 = ω 2 ( X X min ) X max X min X m n = ω k ( X X min ) X max X min
where ω 1 , ω 2 , , ω k represent the weight multiplied by each group during normalization, from the first group to the k group.
For the calculation of weights, we determine the final weight size based on the summed percentage of Pearson as well as Spearman correlations for each group after grouping. The specific formula is as follows:
ω k = i = 1 n ( γ i + ρ i ) i = 1 m ( γ i + ρ i )
where n is the number of indicators in group k and m is the total number of indicators. Through the above calculation, we can obtain the specific weights of each correlation group.

3.3. CNN Architecture Design

Convolutional neural network prediction models usually use multiple hidden-layer structures; multiple hidden-layer convolutional neural networks can extract more complex features, and their representation of the input data is stronger. The hidden layers contain multiple convolutional layers and multiple pooling layers, and the pooling operation is chosen here to use maximum pooling. The ReLU activation function is chosen here for the activation function. The dropout function is introduced before full connectivity to prevent overfitting of the model. The construction of the network model is divided into three main parts: the construction of the convolutional layer, the construction of the pooling layer, and the construction of the fully connected layer. The model equations for the construction of each layer are as follows.
O c = f c ( x t W c + b c )
where O c is the feature outputs obtained from the convolution kernel after convolution, represents the convolution operation, W c is the vector of weights, b c is the bias vector, and f c is the activation function.
O p = Maxpooling ( O c ) = Maxpooling { f c ( x t W c + b c ) }
where O p is the feature outputs obtained from the convolution kernel after pooling, and Maxpooling   ( ) represents maximum pooling operations.
O d = f d ( O p W d + b d )
where O d is the feature outputs obtained from the convolution kernel after full connectivity, W d is the vector of weights, b d is the bias vector, f d is the activation function. They use the ReLU activation function. The CNN schematic is shown in Figure 2.

3.4. Evaluation

This paper uses five evaluation metrics to comprehensively evaluate the model, namely root mean squared error (RMSE), mean absolute error (MAE), mean bias error (MBE), mean absolute percentage error (MAPE), and r-squared (R2). Among them, RMSE and MAE can measure the size of the error between predicted values and true values. MAPE measures the relative size of prediction error in percentage form. MBE can measure the systematic bias of predicted values. R2 is used to measure the explanatory power of a model for data changes, reflecting the goodness of fit of the model. The five evaluation indicators are as follows.
f R 2 = i = 1 n y i y ^ i ² i = 1 n y i y ¯ ²
f RMSE = i = 1 n y i y ^ i 2 n
f MAE = i = 1 n y i y ^ i n
f MBE = i = 1 n y ^ i y i n
f MAPE = 100 % n i = 1 n y i y ^ i y ^ i
where y i is the measured value, y ^ i is the predicted value, y ¯ is the mean measured value, and n is the number of samples.

3.5. The Framework of This Paper

In this paper, the overall framework of the proposed WACNN model for predicting container throughput is shown in Figure 3.
(1)
The framework starts by using four classification criteria to preliminarily select evaluation indicators. Then, it collects monthly data for each indicator group, conducts missing value analysis and outlier analysis, and uses multiple imputation method to complete the data.
(2)
Next, it conducts Pearson and Spearman correlation analyses and uses k-means to group the correlations. Then, it assigns higher weights to the group with higher correlation during normalization.
(3)
It uses quadratic interpolation to convert the collected monthly data into daily data.
(4)
It builds CNN models separately, and then inputs the WA data into the constructed CNN.
(5)
It compares the proposed WACNN with various related models and mainstream models, and validates the analysis using five evaluation criteria.

4. Case Study

In Section 4.1, we analyze and explain the reasons for the selections made in this study. In Section 4.2, we describe the indicator selection methodology, and the specific methods for data processing. In Section 4.3, we detail the overall architecture of and the parameter selection for the seven models. In Section 4.4, we conduct experiments by comparing all the regression models and show the prediction results for each model. Finally, in Section 4.5, we summarize the analysis of the results.

4.1. Data Description

As one of the important ports in China, Yingkou City in Liaoning Province collects and disperses a large number of containers every day. Therefore, container throughput forecasting for Yingkou port is an important logistics and transportation management task. Container transportation involves multiple countries and faces a complex operating environment as well as various uncertainties, which not only increases the complexity of forecasting sea–rail intermodal container transportation for Yingkou port, but also reduces the forecasting performance. Therefore, in order to more accurately predict the container throughput of sea–rail intermodal transportation at Yingkou port, it is necessary to construct a WACNN container throughput prediction model by taking into account the quality of the container throughput data, the correlation between the selected features, and the complexity of the model building to improve the prediction accuracy.

4.2. Data Processing

A combination of quantitative and qualitative analyses was used here for data selection. Qualitative methods were used first. The time series data on sea–rail intermodal container throughput for Yingkou port in Liaoning were selected as the research object for feature selection, and the data were obtained from Liaoning Statistical Yearbook, National Bureau of Statistics, with the relevant statistical data provided by officials for the years from 2014 to 2024. Here, the corresponding evaluation indicators were selected based on four division criteria: the transportation capacity, fixed-asset investment and construction, the economic development level, and international trade and openness. The selection of the evaluation indicators for the four divisional criteria took into account all the public data and indicators for Yingkou City, Liaoning Province, and the whole country, and 33 evaluation indicators were obtained after the statistical organization. The evaluation indicators used under each division standard are shown in Table 2.
The raw data were analyzed and some missing data were found. For the missing data, multiple interpolation was used to interpolate the missing data values. The fixed-assets investment in road, water, and land transportation in Yingkou and Liaoning, which had more missing values, were eliminated. An outlier analysis was performed on the supplemented data; the outliers were excluded and we continued to use multiple interpolation to make up the difference until there were no outliers in the data.
The Pearson and Spearman correlation coefficients between each factor were obtained through the collated data. Because of the simultaneous use of multiple eigenvalues to predict the sea–rail intermodal container throughput, only the correlation coefficient values of the various influencing factors on the sea–rail intermodal container throughput were screened out, and the correlation coefficients of the influencing factors on the sea–rail intermodal container throughput were obtained after their collation, as shown in Table 3, which can be seen as the degree of influence of the different factors on the sea–rail intermodal container throughput at Yingkou port.
Based on the obtained Pearson and Spearman correlation indices, we grouped them using k-means. To visualize the results, we used a PCA and the results are shown in Figure 4.
As can be seen in the figure, the 31 sets of indicators are divided into two groups, and the distribution of the two types of data are relatively concentrated, and there is a certain interval between the clusters, indicating that the clustering effectiveness is better. In order to analyze the clustering results in more detail, we plotted the two types of variables and two clusters into a bar chart and a scatter plot, as shown in Figure 5.
Firstly, in terms of the distribution of variables, there is less overlap in the Pearson correlation distribution in the upper-left corner and the Spearman correlation distribution in the lower-right corner, which indicates that the two clusters have more significant differences in their distribution of the correlation indexes, and that the clustering can effectively differentiate between the data with different correlation characteristics. Second, in terms of the variable correlation, in the scatter plots in the upper-right and lower-left corners, the blue dots and orange dots form relatively independent aggregation areas, indicating that the two clusters, after clustering, have more obvious differences in the association relationship between the Pearson correlation and Spearman correlation, and the clustering is able to effectively differentiate between the different types of data, and the clustering effect is better. The clustering situation is shown in Figure 6.
Bringing the correlation metrics into the equation yielded a weight ratio of 0.39285:0.60715 for the two groups, which was approximated as 1:1.5. The screened eigenvalues were converted from low frequency to high frequency, using a quadratic interpolation to convert the eigenvalue monthly data into daily data. Finally, a 3834 × 31-sized data set consisting of 3834 days of data for 31 evaluation indexes, including the volume of the sea–rail intermodal container transport at Yingkou hub port, was obtained. The first and second sets of data were imported and the stacked diagrams of the imported data were plotted, as shown in Figure 7 and Figure 8.
The following figure shows the second set of stacked data, with the last indicator being the container throughput.
To sum up, the pretreatment in this paper consisted of the following steps. First, 33 indicators were selected according to the classification criteria, and then the indicator data with more missing values were excluded through a missing value analysis, and 31 indicators were obtained. Subsequently, we used multiple imputation and outlier analysis methods many times until there were no outliers in the data. Next, Pearson and Spearman correlation analyses were conducted on the 31 indicators obtained and the container throughput indicators for sea–rail intermodal transport. According to the results obtained, two groups of categories were obtained using k-means clustering. Then, for each group we calculated the sum of the Pearson and Spearman correlations to assign weights. Finally, the monthly data were upsampled into daily data using the quadratic interpolation method to obtain the final input data.

4.3. Parameter Selection

To better test the prediction accuracy and stability of the WACNN model, several models commonly used in regression prediction problems were selected. The CNN, LSTM, BiLSTM, and GRU models for deep learning, and the RF and SVR models for machine learning were used to make a regression prediction of the container throughput of sea–rail intermodal transportation at Yingkou port and compare them at the same time. The structure of the LSTM and BiLSTM models is shown in Figure 9.
−Dividing the data into a training set and a test set, using a portion of the data as the training set for training the model, and then comparing the results with those using the remaining portion of the data as the test set, the results predicted by the trained model were used to assess the accuracy of the model’s predictions. To be able to determine better parameters for training the model, several typical values were set as alternative values for each parameter, by referring to the parameter tuning methods in the existing literature. The number of convolutional layers in the CNN model was set as 1, 2, 3, 4, 5, and 6; the number of convolutional kernels per layer were selected as 16-16-16-32-64, 16-16-32-32-64, 16-32-32-32-64, 16-32-64-64-64, and 16-32-64-64-128; and the dropout was selected as 0, 0.1, 0.2, and 0.3. Through numerical experiments, the effects of the model on the RMSE and MAE of the validation set under different parameter configurations were compared, and the values of each parameter corresponding to the optimization of the evaluation index were selected, as shown in Figure 10.
Because an increase in the number of layers can lead to overfitting of the model, only a single LSTM layer model and double LSTM layer model were created in the LSTM model. The number of neurons was set to 10, 20, 30, 40, 50, 60, and 70 in the single LSTM; the first layer was set to 10, 20, and 30; and the second layer was set to 10, 20, 30, 40, 50, 60, and 70 in the double LSTM layer. Numerical experiments were conducted to compare the effects of the model on the RMSE and MAE of the validation set under different parameter configurations, and the values of each parameter corresponding to the optimum of the evaluation indexes were selected, as shown in Figure 11.
In the BiLSTM model, the neurons were set as 10, 20, 30, 40, and 50. The GRU model was set up with two network structures, a single layer and a double layer, and the parameters were set as 16, 32, 64, 128, 16-32, and 32-32, respectively. The comparison of the results is shown in Figure 12 and Figure 13.
In the RF, the number of decision trees was set to 50, 100, 200, 500, 800, and 1000, and the minimum number of leaves was set to 3, 4, 5, and 6. Numerical experiments were conducted to compare the effects of the model on the validation set’s RMSE and MAE under different parameter configurations and to select the values of each parameter corresponding to the optimization of the evaluation indexes, as shown in Figure 14.
In the SVR, the penalty factor was set to 6, 10, 20, 100, 200, and 500, and the parameters of the radial basis function were 0.2, 0.5, 0.8, and 1.0. Through numerical experiments, we compared the effects of the model on the RMSE and MAE of the validation set under different parameter configurations and selected the values of each parameter corresponding to the optimum of the evaluation indexes, as shown in Figure 15.
In summary, the overall architecture and parameter settings of the CNN model in the experiment are shown in Table 4.
As shown in Figure 7, the use of 40 neurons in the single layer model, and 20-30 and 30-40 neurons in the double layer model produced good results. However, it was found in the experiment that if the number of LSTM layers was less, it would result in the model not being able to fully learn the features of the data; so, from an overall consideration, 30-40 neurons in the two-layer model were selected as the optimal parameter here. The overall frameworks and parameter settings of the LSTM, BiLSTM, and GRU models are shown in Table 5.
The CNN model constructed according to the above process is shown in Figure 16.
In the RF model, 2800 sets of data were selected as the training sets, 1034 sets of data were selected as the test sets, the number of decision trees was set to 500, and the minimum number of leaves was set to 5. In the SVR model, 2800 sets of data were selected as the training sets, 1034 sets of data were selected as the test sets, the penalization factor was set to 500, and the parameter of the radial basis function was set to 0.9.
The network training method can be divided into two stages: signal forward propagation and error backpropagation. In the backpropagation process, the gradient descent method gradually approaches the value that minimizes the loss function by calculating the gradient at the current point and then updating the parameters along the opposite direction of the gradient. In this paper, the two most commonly used gradient descent algorithms, Adam and SGDM, are selected and compared. Using the constructed CNN model and the LSTM model, with both models trained 800 times, the two gradient algorithms, SGDM and Adam, are compared. The RMSE using the SGDM is 21.466, and the RMSE value using the Adam is 51.9197, from which can be seen that with the same choice of parameters, the SGDM is more effective than the Adam. In the LSTM model, the RMSE value of the Adam is 74.5857 and the RMSE value of the SGDM is 198.4812, so the SGDM is selected for the WACNN and CNN models and the Adam is selected for the LSTM, BiLSTM, and GRU models. Moreover, the hyperparameter selection of the BiLSTM and GRU is consistent with that of the LSTM.
To verify the stability of the model, we conducted a hyperparameter sensitivity analysis of the proposed model. The stepwise learning rate decay schedule was selected for use. This strategy balanced the need for rapid convergence in the early stages of training with fine tuning in the later stages by reducing the learning rate in stages. Therefore, the batch size, initial learning rate, and learning rate descent factor were selected for the sensitivity analysis. Among them, a batch size that is too small may introduce more noise, while a batch size that is too large may cause the model to become stuck in local solutions or have poor generalization ability. Choosing too small an initial learning rate and learning rate descent factor may lead to slow convergence, while ones too large may cause a sharp drop in the learning rate and miss the optimal solution. Based on the above considerations, the sensitivity analysis range of the selected hyperparameters is shown in Table 6.
Based on the above parameter combinations, numerical experiments were conducted to obtain the RMSE under different hyperparameter combinations, as shown in Figure 17.
According to the results in the figure, although the selected hyperparameters have a large range, the overall fluctuation is within an acceptable range. Therefore, the model as a whole has high stability. Finally, we selected the best set of hyperparameter combinations for the WACNN, CNN, and LSTM model hyperparameter settings, as shown in Table 7.

4.4. Experimental Results

We input the data into the WACNN, CNN, LSTM, BiLSTM, GRU, RF, SVR, and PLS models with the pre-set parameters. In the models used, the PLS serves as the baseline to validate the predictive performance of the various DL and ML methods. The comparison between the predicted values and the true values of each model on the test set is visualized in Figure 18.
A comparison of the prediction results of the seven models for the training and test sets is shown in Table 8.
From Table 8, the CNN and WACNN performed best on the test set, with RMSEs of 15.596 and 16.3521, respectively, and the WACNN outperformed the CNN, but its training and testing errors were 17.9% and 11.05%, respectively, indicating a certain degree of overfitting. The RF’s overfitting was serious, with a difference of 53.67%, and its generalization ability was poor, whereas the SVR, although stable, with a difference of only 1.04%, had a higher prediction error with an RMSE of 28.6886. The time series models, such as the LSTM, BiLSTM, and GRU, performed worse, with a higher RMSE. The PLS performed the worst.
In order to visualize the predicted values against the true values, we exported the predictions of the seven models and compared them with the true values. As a result of the large amount of data, we zoomed in on the last 100 days of data. The comparison between the predicted and real values of the seven models is shown in Figure 19.
The figure shows the prediction results of the various models for container throughput compared with the real value. Among them, the WACNN model learned the overall trend of the data very well, and the overall trend of the CNN model was more fitting, but with some fluctuations. Among the three time series models, the LSTM, BiLSTM, and GRU, the BiLSTM model had less fluctuations in relation to the data, and the LSTM and GRU models had more ups and downs and the trend was not very fitting. The two ML results were better than the time series models and their overall trends were smoother. It can be seen that the worst was the PLS, the traditional statistical model. Although the model was consistent with the true values in terms of the overall trend, its predicted results differed significantly from the true values.
The comparison of the values of the RMSE, R2, MAE, MAPE, and MBE for each model on the test set and the improvement rate of the CNN compared to the other models are shown in Table 9.

4.5. Cross-Validation

To further validate the robustness and generalization ability of the WACNN, we used a time series k-fold cross-validation. Choosing 6-fold here meant dividing the original dataset into six cycles of five years. We selected four years of data as the training set and one year of data as the testing set for each cycle, as shown in Table 10.
The partitioned dataset was input into the model, and the results are shown in Table 11.
According to the results of the six-fold cross-validation, it can be seen that the model has good prediction results for the six different datasets, and the errors on the training set and testing machine are within an acceptable range, proving that the proposed model has good stability and generalization ability. Meanwhile, the cross-validation results also demonstrate that the size and volatility of the data have a significant impact on the learning of the model.

4.6. Results Analysis

According to the comparative analysis of the models’ performance in the table, it can be seen that the WACNN and CNN models perform optimally in terms of prediction accuracy, with the WACNN having the highest R2 value of 0.99973, as well as the lowest RMSE and MAE of 15.596 and 11.8649, respectively, with an MBE of 0.1194 and a MAPE as low as 0.0081%, which indicates that the prediction results are close to the true values and have high robustness. The ML models RF and SVR have relatively good performance with a higher prediction accuracy. It is worth noting that the LSTM, BiLSTM, and GRU models have larger prediction errors, with a MAPE over 3%, reflecting that the temporal models may have had insufficient feature capture for this task. The worst performing model is PLS, with a test set RMSE of 286.2715 and a prediction accuracy of only 84%. Compared with the other ML and DL models, there is a significant gap, proving that the performance of these models is indeed better than that of traditional statistical models.
The excellent prediction results of the WACNN and CNN models are due to their convolutional structure that can effectively extract local features and capture spatial correlations in the data, while the WACNN further optimizes the feature learning process by introducing a WA. Unlike traditional CNNs that treat all input data equally, WACNNs focuses models more on the key information that is strongly related to the prediction target because the input data are given different weights. Because in time series or spatial data, certain local segments may contain more significant patterns or trends, the WA mechanism enhances the differentiation of feature representations by increasing the expression weight of these high-value features while weakening the interference of noisy or unimportant regions. In contrast, traditional CNNs lack this explicit weight regulation ability, which may lead to insufficient attention to key features or overfitting of redundant information. Therefore, WACNNs achieve more refined feature filtering and fusion through weight allocation to enhance features.
However, the poor prediction performance of the three time series prediction models is mainly due to the mismatch between their model structure and task characteristics. The core advantage of these three models is their capturing of long-term temporal dependencies, but the input data are relatively large and there are complex nonlinear concerns between the different indicator data, making it difficult for these models to effectively extract valuable information.
The RF and SVR, as traditional machine learning models, perform well on this regression prediction problem. The RF performs better in terms of the error stability, with an MAE and MAPE of 14.0925 and 0.85%, respectively, lower than those of the SVR, thanks to its ensemble learning mechanism’s robustness to noise and outliers. The R2 and RMSE of the SVR are 0.99913 and 28.6886, respectively, slightly better than those of the RF, indicating a strong ability to fit the overall trend of the data. However, its MAE is significantly higher, indicating that there are many large error outliers in the prediction results, which may be related to the sensitivity of the kernel function to high-dimensional features and overfitting to local extrema. The prediction results of both methods are inferior to those of the WACNN model, reflecting the limitations of traditional methods at feature automatic extraction and complex pattern capture, especially when dealing with high-dimensional or implicit spatial correlation data, where the limitations are more pronounced.
Among the compared models, the CNN model and its variants have higher prediction accuracy. Compared to models, such as the LSTM, it has significant improvements, especially for basic statistical models like PLS. Although the improvement of the WACNN is relatively small compared to the CNN, at ports with high container throughput, even a small improvement can have a significant impact due to its large base. On the other hand, even though the CNN model achieves high prediction accuracy, the WACNN can still achieve performance breakthroughs through weight allocation, demonstrating the potential space for model design.

5. Conclusions

This article takes Yingkou port’s sea–rail intermodal container throughput prediction as an example, selects 31 indicators related to sea–rail intermodal container throughput, supplements the collected monthly data using the multiple interpolation method, and then conducts Pearson and Spearman correlation analyses. Based on the correlation indicators, k-means grouping is used, and a WA is performed according to the total weight proportion. Next, the quadratic interpolation method is used to upgrade the monthly data into daily data, which horizontally improves the data quality. The WACNN model is used to solve this prediction problem. Simultaneously, three time series prediction models and two ML models are used for prediction, comparison, and validation, proving that the WACNN model has the best overall prediction performance. The results show that the weight assignment can effectively assist a CNN with capturing and learning local spatial features through feature enhancement. And the WACNN has a strong ability to deal with the regression prediction problem and has strong robustness and strong interpretability.
Traditional time series models typically rely on lagged features as inputs to capture temporal dependencies. However, such methods may result in a systematic bias of the predicted results relative to the true values, due to the time interval between the input features and the target prediction. This model avoids cumulative errors caused by multi-step lag by strictly aligning the time steps of the input and target. The experimental results demonstrate that this prediction method achieves high prediction accuracy. Using the WACNN model to predict the container throughput of the port will help enterprises plan the use of transportation resources and paths to cope with various uncertainties. After predicting the volume of container throughput, the relevant departments will also be able to make appropriate emergency plans to prevent a shortage of capacity caused by a surge in freight volume. In addition, although the prediction results of the model have a certain degree of reference, the accuracy needs further improvement. In order to improve the prediction accuracy, the WACNN model can be combined with other deep learning models so that the accuracy of the prediction results can be improved to a certain extent.

Author Contributions

Y.W.: Writing—original draft preparation, Conceptualization, Data curation, Methodology, Writing—reviewing and editing. W.L.: Supervision, Methodology, Validation, Software, Writing—reviewing and editing. X.Q.: Supervision, Writing—reviewing and editing. Y.Y.: Software and Writing—reviewing and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Hubei Provincial Natural Science Foundation Project–Xiangyang innovation and development joint fund (2025AFD064) and the Philosophy and Social Science Research Project of the Hubei Provincial Department of Education (No. 23Q171).

Data Availability Statement

The data that support the findings of this study are available from the authors on reasonable request.

Acknowledgments

The authors acknowledge the Open Fund of the Hubei Key Laboratory of Power System Design and Test for Electrical Vehicle. Finally, the authors are grateful for the useful contributions made by their project partners.

Conflicts of Interest

Author Yinzhang Yu was employed by the company Xiaohe Lingang Development Group Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Li, B.; Wen, C.; Yang, S.; Ma, M.; Cheng, J.; Li, W. Measuring high-speed train delay severity: Static and dynamic analysis. PLoS ONE 2024, 19, e0301762. [Google Scholar] [CrossRef]
  2. Guo, J.; Guo, J.; Kuang, T.; Wang, Y.; Li, W. The short-term economic influence analysis of government regulation on railway freight transport in continuous time. PLoS ONE 2025, 20, e0298614. [Google Scholar] [CrossRef] [PubMed]
  3. Liu, Z.; Zhou, S.; Liu, S. Optimization of Multimodal Transport Paths Considering a Low-Carbon Economy Under Uncertain Demand. Algorithms 2025, 18, 92. [Google Scholar] [CrossRef]
  4. Pellicer, D.S.; Larrodé, E. Analysis of the Effectiveness of a Freight Transport Vehicle at High Speed in a Vacuum Tube (Hyperloop Transport System). Algorithms 2024, 17, 17. [Google Scholar] [CrossRef]
  5. Xie, G.; Zhang, N.; Wang, S. Data characteristic analysis and model selection for container throughput forecasting within a decomposition-ensemble methodology. Transp. Res. E Logist. Transp. Rev. 2017, 108, 160–178. [Google Scholar] [CrossRef]
  6. Milenković, M.; Milosavljevic, N.; Bojović, N.; Val, S. Container flow forecasting through neural networks based on metaheuristics. Oper. Res. 2021, 21, 965–997. [Google Scholar] [CrossRef]
  7. Dragan, D.; Keshavarzsaleh, A.; Intihar, M.; Popović, V.; Kramberger, T. Throughput forecasting of different types of cargo in the Adriatic seaport Koper. Marit. Policy Manag. 2020, 48, 19–45. [Google Scholar] [CrossRef]
  8. Homayoonmehr, R.; Rahai, A.; Ramezanianpour, A.A. Predicting the chloride diffusion coefficient and surface electrical resistivity of concrete using statistical regression-based models and its applicationin chloride-induced corrosion service life prediction of RC structures. Constr. Build. Mater. 2022, 357, 1–25. [Google Scholar] [CrossRef]
  9. Zhang, Y.; Wu, J.; Zhang, S.; Li, G.; Jeng, D.S.; Xu, J.; Tian, Z.; Xu, X. An optimal statistical regression model for predicting wave-induced equilibrium scour depth in sandy and silty seabeds beneath pipelines. Ocean Eng. 2022, 258, 1–14. [Google Scholar] [CrossRef]
  10. Castro-Neto, M.; Jeong, Y.S.; Jeong, M.K.; Han, L.D. Online-SVR for short-term traffic flow prediction under typical and atypical traffic conditions. Expert Syst. Appl. 2009, 36, 6164–6173. [Google Scholar] [CrossRef]
  11. Zeng, Z.; Ying, G.; Zhang, Y.; Gong, Y.; Mei, Y.; Li, X.; Sun, H.; Li, B.; Ma, J.; Li, S. Classification of failure modes, bearing capacity, and effective stiffness prediction for corroded RC columns using machine learning algorithm. J. Build. Eng. 2025, 102, 111982. [Google Scholar] [CrossRef]
  12. Tang, L.; Zhao, Y.; Cabrera, J.; Ma, J.; Tsui, K.L. Forecasting Short-Term Passenger Flow: An Empirical Study on Shenzhen Metro. Trans. Intell. Transp. Syst. 2019, 20, 3613–3622. [Google Scholar] [CrossRef]
  13. Haas, C.; Budin, C.; D’arcy, A. How to select oil price prediction models—The effect of statistical and financial performance metrics and sentiment scores. Energy Econ. 2024, 133, 1–14. [Google Scholar] [CrossRef]
  14. Zhou, W.; Zhou, Y.; Liu, R.; Yin, H.; Nie, H. Predictive modeling of river blockage severity from debris flows: Integrating statistical and machine learning approaches with insights from Sichuan Province, China. Catena 2025, 248, 1–15. [Google Scholar] [CrossRef]
  15. Tan, M.C.; Wong, S.C.; Xu, J.M.; Guan, Z.R.; Zhang, P. An aggregation approach to short-term traffic flow prediction. Trans Intell. Transp. Syst. 2009, 10, 60–69. [Google Scholar] [CrossRef]
  16. Khashei, M.; Bijari, M. A novel hybridization of artificial neural networks and ARIMA models for time series forecasting. Appl Soft Comput. 2011, 11, 2664–2675. [Google Scholar] [CrossRef]
  17. Ruiz-Aguilar, J.; Turias, I.; Jiménez-Come, M. Hybrid approaches based on SARIMA and artificial neural networks for inspection time series forecasting. Transp. Res. E Logist. Transp. Rev. 2014, 67, 1–13. [Google Scholar] [CrossRef]
  18. Temizceri, F.T.; Kara, S.S. Towards sustainable logistics in Turkey: A bi-objective approach to green intermodal freight transportation enhanced by machine learning. Res. Transp. Bus. Manag. 2024, 55, 101145. [Google Scholar] [CrossRef]
  19. Bassiouni, M.M.; Chakrabortty, R.K.; Sallam, K.M.; Hussain, O.K. Deep learning approaches to identify order status in a complex supply chain. Expert Syst. Appl. 2024, 250, 1–28. [Google Scholar] [CrossRef]
  20. Gao, L.; Lu, P.; Ren, Y. A deep learning approach for imbalanced crash data in predicting highway-rail grade crossings accidents. Reliab. Eng. Syst. Saf. 2021, 216, 108019. [Google Scholar] [CrossRef]
  21. Zhuang, X.; Li, W.; Xu, Y. Port planning and sustainable development based on prediction modelling of port throughput: A case study of the deep-water Dongjiakou Port. Sustainability 2022, 14, 4276. [Google Scholar] [CrossRef]
  22. Shankar, S.; Ilavarasan, P.V.; Punia, S.; Singh, S.P. Forecasting container throughput with long short-term memory networks. Ind. Manag. Data Syst. 2020, 120, 425–441. [Google Scholar] [CrossRef]
  23. Shankar, S.; Punia, S.; Ilavarasan, P.V. Deep learning-based container throughput forecasting: A triple bottom line approach. Ind. Manag. Data Syst. 2021, 121, 2100–2117. [Google Scholar] [CrossRef]
  24. Awah, P.C.; Nam, H.; Kim, S. Short term forecast of container throughput: New variables application for the Port of Douala. J. Mar. Sci. Eng. 2021, 9, 720. [Google Scholar] [CrossRef]
  25. Tang, S.; Xu, S.; Gao, J. An optimal model based on multifactors for container throughput forecasting. KSCE J. Civ. Eng. 2019, 23, 4124–4131. [Google Scholar] [CrossRef]
  26. Li, F.; Tong, W.; Yang, X. Short-term forecasting for port throughput time series based on multi-modal fuzzy information granule. Appl. Soft Comput. 2025, 174, 112957. [Google Scholar] [CrossRef]
  27. Semenoglou, A.A.; Spiliotis, E.; Assimakopoulos, V. Data augmentation for univariate time series forecasting with neural networks. Pattern Recognit. 2023, 134, 109132. [Google Scholar] [CrossRef]
  28. Wang, W.; Zhao, C.; Wu, Y. Spatial weighting—An effective incorporation of geological expertise into deep learning models. Geochemistry 2024, 84, 126212. [Google Scholar] [CrossRef]
  29. Dai, Y.; Yu, W.; Leng, M. A hybrid ensemble optimized BiGRU method for short-term photovoltaic generation forecasting. Energy 2024, 299, 131458. [Google Scholar] [CrossRef]
  30. Shukla, D.; Chowdary, C.R. A model to address the cold-start in peer recommendation by using k-means clustering and sentence embedding. J. Comput. Sci. 2024, 83, 102465. [Google Scholar] [CrossRef]
  31. Xiao, J.; Wen, Z.; Liu, B.; Chen, M.; Wang, Y.; Hang, J. A hybrid model based on selective deep-ensemble for container throughput forecasting. Syst. Eng. Theory Pract. 2022, 42, 1107–1128. [Google Scholar] [CrossRef]
  32. Gao, H.; Jia, H.; Huang, Q.; Wu, R.; Tian, J.; Wang, G.; Liu, C. A hybrid deep learning model for urban expressway lane-level mixed traffic flow prediction. Eng. Appl. Artif. Intell. 2024, 133, 1–21. [Google Scholar] [CrossRef]
  33. Pan, Y.A.; Guo, J.; Chen, Y.; Cheng, Q.; Li, W.; Liu, Y. A fundamental diagram based hybrid framework for traffic flow estimation and prediction by combining a Markovian model with deep learning. Expert Syst. Appl. 2024, 238, 122219. [Google Scholar] [CrossRef]
  34. Zhang, Y.; Zhou, Q.; Wang, J.; Kouvelas, A.; Makridis, M.A. CASAformer: Congestion-aware sparse attention transformer for traffic speed prediction. Commun. Transp. Res. 2025, 5, 100174. [Google Scholar] [CrossRef]
  35. Nguyen, T.; Cho, G. Forecasting the Busan container volume using XGBoost approach based on machine learning model. J Internet Things Converg. 2024, 10, 39–45. [Google Scholar] [CrossRef]
  36. Lee, E.; Kim, D.; Bae, H. Container volume prediction using time-series decomposition with a long short-term memory models. Appl. Sci. 2021, 11, 8995. [Google Scholar] [CrossRef]
  37. Cui, J.; Liu, B.; Xu, Y.; Guo, X. Regional collaborative forecast of cargo throughput in China’s Circum-Bohai-Sea Region based on LSTM model. Comput. Intell. Neurosci. 2022, 2022, 5044926. [Google Scholar] [CrossRef]
  38. Hirata, E.; Matsuda, T. Forecasting Shanghai Container Freight Index: A deep-learning-based model experiment. J. Mar. Sci. Eng. 2022, 10, 593. [Google Scholar] [CrossRef]
  39. Cuong, T.N.; Long, L.N.B.; Kim, H.S.; You, S.S. Data analytics and throughput forecasting in port management systems against disruptions: A case study of Busan Port. Marit. Econ. Logist. 2023, 25, 61–89. [Google Scholar] [CrossRef]
  40. Liu, B.; Wang, X.; Liang, X. Neural network-based prediction system for port throughput: A case study of Ningbo-Zhoushan Port. Res. Transp. Bus. Manag. 2023, 51, 101067. [Google Scholar] [CrossRef]
  41. Zhang, L.; Schacht, O.; Liu, Q.; Ng, A.K. Predicting inland waterway freight demand with a dynamic spatio-temporal graph attention-based multi attention network. Transp. Res. Part. E Logist. Transp. Rev. 2025, 199, 104139. [Google Scholar] [CrossRef]
Figure 1. The growth of number of containers and combined rail mileage. (a) Container throughput quantity and growth rate and (b) rail and road mileage and growth rate.
Figure 1. The growth of number of containers and combined rail mileage. (a) Container throughput quantity and growth rate and (b) rail and road mileage and growth rate.
Algorithms 18 00319 g001
Figure 2. Schematic diagram of CNN.
Figure 2. Schematic diagram of CNN.
Algorithms 18 00319 g002
Figure 3. The overall framework of the proposed WACNN model for predicting container throughput.
Figure 3. The overall framework of the proposed WACNN model for predicting container throughput.
Algorithms 18 00319 g003
Figure 4. PCA components of clustered data.
Figure 4. PCA components of clustered data.
Algorithms 18 00319 g004
Figure 5. Comparison of columns in clustered data.
Figure 5. Comparison of columns in clustered data.
Algorithms 18 00319 g005
Figure 6. Grouping results.
Figure 6. Grouping results.
Algorithms 18 00319 g006
Figure 7. Data stacking chart.
Figure 7. Data stacking chart.
Algorithms 18 00319 g007
Figure 8. Data stacking chart.
Figure 8. Data stacking chart.
Algorithms 18 00319 g008
Figure 9. Model structure. (a) LSTM model structure and (b) BiLSTM model structure.
Figure 9. Model structure. (a) LSTM model structure and (b) BiLSTM model structure.
Algorithms 18 00319 g009
Figure 10. CNN model parameter tuning diagram. (a) Number of convolutional layers; (b) dropout; and (c) number of convolutional kernels.
Figure 10. CNN model parameter tuning diagram. (a) Number of convolutional layers; (b) dropout; and (c) number of convolutional kernels.
Algorithms 18 00319 g010
Figure 11. LSTM model parameter tuning diagram.
Figure 11. LSTM model parameter tuning diagram.
Algorithms 18 00319 g011
Figure 12. BiLSTM parameter tuning diagram.
Figure 12. BiLSTM parameter tuning diagram.
Algorithms 18 00319 g012
Figure 13. GRU parameter tuning diagram.
Figure 13. GRU parameter tuning diagram.
Algorithms 18 00319 g013
Figure 14. RF model parameter tuning diagram. (a) Number of decision trees and (b) minimum number of leaves.
Figure 14. RF model parameter tuning diagram. (a) Number of decision trees and (b) minimum number of leaves.
Algorithms 18 00319 g014
Figure 15. SVR model parameter tuning diagram. (a) Punishment factor and (b) radial basis function.
Figure 15. SVR model parameter tuning diagram. (a) Punishment factor and (b) radial basis function.
Algorithms 18 00319 g015
Figure 16. WACNN model.
Figure 16. WACNN model.
Algorithms 18 00319 g016
Figure 17. Sensitivity analysis results.
Figure 17. Sensitivity analysis results.
Algorithms 18 00319 g017
Figure 18. Test set results. (a) WACNN model test set prediction results; (b) CNN model test set prediction results; (c) LSTM model test set prediction results; (d) BiLSTM model test set prediction results; (e) GRU model test set prediction results; (f) RF model test set prediction results; (g) SVR model test set prediction results; and (h) PLS model test set prediction results.
Figure 18. Test set results. (a) WACNN model test set prediction results; (b) CNN model test set prediction results; (c) LSTM model test set prediction results; (d) BiLSTM model test set prediction results; (e) GRU model test set prediction results; (f) RF model test set prediction results; (g) SVR model test set prediction results; and (h) PLS model test set prediction results.
Algorithms 18 00319 g018aAlgorithms 18 00319 g018bAlgorithms 18 00319 g018c
Figure 19. Comparison between predicted results and actual values.
Figure 19. Comparison between predicted results and actual values.
Algorithms 18 00319 g019
Table 1. Summary of literature containing predictive models.
Table 1. Summary of literature containing predictive models.
LiteratureData FilteringData EnhancementPrediction ModelThe Best Model
Dragan [7]SMARIMAX
Shankar [22]MLLSTM
Awah [24]MLRF
Tang [25]DLBP
Lee [36]DLLSTM
Cui [37]DLLSTM
Cuong [39]DLLSTM and GRU
Liu [40]DLBiLSTM
This studyDLWACNN
BP: backpropagation neural network; SM: statistic model; RF: random forest; GRU: gated recurrent unit.
Table 2. Factors affecting container throughput of Yingkou port sea–rail combined transport.
Table 2. Factors affecting container throughput of Yingkou port sea–rail combined transport.
Dividing IndicatorsSetting of Evaluation Indicators
Transportation capacityYingkou port container throughput.
Yingkou port cargo throughput.
Yingkou port container terminal throughput.
Road freight traffic in Liaoning Province.
Liaoning waterway cargo volume.
Liaoning waterway cargo turnover.
Liaoning highway cargo turnover.
Liaoning port cargo throughput.
Liaoning port container throughput.
China coastal container throughput.
China railway cargo delivery.
China railway cargo turnover.
China coastal cargo throughput.
Fixed-asset investment and constructionLiaoning fixed-asset investment.
Fixed-assets investment in road, water, and land transportation in Liaoning.
Fixed-assets investment in road, water, and land transportation in Yingkou.
China highways and waterways fixed investment.
China railway fixed-asset investment.
China fixed-asset investment.
Economic development levelChina primary GDP.
China secondary GDP.
China tertiary GDP.
Yingkou primary GDP.
Yingkou secondary GDP.
Yingkou tertiary GDP.
Yingkou GDP.
Liaoning GDP.
China GDP.
International trade and opennessChina coastal foreign trade cargo throughput.
Yingkou port foreign trade cargo throughput.
Liaoning total imports and exports.
Yingkou total imports and exports.
China total imports and exports.
Table 3. Correlation coefficients of influencing factors on sea–rail intermodal container throughput.
Table 3. Correlation coefficients of influencing factors on sea–rail intermodal container throughput.
Influencing FactorPearsonSpearman
China highways and waterways fixed investment0.5360.495
China GDP0.6890.655
China fixed-asset investment−0.031−0.027
China primary GDP−0.101−0.110
China secondary GDP−0.417−0.423
China tertiary GDP0.4120.347
China total imports and exports0.6980.635
China railway cargo delivery0.7670.776
China railway cargo turnover0.7220.703
China railway fixed-asset investment0.1800.186
China coastal cargo throughput0.6850.651
China coastal foreign trade cargo throughput0.7090.662
China coastal container throughput0.6590.641
Yingkou port cargo throughput−0.571−0.660
Yingkou port foreign trade cargo throughput0.2930.293
Yingkou port container throughput−0.298−0.271
Yingkou total imports and exports0.4650.400
Yingkou port container terminal throughput−0.0110.018
Liaoning highway cargo throughput−0.370−0.356
Liaoning highway cargo turnover−0.309−0.190
Liaoning waterway cargo throughput−0.655−0.744
Liaoning waterway cargo turnover−0.684−0.671
Liaoning port cargo throughput−0.667−0.684
Liaoning port container throughput−0.671−0.685
Yingkou GDP0.2880.329
Yingkou primary GDP0.3480.389
Yingkou secondary GDP0.1310.189
Yingkou tertiary GDP0.2300.200
Liaoning GDP0.4410.445
Liaoning fixed-assets investment−0.035−0.251
Liaoning total imports and exports0.5480.525
Table 4. The overall architecture and parameter settings of the CNN model.
Table 4. The overall architecture and parameter settings of the CNN model.
LayerParametersNumerical Value
Input layerSample sizeTraining set:2800/test set:1034
Step size[1.1]
Number of features16
CNN layerNumber of convolutional layers5
Number of convolution kernels per layer16-16-32-32-64
Size of convolution kernel in each layer[10, 1] [3, 1] [3, 1] [3, 1] [2, 1]
Convolutional layer activation functionReLU
Convolutional layer filling methodSame padding
Number of pooling layers4
Pooling layer pooling window size[5, 1] [2, 1] [2, 1] [2, 1]
Output layerNumber of neurons1
Table 5. The overall frameworks and parameter settings of the LSTM, BiLSTM, and GRU models.
Table 5. The overall frameworks and parameter settings of the LSTM, BiLSTM, and GRU models.
LayerParametersLSTMBiLSTMGRU
Input layerNumber of samplesTraining set: 2800/test set: 10342800/10342800/1034
LSTM layerNumber of LSTM layers212
Number of neurons in the first layer3040128
Number of neurons in the second layer40————
LSTM layer activation functionReLUReLUReLU
Output patternlastlastlast
Output layerNumber of neurons111
Table 6. Sensitivity analysis.
Table 6. Sensitivity analysis.
HyperparametricSelected Hyperparametric Range
Batch size16/32/64/128
Initial learning rate0.005/0.01/0.02
Learning rate descent factor0.2/0.1/0.05
Table 7. The hyperparameter settings for the WACNN, CNN, and LSTM models.
Table 7. The hyperparameter settings for the WACNN, CNN, and LSTM models.
HyperparametricWACNNCNNLSTM
Gradient descent algorithmSGDMSGDMAdam
Normalized interval[0, 1.5] and [0, 1][0, 1][0, 1]
Batch size3232——
Maximum number of training sessions150015001500
Initial learning rate0.010.010.01
Learning rate policyPiecewisePiecewisePiecewise
Learning rate descent factor0.10.10.1
Learning rate drop rounds120012001200
Table 8. Comparison of model training set and test set prediction results.
Table 8. Comparison of model training set and test set prediction results.
ModelTraining Set RMSETest Set RMSEDifferenceDifference Percentage
WACNN13.22815.5962.36817.9%
CNN14.725516.35211.626611.05%
LSTM68.797169.31580.51870.75%
BiLSTM65.065371.74626.680910.27%
GRU56.079661.05314.97358.87%
RF19.478229.931210.45353.67%
SVR28.388428.68860.53841.04%
PLS289.5629286.27153.29141.1%
Table 9. Comparison of model prediction results.
Table 9. Comparison of model prediction results.
ModelR2MBEMAPERMSEPromotion RateMAEPromotion Rate
WACNN0.999730.11940.008115.596——11.8649——
CNN0.99972−2.01770.007916.35214.62%12.11382.05%
LSTM0.9950211.4360.033569.315877.5%51.775977.08%
BiLSTM0.994581.63560.037871.746278.26%53.286777.73%
GRU0.99646−9.87240.031961.053174.46%45.318173.82%
RF0.998921.06520.008529.931247.89%14.092515.81%
SVR0.999130.10670.017528.688645.64%26.144954.62%
PLS0.9126110.01840.1597286.271594.55%228.062694.8%
Table 10. Cross-validation partitioning strategy.
Table 10. Cross-validation partitioning strategy.
Train SetTest Set
Group 11–4 yearsThe fifth year
Group 22–5 yearsThe sixth year
Group 33–6 yearsThe seventh year
Group 44–7 yearsThe eighth year
Group 55–8 yearsThe ninth year
Group 66–9 yearsThe tenth year
Table 11. Cross-validation results.
Table 11. Cross-validation results.
Train-RMSETest-RMSETrain-MAETest-MAETrain-MAPETest-MAPE
113.999917.108610.618212.28261.06%1.27%
214.468117.223611.684713.22061.13%1.27%
313.065513.891310.182710.55850.72%0.75%
421.011725.800216.348218.86780.87%0.98%
515.820122.893712.184216.19850.51%0.68%
618.019621.779913.875316.85840.57%0.71%
Average16.064219.782912.482214.66440.81%0.94%
Different3.71872.18220.13%
Percent23.15%17.48%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Y.; Li, W.; Qi, X.; Yu, Y. A Weight Assignment-Enhanced Convolutional Neural Network (WACNN) for Freight Volume Prediction of Sea–Rail Intermodal Container Systems. Algorithms 2025, 18, 319. https://doi.org/10.3390/a18060319

AMA Style

Wang Y, Li W, Qi X, Yu Y. A Weight Assignment-Enhanced Convolutional Neural Network (WACNN) for Freight Volume Prediction of Sea–Rail Intermodal Container Systems. Algorithms. 2025; 18(6):319. https://doi.org/10.3390/a18060319

Chicago/Turabian Style

Wang, Yuhonghao, Wenxin Li, Xingmin Qi, and Yinzhang Yu. 2025. "A Weight Assignment-Enhanced Convolutional Neural Network (WACNN) for Freight Volume Prediction of Sea–Rail Intermodal Container Systems" Algorithms 18, no. 6: 319. https://doi.org/10.3390/a18060319

APA Style

Wang, Y., Li, W., Qi, X., & Yu, Y. (2025). A Weight Assignment-Enhanced Convolutional Neural Network (WACNN) for Freight Volume Prediction of Sea–Rail Intermodal Container Systems. Algorithms, 18(6), 319. https://doi.org/10.3390/a18060319

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop