Next Article in Journal
Research on Wind Speed Inversion Method for X-Band Networked SAR Satellite
Previous Article in Journal
A Novel Double Layered Hybrid Multi-Robot Framework for Guidance and Navigation of Unmanned Surface Vehicles in a Practical Maritime Environment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Predicting External Influences to Ship’s Average Fuel Consumption Based on Non-Uniform Time Set

Faculty of Maritime Studies, University of Split, Ruđera Boškovića 37, 21000 Split, Croatia
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2020, 8(9), 625; https://doi.org/10.3390/jmse8090625
Submission received: 3 July 2020 / Revised: 7 August 2020 / Accepted: 14 August 2020 / Published: 19 August 2020

Abstract

:
Nowadays, the impact of the ships on the World economy is enormous, considering that every ship needs fuel to sail from source to destination. It requires a lot of fuel, and therefore, there is a need to monitor and predict a ship’s average fuel consumption. However, although there are much models available to predict a ship’s consumption, most of them rely on a uniform time set. Here we show the model of predicting external influences to ship’s average fuel consumption based on a non-uniform time set. The model is based on the numeric fitting of recorded data. The first set of recorded data was used to develop the model, while the second set was used for validation. Statistical quality measures have been used to choose the optimal fitting function for the model. According to statistical measures, the Gaussian 7, Fourier 8, and smoothing spline fitting functions were chosen as optimal algorithms for model development. In addition to extensive data analysis, there is an algorithm for filter length determination for the preprocessing of raw data. This research is of interest to corporate logistics departments in charge of ensuring adequate fuel for fleets when and where required.

1. Introduction

The effects of the marine environment and other causes on fuel consumption can be examined by various parameters and various approaches. Approaches could be model-based, which is usual in references or signal-based, which is considered in this paper. Model-based approaches tend to estimate the exact fuel consumption in the exact engine operating mode, examples of this are [1,2]. Such models are useful in, e.g., dual-fuel engines, when it is possible to calculate how much of fuels are being spent [1]. ANNs (Artificial Neural Network) in [2] need to be trained and re-trained for specific ships over time. However, there is no exact real-time monitoring of the hull status, which could be a problem that increases the error in the calculation during time between maintenances. Namely, during exploitation and over time, the ship hull is affected by fouling—a natural phenomenon where marine/aquatic vegetation and microorganisms attach to the hull, creating bio-layers that have an impact on the ship’s speed and fuel consumption. Other contributing factors may include weather conditions, cargo, propulsion, engine conditions, etc. Data trend analysis is a suitable approach for predicting fuel consumption, depending on the biofouling layers as environmental contributors. In light of current increased efforts to improve energy efficiency, the above-mentioned topic is current, especially since it includes the use of hybrid technologies [3,4]. As stated in [4], fuel consumption reduction cannot be established without first exploring standard fuel consumption prediction models.
Instead of focusing on a specific ship, there are a number of references to fuel demand prediction [5,6,7,8] in which the focus is on energy efficiency and ecology [9,10,11], or predicting global demand and port demand. The dependence of fuel consumption on vessel design was explored in [12,13], and on vessel speed in [14]. There are new trends in addressing the fuel consumption issue that could be divided into several research branches. While the effect of biofouling on ship resistance using CFD (Computational Fluid Dynamics) was the research topic in [15], vessel fuel consumption prediction was examined in [16], and the authors in [17] used a fuel consumption model based on the Vehicle Specific Power distribution. Traffic condition prediction was linked with the fuel consumption model to predict fuel consumption. “Fuel consumption data provided by the On-Board Diagnostic tool was used to verify the proposed application, with a prediction error under 20%” [17]. A statistical approach to the ship’s fuel consumption was presented in [18]. This research could be extended to a speculative approach—can the calculated consumption of the model be subtracted from the total consumption to obtain its environmental impact? Learning approaches to ship fuel consumption with ANNs are the most popular, and include [19,20,21,22,23]. On-line fuel consumption prediction was obtained by machine learning in [19]. Shalow and deep learning was combined in [22]. An outstanding result was the correlation matrix in [20], which correlated various causes of fuel consumption increase (e.g., wind, trim, currents, cargo, etc.). An interesting case study was published in [24], which explored AI (Artificial Intelligence) driven tools to identify fuel usage anomalies across the company’s entire fleet. The gray box model was applied to optimize trim and identify the possibility of decreasing fuel consumption [25]. It was shown that this optimization could decrease fuel consumption up to 2.3% [25].
The subject of this paper is to monitor the daily ships fuel consumption seasonal. The main variable used is the daily fuel consumption, which is the main indicator of fuel consumption. Concerning the previously mentioned research problem of ships fuel consumption, the following hypotheses is that the average daily consumption could be predicted seasonally and yearly. The aim of this paper is to find the fitting function to establish the dynamics of the ships daily fuel consumption, using a simple and processing cost-efficient model. The goal of the paper is to find the fitting function that approximates the ship’s daily fuel consumption.
The paper is organized as follows: the second section defines quality measures and fitting curves used in the research; the third section explains the methodology, while the fourth presents the results. The latter are obtained using the known data that do not pertain to consecutive days, but rather cover an irregular day sequence. The results are produced at yearly and seasonal levels, which is a novelty of the paper. Finally, the discussion and conclusions are presented.

2. Mathematical Background of Curve Fitting and Prediction

This section presents the mathematical foundations used in the paper, together with related references. There are several ways to fit the data. As the dataset in the paper is non-uniform, we used several fitting functions, Equations (1)–(20). First, the data were fitted using the Matlab function “linear fitting”, which can be described with the following equation [26]:
f L F ( x ) = a · sin ( x π ) + b · ( x 10 ) 2 + c
where x is an independent variable, and a , b , and c are constants that have to be determined. The next data fitting function is the so-called exponential of the 1st order, which can be defined as [26]:
f e 1 ( x ) = a e 1 · e b e 1 · x
where x is an independent variable, a e 1 and b e 1 are constant coefficients. The data were tested using the so-called exponential of the 2nd order, defined as follows [26]:
f e 2 ( x ) = a e 2 · e b e 2 · x   +   c e 2 · e d e 2 · x
where a e 2 , b e 2 , c e 2 , and d e 2 are coefficients. The next curve fitting functions considered for research are taken from the Fourier series [26]:
f F ( x ) = a 0 + i = 1 [ a i · cos ( i · ω · x ) + b i · sin ( i · ω · x ) ]
where coefficients from (4) are: a 0 = ( 2 T ) · 0 T f ( x ) · d t , a i = ( 1 T ) · 0 T f ( x ) · cos ( i · ω · x ) · d t , and b i = ( 1 T ) · 0 T f ( x ) · sin ( i · ω · x ) · d t , ω = ( 2 · π ) T , and T is the period or data width. In the analysis, the Fourier series of the 1st, 2nd, and 8th order were considered, presented with Equations (5)–(7) as follows:
Fourier series of the first order:
f F 1 ( x ) = a 0 + a 1 · cos ( ω · x ) + b 1 · sin ( ω · x )
Fourier series of the second order:
f F 2 ( x ) = a 0 + i = 1 2 [ a i · cos ( i · ω · x ) + b i · sin ( i · ω · x ) ]
and Fourier series of the eighth order:
f F 8 ( x ) = a 0 + i = 1 8 [ a i · cos ( i · ω · x ) + b i · sin ( i · ω · x ) ]
We likewise considered the Gaussian function [26], defined as:
f G ( x ) = i = 1 a i · e [ ( x b i ) c i ] 2
where a i , b i , and c i are Gaussian function coefficients. From (8), a Gaussian function with one term (defined in Matlab as Gaussian 1) was considered in the analysis, which could be defined with the following equation [26]:
f G 1 ( x ) = a 11 · e [ ( x b 11 ) c 11 ] 2
where a 11 , b 11 , and c 11 are coefficients of the Gaussian function with one term. Next, a Gaussian function with two terms (Gaussian 2 in Matlab) was considered, defined as [26]:
f G 2 ( x ) = a 12 · e [ ( x b 12 ) c 12 ] 2 + a 22 · e [ ( x b 22 ) c 22 ] 2
where a 12 , b 12 , c 12 , a 22 , b 22 , and c 22 are coefficients of the Gaussian function with two terms. Polynomials-based fitting functions were also considered. First, the polynomial of the first order was defined as follows [26]:
f P 1 ( x ) = p 01 + p 11 · x
where p 01 and p 11 are first-order polynomial coefficients. Next, the second-order polynomial was defined as [26]:
f P 2 ( x ) = p 02 + p 12 · x + p 22 · x 2
where p 02 , p 12 , and p 22 are second-order polynomial coefficients. Third-order polynomials were also used in the analysis, that could be defined as [26]:
f P 3 ( x ) = p 03 + p 13 · x + p 23 · x 2 + p 33 · x 3
where p 03 , p 13 , p 23 , and p 33 are third-order polynomial coefficients. All coefficients from (11), (12), and (13) could be calculated by performing the least square method [26]. In the next two equations, (14) and (15), the so-called first and second power model fitting functions were used, which could be described with the following equations [26]:
f P 1 = α 11 · x β 11
where α 11 and β 11 are coefficients from the so-called first-order power model, and [26]:
f P 2 = δ 21 + α 21 · x β 21
where α 21 , β 21 , and δ 21 are coefficients of the so-called second-order power model. Furthermore, Equations (16)–(20) use rational fitting functions, and can be defined as follows; the rational fitting function 1/1 could be expressed as [26]:
f 1 / 1 ( x ) = z 01 + z 11 · x q 01 + x
where z 01 , z 11 , and q 01 are coefficients of the rational 1/1 fitting function. The rational fitting function 2/1 can be expressed as follows [26]:
f 2 / 1 ( x ) = z 02 + z 12 · x + z 22 · x 3 q 02 + x
where z 02 , z 12 , z 22 , and q 02 are coefficients of the rational 2/1 fitting function. The rational fitting function 3/1 could be expressed as [26]:
f 3 / 1 ( x ) = z 03 + z 13 · x + z 23 · x 3 + z 33 · x 4 q 04 + x
where z 03 , z 13 , z 23 , z 33 , and q 03 are coefficients of the rational 2/1 fitting function. Equation (19) describes the 3/2 rational fitting function as follows [26]:
f 3 / 2 ( x ) = z 05 + z 15 · x + z 25 · x 3 + z 35 · x 4 q 05 + q 15 · x + x 2
where z 05 , z 15 , z 25 , z 35 , q 05 , and q 15 are coefficients of the rational 3/2 fitting function. Finally, the rational 5/3 fitting function could be expressed as [26]:
f 5 / 3 ( x ) = z 08 + z 18 · x + z 28 · x 3 + z 38 · x 4 + z 48 · x 5 q 08 + q 18 · x + q 28 · x 2 + x 3
where z 08 , z 18 , z 28 , z 38 , z 48 , q 08 , q 18 , and q 28 are coefficients of the rational 5/3 fitting function.
In the end, the data were fitted using the smoothing spline s, as in (21), for the specified smoothing parameter p and specified weights w i [26]. The smoothing spline minimizes the expression:
p · i w i · ( y i s ( x i ) ) 2 + ( 1 p ) · ( d 2 s d x 2 ) 2 · d x
If the weights are not specified, they are assumed to be 1 for all the data points. Parameter p is defined between 0 and 1. For p = 0, a least-squares straight-line is produced that fits the data. For p = 1, a cubic spline interpolant is obtained.
The sheer multitude of fitting functions to be tested with the data requires them to be quantified, and quality measures to be introduced. The following quality measures have been used—the first quality measure is the root mean square error (RMSE)—a measure that determines the differences between samples or population values predicted by the fitting function—that can be described as follows [27,28]:
R M S E = X Y 2 N
where X is the observed vector, Y the predicted vector, and N the number of data measured in the observed vector space.
The following three quality measures, as in (23), (24), and (25), are statistical measures that describe the number of variations in the dependent variable. The first quality measure is the sum of the squared estimate errors (SSE), which can be expressed as [27,28]:
S S E = i = 1 N ( y i y i ) 2
where y i is the i-th value of the variable to be predicted, and y i the predicted value of y i . SSE shows the measure of discrepancy between the data and the estimation model. The second quality measure is the total sum of squares (SST), that can be defined as [27,28,29]:
S S T = i = 1 N ( y i y ¯ ) 2
where y i is the ith value of the variable to be predicted, and y ¯ the mean value. SST is defined as a quality measure that equals the sum of the squared differences between observations and their overall mean value. The third quality measure used here was the sum of squares due to regression, that could be defined as [27,28,29]:
S S R = i = 1 N ( y i y ¯ ) 2
The following expression is known to be true [27,28,29]:
S S T = S S E + S S R
Furthermore, the R-square measure could be devised from (23), (24), (25), and (26), and defined as:
r 2 = S S R S S T = 1 S S E S S T
It is known in the literature [20,21,22] as the coefficient of determination that describes the proportion of dependent variable variance predictable from the independent variable. In addition, there is an adjusted R-square quality measure, which corrects the possible error in the R-square measure by increasing the number of samples, described as [27,28,29]:
r ¯ 2 = 1 ( 1 r 2 ) · N 1 N k 1 = 1 ( S S E S S T ) · N 1 N k 1
where N is the number of data samples, and k is the number, which explains the independent variable.
In the next paragraph of this section, the random variable theory used throughout the paper will be presented. If we consider random variables X s u , X a , X w , X s p in the vector notation:
X s u = [ x s u 1 , x s u 2 , , x s u i ] s u i = 1 , 2 , 3 , , N s u i N X a = [ x a 1 , x a 2 , , x a i ] a i = 1 , 2 , 3 , , N a i N X w = [ x w 1 , x w 2 , , x w i ] w i = 1 , 2 , 3 , , N w i N X s p = [ x s p 1 , x s p 2 , , x s p i ] s p i = 1 , 2 , 3 , , N s p i N
where X s u , X a , X w , X s p represent random variables with N data samples measured in the summer (su), autumn (a), winter (w), and spring (sp) for four years. Standard statistical metrics such as expectation or average value, standard deviation, standard error, and correlation coefficient were used to study random variables. The average value of the random variable x or expectation could be represented by equation [27,29]:
E [ x ] = x ¯ = 1 N · i = 1 N x i = 1 N · ( X × X T )
where E represents the expectation for a random variable, is the sign for random variable x, and N is the number of samples measured. The standard deviation of the random variable x could be represented by equation [27,28,29]:
σ x = 1 N 1 · i = 1 N ( x i x ¯ ) 2 = E [ x 2 ] E [ x ] 2
where σ x represents the standard deviation of the random variable x . The transformation of coordinates from one coordinate system to another is described by the following equation [27,28,29]:
x a * = ( x i a x ¯ a ) σ x a = 1 σ x a · ( ( X i a X ¯ a ) × ( X i a X ¯ a ) T ) x w * = ( x i w x ¯ w ) σ x w = 1 σ x w · ( ( X i w X ¯ w ) × ( X i w X ¯ w ) T )
where x a * i x w * are transformed coordinates of random variables. The statistical metric used to quantify the similarity and dependence between variables, x a and x w is the correlation coefficient between the random variables. The correlation coefficient could be calculated using the following equation [27,28,29]:
r = 1 N 1 · i = 1 N ( ( x i a x ¯ a ) · ( x i w x ¯ w ) ) σ x a · σ x w = E [ ( x a E [ x a ] ) · ( x w E [ x w ] ) ] σ x a · σ x w = 1 N 1 × ( X a * × X w * T )
where r represents the correlation coefficient, N is the number of measurements, while σ x a and σ x w represent standard deviations of the random variables x a and x w . Furthermore, a model matrix A could be created using the following equation:
A = [ X s u X a X w X s p ] = [ x s u 1 x s u 2 x s u N x a 1 x a 2 x a N x w 1 x w 2 x w N x s p 1 x s p 2 x s p N ] ( 4 , N )
where X s u , X a , X w , X s p are vectors of random independent variables, together with Equations (29)–(33), transformed in the correlation matrix Cx into the equation:
C X = 1 N 1 × ( A × A T ) = 1 N 1 · ( [ x s u 1 x s u 2 x s u N x a 1 x a 2 x a N x w 1 x w 2 x w N x s p 1 x s p 2 x s p N ] ( 4 x N ) · [ x s u 1 x a 1 x w 1 x s p 1 x s u 2 x a 2 x w 2 x s p 2 . . . . . . . . . x s u N x a N x w N x s p N ] ( N x 4 ) )
In the example from Equation (29), the correlation matrix Cx has dimensions (4 × 4), as follows:
C X = [ c x s u , x s u c x s u , x a c x s u , x w c x s u , x s p c x a , x s u c x a , x a c x a , x w c x a , x s p c x w , x s u c x w , x a c x w , x w c x w , x s p c x s p , x s u c x s p , x a c x s p , x w c x s p , x s p ] ( 4 , 4 )
The correlation matrix, Cx, shows correlations between variables.
To conclude, this section covers the mathematical background of the fitting functions, from (1) to (21) that will be used in the prediction. Equations from (22) to (28) describe the quality measures, which will be used to grade the fitting functions and determine which fitting function is best suited to predicting the average fuel consumption based on the non-uniform data time set. Finally, the theory of random variables is introduced in (29) to (36) that will be useful for identifying the optimal fitting function for prediction.

3. Methodology, Setup, and Preprocessing

This section will cover the data smoothing methodology, otherwise known as data preprocessing. Owing to the non-uniform set of data and the need to identify the best-suited fitting function, choosing the “right” window to perform the moving average operation or filtering is of paramount importance. There are a number of moving average algorithms available that can be used for smoothing, such as a simple moving average (SMA), a weighted moving average (WMA), an exponential moving average (EMA), and a weighted exponential moving average method (WEMA). All the mentioned moving average algorithms have their advantages and disadvantages, but they all require choosing optimal filter window size, i.e., the number of past and future points that will determine the current point. The preprocessing procedure can be explained in steps, as shown in Figure 1.
The first step in the signal processing is noise removal. Although the noise does not exist physically within the average daily fuel consumption data, it exists from the signal processing point of view. Namely, as there are sudden spikes in fuel consumption, comparing raw data would be somewhat misleading. Instead, the moving average or filtering operation is performed on raw data. The first step is to choose the size of the moving average filter (filter length). It has to be noted that ship data are not uniformly collected. Hence, there is no year, season, or month with an equal number of samples. Smoothing at the season level has been performed. The optimal filter length was identified by developing and performing an algorithm.
First, the R-square analysis of the raw data was performed. In the raw data, R-square is, e.g., 0.01636 for the Gauss 1 fitting function. As a low R-square measure implies a high noise level, raw data should be preprocessed by eliminating the noise. Since we failed to obtain satisfactory results, we proceeded according to the flow diagram given in Figure 1. First, we detected the filter length necessary for moving the average operation, and then performed the moving average to denoise the raw data and obtain enhanced input to the curve fitting operation.
The influence of filter length can be seen in the correlation matrix. The results for the length of 52 points over four years of the moving average filter are as follows (including trends):
[ 1.0000 0.3987 0.3963 0.4210 0.3987 1.0000 0.3532 0.3311 0.3963 0.3532 1.0000 0.6988 0.4210 0.3311 0.6988 0.9999 ]
The results for the length of 12 points of the moving average filter are:
[ 0.9999 0.0235 0.2615 0.1486 0.0235 0.9999 0.1758 0.2552 0.2615 0.1758 1.0000 0.2785 0.1486 0.2552 0.2785 0.9999 ]
The results for the length of 24 points of the moving average filter are:
[ 1.0000 0.2436 0.4057 0.2353 0.2436 1.0000 0.2725 0.4084 0.4057 0.2725 1.0000 0.5211 0.2353 0.4084 0.5211 1.0000 ]
The closer the correlation is to 1, the higher the similarity, and the closer the correlation is to 0, the lower the similarity. As various results for the correlation matrix are obtained by using different moving average filter lengths, we implemented an algorithm for identifying the best solution (see below: Algorithm 1 for optimum moving average filter length identification).
Algorithm 1 for optimum moving average filter length identification.
for kk from 3 to 52
  data1_filtered=movemean(data1,kk)
  data2_filtered=movemean(data2,kk)
  A = [data1(1:length(min(data1, data2)))’
  data2(1:length(min(data1, data2)))’];
  % alternatively zero padding can be used that all vectors have the same length.
  d = (A*A’)/(N − 1);
  e = d/max(max(d));
  zb(k) = sum(sum(dist(e-ones(size(e)))));
end;
  find(zb == min(zb))
where N = length(min(data1, data2)).
Different results were obtained for different seasons (years). An optimum filter length was used for corresponding data in further analysis.
Data are presented by years (in the first part of the research) and by seasons (in the second part of the research). Half of the data were used to generate the equation and the other half to test the prediction hypothesis.
All the calculations were performed in the Matlab application Curve fitting tool [26]. The variables of interest were input as y-data.

4. Results

The data obtained consist of 829-time samples that are not uniformly sampled. They are obtained from a single ship (5 October 2008–28 August 2012), and refer to fuel consumption in 24 h. We used the first 434 samples (two years) to obtain the fit function, while the remaining 395 samples (two years) were used for the identification of the fitting function best suited to predict future fuel consumption. Similar analysis was performed by seasons. The analysis was divided into two parts: analysis by years known as the vertical analysis and horizontal analysis, where one year was divided into four seasons, namely, summer, autumn, winter, and spring.

4.1. Analysis by Year

The correlation matrix for the first two years is:
[ 1.0000 0.9661 0.9661 0.9361 ]
From (40), it can be seen that the correlation between the first, and the second year was 96.61%. For four years, the smoothed data had the following correlation:
[ 1.0000 0.9600 0.9595 0.8379 0.9600 0.9237 0.9231 0.8064 0.9595 0.9231 0.9235 0.8064 0.8379 0.8064 0.8064 0.7048 ]
As the lowest correlations were between the second and the fourth year, the third and the fourth, and within the fourth year itself, there was obviously a fundamental occurrence in the 4th year, as can be seen from (41). The results obtained suggest that denoised data are suitable for further research.
Table A1 indicates the coefficients obtained. As in the case of the fitting function called Rational 1/1 (16), the fit computation did not converge, Matlab stopped fitting, because the number of iterations or function evaluations exceeded the specified maximum. The results for the other fitting functions are presented in Table A2. Table A2 shows the results for years 1 and 2, where the fit was examined by SSE, R-square, RMSE, and the adjusted R-square. The table is sorted by R-square (the best on the top). Negative results denoted with “*” imply the invalidity of the model. It can be seen that the best fit (other than the smoothing spline curves) was obtained by Fourier 8. Table A3 shows the results of the prediction for various fitting functions. Table A3 is sorted by R-square in the prediction interval (the best on the top). The measure of quality is the R-square coefficient. The best result after smoothing splines was obtained by Gaussian 7.
Figure 2 and Figure 3 show examples of the results. The dots are real samples, and the full lines are the estimated curve. Since the data are obtained by non-uniform daily sampling, the x-axes are marked as “indexed time” in all figures in the Results section. Hence, the numbers on the x-axes are not time units, but the number of daily samples, which have no units. Figure 2a and Figure 3a present the first two years, denoised ( y 12 m ), and the fitted data ( y _ y e a r s 12 _ f i t ). Figure 2b and Figure 3b show the same for years 3 and 4, and Figure 2c and Figure 3c the same for all data (four years).

4.2. Analysis by Seasons

Optimal filter length for the available data set in the summer is 21. By using the result from the algorithm, the correlation matrixes for summers 2 and 4 are:
[ 1.0000 0.8915 0.8915 0.7984 ] ,
[ 1.0000 0.9007 0.9859 0.8006 0.9007 0.8132 0.8931 0.7200 0.9859 0.8931 0.9858 0.7863 0.8006 0.7200 0.7863 0.6419 ] .
The obtained coefficients for the considered fitting functions are given in Table A4 (see Appendix A). As the fit computation for the Gaussian 7 function did not converge, the fitting was discontinued since the number of iterations or function evaluations exceeded the specified maximum.
The obtained measures of similarity with the fitting functions are given in Table A5, sorted by the R-square (the best on the top row). In spite of the calculation problems, Gaussian 7 yielded the best fit under the R-square criterion. Table A6 shows R-square results for the domain used to extrapolate the fitting function parameters, predicted curve, and total range, sorted by the predicted R-square domain. The Gaussian 7 fitting function could be observed to be the best fit in the prediction interval. Figure 4 and Figure 5 show examples of the results for the summer.
After the summers, the analysis by season turned to autumns. The algorithm for the optimal length of the moving average filter gave the length of 42 in case of autumn. The correlation matrix for the first two autumns is:
[ 0.9734 0.9828 0.9828 1.0000 ]
In case of all four autumns, the correlation matrix is:
[ 0.9734 0.9828 0.9432 0.8338 0.9828 1.0000 0.9557 0.8436 0.9432 0.9557 0.9155 0.8088 0.8338 0.8436 0.8088 0.7147 ]
The obtained coefficients are given in Table A7 (see Appendix A). There was a problem with the Gaussian 7 function, where the fit computation did not converge and Matlab stopped fitting because the number of iterations or function evaluations exceeded the specified maximum.
Table A8 shows the results for level of fit. Data were sorted by the R-square parameter. The best results (after the smoothing splines) were obtained by the Gaussian 8 and Fourier 7. Table A9 shows R-square for autumns 1 and 2, autumns 3 and 4, and autumns 1–4. Data are sorted by prediction interval, with the Gaussian 7 fitting function giving the best prediction (after the smoothing spline functions).
Figure 6 and Figure 7 show examples of the results, with real data illustrated with dots, and fitted curves with full lines. In this case, Gaussian 7 is chosen for Figure 6, and Rational 3/2 function for Figure 7. Figure 6b,c and Figure 7b,c show discontinuities in the middle of the graph.
Correlations for two winters are:
[ 1.0000 0.9513 0.9513 0.9049 ]
and for all winters:
[ 1.0000 0.9513 0.9647 0.8458 0.9513 0.9049 0.9176 0.8045 0.9647 0.9176 0.9306 0.8160 0.8458 0.8045 0.8160 0.7157 ]
The obtained coefficients (for fitting functions) for winters 1 and 2 are given in Table A10 (see Appendix A). For the Gaussian 7 fitting function, the fit computation did not converge and the fitting was halted. The results in the table are the last obtained in this case.
Table A11, sorted by the R-square parameter, shows quality measures for winters 1 and 2. The best fit was obtained with Gaussian 7 (if the smoothing splines were excluded). Table A12, sorted by the predicted domain, shows R-squares for winters 1–2, 3–4, and 1–4. Gaussian 7 is the best for the prediction (when smoothing splines are excluded).
Figure 8 shows an example of the results in the case of the Gaussian 7 fitting function.
In the case of the two springs, the optimal length for the moving average filter was 47. Correlations for the two springs are:
[ 0.9780 0.9885 0.9885 1.0000 ]
and for four springs:
[ 0.8157 0.8332 0.9025 0.6972 0.8332 0.8515 0.9226 0.7113 0.9025 0.9226 1.0000 0.7700 0.6972 0.7113 0.7700 0.5974 ]
The results for two springs are given. The obtained coefficients for the fitting functions are presented in Table A13 (see Appendix A).
The results for quality measures in case of springs 1 and 2 are shown in Table A14, which is sorted by the best R-square. Better results are for rougher smoothing spline (p closer to 1), which is not favorable. On the contrary, smoother splines are welcome and the achieved results are not as good. Apart from the smoothing splines, the best result was obtained with Gaussian 7. Table A15, sorted by the best fits in the prediction interval, shows the results for springs 1–2, 3–4, 1–4. Fourier 8 was shown as the best for the prediction purposes in the case of springs.
Figure 9 and Figure 10 show examples of the results for springs. Fourier 8 clearly exhibits discontinuity around the middle of the dataset.

5. Discussion and Conclusions

The paper presents a black-box approach to ship fuel consumption based on signal (daily fuel consumption) analysis. The real data were analyzed. The fitting function should enable the easy build in of the possible tools/software in the future, which is simple and requires low processor power. This fitting function should reflect the coupling between data and sources. Due to nonlinearities, various data were transformed to the same coordinates. This is important, because monthly, seasonally, or yearly data do not have the same length. The explored approach should include aspects of de-trending, coupling, nonlinearities, unreliable data, and different lengths of data within the same time period. Finally, the fitting function would enable the prediction of fuel consumption and differentiate it from the actual consumption. If an anomaly is detected, the company could investigate whether the cause is the weather, fouling, or even fraud. During the research, it was assumed that fuel consumption would grow linearly over time. Such dependence would make possible to use some linear regression methods (e.g., [20,29]). However, the results did not show that consumption had actually decreased, which could be associated with the following factors:
  • non-uniform time sampling (leading to wrong curve angle between the interpolating points), and
  • average (which depends on the route the ship was sailing at the time of data acquisition, and the sailing hours on a specific day).
It was obvious from the results of the R-square and the adjusted R-square measures that the data were not well conditioned for the correct analysis. Hence, different approaches were used. Firstly, the data were preprocessed with filters, which resulted in better R-square and adjusted R-square metrics. The next step was to use a bank of fitting functions to find the best match. Higher-order functions were observed to result in a better fit. The best functions to choose for the model were the smoothing splines, but in the so-called rougher (not smoother) versions (parameter p closer to 1), as rougher splines fit better sampling points, although they were not suitable for real-world applications. Instead, the Gaussian, Fourier or rational functions were chosen for the prediction model. In this case, the best fit of the predicted data (samples for 3rd and 4th year/season) was obtained by Gaussian, Fourier, and rational functions that generally tend to be represented by the first-order polynomial. Higher order functions tend to oscillate around the sampled data more tightly. That reduces the error, and it could lead to the conclusion that such functions (i.e., Gaussian 7) could be used in any ship by simply adjusting the function parameters, which is the advantage in our signal-based approach. The result was obtained easily and fast, which is an advantage over methods that attempt to estimate real fuel consumption by examining engine characteristics and parameters.
In the analysis, we explored comparisons between years, and comparisons between seasons. The worst correlations were obtained between the second and the fourth year, the third and the fourth, and within the fourth year itself, suggesting a relevant event of some sort in the fourth year. Likewise, the correlation between summers, autumns, and winters was above 80%. On the other hand, springs were less correlated, presumably due to the state of the ship’s hull after wintertime, and the great oscillations in weather conditions that typically occur in springtime.
Finally, there were two contributions of this paper. The main contribution, as far as the authors are aware, was that this was the first time that seasons (so called horizontal analysis) were considered for prediction purposes. A minor contribution of the paper was the algorithm for identifying the optimum moving average filter length.
The prediction formula takes into account environmental and biological influences, as well as cargo mass, and all other possible fuel consumption factors. We strongly believe that it could be used to detect potential frauds (fuel theft), which may be of interest to various authorities. Furthermore, the proposed prediction model is simple and fast to use, and can be used to check deviations in fuel consumption by comparing the predicted and deviated consumption. A deviation could be caused by fraud or by environmental factors such as fouling.
There are additional differences between this paper and references. For example, the results in [19] are obtained by excluding suspicious data in the preprocessing stage. We used all available data to obtain our results. Most of the references used a deep learning approach with ANNs [19,20,21,22,23]. On the other hand, it was not necessary to use ANNs always. It was popular, but not necessary. Hence, further research could include the identification of factors that cause unplanned fuel consumption and modeling fuel consumption by artificial neural networks (ANN) and hybrid models of velocity and fuel consumption.

Author Contributions

Conceptualization, I.V.; methodology, I.V., J.Š. and I.K.; software, I.V. and M.P.; validation, J.Š. and I.V.; formal analysis, I.K., and I.V.; investigation, I.V. and M.P.; writing—original draft preparation, I.V.; Writing—Review & Editing, I.V. and J.Š.; visualization, I.V. and M.P.; supervision, I.V. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

The obtained coefficients of the fitting functions are given in this Appendix. To simplify the tables, several additional equations are used. The expression for Fourier 8 is:
f ( x ) = a 0 + a 1 · cos ( x · ω ) + b 1 · sin ( x · ω ) + a 2 · cos ( 2 x · ω ) + b 2 · sin ( 2 x · ω ) + + a 3 · cos ( 3 x · ω ) + b 3 · sin ( 3 x · ω ) + a 4 · cos ( 4 x · ω ) + b 4 · sin ( 4 x · ω ) + + a 5 · cos ( 5 x · ω ) + b 5 · sin ( 5 x · ω ) + a 6 · cos ( 6 x · ω ) + b 6 · sin ( 6 x · ω ) + + a 7 · cos ( 7 x · ω ) + b 7 · sin ( 7 x · ω ) + a 8 · cos ( 8 x · ω ) + b 8 · sin ( 8 x · ω )
The expression for Gaussian 3 is:
f ( x ) = a 1 e ( ( x b 1 ) / c 1 ) 2 + a 2 e ( ( x b 2 ) / c 2 ) 2 + a 3 e ( ( x b 3 ) / c 3 ) 2
The expression for Gaussian 7 is:
f ( x ) = a 1 e ( ( x b 1 ) / c 1 ) 2 + a 2 e ( ( x b 2 ) / c 2 ) 2 + a 3 e ( ( x b 3 ) / c 3 ) 2 + a 4 e ( ( x b 4 ) / c 4 ) 2 + + a 5 e ( ( x b 5 ) / c 5 ) 2 + a 6 e ( ( x b 6 ) / c 6 ) 2 + a 7 e ( ( x b 7 ) / c 7 ) 2
If we take Table A1, the first row, for the fitting function described by (1), it was found that the coefficients (with 95% confidence bounds) are: a = 0.01021 (−0.1528, 0.1732), b = 2.519·10−6 (confidence bounds: 3.792·10−7, 4.658·10−6), c = 28.91 (28.74, 29.08) and the test function is therefore:
f ( t ) = 0.01021 * ( s i n ( t π ) ) + 2.519 · 10 6 · ( t 10 ) 2 + 28.91
for the first two years. Variable t denotes samples, not physical time.
For the fitting exponential function described by (2), the coefficients are: a = 29.25 (bounds: 29.02, 29.48), b = −3.011·10−5 (95% confidence bounds: −6.18·10−5, 1.573·10−6):
f ( t ) = 29.25 e 3.011 · 10 5
For the fitting function in (3), the coefficients are: a = 30.39 (30.22, 30.57), b = −0.00035 (−0.0003961, −0.0003039), c = 0.002414 (−0.0005264, 0.005354), d = 0.01832 (0.01554, 0.02109):
f ( t ) = 30.39 e 0.00035 t + 0.002414 e 0.01832 t
For Fourier 1 (12), the fitting function coefficients are: a0 = 6.872·107 (−1.162·1015, 1.162·1015), a1 = −6.872·107 (−1.162·1015, 1.162·1015), b1 = 2.12·104 (−1.792·1011, 1.7921011), ω = −1.377·10−6 (−11.64, 11.64). The final fitting function is:
f ( t ) = 6.872 · 10 7 6.872 · 10 7 cos ( t · ω ) + 2.12 · 10 4 sin ( t · ω )
The detailed results on the coefficients and fitting are presented in the appendix. Table A2, Table A3, Table A5, Table A6, Table A8, Table A9, Table A11, Table A12, Table A14 and Table A15 present the results of the fitting, and Table A1, Table A4, Table A7, Table A10 and Table A13 results for the obtained coefficients.
Table A1. Obtained coefficients for the considered functions (year 1–2 interval), values in brackets are confidence bounds for 95%.
Table A1. Obtained coefficients for the considered functions (year 1–2 interval), values in brackets are confidence bounds for 95%.
Linear fitting (1)a = 0.01021 (−0.1528, 0.1732), b = 2.519·10−6 (3.792·10−7, 4.658·10−6), c = 28.91 (28.74, 29.08)
Exponential of 1st order (2)a = 29.25 (29.02, 29.48), b = −3.011·10−5 (−6.18·10−5, 1.573·10−6)
Exponential of 2nd order (3)a = 30.39 (30.22, 30.57), b = −0.00035 (−0.0003961, −0.0003039), c = 0.002414 (−0.0005264, 0.005354), d = 0.01832 (0.01554, 0.02109)
Fourier 1 (5)a0 = 6.872·107 (−1.162·1015, 1.162·1015), a1 = −6.872·107 (−1.162·1015, 1.162·1015), b1 = 2.12·104 (−1.792·1011, 1.792·1011), ω = −1.377·10−6 (−11.64, 11.64)
Fourier 2 (6)a0 = 4.758·109 (−2.346·1014, 2.346·1014), a1 = −6.344e+09 (−3.128·1014, 3.127·1014),
b1 = −5.466·107 (−2.021·1012, 2.021·1012), a2 = 1.586·109 (−7.818·1013, 7.818·1013),
b2 = 2.733·107 (−1.011·1012, 1.011·1012), ω = 4.293·10−5 (−0.529, 0.5291)
Fourier 8 (7)a0 = −8.62·106 (−5.631·107, 3.907·107), a1 = 6.305·106 (−3.368·107, 4.629·107),
b1 = 1.433·107 (−6.25·107, 9.116·107), a2 = 7.899·106 (−2.917·107, 4.497·107),
b2 = −8.618·106 (−6.156·107, 4.432·107), a3 = −6.728·106 (−4.564·107, 3.218·107),
b3 = −2.288·106 (−7.388·106, 2.812·106), a4 = 2.967·105 (−6.255·106, 6.848·106),
b4 = 3.439·106 (−1.443·107, 2.131·107), a5 = 1.139·106 (−3.655·106, 5.932·106),
b5 = −6.212·105 (−5.908·106, 4.665·106), a6 = −2.82·105 (−2.176·106, 1.612·106),
b6 = −2.183·105 (−7.059·105, 2.694·105), a7 = −1.572·104 (−9.789·104, 6.645·104),
b7 = 6.237·104 (−2.793·105, 4.041·105), a8 = 5689 (−1.772·104, 2.909·104),
b8 = −940.6 (−2.177·104, 1.988·104), ω = 0.00524 (0.003601, 0.006878)
Gaussian 1 (9)a1 = 3.427·1014 (−3.9·1020, 3.9·1020), b1 = −2.005·106 (−7.582·1010, 7.581·1010),
c1 = 3.655·105 (−6.909·109, 6.91·109)
Gaussian 2 (10)a1 = 2.511·1013 (−7.073·1016, 7.078·1016), b1 = 5051 (−4.541·105, 4.642·105),
c1 = 861.5 (−4.259·104, 4.431·104), a2 = 31.26 (23, 39.52), b2 = −239 (−1875, 1397), c2 = 1316 (−3279, 5910)
Gaussian 3 (A2)a1 = 4.832·1013 (−2.725·1017, 2.726·1017), b1 = 6316 (−1.167·106, 1.18·106),
c1 = 1096 (−1.102·105, 1.124·105), a2 = 4.068 (0.9324, 7.204), b2 = 1.883 (−16.1, 19.86), c2 = 43.76 (23.13, 64.38), a3 = 28.29 (9.087, 47.5), b3 = 104 (−323, 531.1), c3 = 473.6 (−422.7, 1370)
Gaussian 7 (A3)a1 = 67.24 (−1483, 1617), b1 = 1033 (−1.778·104, 1.985·104), c1 = 705.8 (−1.017·104, 1.159·104), a2 = 23.55 (−64.26, 111.4), b2 = −9.979 (−35.28, 15.32), c2 = 138.8 (−90.6, 368.3), a3 = 10.09 (−10.02, 30.2), b3 = 171.9 (140.2, 203.6), c3 = 58.59 (9.888, 107.3),
a4 = −7.617 (−2314, 2298), b4 = 284.8 (−70.82, 640.5), c4 = 43.15 (−808.4, 894.7),
a5 = 2.474 (−0.5536, 5.502), b5 = 100.5 (93.52, 107.4), c5 = 29.94 (17.49, 42.39), a6 = 1.201 (−19.26, 21.66), b6 = 349.6 (252.2, 447), c6 = 30.59 (−61.28, 122.5),
a7 = 14.42 (−2241, 2269), b7 = 277.2 (−838.6, 1393), c7 = 48.2 (−165.1, 261.5)
Polynomial 1 (11)p1 = −0.0008497 (−0.001771, 7.112·10−5), p2 = 29.25 (29.01, 29.48)
Polynomial 2 (12)p1 = 6.516·10−5 (5.972·10−5, 7.061·10−5), p2 = −0.0292 (−0.03164, −0.02675), p3 = 31.31 (31.07, 31.54)
Polynomial 3 (13)p1 = 1.807·10−7 (1.343·10−7, 2.272·10−7), p2 = −5.277·10−5 (−8.352·10−5, −2.202·10−5),
p3 = −0.008652 (−0.01441, −0.002893), p4 = 30.56 (30.27, 30.85)
Power 1 (14)a = 31.41 (30.8, 32.01), b = −0.01529 (−0.01903, −0.01156)
Power 2 (15)a= 5.338 (4.078, 6.597), b = −0.3466 (−0.5568, −0.1364), c = 28.08 (27.17, 28.98)
Rational 1/1 (16)p1 = 26.56 (25.81, 27.31), p2 = −14.02 (−19.49, −8.555), q1 = −2.457 (−2.51, −2.404)
Rational 2/1 (17)p1 = −0.0007214 (−0.001646, 0.0002029), p2 = 29.21 (28.97, 29.44),
p3 = −80.3 (−95.63, −64.96), q1 = −2.77 (−3.257, −2.283)
Rational 3/1 (18)p1 = 6.501·10−5 (5.95·10−5, 7.051·10−5), p2 = −0.02954 (−0.03205, −0.02703), p3 = 31.49 (31.24, 31.73), p4 = −208.4 (−255.8, −161.1), q1 = −6.665 (−8.17, −5.161)
Rational 3/2 (19)p1 = −0.0006458 (−0.001574, 0.0002824), p2 = 29.19 (28.94, 29.43), p3 = −279.9 (−305.2, −254.6), p4 = 645.2 (522.9, 767.4), q1 = −9.623 (−10.42, −8.825), q2 = 22.25 (18.36, 26.14)
Rational 5/3 (20)p1 = 6.325·10−5 (5.593·10−5, 7.057·10−5), p2 = −0.02889 (−0.03347, −0.02431), p3 = 31.51 (30.55, 32.47), p4 = −371.8 (−993.9, 250.3), p5 = 1140 (−5307, 7586),
p6 = 184.8 (−1.71·104, 1.747·104), q1 = −11.96 (−31.33, 7.415), q2 = 37.02 (−166.3, 240.4), q3 = 5.183 (−543.4, 553.8)
Table A2. Quality measures for the considered functions (year 1–2 interval).
Table A2. Quality measures for the considered functions (year 1–2 interval).
Fitness FunctionSSER-SquareRMSEAdjusted R-Square
Smoothing spline, p = 0.998767188750.000358210.0069831
p = 0.90.19070.99970.033090.9993
p = 0.3094320.92270.99860.054770.998
Fourier 839.790.93890.30930.9364
Gaussian 744.330.93190.32760.9286
Gaussian 3152.10.76630.59830.7619
Fourier 2159.10.75560.60970.7527
Exponential of 2nd order205.20.68480.69070.6826
Gaussian 2217.10.66650.71210.6626
Polynomial 3248.90.61760.76080.6149
Rational 5/3281.30.56790.81350.5598
Rational 3/1282.40.56610.81130.5621
Fourier 1282.70.56560.81090.5626
Polynomial 2282.70.56560.80990.5636
Power 2549.40.15591.1290.152
Power 1569.20.12551.1480.1235
Rational 3/26340.026021.2170.01464
Rational 2/1638.20.019471.2180.01263
Linear fitting642.90.012311.2210.007731
Exponential of 1st order645.80.0077761.2230.005479
Gaussian 1645.80.0077721.2240.003168
Polynomial 16460.0075571.2230.00526
Rational 1/12.66 × 104−39.87 *7.856−40.06 *
* model results are not well conditioned by Matlab because the model is not good for this data.
Table A3. R-square for the domain used to fit, prediction interval, and total range.
Table A3. R-square for the domain used to fit, prediction interval, and total range.
Fitness FunctionFit-Domain (Samples 1–434)Prediction Interval (Samples 435–829)Total Range (Samples 1–829)Comment
Smoothing spline, p = 0.99876718875111The best fit
p = 0.90.99970.99870.9919Near the best fit
p = 0.3094320.99860.99550.9501Near the best fit
Gaussian 70.93190.95990.6899The best fit if smoothing splines are not taken into account
Fourier 80.93890.9340.7713
Gaussian 30.76630.75830.5583
Rational 5/30.560.69490.5008
Rational 3/20.56560.6666−3.154 *Not good for calculations
Fourier 20.75560.65130.6933
Gaussian 20.66650.64620.504
Polynomial 30.61760.62470.4774
Fourier 10.56560.62360.5708
Exponential 20.68480.57630.5848
Power 20.15590.53350.4715
Gaussian 10.0077720.52970.5667
Rational 3/10.56590.5290.4613
Rational 2/10.017680.52830.5817Too low
Polynomial 20.56560.52820.4597
Polynomial 10.0075570.52720.3788Too low for column1
Exponential 10.0077760.52560.486Too low for column1
Linear fitting0.012310.50230.5635Too low for column1
Power 10.12550.30730.1848
Rational 1/10.0088080.0032050.1977Too low
* Matlab warning: A negative R-square is possible if the model does not contain a constant term, and the fit is poor (worse than just fitting the mean). Try changing the model or using a different start point.
Table A4. Obtained coefficients for the considered functions (summers 1–2 interval), values in brackets are confidence bounds for 95%.
Table A4. Obtained coefficients for the considered functions (summers 1–2 interval), values in brackets are confidence bounds for 95%.
Linear fitting (1)a = 0.01683 (−0.4147, 0.4484), b = −0.000439 (−0.0005423, −0.0003357),
c = 30.13 (29.7, 30.57)
Exponential of 1st order (2)a = 31.25 (30.63, 31.87), b = −0.001501 (−0.001827, −0.001175)
Exponential of 2nd order (3)a = 31.57 (30.81, 32.33), b = −0.00183 (−0.002467, −0.001192),
c = 7.661·10−5 (−0.001465, 0.001618), d = 0.09373 (−0.09131, 0.2788)
Fourier 1 (5)a0 = 28.8 (28.6, 29.01), a1 = 0.09534 (−0.4701, 0.6608), b1 = 2.519 (2.217, 2.821),
ω = 0.05844 (0.05501, 0.06188)
Fourier 2 (6)a0 = 28.82 (28.67, 28.96), a1 = 0.1301 (−0.1963, 0.4566), b1 = 2.519 (2.301, 2.736), a2 = −1.042 (−1.245, −0.8392), b2 = 0.02157 (−0.2299, 0.273), ω = 0.05837 (0.05573, 0.061)
Fourier 8 (7)a0 = −1.122·1012 (−7.808·1013, 7.584·1013), a1 = 1.699·1012 (−1.173·1014, 1.207·1014), b1 = 1.061·1012 (−6.774·1013, 6.986·1013), a2 = −6.233·1011 (−4.929·1013, 4.805·1013),
b2 = −1.278·1012 (−8.583·1013, 8.327·1013), a3 = −8.302·1010 (−1.999·1011, 3.384·1010), b3 = 7.891·1011 (−5.367·1013, 5.525·1013), a4 = 2.104·1011 (−1.157·1013, 1.199·1013),
b4 = −2.69·1011 (−2.053·1013, 1.999·1013), a5 = −1.03·1011 (−6.63·1012, 6.424·1012), b5 = 3.73·1010 (−3.696·1012, 3.771·1012), a6 = 2.414·1010 (−1.672·1012, 1.72·1012), b6 = 5.209·109 (−4.283·109, 1.47·1010), a7 = −2.504·109 (−2.098·1011, 2.047·1011),
b7 = −2.443·109 (−1.241·1011, 1.192·1011), a8 = 5.484·107 (−7.873·109, 7.982·109), b8 = 2.28·108 (−1.38·1010, 1.426·1010), ω = 0.009698 (−0.03157, 0.05097)
Gaussian 1 (9)a1 = 34.57 (12.13, 57), b1 = −183.4 (−1057, 689.9), c1 = 561.6 (−470.3, 1594)
Gaussian 2 (10)a1 = 26.05 (15.26, 36.83), b1 = 14.46 (8.582, 20.34), c1 = 44.1 (35.12, 53.08), a2 = 27.11 (25.91, 28.31), b2 = 103.9 (98.54, 109.3), c2 = 70.06 (27.97, 112.2)
Gaussian 3 (A2)a1 = 5.155 (4.873, 5.437), b1 = 26.16 (25.76, 26.57), c1 = 9.728 (9.011, 10.44),
a2 = −1.671 (−1.947, −1.396), b2 = 72.03 (70.63, 73.44), c2 = 12.67 (9.922, 15.43), a3 = 28.78 (28.59, 28.98), b3 = 14.29 (−15.42, 44), c3 = 410.3 (269.1, 551.6)
Gaussian 7 (A3)a1 = 4.437 (3.671, 5.202), b1 = 26.49 (25.87, 27.11), c1 = 8.597 (7.324, 9.869), a2 = 29.6 (28.86, 30.34), b2 = 22.63 (13.09, 32.18), c2 = 107 (68.95, 145.1), a3 = 0.4699 (−0.03077, 0.9705), b3 = 42.14 (40.67, 43.6), c3 = 1.975 (−0.6027, 4.552), a4 = 1.863 (0.6962, 3.029), b4 = 55.07 (54.4, 55.75), c4 = 3.957 (2.218, 5.697), a5 = 12.85 (−13.58, 39.29),
b5 = 119 (62.36, 175.7), c5 = 29.6 (−249.7, 308.9), a6 = 2.819 (−70.61, 76.25), b6 = 85.94 (23.4, 148.5), c6 = 17.51 (−73.16, 108.2), a7 = 0.6927 (−0.2962, 1.682), b7 = 65.57 (63.04, 68.1), c7 = 4.351 (0.2728, 8.43)
Polynomial 1 (11)p1 = −0.04331 (−0.05269, −0.03393), p2 = 31.19 (30.6, 31.78)
Polynomial 2 (12)p1 = −7.924·10−6 (−0.0003428, 0.000327), p2 = −0.04244 (−0.08046, −0.004409), p3 = 31.17 (30.27, 32.08)
Polynomial 3 (13)p1 = 4.488·10−5 (3.633·10−5, 5.342·10−5), p2 = −0.007412 (−0.008842, −0.005983), p3 = 0.2848 (0.217, 0.3527), p4 = 28.11 (27.24, 28.97)
Power 1 (14)a = 32.62 (31.08, 34.16), b = −0.03351 (−0.04599, −0.02103)
Power 2 (15)a = −0.0117 (−0.05463, 0.03122), b = 1.274 (0.5054, 2.042), c = 30.86 (29.9, 31.81)
Rational 1/1 (16)p1 = 28.8 (28.39, 29.2), p2 = −45.01 (−143.4, 53.39), q1 = −1.57 (−4.969, 1.828)
Rational 2/1 (17)p1 = −0.04398 (−0.05356, −0.0344), p2 = 31.59 (30.91, 32.27),
p3 = −257.8 (−277.1, −238.4), q1 = −8.275 (−8.957, −7.593)
Rational 3/1 (18)p1 = 4.597e−05 (−0.0002944, 0.0003864), p2 = −0.04994 (−0.08985, −0.01003), p3 = 31.54 (30.5, 32.58), p4 = −91.31 (−111.2, −71.39), q1 = −2.9 (−3.59, −2.21)
Rational 3/2 (19)p1 = −0.04523 (−0.05495, −0.03552), p2 = 31.92 (31.18, 32.66), p3 = −416.9 (−443.9, −390), p4 = 1157 (972.2, 1342), q1 = −13.28 (−14.23, −12.32), q2 = 37.01 (30.4, 43.63)
Rational 5/3 (20)p1 = 0.0004295 (0.0002086, 0.0006504), p2 = −0.08024 (−0.1215, −0.03895), p3 = 31.99 (29.48, 34.49), p4 = −1545 (−1626, −1465), p5 = 2.44·104 (2.175·104, 2.704·104),
p6 = −4.617·104 (−6.472·104, −2.762·104), q1 = −52.8 (−55.81, −49.79), q2 = 841.7 (747.9, 935.5), q3 = −1592 (−2210, −973.6)
Table A5. Quality measures for the considered functions (summers 1 and 2).
Table A5. Quality measures for the considered functions (summers 1 and 2).
Fitness FunctionSSER-SquareRMSEAdjusted R-Square
Smoothing spline, p = 0.998767188750.000864610.021830.9999
Smoothing spline, p = 0.90.44740.9990.10180.9976
p = 0.3094321.9940.99570.16140.9939
Gaussian 75.8860.98720.25860.9843
Fourier 87.8230.9830.29320.9799
Gaussian 311.010.97610.33190.9742
Rational 5/319.510.95770.44180.9543
Fourier 259.30.87130.75880.8651
Gaussian 264.520.860.79150.8532
Fourier 1119.10.74151.0650.7341
Polynomial 3127.10.72411.10.7163
Rational 3/2245.10.46831.5420.4425
Exponential of 2nd order246.90.46421.5330.4489
Rational 2/1251.70.45381.5480.4382
Rational 3/1251.90.45331.5560.4323
Power 2254.90.4471.5510.4365
Gaussian 12580.44021.560.4297
Polynomial 1258.50.43921.5540.4339
Polynomial 2258.50.43921.5610.4286
Exponential of 1st order258.70.43871.5550.4334
Linear fitting275.90.40141.6130.3901
Power 1363.90.21031.8440.2029
Rational 1/1460.40.00090772.084−0.01794
Table A6. R-square for the domain used to fit, prediction interval, and total range in case of summers.
Table A6. R-square for the domain used to fit, prediction interval, and total range in case of summers.
Fitness FunctionFit-Domain (Summers 1 and 2)Prediction Interval (Summers 3 and 4)Total Range (all 4 Summers)Comment
Smoothing spline p = 0.9987671110.9995The best fit
Smoothing spline p = 0.90.9990.99780.996Near best fit
Gaussian 70.98720.99310.9267Best results when smoothing splines are excluded
Smoothing spline p = 0.30940.99570.99220.9858Near best fit
Fourier 80.9830.98730.9442Best results when smoothing splines are excluded
Gaussian 30.97610.95550.8218
Fourier 20.87130.94180.8169
Exponential of 2nd order0.46420.88280.5431
Gaussian 20.860.88280.7003
Rational 2/10.45160.88230.5421
Polynomial 30.72410.88040.5431
Rational 5/30.79610.8794−4.337 **
Rational 3/10.45390.87850.5471
Fourier 10.74150.87840.5373
Polynomial 20.43920.87840.5431
Rational 1/10.00090760.82440.004622
Power 10.21030.78840.3223
Power 20.4470.78840.5442
Exponential of 1st order0.43870.56040.5334
Rational 3/20.54410.55830.5466
Polynomial 10.43920.52890.5376
Linear fitting0.40140.24540.5263
Gaussian 10.4402N/A *0.543
* Infinity computed by the model function; fitting cannot continue. ** Results are not reliable due to data being not suitable for the function.
Table A7. Obtained coefficients for the considered functions (autumns 1–2 interval), values in brackets are confidence bounds for 95%.
Table A7. Obtained coefficients for the considered functions (autumns 1–2 interval), values in brackets are confidence bounds for 95%.
Linear fitting (1)a = −0.04539 (−0.3596, 0.2688), b = 0.0002863 (0.000223, 0.0003495), c = 28.74 (28.42, 29.06)
Exponential of 1st order (2)a = 28.59 (28.07, 29.1), b = 0.0006785 (0.0004205, 0.0009364)
Exponential of 2nd order (3)a = 20.28 (13.76, 26.8), b = −0.01133 (−0.01737, −0.005289), c = 11.39 (4.565, 18.22), and d = 0.007709 (0.004173, 0.01125)
Fourier 1 (5)a0 = 30.23 (29.9, 30.56), a1 = 1.47 (1.117, 1.824), b1 = −1.839 (−2.402, −1.276), ω = 0.04373 (0.03873, 0.04873)
Fourier 2 (6)a0 = 30.1 (29.93, 30.27), a1 = 1.146 (0.4727, 1.819), b1 = −1.75 (−2.099, −1.401), a2 = −0.4874 (−0.729, −0.2458), b2 = 0.3735 (−0.194, 0.9409), ω = 0.04314 (0.03784, 0.04844)
Fourier 8 (7)a0 = 29.82 (29.8, 29.84), a1 = 1.851 (1.794, 1.907), b1 = −0.9324 (−1.015, −0.8492), a2 = −0.3776 (−0.4395, −0.3156), b2 = −0.3376 (−0.3814, −0.2937), a3 = 0.1491 (0.07512, 0.2231), b3 = −0.4124 (−0.4374, −0.3874), a4 = −0.04963 (−0.1024, 0.003145), b4 = −0.07979 (−0.1323, −0.02726), a5 = −0.2298 (−0.2614, −0.1982), b5 = 0.1685 (0.101, 0.236), a6 = −0.01184 (−0.06022, 0.03655), b6 = −0.1228 (−0.1505, −0.09501), a7 = −0.1947 (−0.2209, −0.1685), b7 = −0.005933 (−0.09148, 0.07961), a8 = −0.07079 (−0.1274, −0.0142), b8 = 0.09574 (0.05283, 0.1386), ω = 0.05152 (0.05085, 0.05219)
Gaussian 1 (9)a1 = 4.026·1096 (−1.429·10103, 1.429·10103), b1 = 6.461·105 (−1.047·1010, 1.047·1010), c1 = 4.365·104 (−3.536·108, 3.536·108)
Gaussian 2 (10)a1 = 25.59 (17.38, 33.79), b1 = 123.3 (120.4, 126.3), c1 = 55.5 (43.33, 67.67), a2 = 30.31 (29.55, 31.07), b2 = 1.598 (−2.844, 6.039), c2 = 95.27 (56.39, 134.1)
Gaussian 3 (A2)a1 = 32.85 (32.26, 33.43), b1 = 130 (116.7, 143.3), c1 = 155 (126, 184), a2 = 12.88 (10.6, 15.15), b2 = −1.083 (−3.304, 1.138), c2 = 17.47 (11.68, 23.26), a3 = 7.65 (4.925, 10.38), b3 = 27.5 (21.28, 33.73), c3 = 20.62 (15.55, 25.69)
Gaussian 7 (A3)a1 = 32.69 (30.32, 35.06), b1 = 112.7 (108.1, 117.4), c1 = 27.42 (18.18, 36.67), a2 = 30.76 (29.99, 31.53), b2 = 3.096 (−2.256, 8.448), c2 = 45.15 (8.354, 81.94), a3 = 10.8 (−35.29, 56.89), b3 = 82.33 (80.03, 84.63), c3 = 11.69 (2.008, 21.37), a4 = 1.048 (−92.15, 94.25), b4 = 26.08 (−82.89, 135.1), c4 = 11.07 (−89.58, 111.7), a5 = 20.26 (−1.597, 42.11), b5 = 63.03 (56, 70.06), c5 = 20.46 (−48.7, 89.62), a6 = 7.41 (−95.68, 110.5), b6 = 37.05 (−22.28, 96.38), c6 = 14.19 (−81.59, 110), a7 = 4.07 (1.256, 6.884), b7 = 95.85 (94.26, 97.43), c7 = 7.985 (6.273, 9.697)
Polynomial 1 (11)p1 = 0.01943 (0.01172, 0.02714), p2 = 28.62 (28.09, 29.15)
Polynomial 2 (12)p1 = 0.001234 (0.001121, 0.001347), p2 = −0.1274 (−0.1412, −0.1136), p3 = 31.56 (31.2, 31.91)
Polynomial 3 (13)p1 = −1.947·10−6 (−5.712·10−6, 1.818·10−6), p2 = 0.001581 (0.0009001, 0.002263),
p3 = −0.144 (−0.179, −0.109), p4 = 31.72 (31.24, 32.21)
Power 1 (14)a = 29.11 (27.92, 30.3), b = 0.005937 (−0.004536, 0.01641)
Power 2 (15)a = 1.284·10−10 (−9.654·10−10, 1.222·10−10), b = 5.096 (3.291, 6.901), c = 28.99 (28.72, 29.27)
Rational 1/1 (16)p1 = 29.69 (29.3, 30.08), p2 = 17.66 (−236.1, 271.4), q1 = 0.5237 (−7.731, 8.778)
Rational 2/1 (17)p1 = 1.366 (−3.127, 5.859), p2 = −110.5 (−574.6, 353.7), p3 = 3.269·104 (−8.198·104, 1.474·105), q1 = 1033 (−2601, 4666)
Rational 3/1 (18)p1 = 0.001356 (0.001122, 0.001589), p2 = −0.142 (−0.166, −0.1181), p3 = 31.82 (30.7, 32.94), p4 = 111.7 (−392.1, 615.4), q1 = 3.678 (−12.62, 19.98)
Rational 3/2 (19)p1 = 0.01698 (0.004851, 0.02911), p2 = 29.55 (27.51, 31.58), p3 = −3671 (95% confidence bounds: −4079, −3263), p4 = 1.327·105 (1.116·105, 1.539·105), q1 = −118.8 (−129.5, −108.1), q2 = 4350 (3678, 5021)
Rational 5/3 (20)p1 = 0.00135 (0.001112, 0.001588), p2 = −0.1589 (−0.1877, −0.1301), p3 = 33.7 (31.57, 35.84), p4 = −317.1 (−1029, 394.7), p5 = 38.35 (−6943, 7020), p6 = 3728 (−1.678e+04, 2.424e+04), q1 = −9.637 (−32.02, 12.74), q2 = −0.6917 (−224.9, 223.5), q3 = 123 (−541, 786.9)
Table A8. Quality measures for the considered functions (autumns 1 and 2).
Table A8. Quality measures for the considered functions (autumns 1 and 2).
Fitness FunctionSSER-SquareRMSEAdjusted R-Square
Smoothing spline, p = 0.998767188750.000133810.0082461
p = 0.90.073890.99970.039720.9994
p = 0.3094320.40070.99860.06950.9981
Fourier 80.94640.99680.097280.9962
Gaussian 70.92490.99680.097650.9962
Gaussian 39.530.96740.29570.965
Rational 3/214.80.94940.36350.9471
Fourier 217.10.94150.39070.9389
Gaussian 219.350.93380.41570.9308
Fourier 132.920.88740.53740.8844
Rational 5/345.340.84490.64490.8335
Rational 3/145.430.84460.63410.8391
Polynomial 346.80.83990.64070.8357
Rational 2/146.980.83930.64190.8351
Polynomial 247.230.83840.64090.8356
Exponential of 2nd order47.870.83620.6480.8319
Power 2126.30.56781.0480.5603
Linear fitting172.10.41141.2230.4011
Exponential of 1st order238.60.18381.4340.1767
Gaussian 1238.60.18381.440.1696
Polynomial 1240.70.17671.440.1696
Rational 1/1289.10.010991.586−0.00621
Power 1289.30.010421.5790.001884
Table A9. R-square for domain used to fit, prediction interval, and total range (autumns).
Table A9. R-square for domain used to fit, prediction interval, and total range (autumns).
Fitness FunctionFit-Domain (Autumns 1 and 2)Prediction Interval (Autumns 3 and 4)Total Range (all 4 Autumns)Comment
Smoothing spline, p = 0.99876718875111The best fit
p = 0.90.99970.99710.9984
p = 0.3094320.99860.98970.9942
Gaussian 70.99680.98590.9742Best choice when smoothing splices are excluded.
Rational 3/20.94940.97410.6388
Fourier 80.99680.97330.9494Near best choice when smoothing splices are excluded.
Gaussian 30.96740.94840.8847
Rational 5/30.84490.89140.6214
Fourier 20.94150.88140.791
Gaussian 20.93380.87950.7352
Fourier 10.88740.87130.6388
Polynomial 30.83990.8650.6507
Exponential of 2nd order0.83620.8287−4.861 ** Matlab warning
Power 20.56780.79950.6424
Gaussian 10.18380.79590.6381
Rational 3/10.84460.79530.6968
Polynomial 20.83840.79310.6388
Rational 2/10.83930.79090.4735
Polynomial 10.17670.78580.473
Exponential of 1st order0.18380.78120.4593
Linear fitting0.41140.76090.5964
Power 10.010420.48930.2602
Rational 1/10.010990.029380.4729
Table A10. Obtained coefficients for the considered functions (winters 1–2 interval), values in brackets are confidence bounds for 95%.
Table A10. Obtained coefficients for the considered functions (winters 1–2 interval), values in brackets are confidence bounds for 95%.
Linear fitting (1)a = −0.02188 (−0.2189, 0.1752), b = 6.907·10−5 (1.449·10−5, 0.0001237), c = 28.12 (27.92, 28.32)
Exponential of 1st order (2)a = 28.31 (bounds: 28.02, 28.6), and b = −6.32·10−6 (−0.0001787, 0.0001661)
Exponential of 2nd order (3)a = 28.79 (27.71, 29.87), b = −0.00243 (−0.003808, −0.001053), c = 0.6529 (−0.5924, 1.898), d = 0.02388 (0.01006, 0.0377)
Fourier 1 (5)a0 = 28.16 (28.09, 28.23), a1 = 0.5406 (0.3579, 0.7233), b1 = 0.7318 (0.5792, 0.8843), ω = 0.07365 (0.06925, 0.07804)
Fourier 2 (6)a0 = 28.44 (28.31, 28.56), a1 = 0.8091 (0.4632, 1.155), b1 = −0.3872 (−0.8644, 0.09007), a2 = −0.3317 (−0.4104, −0.2531), b2 = 0.1166 (−0.3926, 0.6258), ω = 0.05057 (0.04012, 0.06102)
Fourier 8 (7)a0 = 28.4 (28.31, 28.49), a1 = 0.9113 (0.755, 1.068), b1 = −0.2112 (−0.4711, 0.04881), a2 = −0.2993 (−0.4587, −0.1399), b2 = −0.06535 (−0.2176, 0.08691), a3 = −0.04238 (−0.1884, 0.1037), b3 = −0.1498 (−0.2659, −0.03362), a4 = 0.09473 (−0.03816, 0.2276),
b4 = −0.01995 (−0.1303, 0.09044), a5 = −0.1726 (−0.2623, −0.0828), b5 = 0.1299 (−0.1434, 0.4033), a6 = −0.02845 (−0.09781, 0.04091), b6 = 0.0155 (−0.07253, 0.1035), a7 = 0.07436 (−0.0176, 0.1663), b7 = −0.004225 (−0.09292, 0.08447), a8 = −0.06261 (−0.2146, 0.08941), b8 = 0.1033 (−0.0578, 0.2644), ω = 0.05442 (0.05006, 0.05879)
Gaussian 1 (9)a1 = 28.35 (−60.17, 116.9), b1 = −637.7 (−1.061·106, 1.06·106), c1 = 1.529·104 (−1.175·107, 1.178·107)
Gaussian 2 (10)a1 = 19.57 (7.279, 31.86), b1 = 116.5 (110.4, 122.6), c1 = 51.55 (32.7, 70.41), a2 = 28.71 (27.94, 29.48), b2 = 7.058 (0.8335, 13.28), c2 = 97.94 (51.13, 144.8)
Gaussian 3 (A2)a1 = 29.18 (26.94, 31.43), b1 = 113 (106.9, 119.2), c1 = 102.8 (15.34, 190.2), a2 = 21.12 (3.491, 38.75), b2 = −13.39 (−19.73, −7.037), c2 = 61.1 (34.01, 88.18), a3 = 1.117 (0.893, 1.34), b3 = 31.73 (30.94, 32.53), c3 = 5.535 (4.146, 6.924)
Gaussian 7 (A3)a1 = 29.83 (28.1, 31.57), b1 = 117.1 (64.29, 170), c1 = 140.8 (−85.14, 366.8), a2 = 14.57 (−13.05, 42.2), b2 = −9.303 (−34.14, 15.53), c2 = 44.71 (−6.636, 96.06), a3 = 1.096 (0.7929, 1.4), b3 = 34.79 (34.52, 35.07), c3 = 1.847 (1.272, 2.422), a4 = 0.5781 (−0.6638, 1.82), b4 = 46.14 (42.81, 49.47), c4 = 6.095 (0.2361, 11.95), a5 = 1.61 (−0.7591, 3.98), b5 = 29.99 (26.84, 33.14), c5 = 9.895 (3.158, 16.63), a6 = 0.8952 (−1.199, 2.99), b6 = 56.75 (46.22, 67.27), c6 = 11.02 (−5.995, 28.03), a7 = 0.6716 (0.3439, 0.9993), b7 = 79 (78.26, 79.75), c7 = 4.245 (2.401, 6.089)
Polynomial 1 (11)p1 = −0.0001754 (−0.005053, 0.004702), p2 = 28.31 (28.02, 28.6)
Polynomial 2 (12)p1 = 0.0007765 (0.0006731, 0.0008799), p2 = −0.08015 (−0.09115, −0.06916), p3 = 29.69 (29.45, 29.94)
Polynomial 3 (13)p1 = 8.683e−06 (5.058e−06, 1.231e−05), p2 = −0.000565 (−0.001133, 2.833e−06),
p3 = −0.02461 (−0.04985, 0.0006242), p4 = 29.2 (28.9, 29.51)
Power 1 (14)a = 28.99 (28.41, 29.57), b = −0.006629 (−0.01191, −0.001349)
Power 2 (15)a = 1.526 (bounds: 0.4475, 2.604), b = −0.4284 (−1.337, 0.4801), and c = 27.95 (26.86, 29.03)
Rational 1/1 (16)p1 = 28.28 (28.14, 28.43), p2 = −50.27 (−80.95, −19.59), q1 = −1.783 (−2.842, −0.7249)
Rational 2/1 (17)p1 = 0.0004235 (bounds: −0.00459, 0.005437), p2 = 28.26 (27.95, 28.57),
p3 = −50.2 (−80.05, −20.35), q1 = −1.782 (−2.812, −0.7532)
Rational 3/1 (18)p1 = 0.001457 (0.0008546, 0.002059), p2 = −0.1586 (−0.2211, −0.09621), p3 = 31.29 (30.13, 32.46), p4 = 735.3 (−433.5, 1904), q1 = 25.65 (−14.72, 66.02)
Rational 3/2 (19)p1 = 0.0005795 (−0.004503, 0.005662), p2 = 28.25 (27.88, 28.61),
p3 = −338.4 (−388.7, −288.1), p4 = 894 (bounds: 596.2, 1192),
q1 = −11.98 (−13.69, −10.27), q2 = 31.66 (21.45, 41.86)
Rational 5/3 (20)p1 = 0.001389 (0.0009086, 0.001869), p2 = −0.1762 (−0.2345, −0.1179), p3 = 34.02 (32.02, 36.01), p4 = −70.51 (−935, 793.9), p5 = −2716 (−1.161·104, 6177), p6 = 1.366·104 (bounds: −1.202·104, 3.934·104), q1 = −0.4298 (−30.31, 29.45), q2 = −103.4 (−412.6, 205.8), q3 = 487.4 (−412.4, 1387)
Table A11. Quality measures for the considered functions (winters 1 and 2).
Table A11. Quality measures for the considered functions (winters 1 and 2).
Fitness FunctionSSER-SquareRMSEAdjusted R-Square
Smoothing spline, p = 0.998767188750.000571910.018370.9994
p = 0.90.26650.9950.081240.9875
p = 0.3094320.96640.98190.11620.9745
Gaussian 71.1020.97940.11670.9743
Fourier 81.9970.96260.15420.9551
Fourier 27.0850.86750.27170.8605
Gaussian 28.0460.84950.28950.8416
Fourier 18.6010.83910.29620.8342
Rational 5/311.540.78410.35230.7655
Rational 3/111.60.78310.34570.7741
Polynomial 313.40.74940.36970.7417
Exponential of 2nd order14.60.72680.3860.7184
Polynomial 216.490.69160.40810.6853
Linear fitting50.230.060320.71230.04134
Power 150.380.057410.70980.04798
Rational 3/252.470.018340.7393−0.03279 *
Rational 2/152.780.012540.7339−0.01768 *
Rational 1/152.80.012260.7303−0.007692 *
Gaussian 33.7940.9290.2020.9229
Power 249.40.07580.70640.05713
Exponential of 1st order53.455.189·10−50.7311−0.009948 *
Polynomial 153.455.09·10−50.7311−0.009949 *
Gaussian 153.46−0.0001639 *0.7348−0.02037 *
* only calculated data by Matlab without interpretation (negative numbers should be impossible).
Table A12. R-square for the domain used to fit, prediction interval, and total range for winters.
Table A12. R-square for the domain used to fit, prediction interval, and total range for winters.
Fitness FunctionFit-Domain (Winters 1–2)Prediction Interval (Winters 3–4)Total Range (all 4 Winters)Comment
Smoothing spline, p = 0.99876718875111The best fit
p = 0.90.9950.99780.9985
Gaussian 70.97940.99310.9267
Smoothing spline p = 0.3094320.98190.99220.9945
Fourier 80.96260.98730.9761
Gaussian 30.9290.95550.9466
Fourier 20.86750.94180.9354
Rational 5/30.78410.88290.9429
Exponential of 2nd order0.72680.88280.8571
Gaussian 20.84950.88280.9104
Polynomial 30.74940.88040.8665
Fourier 10.83910.87840.8458
Polynomial 20.69160.87840.8458
Rational 3/10.78310.8810.8477
Power 10.057410.78840.3671
Power 20.07580.78840.8614
Exponential of 1st order5.189·10−50.56040.656
Rational 3/20.018340.53970.9403
Rational 2/10.012540.53320.6771
Polynomial 15.09·10−50.52890.6745
Linear fitting0.060320.24540.816
Rational 1/10.012260.013290.0109
Gaussian 1−0.0001639 *N/A0.841Inf computed by model function, fitting cannot continue.
* Matlab warning: A negative R-square is possible if the model does not contain a constant term and the fit is poor (worse than just fitting the mean). Try changing the model or using a different start point.
Table A13. Obtained coefficients for the considered functions (springs 1–2 interval), values in brackets are confidence bounds for 95%.
Table A13. Obtained coefficients for the considered functions (springs 1–2 interval), values in brackets are confidence bounds for 95%.
Linear fitting (1)a = −0.001745 (−0.1453, 0.1418), b = −2.716·10−5 (−6.444·10−5, 1.011·10−5), c = 29.09 (28.94, 29.23)
Exponential of 1st order (2)a = 28.97 (28.77, 29.18), b = 2.768·10−5 (−8.86·10−5, 0.000144)
Exponential of 2nd order (3)a = −0.01843 (−0.07621, 0.03936), b = 0.04838 (0.02118, 0.07558), c = 28.47 (28.28, 28.65), d = 0.0007368 (0.0003567, 0.001117)
Fourier 1 (5)a0 = 29.09 (29.02, 29.16), a1 = 0.02635 (−0.1644, 0.2171),
b1 = −0.5724 (−0.6691, −0.4757), ω = 0.07486 (0.06923, 0.08049)
Fourier 2 (6)a0 = 29.12 (29.06, 29.17), a1 = 0.0604 (−0.05226, 0.1731), b1 = −0.5426 (−0.6154, −0.4699), a2 = −0.1226 (−0.2415, −0.003689), b2 = −0.3169 (−0.4064, −0.2273), ω= 0.07436 (0.07171, 0.077)
Fourier 8 (7)a0 = 28.99 (28.94, 29.04), a1 = −0.502 (−0.5794, −0.4245), b1 = −0.1689 (−0.3016, −0.03619), a2 = −0.0574 (−0.2131, 0.09831), b2 = 0.3611 (0.3103, 0.412), a3 = 0.199 0.1409, 0.2571), b3 = −0.03921 (−0.1616, 0.08319), a4 = −0.03603 (−0.08402, 0.01196), b4 = 0.09401 (0.05168, 0.1363), a5 = 0.08342 (−0.07823, 0.2451), b5 = −0.1118 (−0.17, −0.05353), a6 = −0.1135 (−0.2218, −0.005219), b6 = 0.1204 (−0.04748, 0.2884), a7 = 0.06244 (0.02226, 0.1026), b7 = −0.004121 (−0.08272, 0.07448), a8 = −0.01616 (−0.09508, 0.06275), b8 = 0.07005 (0.02402, 0.1161), ω = 0.05696 (0.05373, 0.06018)
Gaussian 1 (9)a1 = 29.38 (29.26, 29.5), b1 = 54.13 (50.75, 57.51), c1 = 270 (236.6, 303.5)
Gaussian 2 (10)a1 = 29.57 (29.36, 29.77), b1 = 68.9 (59.9, 77.91), c1 = 144.9 (107.5, 182.2), a2 = 6.047 (−1.507, 13.6), b2 = −15.25 (−45.87, 15.36), c2 = 36.4 (8.628, 64.18)
Gaussian 3 (A2)a1 = 1.419 (0.35, 2.487), b1 = 70.92 (66.05, 75.8), c1 = 14.86 (9.047, 20.67), a2 = 78.21 (−4.145·105, 4.147·105), b2 = 6155 (−3.252·107, 3.253·107), c2 = 6144 (−1.623·107, 1.625·107), a3 = −1.338 (−44.12, 41.44), b3 = 101.9 (−435.5, 639.2), c3 = 41.61 (−449.4, 532.6)
Gaussian 7 (A3)a1 = 1.073 (0.8171, 1.329), b1 = 60.96 (60.51, 61.42), c1 = 3.182 (2.307, 4.057), a2 = 29.82 (29.74, 29.9), b2 = 71.53 (70.08, 72.99), c2 = 79.17 (66.86, 91.48),
a3 = 1.615 (−19.11, 22.34), b3 = 43.72 (27.25, 60.19), c3 = 7.56 (−7.447, 22.57), a4 = 15.4 (−58.75, 89.55), b4 = −4.202 (−90.21, 81.8), c4 = 18.05 (−271.1, 307.2),
a5 = 6.021 (−266.5, 278.5), b5 = 18.45 (−40.3, 77.2), c5 = 13.43 (−288.4, 315.2),
a6 = 3.458 (1.835, 5.081), b6 = 107.5 (103.7, 111.3), c6 = 10.99 (7.16, 14.81),
a7 = 3.096 (−142.1, 148.3), b7 = 32.69 (−91.44, 156.8), c7 = 10.67 (−113.4, 134.8)
Polynomial 1 (11)p1 = 0.0008112 (−0.002563, 0.004185), p2 = 28.97 (28.77, 29.18)
Polynomial 2 (12)p1 = −0.0003957 (−0.0004937, −0.0002976), p2 = 0.04275 (0.03203, 0.05348), p3 = 28.22 (27.98, 28.47)
Polynomial 3 (13)p1 = −1.111·10−5 (−1.41·10−5, −8.131·10−6), p2 = 0.001371 (0.0008905, 0.001852),
p3 = −0.03252 (−0.05451, −0.01053), p4 = 28.9 (28.63, 29.17)
Power 1 (14)a = 28.65 (28.23, 29.06), b = 0.003467 (−0.0003017, 0.007236)
Power 2 (15)a = −0.8686 (−3.288, 1.551), b = −0.2256 (−1.632, 1.181), c = 29.4 (26.49, 32.32)
Rational 1/1 (16)p1 = 29.02 (confidence bounds: 28.91, 29.12), p2 = −50.97 (−125.8, 23.84), q1 = −1.755 (−4.35, 0.8403)
Rational 2/1 (17)p1 = 0.0006562 (confidence bounds: −0.002789, 0.004101), p2 = 28.98 (28.74, 29.22), p3 = −257 (−303.3, −210.7), q1 = −8.865 (−10.48, −7.253)
Rational 3/1 (18)p1 = −0.000405 (−0.0005055, −0.0003045), p2 = 0.0456 (0.03414, 0.05707), p3 = 28.01 (27.72, 28.31), p4 = −116.4 (−147.3, −85.38), q1 = −4.125 (−5.196, −3.054)
Rational 3/2 (19)p1 = −18.77 (−8290, 8252), p2 = 2122 (−9.201·105, 9.244·105), p3 = 1.299·106 (−5.72·108, 5.746·108), p4 = −3.706·106 (−1.638·109, 1.631·109), q1 = 4.631·104 (−2.038·107, 2.048·107), q2 = −1.316·105 (−5.817·107, 5.791·107)
Rational 5/3 (20)p1 = −8.931·10−5 (−0.0003049, 0.0001263), p2 = 0.01715 (−0.03251, 0.0668), p3 = 27.42 (23.49, 31.35), p4 = −4117 (−4561, −3673), p5 = 1.603e+05 (1.304·105, 1.902·105),
p6 = −2.4·105 (−4.535·105, −2.649·104), q1 = −145 (bounds: −158.3, −131.6), q2 = 5597 (4569, 6624), q3 = −8388 (−1.586·104, −918.6)
Table A14. Quality measures for the considered functions (domain used for prediction) for springs 1–2.
Table A14. Quality measures for the considered functions (domain used for prediction) for springs 1–2.
Fitness FunctionSSER-SquareRMSEAdjusted R-Square
Smoothing spline, p = 0.998767188750.000504210.016990.999
p = 0.90.24880.99140.077350.9784
p = 0.3094320.82440.97140.10580.9596
Gaussian 71.4070.95120.12940.9395
Fourier 81.7050.94080.140.9293
Gaussian 35.6260.80470.24210.7885
Rational 5/36.3910.77820.2580.7598
Fourier 26.4450.77630.25520.765
Gaussian 210.640.63070.32780.6121
Polynomial 311.460.60230.33680.5905
Fourier 112.160.57810.3470.5655
Exponential of 2nd order13.640.52650.36750.5125
Rational 3/117.450.39450.41770.3703
Rational 3/217.480.39330.42020.3627
Gaussian 117.630.38820.41570.3762
Polynomial 217.660.38720.41610.3752
Power 227.830.034060.52240.01512
Power 127.910.031530.52050.02213
Linear fitting28.240.02010.52610.0008901
Rational 2/128.620.0067460.5323−0.02276
Polynomial 128.750.0022030.5284−0.007485
Exponential of 1st order28.750.0021810.5284−0.007507 *
Rational 1/128.760.0019360.531−0.01763
* Matlab warning: A negative R-square is possible if the model does not contain a constant term and the fit is poor (worse than just fitting the mean). Try changing the model or using a different start point.
Table A15. R-square for the domain used to fit, prediction interval, and total range in case of springs.
Table A15. R-square for the domain used to fit, prediction interval, and total range in case of springs.
Fitness FunctionFit-Domain (Springs 1–2)Prediction Interval (Springs 3–4)Total Range (all 4 Springs)Comment
Smoothing spline, p = 0.99876718875111The best fit
p = 0.90.99140.9950.9918
p = 0.3094320.97140.98270.9717
Fourier 80.94080.95960.8896The best for spring 3–4 when smoothing splines are excluded
Gaussian 70.95120.95010.9062The best for whole dataset and springs 1–2 when smoothing splines are excluded
Gaussian 30.80470.93450.5874
Rational 3/20.39330.91740.8254
Fourier 20.77630.83650.5747
Gaussian 20.63070.82520.5659
Fourier 10.57810.73250.3169
Rational 5/30.77820.7256−66.28 *
Rational 3/10.39450.71850.2409
Polynomial 30.60230.71020.2664
Polynomial 20.38720.710.2399
Exponential of 2nd order0.52650.7060.2258
Power 20.034060.52150.2072
Power 10.031530.51810.04345
Exponential of 1st order0.0021810.37520.1342
Gaussian 10.38820.37520.2458
Rational 2/10.0067460.35580.1383
Polynomial 10.0022030.35180.1375
Linear fitting0.02010.14750.1975
Rational 1/10.0019360.027940.0001715
* A negative R-square is possible if the model does not contain a constant term and the fit is poor (worse than just fitting the mean). Try changing the model or using a different start point.

References

  1. Marques, C.H.; Caprace, J.-D.; Belchior, C.R.P.; Martini, A. An Approach for Predicting the Specific Fuel Consumption of Dual-fuel Two-stroke Marine Engines. J. Mar. Sci. Eng. 2019, 7, 20. [Google Scholar] [CrossRef] [Green Version]
  2. Tran, T.A. Design the Prediction Model of Low-Sulfur-Content Fuel Oil Consumption for M/V NORD VENUS 80,000 DWT Sailing on Emission Control Areas by Artificial Neural Networks. In Proceedings of the Institution of Mechanical Engineers, Part M. J. Eng. Marit. Environ. 2019, 233, 345–362. [Google Scholar]
  3. Krčum, M.; Gudelj, A.; Tomas, V. Optimal Design of Ship’s Hybrid Power System for Efficient Energy. Trans. Marit. Sci. 2018, 7, 23–32. [Google Scholar] [CrossRef] [Green Version]
  4. Geertsma, R.D.; Visser, K.; Negenborn, R.R. Adaptive pitch control for ships with diesel mechanical and hybrid propulsion. Appl. Energy 2018, 228, 2490–2509. [Google Scholar] [CrossRef]
  5. Aronietis, R.; Sys, C.; van Hassel, E.; Vanelslander, T. Forecasting Port-level Demand for LNG as a Ship Fuel: The Case of the Port of Antwerp. J. Shipp. Trade 2016, 1, 2. [Google Scholar] [CrossRef] [Green Version]
  6. Tsujimoto, M.; Sogihara, N. Prediction of Fuel Consumption of Ships in Actual Seas. Available online: http://www.naoe.eng.osaka-u.ac.jp/kashi/SOEMeeting/PPT/SOE1_Tsujimoto.pdf (accessed on 5 November 2019).
  7. Górski, W.; Abramowicz-Gerigk, T.; Burciu, Z. The Influence of Ship Operational Parameters on Fuel Consumption. Zesz. Nauk. 2013, 36, 49–54. [Google Scholar]
  8. Bialystocki, N.; Konovessis, D. On the Estimation of Ship’s Fuel Consumption and Speed Curve: A Statistical Approach. J. Ocean Eng. Sci. 2016, 1, 157–166. [Google Scholar] [CrossRef] [Green Version]
  9. Lu, R.; Turan, O.; Boulougouris, E. Voyage Optimisation: Prediction of Ship Specific Fuel Consumption for Energy Efficient Shipping. In Proceedings of the 3rd International Conference on Technologies, Operations, Logistics and Modelling for Low Carbon Shipping, London, UK, 9–10 September 2013. [Google Scholar]
  10. Banawan, A.A.; Mosleh, M.; Seddiek, I.S. Prediction of the Fuel Saving and Emissions Reduction by Decreasing Speed of a Catamaran. J. Mar. Eng. Technol. 2014, 12, 40–48. [Google Scholar]
  11. Larsen, U.; Pierobon, L.; Baldi, F.; Haglind, F.; Ivarsson, A. Development of a Model for the Prediction of the Fuel Consumption and Nitrogen Oxides Emission Trade-off for Large Ships. Energy 2015, 80, 545–555. [Google Scholar] [CrossRef] [Green Version]
  12. Gheriani, E. Fuel Consumption Prediction Methodology for Early Stages of Naval Ship Design. Master’s Thesis, MIT, Massachusetts Institute of Technology, Cambridge, MA, USA, 2012. [Google Scholar]
  13. Pétursson, S. Predicting Optimal Trim Configuration of Marine Vessels with Respect to Fuel Usage. Master’s Thesis, University of Iceland, Reykjavik, Iceland, 2009. [Google Scholar]
  14. Borkowski, T.; Kasyk, L.; Kowalak, P. Assessment of Ship’s Engine Effective Power Fuel Consumption and Emission using the Vessel Speed. J. KONES Powertrain Transp. 2011, 18, 31–39. [Google Scholar]
  15. Demirel, A.K.; Turan, O.; Incecik, A. Predicting the effect of biofouling on ship resistance using CFD. Appl. Ocean Res. 2017, 62, 100–118. [Google Scholar] [CrossRef] [Green Version]
  16. Kan, Z.; Tang, L.; Kwan, M.P.; Zhang, X. Estimating Vehicle Fuel Consumption and Emissions Using GPS Big Data. Int. J. Environ. Res. Public Health 2018, 15, 566. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Zhao, Q.; Chen, Q.; Wang, L. Real-Time Prediction of Fuel Consumption Based on Digital Map API. Appl. Sci. 2019, 9, 1369. [Google Scholar] [CrossRef] [Green Version]
  18. Kee, K.K.; Simon, B.Y.L.; Renco, K.H.Y. Prediction of Ship Fuel Consumption and Speed Curve by Using Statistical Method. J. Comp. Sci. Comput. Math. 2018, 8, 19–24. [Google Scholar] [CrossRef] [Green Version]
  19. Hu, Z.; Jin, Y.; Hu, Q.; Sen, S.; Zhou, T.; Osman, M.T. Prediction of Fuel Consumption for Enroute Ship Based on Machine Learning. IEEE Access 2019, 7, 119497–119505. [Google Scholar] [CrossRef]
  20. Wang, S.; Ji, B.; Zhao, J.; Liu, W.; Xu, T. Predicting Ship Fuel Consumption based on LASSO Regression. Transp. Res. D Transp. Environ. 2018, 65, 817–824. [Google Scholar] [CrossRef]
  21. Ahlgren, F.; Mondejar, M.E.; Thern, M. Predicting Dynamic Fuel Oil Consumption on Ships with Automated Machine Learning. Energy Procedia 2019, 158, 6126–6131. [Google Scholar] [CrossRef]
  22. Panapakidis, I.; Sourtzi, V.-M.; Dagoumas, A. Forecasting the Fuel Consumption of Passenger Ships with a Combination of Shallow and Deep Learning. Electronics 2020, 9, 776. [Google Scholar] [CrossRef]
  23. Uyanik, T.; Arslanoglu, Y.; Kalenderli, O. Ship Fuel Consumption Prediction with Machine Learning. In Proceedings of the 4th International Mediterranean Science and Engineering Congress, Antalya, Turkey, 25–27 April 2019; pp. 757–759. [Google Scholar]
  24. Mosaic Data Science, Predicting Fuel Usage Anomalies, Case Study. Available online: http://www.mosaicdatascience.com/2018/05/11/predicting-fuel-usage-anomalies/ (accessed on 20 October 2019).
  25. Coraddu, A.; Oneto, L.; Baldi, F.; Anguita, D. Vessels Fuel Consumption: A Data Analytics Perspective to Sustainability. In Soft Computing for Sustainability Science; Studies in Fuzziness and Soft Computing; Cruz Corona, C., Ed.; Springer: Cham, Switzerland, 2018; pp. 11–48. [Google Scholar]
  26. MathWorks Inc. Curve Fitting Toolbox™ User’s Guide R2018b; MathWorks Inc.: Natick, MA, USA, 2018. [Google Scholar]
  27. Oppenheim, A.V.; Verghese, G.C. Signals, Systems and Inference; Pearson Education Limited: New York, NY, USA, 2016. [Google Scholar]
  28. Wang, Z.; Bovik, A.C. Mean Squared Error: Love It or Leave It? IEEE Signal Proc. Mag. 2009, 26, 98–117. [Google Scholar] [CrossRef]
  29. Botchkarev, A. A New Typology Design of Performance Metrics to Measure Errors in Machine Learning Regression Algorithms. Interdiscip. J. Inf. Knowl. Manag. 2019, 4, 45–76. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Preprocessing and overall process flow.
Figure 1. Preprocessing and overall process flow.
Jmse 08 00625 g001
Figure 2. Smoothing spline for p = 0.309432: (a) years 1–2, (b) years 3–4, (c) years 1–4.
Figure 2. Smoothing spline for p = 0.309432: (a) years 1–2, (b) years 3–4, (c) years 1–4.
Jmse 08 00625 g002
Figure 3. Gaussian 3 fitting function: (a) years 1–2, (b) years 3–4, (c) years 1–4.
Figure 3. Gaussian 3 fitting function: (a) years 1–2, (b) years 3–4, (c) years 1–4.
Jmse 08 00625 g003
Figure 4. Gaussian 7 fitting function: (a) summers 1–2, (b) summers 3–4, (c) summers 1–4.
Figure 4. Gaussian 7 fitting function: (a) summers 1–2, (b) summers 3–4, (c) summers 1–4.
Jmse 08 00625 g004
Figure 5. Smoothing spline for p = 0.9987: (a) summers 1–2, (b) summers 3–4, (c) summers 1–4.
Figure 5. Smoothing spline for p = 0.9987: (a) summers 1–2, (b) summers 3–4, (c) summers 1–4.
Jmse 08 00625 g005
Figure 6. Gaussian 7 fitting function results: (a) autumns 1–2, (b) autumns 3–4, (c) autumns 1–4.
Figure 6. Gaussian 7 fitting function results: (a) autumns 1–2, (b) autumns 3–4, (c) autumns 1–4.
Jmse 08 00625 g006
Figure 7. Rational3/2 fitting function results: (a) autumns 1–2, (b) autumns 3–4, (c) autumns 1–4.
Figure 7. Rational3/2 fitting function results: (a) autumns 1–2, (b) autumns 3–4, (c) autumns 1–4.
Jmse 08 00625 g007
Figure 8. Gaussian 7 fitting function results: (a) winters 1–2, (b) winters 3–4, (c) winters 1–4.
Figure 8. Gaussian 7 fitting function results: (a) winters 1–2, (b) winters 3–4, (c) winters 1–4.
Jmse 08 00625 g008
Figure 9. Fourier 8 fitting function results: (a) springs 1–2, (b) springs 3–4, (c) springs 1–4.
Figure 9. Fourier 8 fitting function results: (a) springs 1–2, (b) springs 3–4, (c) springs 1–4.
Jmse 08 00625 g009
Figure 10. Gaussian 7 fitting function results: (a) springs 1–2, (b) springs 3–4, (c) springs 1–4.
Figure 10. Gaussian 7 fitting function results: (a) springs 1–2, (b) springs 3–4, (c) springs 1–4.
Jmse 08 00625 g010

Share and Cite

MDPI and ACS Style

Vujović, I.; Šoda, J.; Kuzmanić, I.; Petković, M. Predicting External Influences to Ship’s Average Fuel Consumption Based on Non-Uniform Time Set. J. Mar. Sci. Eng. 2020, 8, 625. https://doi.org/10.3390/jmse8090625

AMA Style

Vujović I, Šoda J, Kuzmanić I, Petković M. Predicting External Influences to Ship’s Average Fuel Consumption Based on Non-Uniform Time Set. Journal of Marine Science and Engineering. 2020; 8(9):625. https://doi.org/10.3390/jmse8090625

Chicago/Turabian Style

Vujović, Igor, Joško Šoda, Ivica Kuzmanić, and Miro Petković. 2020. "Predicting External Influences to Ship’s Average Fuel Consumption Based on Non-Uniform Time Set" Journal of Marine Science and Engineering 8, no. 9: 625. https://doi.org/10.3390/jmse8090625

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop