Next Article in Journal
A Hierarchical PSMC–LQR Control Framework for Accurate Quadrotor Trajectory Tracking
Previous Article in Journal
Gaussian Process-Based Multi-Fidelity Bayesian Optimization for Optimal Calibration Point Selection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimizing Fuel Consumption Prediction Model Without an On-Board Diagnostic System in Deep Learning Frameworks

1
Department of Electrical Electronics Engineering, Zonguldak Bulent Ecevit University, Zonguldak 67100, Türkiye
2
Department of Geomatics Engineering, Zonguldak Bulent Ecevit University, Zonguldak 67100, Türkiye
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(22), 7031; https://doi.org/10.3390/s25227031
Submission received: 17 October 2025 / Revised: 12 November 2025 / Accepted: 13 November 2025 / Published: 18 November 2025
(This article belongs to the Section Electronic Sensors)

Abstract

Real-time prediction of the instantaneous fuel consumption rate (FCR) of any vehicle is the key to improving energy efficiency and reducing emissions. The conventional prediction methods, which include an on-board diagnostic (OBD) system, require the specific vehicle parameters and environmental conditions such as air density. We propose a data-driven Bayesian optimization and Monte Carlo (MC) Dropout methods-based long short-term memory (BMC-LSTM) network FCR prediction model using only the vehicle’s throttle position, velocity, and acceleration data. The cost-effective LSTM network-based solution enhances the high-resolution prediction accuracy within a deep learning framework. The network is integrated with the Bayesian optimization and MC-Dropout methods to ensure a probabilistically optimal hyperparameter set and robust networks. The proposed method presents an FCR model that provides calibrated predictions and reliability against distribution drift by probabilistically tuning hyperparameters with Bayesian optimization and quantifying epistemic uncertainty with the MC-Dropout. Our approach requires only vehicle speed, longitudinal acceleration, and throttle position at inference time. Note, however, that the reference FCR used to train and validate the models was obtained from OBD during data acquisition. The performance of the proposed method is compared with a conventional LSTM and Bidirectional LSTM-based multidimensional models, XGBoost and support vector regression-based models, and first- and fourth-order polynomials, which are derived using the least-squares method. The prediction performance of the method is evaluated using Mean Squared Error, Root Mean Squared Error, Mean Absolute, and R-squared statistical metrics. The proposed method achieves a superior R 2 score and substantially reduces the conventional error metrics.

1. Introduction

The instantaneous fuel consumption (FC) of an internal combustion engine-powered vehicle is a metric that represents the fuel economy and the vehicle’s emissions to achieve clean, economic, and sustainable city environments [1,2,3]. Fuel consumption is traditionally measured with an on-board diagnostic (OBD) system; however, some buses, motorcycles, and heavy vehicles are not included [4,5]. Such a system offers the measurement of vehicle parameter IDs, i.e., instantaneous engine revolution, fuel consumption, velocity, temperature, and acceleration, etc. The system can send on-road and off-road parameter IDs to a user device in a wireless environment [6,7,8]. The system, however, requires multiple sensors, which may cause high cost and unreliability, to render the online measurement available in long-term vehicle operation [9]. The key factors affecting the fuel consumption and emission of the vehicles are summarized [10,11,12].
The instantaneous FC of the vehicle is conventionally modeled based on vehicle-specific power (VSP) [13,14]. This method requires multiple specific parameters of a vehicle and the air, i.e., vehicle mass, the translational mass of the rotating components, road grade, coefficient of rolling resistance, ambient air density, headwind into the vehicle, etc., which makes the prediction specific to the vehicle [15]. A one-dimensional polynomial was proposed to estimate the instantaneous power demand of the vehicle using only the vehicle’s specific parameters [16]. One-dimensional motorcycle and bicycle polynomials were constructed to characterize energy consumption rates based on on-road data in Lisbon, Portugal [17]. However, the VSP-based methods evaluate the FC rate of a vehicle using discrete vehicle velocity, which reduces the prediction of the real FC rate in on-road and off-road conditions. In particular, the methods assume that the FC rate is constant during standstill, which can be unrealistic in traffic conditions [18]. A one-dimensional fuel consumption model was proposed using idling energy consumption, in which the energy depends on the movement, the acceleration, the deceleration, and the energy use of the total distance [19]. A one-dimensional polynomial was constructed to represent the fuel consumption and gas emission rates of a vehicle [20,21]. With the emergence of the big data era, a substantial transformation in the field of computational technology has been precipitated. This era enables multidimensional and multi-layer nonlinear mapping to predict the velocity, fuel consumption, and emission characterization of a vehicle [22]. A support vector regression- or artificial neural network (ANN)-based FC rate prediction method was proposed using multiple vehicles’ on-road data [23,24,25]. A standard feed-forward ANN does not address vanishing gradients and struggles with long-range temporal learning. A backpropagation neural network was utilized for the prediction of the FC rate, where a data augmentation method was proposed to obtain a high-precision FC rate [22]. A recurrent neural network (RNN)-based FC rate prediction method was proposed using smartphone measurements [26]. The method results could be insufficient owing to the disappearing and expanding gradient issues encountered during backpropagation, hindering the effective learning of long-term dependencies. In tasks necessitating long-horizon temporal modeling, the absence of gating mechanisms in conventional RNNs results in a substantial degradation due to the limited memory capacity. A comparative study of machine learning methods-based vehicle fuel efficiency prediction methods was proposed to obtain high prediction accuracy in capturing nonlinear mapping [27]. Influencing factors of fuel consumption in heavy-duty trucks were investigated [28,29]. The work collected the data from simulated/lab data, which did not incorporate dynamic factors such as driving behavior and traffic conditions. CatBoost-based fuel consumption rate prediction was proposed for heavy-duty diesel trucks [30]. A long short-term memory (LSTM) network-based fuel consumption prediction method was proposed [31], where the prediction was constructed using Electronic Control Unit (ECU) and simulation-based data. The traffic conditions were not included in the evaluation of the method’s performance. An artificial rabbit optimization algorithm was proposed to optimize the hyperparameters of the LSTM network, i.e., the number of hidden layers, learning rate, and learning decay rate [32]. There have been some works on optimizing the hyperparameters of the RNN networks based on metaheuristic methods [33,34]. The methods may be time-consuming, although such methods do not increase the robustness of the training phase against the network uncertainties. Grid search and random sampling methods are proposed to tune the parameters of the networks [35]. However, these methods may be insufficient for high-dimensional systems. In summary, FCR prediction models require real-time measurements of vehicle parameter IDs from the OBD system to predict the fuel consumption of a vehicle with high precision. The methods that predict the fuel consumption using vehicle velocity, orientation, and acceleration data without requiring an OBD system ignore traffic conditions. As far as we know, no prior work predicts the vehicle FCR in heavy traffic (stop-and-go) and on rough terrain without OBD.
In this context, we propose a parameter-optimized robust LSTM-based fuel consumption model. We have integrated the method with the Monte Carlo (MC) Dropout method to increase the robustness against network uncertainties. We collect real-world driving data from two gasoline vehicles across rough terrain, stop-and-go traffic, and free-flow conditions, yielding a high-quality fuel-consumption dataset. The reference FCR data were obtained from the OBD during the data acquisition process. The proposed method is compared with the FCR prediction models, which are constructed utilizing XGBoost and support vector regression (SVR) methods, a conventional LSTM network, and a least-squares method based on first- and fourth-order polynomials. The contributions of the paper are as follows:
  • We propose a method that combines Bayesian optimization with the LSTM network to optimize probabilistically the hyperparameters of the network, i.e., the learning rate, learning rate decay, batch size, the number of hidden layers, and the number of nodes in each hidden layer. The method models the outcome of each hyperparameter combination as a probabilistic function.
  • The method presents an FCR prediction model trained using the MC-Dropout method to quantify epistemic uncertainty, improving robustness to distribution shift.
  • The method requires only the vehicle translational speed, longitudinal acceleration, and throttle position at inference time.
The remainder of this paper is structured as follows: Section 2 presents the data logger process of the method on the two different vehicles. Section 3 introduces the LSTM network to construct the prediction model. Section 4 presents a summary of the data collection process. Section 5 presents the data-driven FCR model’s performance results using statistical metrics. Section 6 concludes the paper.

2. System Description and Problem Formulation

The OBD system gives the instantaneous fuel consumption rate in gasoline- and diesel-powered vehicles. The system gives the parameter IDs based on the specific vehicle. Access to parameter IDs may not be available for certain vehicle types: motorcycles and heavy-duty vehicles. In addition, the OBD-dependent system may not be available in electric vehicles. The fuel flow rate of the engine is given as
F C R = 360 m ˙ a i r λ B v ,
where F C R is the fuel consumption rate ( g / 100 km ), m ˙ a i r is the air mass ( g / s ) obtained from a mass air flow sensor, λ is the actual air/fuel ratio, B is the stoichiometric AFR, and v is the translational velocity of the vehicle ( km / h ). The real FCR data, which are obtained from (1), are obtained via the OBD system. The proposed data-driven FCR model is presented in Figure 1, where the system is integrated on two different vehicles: 2018 Opel Astra edition plus and 2014 Nissan Juke. The normalized throttle position (r) is obtained using an ultrasonic distance sensor ( r [ r i n t , 1 ] ), where r i n t is the initial throttle position including the stationary operation of the vehicle. The data capture stop-and-go behavior and rough-terrain operation. The RCWL-1670 ultrasonic sensor is preferred to measure the throttle/pedal position of the vehicles, since it has high measurement accuracy and stability, capable of measuring distances between 2 cm and 400 cm. It offers high precision of up to ±1 cm beyond the 2 cm blind spot. When calibrating the sensor’s sensitivity, data obtained at known distances can be adjusted using linear regression or software-based correction coefficients. The norm of the vehicle’s position vector, p = [ p x p y p z ] , is given as
p 2 = p x 2 + p y 2 + p z 2 ,
where x , y , z are the Cartesian coordinates. The position data are obtained using the NEO-6M global positioning system (GPS) sensor via an Arduino microprocessor and sent to a host computer to process the data in the Python 3.7 environment. The translational velocity v and acceleration a are obtained by differentiating position along time.
We define twenty (20) closed-loop scenarios to consider the traffic conditions: the roads have uphill and downhill parts, and there are road behaviors where pedestrians are more dense or less dense for every vehicle. The targeted area of the trajectories is presented in Figure 2, where the traffic lights, heavy traffic, and road conditions are included. The fuel consumption rates of test vehicles for half of the closed-loop trajectories are presented in Figure 3. Vehicle fuel consumption rates are presented as percentages in the adjacent bar. Green lines represent minimum fuel consumption, while red lines represent maximum fuel consumption. The fuel consumption rates of the test vehicles in three-dimensional space are presented in Figure 4. Lower fuel consumption is naturally expected on flat routes and higher fuel consumption rates on slopes and rough roads, because the traction demand increases, and the engine generates more power when climbing. However, the internal dynamics of the powertrain and the driving strategy can reverse this general trend. The engine’s ability to stay in the efficient zone on the specific fuel consumption map is closely related to the appropriate gear selection, constant acceleration at low but torque-favorable speeds, avoiding unnecessary throttle–brake transitions, and smooth traffic conditions. As a result, the vehicle’s internal dynamics (engine efficiency zone, gear levels, transmission efficiency, and internal losses), the driver’s throttle-to-gear strategy, and the continuity of the traffic flow can override the expected effect of the topography, leading to lower fuel consumption even on rough tracks than on a flat road. Therefore, we propose a deep learning method-based model to predict the instantaneous fuel consumption rate of the test vehicles.
We have constructed three different learning scenarios: (i) Scenario 1: the learning process, which is based on the leave-one-route-out principle, (ii) the learning process, which is based on the train-on-vehicle-A/test-on-vehicle-B principle, (iii) train-on-vehicle-B/test-on-vehicle-A, where vehicle A is the Nissan/Juke model, and vehicle B is the Opel/Astra model, to obtain a leakage-safe learning process.

3. Deep Learning-Based Prediction Model

This section introduces the proposed bidirectional LSTM network and ridge regression method-based prediction model construction. We construct first- and fourth-order one-dimensional ridge-regularized polynomials as a combination of the vehicle’s translational velocity, acceleration, and the throttle’s position to compare the performance of the proposed method. The flowchart of the proposed data-driven fuel consumption rate predic- tion methodology is presented in Figure 5.

3.1. Long Short-Term Memory

RNNs can learn short-term dependencies in the input vectors of a deep learning structure [36]. However, they may cause some gradient problems in the loss function when long-term dependencies are present. The weights of the networks cannot be defined or updated, when there is a dramatic growth in the gradient, which significantly limits the performance of RNN networks, especially in problems involving long-term dependencies. The LSTM networks introduced a solution to this undefined gradient problem in the backpropagation algorithm, since the LSTM has an internal iteration technique without using a nonlinear function. The network is included in the framework of the RNN network. The full structure of an LSTM cell is presented in Figure 6, where t defines the current index of the structure, c t is the internal state, c t 1 is the previous internal state, h t 1 is the previous hidden state, x t is the input pattern, f t is the forgetting gate, i t is the input gate, o t is the output gate, c ^ t is the input node, W h i is the weight from the hidden gate to the input, and W x i is the weight from the input pattern to the input gate. The LSTM cell structure consists of three gates: the forgetting, input, and output gates, which are defined as
f t = σ W f x x t + W h f h t 1 + b f ,
i t = σ W i x x t + W h i h t 1 + b i ,
o t = σ W o x x t + W h o h t 1 + b o ,
where σ ( . ) is the sigmoid activation function, W f x is the weight vector from the forgetting gate to the input pattern, W h f is the weight from the hidden state to the forgetting gate, W o x is the weight from output gate to the input pattern, and W h o is the weight from the hidden state to the output gate [37]. The terms b f , b i , and b o are the bias terms of the relative gates. The sigmoid function constrains the outputs of each gate within [0, 1] in the LSTM cell structure. The forget gate determines which information is discarded from the cell state. The input gate specifies which new information is added to the cell state. The output gate controls which information is transferred to the cell output based on the current cell state. The input node is given as
c ˜ t = tanh W c x x t + W h c h t 1 + b c ,
which is a function of the previous hidden state and input. The term t a n h ( . ) is the hyperbolic tangent function. The forgetting and update mechanism of the memory cell of the LSTM network is defined as
c t = f t c t 1 + i t c ˜ t ,
where the term ⊙ denotes element-wise multiplication. The differentiation of (7) is given as
c t c t 1 = c t 1 f t c t 1 + i t c ˜ t ,
which yields
c t c t 1 = diag f t .
Therefore, the network can prevent the vanishing problem in the RNN networks. The final main component of the LSTM cell is the output gate, which determines the cell output [38]. This output gate controls which information within the cell is transferred to the outside. The hidden state of the LSTM cell is given as
h t = o t tanh ( c t ) .

3.2. Bidirectional LSTM

Bidirectional long short-term memory (BiLSTM), with sequences in both forward and backward directions, is an extension of the classical LSTM to exploit past and future context [39]. Standard LSTMs achieve strong performance on sequence modeling tasks. This LSTM network is capable of learning long-term dependencies in data. However, during the learning process of the network, training is performed only on the information from the past. The BiLSTM deep learning network could complete the learning process by training data from the past and the future simultaneously. The BiLSTM deep learning structure feeds the input data simultaneously to two separate LSTM networks. While one LSTM network processes the input data in the forward direction, the other LSTM network could process the same data in the reverse direction. As a result, the comprehensive learning process is performed in both forward and backward directions at each time step. The forward hidden state in BiLSTM is given as
h t = σ W h x x t + W h h h t 1 + b h ,
where W h x and W h h are the forward weight matrices, and b h is the forward bias vector. The backward hidden state in BiLSTM is defined as
h t = σ W h x x t + W h h h t + 1 + b h ,
where W h x and W h h are the backward weight vectors, and b h is the backward bias vector [40]. Therefore, the forward and backward hidden layers in BiLSTM are evaluated to increase the prediction performance. The output of the network is given as
y t = W y h h t + W y h h t + b y ,
where W y h and W y h are weights of the output layer from the forward and backward hidden layers, and b y is the bias term of the network [41].

3.3. Bayesian Method-Based Hyperparameter Optimization

In RNN methods, the hyperparameters of the network are typically fixed and adjusted manually based on the designer’s prior experience. There are some methods to optimize these parameters using metaheuristic optimization methods, i.e., gray wolf, particle swarm, genetic algorithm, etc. [26]. However, such an approach dramatically increases the time of the training process. Such methods can be slow and do not address model uncertainty. There are basic methods to optimize the parameters in a lower time-consuming context: grid search and random sampling. The Bayesian method treats the verification error as a black-box cost and models this function probabilistically with a Gaussian process. In each iteration, a hyperparameter vector is chosen to train the model, and a probabilistic surrogate model is fitted. The posterior, which is a distribution defined by the Bayesian method, is given by
ρ ( ϑ | D ) = ρ ( D | ϑ ) ρ ( ϑ ) ρ ( D ) ,
where
ρ ( D ) = ρ ( D | ϑ ) ρ ( ϑ ) d ϑ ,
where D is the input-output dataset, ϑ is the the layer parameter space W ϑ , and ρ ( · ) is the distribution. The predictive posterior distribution of the next desired output y k + 1 is given by
ρ ( y k + 1 | x k + 1 , D ) = ρ ( y k + 1 | x k + 1 , ϑ ) ρ ( ϑ | D ) d ϑ ,
where k is the index of samples. The posterior is updated to determine the next hyperparameter vector. The method utilizes posterior uncertainty to select new evaluation points and approach the global optimum with minimal attempts. For a given set of hyperparameters, θ , the optimal set is given by
θ o p t = arg max θ l o g ρ ( D , θ ) ,
where θ o p t is the optimal hyperparameter set [38]. Dropout entails the exclusion of random groups of graph nodes during training, with a new random subset selected for each forward pass in graph neural networks [42]. MC-dropout (MC) approximately marginalizes the weighted sequence by keeping dropout enabled in the test on this chosen structure; therefore, generalization and calibration are improved due to epistemic uncertainty [43]. The approximate posterior function of the MC-Dropout is given by
q ( ϑ ) = 1 k L z k = 0 , 1 δ ϑ ^ 1 z 1 , , ϑ ^ L z L ( ϑ 1 , , ϑ L ) · ρ z k ( 1 ρ ) 1 z k q ( z ) ,
where ϑ ^ represents the learned weights, L is the dimension of the parameter vector, δ c ( ϑ ) is the point mass at c, and z is the Bernoulli dropout mask [44]. Substituting p ( ϑ , X , Y ) of (16) with (18), the closed form of the MC-Dropout predictive posterior is given by
p M C ( y k + 1 | x k + 1 , D ) = z [ 0 , 1 ] L ρ ( y k + 1 | x k + 1 , ϑ ^ z ) · q ( z ) .

3.4. Ridge-Regularized Polynomial Model

A feature input vector is chosen as a combination of the velocity, acceleration, throttle position, and the time derivative of the throttle position x k = [ v k a k r k r ˙ k ( v a ) k ] , where k = 1 , . . . , N is the data point. A predicted ridge polynomial is given by
y ^ k = β 0 + j = 1 m β j ϕ j ( x k ) ,
where m is the number of features, j is the feature index, { ϕ j ( · ) } j = 1 m are polynomial features, β 0 is the bias term of the polynomial, and β is the coefficient of the polynomial β = [ β 1 , , β m ] . The loss function of the ridge polynomial is defined by
J ( β 0 , β ) = k = 1 N ( y k β 0 j = 1 m β j ϕ j ( x k ) 2 ) + γ j = 1 m β j 2 ,
where N is the number of samples, y k is the k-th real value, and γ is a non-dimensional positive coefficient. The optimal coefficients of the polynomial are given by
( β 0 , β ) = arg min β 0 , β J ( β 0 , β ) .

4. Data Collection

In this study, data were collected as a time series with a 1 s sampling time to propose a deep learning-based vehicle fuel consumption estimation model. The sensor data, including the fuel consumption rate, GPS positions, and vehicle accelerator throttle/pedal position, are fused on the Arduino Uno card to be streamed to the host computer via serial communication interface at a band rate of 230,400. Thus, since all streams were time-stamped with the same Arduino time base, channel alignment is assumed to be achieved within the 1 Hz common sampling grid; no additional delay model was used. The internal sampling rates of the sensors and the internal interface delays were not measured separately in this study; the decisive end-to-end sampling rate used in the experimental setup was 1 Hz. This dataset, which consists of N = 23,701 samples, is processed using a low-pass filter. We have designed a second-order discrete-time Butterworth filter which has a 0.12 cut-off frequency. The filter order is chosen as two to avoid undesired signal delays. The filter is applied on a Python 3.7 in Jupyter Notebook. environment using the filtfilt command. For the filtering part of the experimental setup, we did not filter the signals during the real-time operation (the filtfilt command can not be used in real-time operation), Therefore, we have filtered the dataset in the offline data processing part of the study. The filter is given by
H ( z ) = 0.09 z 2 + 0.18 z 0.0913149 z 2 0.98240579 z + 0.34766539 .
In real-time operation, the Butterworth filter can be implemented in causal processors without the zero-phase advantage, since the operating/resampling frequency is chosen as the sampling frequency of the serial port. This process adds some delays to the dataset; however, the additional delay of the filtering process is assumed to be under the acceptable tolerance limit, since the filter is applied to the whole dataset. The networks are trained using the dataset to predict the model using the Python Keras 1.4.7 and TensorFlow 2.13.0 Libraries. The data information of the acquired dataset, which is presented in Figure 3 and Figure 4, is given in Table 1.

5. Performance Results of the Data-Driven Prediction Model

This section presents the performance results of the prediction method to evaluate the effectiveness of the proposed methodology. The statistical performance metrics are given as
RMSE = 1 N k = 1 N y k y ^ k 2 ,
MSE = 1 N k = 1 N y k y ^ k 2 ,
MAE = 1 N k = 1 N y k y ^ k ,
R 2 = 1 k = 1 N y k y ^ k 2 k = 1 N y k y ¯ 2 ,
Loss = 1 N k = 1 N log cosh y k y ^ k ,
where y ^ k is the k -th predicted value of the model, y ¯ is the mean of the real values, c o s h ( x ) is the hyperbolic cosine function, M S E is the mean square error, M A E is the mean absolute error, R 2 is the coefficient of determination, L o s s is the loss function of the deep learning methods, and R M S E is the root mean square error. Table 2 presents the capacity and regularization regimes of the three deep learning methods. For BiLSTM and LSTM, a deterministic and medium-capacity setup is adopted with fixed 128–128 units in the two hidden layers, zero dropout, and learning rate 10 3 (total number of layers 4, epoch 60, batch size 16 fixed). In contrast, in BMC-LSTM, the first and second hidden layer value ranges are defined as [ 64 , 180 ] , dropout [ 0.1 , 0.5 ] , and learning rate [ 10 4 , 10 2 ] , thus adapting the model complexity, regularization power, and optimization speed together. This design allows BMC-LSTM to adjust the bias–variance balance according to the scenario and data distribution, control the risk of overfitting (dropout), and keep the training dynamics (learning rate) in a stable region, while fixing the epoch and batch size ensures a fair comparison between different architectures.
For the parameters of the XGBoost method, the learning rate is chosen as 0.05, the number of estimators is chosen as 300, and the number of states is chosen as 42. For the parameters of the SVR method, the regularization coefficient is chosen as 100, the gamma coefficient is chosen as 0.1, epsilon is chosen as 0.1, and the kernel of the SVR model is defined as a radial basis function.
The training loss performance of the deep learning-based method is presented in Figure 7. The deep learning method is executed in 60 iterations. The proposed deep learning methodology has a minimum loss function value at the 60 i-th iteration. The BiLSTM and the proposed deep learning method have the loss value at 38 i-th iteration. However, the fluctuation of the loss function is the maximum in the BiLSTM. This situation causes the overfitting problem during the training process. The proposed BMC-LSTM method has a roughly monotonically decreasing learning curve during the training process. Furthermore, the proposed deep learning model presents robustness against network uncertainties and the overfitting problem. The FCR prediction performance of the models is presented in Figure 8 and Figure 9. The prediction performance of the fuel consumption rate in this model is tested on 200 samples. The results indicate that the BMC-LSTM based model has a superior FCR prediction performance. In Figure 9, the PR denotes the fourth-order polynomial ridge-based polynomial, and LR denotes the first-order linear ridge-based polynomial. It is emphasized that the deep learning-based methods indicate more than sufficient FCR prediction performances. The R 2 scatter performance of the methods is presented in Figure 10. It is stated that the prediction performances of BiLSTM and the proposed deep learning approach are close to each other. However, the prediction performance of the proposed method is slightly better than that of BiLSTM in Scenario 1.
Table 3 shows that BMC–LSTM achieves the lowest error (RMSE/MSE/MAE) and highest agreement ( R 2 ) values in all three scenarios. Specifically, Scenario 3 shows the best performance by far with RMSE = 4.57 , MSE = 20.89 , MAE = 2.14 , and R 2 = 0.95 , while Scenario 2 maintains strong results, with RMSE = 5.40 and R 2 = 0.91 . In Scenario 1, BMC–LSTM produces an RMSE close to BiLSTM (5.54 vs. 5.76), while exhibiting a more robust error distribution with a significantly lower MAE (2.04). While the two-layer LSTM is competitive in some cases (e.g., RMSE = 4.97 , R 2 = 0.92 in Scenario 3), BMC–LSTM consistently outperforms all metrics. Classical linear/multinomial regressions (LR/PR) are significantly weaker, while SVR and XGBoost are moderate; these results confirm that deep Bayesian uncertainty modeling provides a significant advantage in terms of both accuracy and stability. The inference times of the BiLSTM, LSTM, SVR, XGBoost, and BMC-LSTM are 39.047 ms/sample, 36.292 ms/sample, 87.97 ms/sample, 1.8 ms/sample, and 38.171 ms/sample, respectively. When we evaluate the performance of these learning models in the context of the requirements of a typical telematics flow, an inference time of 38.171 ms/sample means our model can process 26 samples per second. This rate meets the requirements of the most standard OBD-II, GPS, and ultrasonic distance sensor acquisition rates of 10 Hz (10 samples per second) or 20 Hz (20 samples per second). As a result, although our BMC-LSTM, BiLSTM, and LSTM models have higher inference times than XGBoost, these models remain within the maximum latency limits required by real-time telematics systems, in exchange for their more complex pattern capture and high accuracy. This situation indicates that our BMC-LSTM model achieves both high accuracy and practical applicability.
The three scenarios presented in Table 4 consistently validate the cross-vehicle generalization ability and uncertainty calibration of the BMC-LSTM model. The prediction interval coverage probability (PICP) values (S1: 94.12%, S2: 96.51%, S3: 95.86%) are very close to the nominal 95% coverage target; in particular, the 96.51% value at S2 indicates that the intervals are slightly conservative and well-calibrated, while S1 and S3 show neutral calibration close to the target. The low out-of-distribution (OOD) error rates (S1: 5.88%, S2: 2.96%, S3: 3.89%) indicate stable performance even on ODD samples, while confirming the robustness in the cross-vehicle test (S2: training on Nissan → Astra; S3: training on Astra → Nissan Juke); Furthermore, the lower OOD ratio of S2 compared to S3 suggests a stronger overlap of feature spaces in the Nissan→Astra direction. Although the NLL values (S1: 1.99, S2: 3.10, S3: 2.94) increased to an expected extent in scenarios involving vehicle transfers (moderate difficulty in probabilistic fitting), when considered together with PICP and OOD, the prediction distributions appear to produce a reliable uncertainty signal for practical use. In conclusion, in the trip-by-trip evaluation, BMC-LSTM provides a fuel prediction model with well-calibrated prediction intervals, low out-of-distribution error, and acceptable probabilistic fit in both within-vehicle and cross-vehicle tests. The calibration and prediction curves for the proposed BMC-LSTM method in three scenarios are presented in Figure 11.
The residual distribution of the proposed method is presented for scenarios 1–3 in Figure 12. In general, the residual errors are centralized and largely symmetrical around zero, indicating that the model makes reliable and consistent predictions.
The current dataset is limited to two gasoline vehicles (Astra, Juke), predominantly urban routes in a single region, and predominantly mild/dry weather conditions; therefore, vehicle diversity (engine displacement, gear type, age/weight/powertrain architecture), road/topography spectrum (highway, high altitude, long ascents/descendants), driving styles, and seasonality are underrepresented. The 1 Hz sampling may partially under-sample short-term transition regimes (hard throttle/brake, gear shifts); GPS/ultrasonic sensor-induced noise/calibration errors and heteroskedasticity measurement uncertainty may impact data quality. Additionally, systematic testing has not been conducted with different fuel types (diesel/hybrid) and under different climate/terrain conditions; no transfer/adaptive learning framework explicitly designed for transfer to unknown vehicle models has been implemented. These limitations impose inherent limits on the generalization and external validity of the model under dispersion-shift conditions.

6. Conclusions

In this paper, we have proposed the Bayesian and MC-Dropout-based LSTM deep learning methodology to predict the fuel consumption rate of vehicles, where the ground-truth FCR data are obtained from OBD in the data acquisition process. Bayesian optimization probabilistically optimized the LSTM’s hyperparameters to find the optimal structure and learning rates. MC-Dropout reduced the model’s epistemic uncertainty, improving the predictive stability and generalization ability of this optimized LSTM. Therefore, the method provides calibrated predictions and improved reliability under distribution shift. The VSP-like first- and fourth-order polynomials are constructed for comparison purposes. The proposed method is compared with advanced learning methods, i.e., XGBoost, SVR, LSTM, and BiLSTM to verify the effectiveness and robustness of the method. Four different performance criteria are selected to analyze the effectiveness of the proposed deep learning methodology. In the context of the R 2 score, the one-dimensional LR- and PR-based models are insufficient to predict the FCR rate, since such models could achieve prediction performance in all scenarios. The XGBoost, SVR, LSTM, and BiLSTM network-based FCR models achieve 58 % , 44 % , 60 % , and 85 % prediction performance in Scenario 1. However, it is observed that the networks may be sensitive to network uncertainties and the overfitting problem, as can be seen from the training loss curve. The BMC-LSTM deep learning model has the minimum RMSE, MSE, and MAE values in all three scenarios. The R 2 value of the proposed BMC-LSTM deep learning model in Scenarios 1, 2, and 3 has been obtained as 0.86, 0.91, and 0.95, respectively.

7. Future Work

We are planning to expand the dataset to include multi-city/multi-climate data, different road classes, and vehicle types (diesel, hybrid). We will focus on Internet of Things (IoT) and mobile application integrations to translate the developed model into practical applications. Future goals include exploring structures based on combining data from phone sensors with sensors other than the OBD system (sensor fusion) and privacy-preserving federated learning infrastructures that allow the model to perform real-time inference directly on edge devices.

Author Contributions

Conceptualization, R.K., E.B., and S.H.K.; methodology, R.K., E.B., and S.H.K.; software, R.K. and E.B.; writing—original draft preparation, R.K. and E.B.; writing—review and editing, R.K., E.B., and S.H.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original data supporting the findings of this study is openly available on GitHub at a permanent repository https://github.com/rkeskin/LSTM-GRU-Fuel-consumption-Vehicle-emission-prediction/issues/1#issue-3625340459 (accessed on 12 November 2025).

Acknowledgments

This work was supported by the Research Fund of Zonguldak Bulent Ecevit University (Project Number: 2025-75737790-01).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kabir, R.; Remias, S.M.; Waddell, J.; Zhu, D. Time-Series fuel consumption prediction assessing delay impacts on energy using vehicular trajectory. Transp. Res. Part D Transp. Environ. 2023, 117, 103678. [Google Scholar]
  2. Öztürk, A.T.; Kasliwal, A.; Fitzmaurice, H.; Kavvada, O.; Calvez, P.; Cohen, R.C.; González, M.C. A mesoscopic model of vehicular emissions informed by direct measurements and mobility science. Sustain. Cities Soc. 2025, 129, 106421. [Google Scholar] [CrossRef]
  3. Singh, M.; Dubey, R.K. Deep learning model based CO2 emissions prediction using vehicle telematics sensors data. IEEE Trans. Intell. Veh. 2021, 8, 768–777. [Google Scholar] [CrossRef]
  4. Ammar, M.; Janjua, H.; Thangarajan, A.S.; Crispo, B.; Hughes, D. Securing the on-board diagnostics port (obd-ii) in vehicles. SAE Int. J. Transp. Cybersecur. Priv. 2020, 2, 83–106. [Google Scholar]
  5. Jiansen, Y.; Konghui, G.; Haitao, D.; Jianwei, Z.; Bin, X. The application of sae j1939 protocol in automobile smart and integrated control system. In Proceedings of the 2010 International Conference on Computer, Mechatronics, Control and Electronic Engineering, Changchun, China, 24–26 August 2010; IEEE: New York, NY, USA, 2010; Volume 3, pp. 412–415. [Google Scholar]
  6. Wang, J.; Wang, R.; Yin, H.; Wang, Y.; Wang, H.; He, C.; Liang, J.; He, D.; Yin, H.; He, K. Assessing heavy-duty vehicles (HDVs) on-road NOx emission in China from on-board diagnostics (OBD) remote report data. Sci. Total Environ. 2022, 846, 157209. [Google Scholar] [CrossRef]
  7. Zhao, D.; Li, H.; Hou, J.; Gong, P.; Zhong, Y.; He, W.; Fu, Z. A review of the data-driven prediction method of vehicle fuel consumption. Energies 2023, 16, 5258. [Google Scholar] [CrossRef]
  8. Campos-Ferreira, A.E.; Lozoya-Santos, J.d.J.; Tudon-Martinez, J.C.; Mendoza, R.A.R.; Vargas-Martínez, A.; Morales-Menendez, R.; Lozano, D. Vehicle and driver monitoring system using on-board and remote sensors. Sensors 2023, 23, 814. [Google Scholar] [CrossRef]
  9. Li, S.; Li, G.; He, L.; Wang, H.; Ji, L.; Wang, J.; Chen, W.; Kang, Y.; Yin, H. Heavy-duty diesel vehicle NOX emissions and fuel consumption: Remote on-board monitoring analysis. Transp. Res. Part D Transp. Environ. 2025, 146, 104875. [Google Scholar] [CrossRef]
  10. Qu, L.; Wang, W.; Li, M.; Xu, X.; Shi, Z.; Mao, H.; Jin, T. Dependence of pollutant emission factors and fuel consumption on driving conditions and gasoline vehicle types. Atmos. Pollut. Res. 2021, 12, 137–146. [Google Scholar] [CrossRef]
  11. Molina-Campoverde, J.J.; Zurita-Jara, J.; Molina-Campoverde, P. Driving Pattern Analysis, Gear Shift Classification, and Fuel Efficiency in Light-Duty Vehicles: A Machine Learning Approach Using GPS and OBD II PID Signals. Sensors 2025, 25, 4043. [Google Scholar] [CrossRef]
  12. Udoh, J.; Lu, J.; Xu, Q. Application of Machine Learning to Predict CO2 Emissions in Light-Duty Vehicles. Sensors 2024, 24, 8219. [Google Scholar] [CrossRef]
  13. Jimenez-Palacios, J.L. Understanding and Quantifying Motor Vehicle Emissions with Vehicle Specific Power and TILDAS Remote Sensing. Ph.D.Thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 1998. [Google Scholar]
  14. Frey, H.C.; Unal, A.; Chen, J.; Li, S. Modeling mobile source emissions based upon in-use and second-by-second data: Development of conceptual approaches for EPA’s new moves model. In Proceedings of the Proceedings, Annual Meeting of the Air and Waste Management Association, Pittsburgh, PA, USA, 26 June 2003. [Google Scholar]
  15. Wang, H.; Chen, C.; Huang, C.; Fu, L. On-road vehicle emission inventory and its uncertainty analysis for Shanghai, China. Sci. Total Environ. 2008, 398, 60–67. [Google Scholar]
  16. Lindhjem, C.E.; Pollack, A.K.; Slott, R.S.; Sawyer, R.F. Analysis of EPA’s Draft Plan for Emissions Modeling in MOVES and MOVES GHG; Research Report# CRC Project E-68; ENVIRON International Corporation: Novato, CA, USA, 2004. [Google Scholar]
  17. Mendes, M.; Duarte, G.; Baptista, P. Introducing specific power to bicycles and motorcycles: Application to electric mobility. Transp. Res. Part C Emerg. Technol. 2015, 51, 120–135. [Google Scholar] [CrossRef]
  18. Topić, J.; Škugor, B.; Deur, J. Neural network-based prediction of vehicle fuel consumption based on driving cycle data. Sustainability 2022, 14, 744. [Google Scholar] [CrossRef]
  19. Oehlerking, A.L. StreetSmart: Modeling Vehicle Fuel Consumption with Mobile Phone Sensor Data Through a Participatory Sensing Framework. Ph.D. Thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 2011. [Google Scholar]
  20. Varga, B.; Lulić, Z.; Tettamanti, T. Systematic Error Correction of SUMO Traffic Simulator’s HBEFA Vehicle Emission Model. In Proceedings of the 2024 IEEE International Conference on Electrical Systems for Aircraft, Railway, Ship Propulsion and Road Vehicles & International Transportation Electrification Conference (ESARS-ITEC), Naples, Italy, 26–29 November 2024; IEEE: New York, NY, USA, 2024; pp. 1–6. [Google Scholar]
  21. Stokić, M.; Momčilović, V.; Dimitrijević, B. A bilinear interpolation model for estimating commercial vehicles’ fuel consumption and exhaust emissions. Sustain. Futur. 2023, 5, 100105. [Google Scholar] [CrossRef]
  22. Zeng, X.; Jiang, X.; Song, D.; Bi, D.; Xiang, R. High-precision vehicle engine fuel consumption model based on data augmentation. Energy 2025, 333, 137400. [Google Scholar] [CrossRef]
  23. Ansari, A.; Abediasl, H.; Patel, P.R.; Hosseini, V.; Koch, C.R.; Shahbakhti, M. Estimating instantaneous fuel consumption of vehicles by using machine learning and real-time on-board diagnostics (OBD) data. In Proceedings of the Canadian Society for Mechanical Engineering International Congress 2022, Edmonton, AB, Canada, 5–8 June 2022. [Google Scholar]
  24. Moradi, E.; Miranda-Moreno, L. Vehicular fuel consumption estimation using real-world measures through cascaded machine learning modeling. Transp. Res. Part D Transp. Environ. 2020, 88, 102576. [Google Scholar]
  25. Abediasl, H.; Ansari, A.; Hosseini, V.; Koch, C.R.; Shahbakhti, M. Real-time vehicular fuel consumption estimation using machine learning and on-board diagnostics data. Proc. Inst. Mech. Eng. Part D J. Automob. Eng. 2024, 238, 3779–3793. [Google Scholar] [CrossRef]
  26. Kanarachos, S.; Mathew, J.; Fitzpatrick, M.E. Instantaneous vehicle fuel consumption estimation using smartphones and recurrent neural networks. Expert Syst. Appl. 2019, 120, 436–447. [Google Scholar] [CrossRef]
  27. Yoo, S.r.; Shin, J.w.; Choi, S.H. Machine learning vehicle fuel efficiency prediction. Sci. Rep. 2025, 15, 14815. [Google Scholar] [CrossRef]
  28. Zhu, X.; Shen, X.; Chen, K.; Zhang, Z. Research on the prediction and influencing factors of heavy duty truck fuel consumption based on LightGBM. Energy 2024, 296, 131221. [Google Scholar] [CrossRef]
  29. Gong, J.; Shang, J.; Li, L.; Zhang, C.; He, J.; Ma, J. A comparative study on fuel consumption prediction methods of heavy-duty diesel trucks considering 21 influencing factors. Energies 2021, 14, 8106. [Google Scholar] [CrossRef]
  30. Liu, A.; Fan, P.; He, X.; Lu, H.; Yu, L.; Song, G. CatBoost-based fuel consumption modeling and explainable analysis for heavy-duty diesel trucks: Impact of engine, driving behavior, and vehicle weight. Energy 2025, 333, 137485. [Google Scholar] [CrossRef]
  31. Zhao, J.; Liang, K.; Guan, W.; Sang, H.; Zhou, S.; Deng, L.; Pan, M. Energy management strategy for diesel vehicles based on fuel consumption prediction. Transp. Res. Part D Transp. Environ. 2025, 146, 104896. [Google Scholar] [CrossRef]
  32. Gülmez, B. Stock price prediction with optimized deep LSTM network with artificial rabbits optimization algorithm. Expert Syst. Appl. 2023, 227, 120346. [Google Scholar] [CrossRef]
  33. Ewees, A.A.; Al-qaness, M.A.; Abualigah, L.; Abd Elaziz, M. HBO-LSTM: Optimized long short term memory with heap-based optimizer for wind power forecasting. Energy Convers. Manag. 2022, 268, 116022. [Google Scholar] [CrossRef]
  34. Liu, L.; Li, Y. Research on a photovoltaic power prediction model based on an IAO-LSTM optimization algorithm. Processes 2023, 11, 1957. [Google Scholar] [CrossRef]
  35. Bergstra, J.; Bengio, Y. Random search for hyper-parameter optimization. J. Mach. Learn. Res. 2012, 13, 281–305. [Google Scholar]
  36. Waqas, M.; Humphries, U.W. A critical review of RNN and LSTM variants in hydrological time series predictions. MethodsX 2024, 13, 102946. [Google Scholar] [CrossRef]
  37. awryńczuk, M.; Zarzycki, K. LSTM and GRU type recurrent neural networks in model predictive control: A Review. Neurocomputing 2025, 632, 129712. [Google Scholar] [CrossRef]
  38. Martínez-Ramón, M.; Ajith, M.; Kurup, A.R. Deep Learning: A Practical Introduction; John Wiley & Sons: Hoboken, NJ, USA, 2024. [Google Scholar]
  39. Singla, P.; Duhan, M.; Saroha, S. An ensemble method to forecast 24-h ahead solar irradiance using wavelet decomposition and BiLSTM deep learning network. Earth Sci. Inform. 2022, 15, 291–306. [Google Scholar] [CrossRef]
  40. Peng, S.; Zhu, J.; Wu, T.; Yuan, C.; Cang, J.; Zhang, K.; Pecht, M. Prediction of wind and PV power by fusing the multi-stage feature extraction and a PSO-BiLSTM model. Energy 2024, 298, 131345. [Google Scholar] [CrossRef]
  41. Michael, N.E.; Hasan, S.; Al-Durra, A.; Mishra, M. Short-term solar irradiance forecasting based on a novel Bayesian optimized deep Long Short-Term Memory neural network. Appl. Energy 2022, 324, 119727. [Google Scholar] [CrossRef]
  42. Gal, Y.; Ghahramani, Z. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016; PMLR: Cambridge, MA, USA, 2016; pp. 1050–1059. [Google Scholar]
  43. Seoh, R. Qualitative analysis of monte carlo dropout. arXiv 2020, arXiv:2007.01720. [Google Scholar] [CrossRef]
  44. Folgoc, L.L.; Baltatzis, V.; Desai, S.; Devaraj, A.; Ellis, S.; Manzanera, O.E.M.; Nair, A.; Qiu, H.; Schnabel, J.; Glocker, B. Is MC dropout bayesian? arXiv 2021, arXiv:2110.04286. [Google Scholar] [CrossRef]
Figure 1. Experimental setup of the proposed fuel consumption rate prediction methodology.
Figure 1. Experimental setup of the proposed fuel consumption rate prediction methodology.
Sensors 25 07031 g001
Figure 2. The area that covers the trajectories of the test vehicles.
Figure 2. The area that covers the trajectories of the test vehicles.
Sensors 25 07031 g002
Figure 3. Spatial distribution of the fuel consumption rate of the test vehicles in various trajectory scenarios in 2D space.
Figure 3. Spatial distribution of the fuel consumption rate of the test vehicles in various trajectory scenarios in 2D space.
Sensors 25 07031 g003aSensors 25 07031 g003b
Figure 4. Spatial distribution of fuel consumption rate of the test vehicles in various trajectory scenarios in 3D space; the road conditions of the vehicles (conditions = con.).
Figure 4. Spatial distribution of fuel consumption rate of the test vehicles in various trajectory scenarios in 3D space; the road conditions of the vehicles (conditions = con.).
Sensors 25 07031 g004aSensors 25 07031 g004b
Figure 5. Flowchart representation of the proposed data-driven fuel consumption rate prediction methodology.
Figure 5. Flowchart representation of the proposed data-driven fuel consumption rate prediction methodology.
Sensors 25 07031 g005
Figure 6. The full structure of the LSTM gates, where the four NN layers are placed in parallel.
Figure 6. The full structure of the LSTM gates, where the four NN layers are placed in parallel.
Sensors 25 07031 g006
Figure 7. The training loss performance of the deep learning-based methods.
Figure 7. The training loss performance of the deep learning-based methods.
Sensors 25 07031 g007
Figure 8. The prediction results on the sample data of the deep learning- and the machine learning-based methods.
Figure 8. The prediction results on the sample data of the deep learning- and the machine learning-based methods.
Sensors 25 07031 g008
Figure 9. The prediction results on the sample data of the linear regression-based polynomials.
Figure 9. The prediction results on the sample data of the linear regression-based polynomials.
Sensors 25 07031 g009
Figure 10. The scatter metric performance of the methods: (a) R 2 performance of the proposed method, (b) R 2 performance of the BiLSTM method, (c) R 2 performance of the LSTM method, (d) R 2 performance of the XGBoost method, (e) R 2 performance of the SVR method, (f) R 2 performance of the P R method.
Figure 10. The scatter metric performance of the methods: (a) R 2 performance of the proposed method, (b) R 2 performance of the BiLSTM method, (c) R 2 performance of the LSTM method, (d) R 2 performance of the XGBoost method, (e) R 2 performance of the SVR method, (f) R 2 performance of the P R method.
Sensors 25 07031 g010
Figure 11. Calibration and prediction views for BMC–LSTM in three scenarios: (a,c,e) residual/distribution fit with standard normal Q–Q diagrams (Red lines are the standard distribution and blue lines are the input samples); (b,d,f) comparison of actual and predicted FCR with 95% uncertainty band.
Figure 11. Calibration and prediction views for BMC–LSTM in three scenarios: (a,c,e) residual/distribution fit with standard normal Q–Q diagrams (Red lines are the standard distribution and blue lines are the input samples); (b,d,f) comparison of actual and predicted FCR with 95% uncertainty band.
Sensors 25 07031 g011
Figure 12. The residual distribution of the BMC-LSTM model in all scenarios.
Figure 12. The residual distribution of the BMC-LSTM model in all scenarios.
Sensors 25 07031 g012aSensors 25 07031 g012b
Table 1. The data info of each test drive presented in Figure 3 and Figure 4 (Dur. = duration, cond. = conditions, S = scenario.
Table 1. The data info of each test drive presented in Figure 3 and Figure 4 (Dur. = duration, cond. = conditions, S = scenario.
Drive NameSamples (N)Dur. Time (min)Road Cond.Role in the S1/S2/S3
Juke/Route-1116019.3FlatTraining/Training/Test
Juke/Route-2105117.52Extra UrbanTraining/Training/Test
Juke/Route-3101816.97Semi-flatTraining/Training/Test
Juke/Route-4139523.25FlatTest/Training/Test
Juke/Route-5120220UrbanTest/Training/Test
Astra/Route-171411.9FlatTraining/Test/Training
Astra/Route-2133522.25Semi-FlatTraining/Test/Training
Astra/Route-386114.35UrbanTraining/Test/Training
Astra/Route-477612.93UrbanTest/Test/Training
Astra/Route-5120920.15Extra UrbanTest/Test/Training
Table 2. The hyperparameters of the deep learning networks.
Table 2. The hyperparameters of the deep learning networks.
ModelHidden Layers (First and Second)LayersDropoutLearning RateEpochsBatch Size
BiLSTM128–128400.0016016
LSTM128–128400.0016016
BMC-LSTM[64–180], [64–180]4[0.1–0.5][0.0001–0.01]6016
Table 3. The performance comparison of vehicle emission estimation models.
Table 3. The performance comparison of vehicle emission estimation models.
Scenario NumberModelRMSEMSEMAE R 2
1BiLSTM5.7633.152.920.85
2-Layer LSTM9.4288.754.940.60
PR13.77189.717.340.15
LR14.52210.758.130.06
XGBoost9.5791.624.820.58
SVR11.05122.085.910.44
BMC-LSTM5.5430.662.040.86
2BiLSTM8.1265.953.650.79
2-Layer LSTM5.7533.112.370.90
PR12.67160.495.910.09
LR12.85165.157.370.06
XGBoost13.48181.625.890.45
SVR14.07198.096.610.40
BMC-LSTM5.4029.142.230.91
3BiLSTM5.7032.462.690.91
2-Layer LSTM4.9724.742.550.92
PR14.01196.377.680.22
LR15.60243.309.110.03
XGBoost11.61134.795.170.65
SVR13.55183.556.460.52
BMC-LSTM4.5720.892.140.95
Table 4. The metrics of the proposed emission estimation model.
Table 4. The metrics of the proposed emission estimation model.
MetricScenario 1Scenario 2Scenario 3
PICP94.12%96.51%95.86%
NLL1.98553.10492.9385
OOD5.88%2.96%3.89%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Keskin, R.; Belge, E.; Kutoglu, S.H. Optimizing Fuel Consumption Prediction Model Without an On-Board Diagnostic System in Deep Learning Frameworks. Sensors 2025, 25, 7031. https://doi.org/10.3390/s25227031

AMA Style

Keskin R, Belge E, Kutoglu SH. Optimizing Fuel Consumption Prediction Model Without an On-Board Diagnostic System in Deep Learning Frameworks. Sensors. 2025; 25(22):7031. https://doi.org/10.3390/s25227031

Chicago/Turabian Style

Keskin, Rıdvan, Egemen Belge, and Senol Hakan Kutoglu. 2025. "Optimizing Fuel Consumption Prediction Model Without an On-Board Diagnostic System in Deep Learning Frameworks" Sensors 25, no. 22: 7031. https://doi.org/10.3390/s25227031

APA Style

Keskin, R., Belge, E., & Kutoglu, S. H. (2025). Optimizing Fuel Consumption Prediction Model Without an On-Board Diagnostic System in Deep Learning Frameworks. Sensors, 25(22), 7031. https://doi.org/10.3390/s25227031

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop