Next Article in Journal
A Managed Volatility Investment Strategy for Pooled Annuity Products
Previous Article in Journal
Analyzing How the Social Security Reserve Fund in Spain Affects the Sustainability of the Pension System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimization in Item Delivery as Risk Management: Multinomial Case Using the New Method of Statistical Inference for Online Decision

by
Sapto Wahyu Indratno
1,2,
Kurnia Novita Sari
2 and
Mokhammad Ridwan Yudhanegara
3,*
1
University Center of Excellence on Artificial Intelligence for Vision, Institut Teknologi Bandung, Natural Language Processing & Big Data Analytics (U-CoE AI-VLB), Bandung 40132, West Java, Indonesia
2
Statistics Research Division, Faculty of Mathematics and Natural Sciences, Institut Teknologi Bandung, Bandung 40132, West Java, Indonesia
3
Mathematics Education Department, Faculty of Teacher Training and Education, Universitas Singaperbangsa Karawang, Karawang 41361, West Java, Indonesia
*
Author to whom correspondence should be addressed.
Risks 2022, 10(6), 122; https://doi.org/10.3390/risks10060122
Submission received: 15 April 2022 / Revised: 26 May 2022 / Accepted: 4 June 2022 / Published: 10 June 2022

Abstract

:
Online activity increasing spreads with the power of technological development. Many studies reported the impact of online activities on decision making. From the statistical perspective, decision making is related to statistical inference. In this regard, it is interesting to propose a new method of statistical inference for online decisions. This method is built by the logarithm distribution of the likelihood function, which allows us to determine statistics using the normal statistical test approach iteratively. It means that the inference can be made in an online way every time new data arrive. Compared to classical methods (commonly, chi-squared), the advantage of this method is that it allows us to make decisions without storing large data. In particular, the novelty of this research is expressed in the algorithm, theorem, and corollary for the statistical inference procedure. In detail, this paper’s simulation discusses online statistical tests for multinomial cases and applies them to transportation data for item delivery, namely traffic density. Changes in traffic density resulted in changes to the strategy of item delivery. The goal is to obtain a minimum delivery time for the risk of losses.

1. Introduction

This paper is a continuation of research from Yudhanegara et al. (2021a) on network clustering using item delivery strategies and predictive distribution, as well as Yudhanegara et al. (2021b) on the importance of network clustering, which affects the optimization of item delivery time, and Yudhanegara et al. (2022), which proposed a new strategy in the item delivery process. The focus of the problem presented in this paper is a unique alternative related to statistical inference that is needed in online decision making.
The results of the statistical inference become the basis for carrying out an item delivery strategy, which includes the network clustering process and time optimization. The method formulated in this study provides an efficient theoretical contribution in dealing with the challenges of extensive data and the dynamics of incoming information.
The knowledge required is the involvement of statistics and graph theory (network) in solving research problems. This research problem is raised from the latest natural phenomenon: the item delivery system, which involves the traffic system, time, and distance traveled. The role of statistics in answering real-time data (online) is packaged by utilizing the Bayesian method to determine predictive distributions and statistical inference in decision making. In addition, the multinomial distribution, which has the parameter Dirichlet distribution, is briefly examined. The distribution is considered the most representative distribution in the item delivery process in a traffic system with many path choices. As for the construction of the network and its clustering, graph theory can be used.
The subject of this research consists of four parts: network clustering, optimization, predictive distribution, and new statistical inference procedures for online decision making. The ultimate goal of integrating the four subjects is to produce an item delivery strategy that can overcome dynamic network conditions. This is important because traffic density changes over time, especially in urban areas or densely populated areas. Furthermore, optimization in item delivery can be used as risk management for goods delivery companies.
We need a precise and efficient statistical inference method. The contribution of this paper is to provide a new alternative for inferential statistics required to make decisions online. Thus, we can make inferences with every arrival of further information or data. The advantage of the method, i.e., it does not require historical data storage, which requires low memory. We make inferences using the previous results. Unlike the classic method (Oosterhoff and van Zwet 1972; Owen 2018), we need historical data storage, which requires a large memory.
The type of data distribution discussed in this paper is assumed to have a multinomial distribution. Classically, we use chi-squared for the multinomial goodness-of-fit test (Azen and Walker 2010; Walpole et al. 2012). This paper promotes the new method for the multinomial goodness-of-fit test. The method uses a normal distribution as inferential statistics for online decisions. Thus, the hypothesized assumptions for model fit use the Z statistical test. In the future, there will be many issues regarding big data, so that the role of statistics using classical methods is no longer efficient. The new method of inferential statistics for online decisions in this era of big data gives statistics that are still needed.
The new method aims to help practitioners, researchers, and users solve problems in cases of the multinomial distribution. The characteristics of the multinomial distribution, i.e., each experiment has more than two possible events to occur; each experiment is statistically independent so that the events resulting from one experiment do not affect subsequent trials; and the probability of each event in each experiment does not change. The outline of the multinomial discusses data with the number of frequencies in a data category. These multinomial cases have been widely used by practitioners, researchers, and users covering various disciplines, such as politics (Nownes 1992), social (Brock and Durlauf 2003), engineering (Rith et al. 2019), transportation (Reddy et al. 2015), medicine (Nelissen et al. 2020), and other exact sciences. For this reason, the presence of this new method is expected to add new references.
The simulation of the number of vehicles on each road segment is assumed to have a multinomial distribution. Suppose we represent a road map as a network—traffic conditions with the number of vehicles changing every time cause network conditions to be not static. Hence, we need a distribution to describe the dynamic situation of the network. Thus, the predictive distribution function and the estimation of the required parameters are of great importance. For prediction, the predictive distribution approach is used. Predictive distribution is appropriate for cases with known data distribution (Klugman et al. 2012).
In general, the Section 1 (Introduction) of this paper describes the importance of the methods proposed in this paper. The Section 2 (Literature Review) explains the literature on optimization in risk management. The theoretical study of predictive distribution, prediction accuracy, network clustering, and item delivery strategy are described in the Section 3 (Material and Methods). The new methods, limitations, simulation, and result optimization as risk management are discussed in the Section 4 (Result and Discussion). The conclusions are described in the Section 5 (Conclusions).

2. Literature Review

Mathematical optimization, or optimization, is part of operations research. Optimization is a way to determine suitable solutions by minimizing or maximizing one or more objective functions that satisfy all constraints. Optimization is widely used in industrial engineering. Examples of optimization methods include traveling salesman problems, vehicle routing problems, etc.
Optimization is also part of risk management (Aranburu et al. 2016). Risk management is an iterative process and assists in setting strategy, achieving goals, and making informed decisions (Degtereva et al. 2022). Decision making and optimization under uncertainty constitute a broad and popular area of operations research and management sciences (Krokhmal et al. 2011).
The role of risk management is essential for mathematical optimization under uncertainty. Whenever uncertainty exists, there is a risk. Uncertainty is present when there is a possibility that the outcome of a particular event will deviate from what is expected (Better et al. 2008).
There are not many studies that examine optimization in risk management until now. One of the old studies that examine optimization in risk management is in civil engineering conducted by Cooke and Pinter (1989). The information that can be obtained from this research, i.e., the optimization concept and technique can be combined with probabilistic analysis to structure and solve risk management problems. Other research on optimization in risk management can be seen in the study of Better et al. (2008); Krokhmal et al. (2011); Aranburu et al. (2016); and Lam (2016).
Based on the literature study, optimization in Item Delivery as Risk Management has not been found. For this reason, this study is one part of the contents of this paper. This research resulted in how to make online decisions on item delivery strategies to reduce the risk of loss. These losses can be identified through changes in optimizing the delivery time obtained based on the dynamic network.

3. Material and Methods

3.1. Predictive Distribution

Another term for the predictive distribution is the posterior predictive distribution (Klugman et al. 2012). In general, the predictive distribution is the distribution of the possible unobserved values conditional on the observed value. In this case, we will discuss the predictive distribution of the multinomial case. Of course, it involves the Dirichlet distribution because it is prior for the multinomial. The multinomial distribution is denoted Mult ( θ 1 , θ 2 , , θ m ; n ) , with the probability mass function as follows:
p ( x ; θ ) = n ! i = 1 m x i ! i = 1 m θ i x i ,
where x = [ x 1   x 2 x m ] T ,   i = 1 m x i = n ,   n ,   θ = [ θ 1   θ 2     θ m ] T ,   i = 1 m θ i = 1 , θ i > 0 and θ 1 , θ 2 , , θ m is the probability parameter. The expected value of the multinomial distribution, namely:
E [ X ] = [ E [ X 1 ] , E [ X 2 ] , , E [ X m ] ] T = [ n θ 1 , n θ 2 , , n θ m ] T .
Then, the Dirichlet distribution is denoted Dir ( α 1 , α 2 , , α m ) ,   with the probability density function as follows:
f ( θ ; α ) = Γ ( i = 1 m α i ) i = 1 m Γ ( α i ) i = 1 m θ i α i 1 ,
where α = [ α 1   α 2   α m ] T ,   Γ ( α ) = 0 θ α 1 e θ d θ , m 2 , and α 1 , α 2 , , α m is the concentration parameter, α i > 0 . The expected value of the Dirichlet distribution, namely:
E [ Θ ] = [ E [ θ 1 ] , E [ θ 2 ] , , E [ θ m ] ] T = [ α 1 i = 1 m α i , α 2 i = 1 m α i , , α m i = 1 m α i ] T .
Let X t = [ X 1 , t   X 2 , t X m , t ] T be a random vector that stating the number of events at tth time, where X i , t is a random variable stating the number of events at category i at tth time. Then, assume X t follows a multinomial distribution with the parameter θ . Next, we will find the conditional probability distribution for the x t + k observation if given D = ( x 1 , x 2 , , x t + ( k 1 ) :
p ( x t + k | D t + ( k 1 ) ) =   p ( x t + k | θ ) f ( θ | D ) d θ .
Thus, we obtain Equation (6) (Yudhanegara et al. 2021a):
p ( x t + k | D t + ( k 1 ) ) = n ! Γ ( i = 1 m α i ) Γ ( n + i = 1 m α i ) i = 1 m Γ ( x i + α i ) x i ! Γ ( α i ) .
Equation (6) is a predictive distribution probability function which is a probability mass function of the Dirichlet-multinomial (Avetisyan and Fox 2012). The expected value of Equation (6) can be found by utilizing the multinomial expected value and Theorem 1.
Theorem 1.
(Hogg et al. 2005). Let Y and X be random variables, then E [ Y ] = E [ E [ Y | X ] ] .
Based on Theorem 1, we can find E [ X t + k ] = E [ E [ X t + k | Θ ] ] . In the other, we have ( X t + k | Θ ) ~   Mult ( θ 1 , θ 2 , , θ m , n ) ; thus:
E [ X t + k ] = n E [ Θ ] = n   [ α 1 α 0 , α 2 α 0 , , α m α 0 ] T .
In the Dirichlet-multinomial case, we can take the value α 0 = i = 1 m α i dan θ i = α i i = 1 m α i = α i α 0 (Johnson et al. 1996), so that the following relationship can be seen:
E [ X i , t + k ] = n   θ i = n   α i α 0 .
Thus, we can find θ i from the multinomial estimated parameter, namely:
θ ^ i = t = 1 l x i , t l   n ,   i = 1 , 2 , , m ,
where i = 1 m x i = n ,   n .

3.2. Prediction Accuracy

There are several calculations for prediction errors, one of which is the mean absolute percentage error (MAPE), which is the absolute average percentage error. MAPE is a statistical measure of predictive accuracy in predictive models that provides information on how much prediction error is compared to the observed value based on the sequence (Swamidas 2000). According to Lewis (1982), MAPE scores can be interpreted into four criteria. There are predetermined criteria as described in Table 1.
The smaller the MAPE score, the smaller the error of the prediction results, and the greater the MAPE score, the greater the error of the prediction results. The results of a predictive model have an excellent predictive ability if the MAPE score is below 10 % and they have a good predictive ability if the MAPE score is between 10 % and 20 % .  Table 1 shows the meaning of the error percentage score in MAPE, where the MAPE score can still be used if it does not exceed 50 % ; if the MAPE score is above 50 % , then the prediction model cannot be used. The MAPE formula is:
MAPE = 1 m i = 1 m | ( x i , t x ^ i x i , t ) | × 100 % ,
where x ^ i = n α i α 0 ,   α 0 = i = 1 m α i ,   α i i = 1 m α i = θ i , and i = 1 m θ i = 1 , θ i > 0 .

3.3. Network Clustering

In the item delivery strategy, the purpose of network clustering is to determine delivery zones (Yudhanegara et al. 2021a, 2021b, 2022). In graph theory, network clustering is a graph partition that divides a graph into more than one subgraph (Elsner 1997; Newman and Girvan 2004). The clustering used in this paper is a recursive spectral bisection (Yudhanegara et al. 2021a, 2022). The network is constructed from the map of Bandung City-Indonesia, in Figure 1, see Yudhanegara et al. (2021a, 2022).
In this study, the map is represented as a dynamic network. Dynamic networks experience changes in structure and weight from time to time (Casteigts et al. 2011). The dynamic network concept in this study is defined as a network structure with a change in weight, not a change in the network structure. This dynamic network condition is obtained based on the observation process in alternating periods accompanied by changes in weight. The weight of the network is a random vector with a specific distribution. Unlike the static network, the weights have no distribution.
In Figure 1, the number in the orange circles is the area or location of the destination for the delivery of items, and the blue symbol is the depot. Based on Figure 1, we have the network in Figure 2.
In Figure 2, the number in the orange circles is the area or location (node) of the destination for the delivery of items, except for the orange circle with zeros; it’s a depot (starting node). Furthermore, Figure 3 is an example of a network clustering into four clusters using a recursive spectral bisection method used in this study (Yudhanegara et al. 2021a).
The network in Figure 3 consists of four zones, namely the green zone, gray zone, yellow zone, and purple zone. Green nodes are locations in the green zone, gray nodes in the gray zone, yellow nodes in the yellow zone, and purple nodes in the purple zone.

3.4. Optimization

In this study, the optimization procedure in item delivery uses the Vehicle Routing Problem model, which has been relaxed. This relaxation allows more than one vehicle to pass each road segment. The network is divided into several zones and consists of only one depot. The definition of a delivery zone is a subnet resulting from the clustering of the spectral bisection method. As described previously, the objective function of time optimization in this research problem is used to analyze the structure of the item delivery problem.
Referring to Dror and Trudeau (1990), Reinelt (2012), and Bernardino and Paias (2018), the flow problems with the smallest delivery time used in this study are as follows:
  • One unit of vehicle delivers items to several locations (nodes), but each node is visited only once;
  • The vehicle must return to the depot;
  • The goal is to find the path with the shortest delivery time.
They are adjusted to traffic conditions with changing road density and delivery time limits. The focus of this item delivery process is to get the minimum total time for each zone. The existing data is in the form of the number of vehicles as edge weights. Therefore, it is necessary to convert the edge weights. The conversion is a change from the number of vehicle features to each road segment’s travel time feature. The time weight between nodes is obtained by paying attention to traffic parameters, namely speed and traffic density, which refers to the Direktorat Jenderal Bina Marga or Department of Highways (Yudhanegara et al. 2022).
The variables used in the objective function are:
η i , j ( q ) : { 1 ,   if   vehicle   q   to   node   v j     from   node   v i 0 ,   elsewhere .
Next, we have variable t i , j , it is the delivery time from v i to v j . The goal is to minimize the following function:
Z = i = 0 𝓃 j = 0 𝓃 q = 1 h t i , j η i , j ( q ) ,
subject to
i = 0 𝓃 q = 1 h η i , j ( q ) = 1 , j = 1 , 2 , , n ,   i j ,
and
j = 0 𝓃 q = 1 h η j , k ( q ) = 1 , k = 1 , 2 , , 𝓃 ,   j k ,
Equation (12) guarantees that the vehicle that will deliver items to a node originates from the previous node. Equation (13) guarantees that a vehicle that has delivered items from one node must deliver items to the next node. Equations (12) and (13) also indicate that each node can only be stopped precisely once by one vehicle. The other constraints used are:
j = 1 𝓃 η 0 , j ( q ) = 1 ,   q = 1 , 2 , , h ,
and
j = 1 𝓃 η j , 0 ( q ) = 1 ,   q = 1 , 2 , , h ,
where 0 is the depot. Equation (14) indicates that every vehicle’s travel route begins at the depot, while Equation (15) indicates that every vehicle’s travel route ends at the depot.
Let y = { y 1 , y 2 , , y 𝓃 } be a vector set consisting of 𝓃 items with y y j = [ y j , 1   y j , 2   y j , a ] T ,   j = 1 , 2 , ,   𝓃 and a are features of the item. Suppose we take a where y j , a = τ j , which is the delivery time feature. When operating, each vehicle has a different total delivery time, namely T . Suppose that vehicle q has a delivery time capacity of T ( q ) . Each object j to be sent to node v j , j = 1 , 2 , , 𝓃 , has the delivery time of τ j maks q { T ( q ) } , and each edge ( v i , v j ) , i j ; there is a weight of delivery time between nodes, namely t i , j . Thus, the time constraint for model Z is:
j = 1 𝓃 τ j ( q ) T ( q )   ,   q = 1 , 2 , , h ,
where
τ j ( q ) = i = 0 𝓃 j = 0 𝓃 t i , j ,
and τ j ( q ) is the delivery time of object j to node v j by vehicle q . Equation (17) guarantees that the delivery time for each vehicle may not exceed the delivery time capacity.

3.5. Item Delivery Strategy

Network clustering aims to divide the network into several delivery zones. The distribution of delivery zones will simplify the delivery process to produce a minimum delivery time according to the available vehicles. In theory, this process avoids deliveries between points that are too far apart, which ensures the route of connected vehicles in the same cluster, and obtains an effective delivery route.
In practice, the item delivery process is influenced by traffic conditions. This condition consists of various factors, one of which is changes in road density, which tend to change or are dynamic, making it difficult to control or predict. These dynamic conditions can be anticipated through a predictive distribution approach. This approach involves statistical inference that predicts delivery zones, as presented in Figure 4.
Figure 4 shows that predictive distribution plays a role in predicting delivery zones, while statistical inference is used to test new data distribution.

4. Result and Discussion

4.1. New Method

To get the distribution of the logarithm of the multinomial likelihood function, we can use Theorem 2.
Theorem 2.
(Central Limited Theorem) (Athreya and Lahiri 2006). Let X 1 , X 2 ,   , X m be i.i.d. random samples with mean μ and variance σ 2 > 0 . When m is large, the sample mean X ¯ m = t = 1 m X t m can be approximated by the normal distribution N ( μ , σ 2 m ) .
Based on Theorem 2, for the multinomial case, we have Theorem 3. It is the distribution of the logarithm of the multinomial likelihood function.
Theorem 3.
Let L l ( θ ) = t = 1 l ( n ! i = 1 m x i , t ! i = 1 m θ i x i , t ) be a multinomial likelihood function, for n , m , l 1 , then random variable log L l ( θ ) has a normal distribution with mean:
μ ¨ = t = 1 l i = 1 m x i , t log θ i t = 1 l i = 1 m log x i , t ! + t = 1 l log n !
and variance:
σ ¨ 2 = m ( m 1 ) t = 1 l i = 1 m ( x i , t log θ i i = 1 m x i , t log θ i m ) 2 t = 1 l i = 1 m ( log x i , t ! i = 1 m log x i , t ! m ) 2 .  
Proof of Theorem 3.
We can find:
log L l ( θ ) = log ( t = 1 l ( n ! i = 1 m x i , t ! i = 1 m θ i x i , t ) ) = t = 1 l log ( n ! i = 1 m x i , t ! i = 1 m θ i x i , t ) = t = 1 l ( log n ! i = 1 m log x i , t ! + i = 1 m x i , t log θ i ) = t = 1 l ( log n ! i = 1 m log x i , t ! + i = 1 m x i , t log θ i ) = t = 1 l ( log n ! m i = 1 m log x i , t ! m + m i = 1 m x i , t log θ i m ) .
Let H t = i = 1 m log X i , t ! m = log X i , t ! ¯   and O t = i = 1 m X i , t log θ i m = X i , t log θ i ¯   be random variables that represent the mean, called the sample mean. Since the sampling is assumed to be a random variable, it has a distribution. The sample mean distribution is the distribution of the means obtained from all possible samples of a population, where the sample size is the same.
The center limit theorem applies to the average sampling distribution. When sampling with a simple random sample size m from a population that comes from any distribution, the distribution of the sample mean can be approximated by a normal distribution provided that the sample size is large—thus, based on the central limit theorem, H t ~ N ( μ H t , σ H t 2 ) , where the mean μ H t 𝓀 ¯ = i = 1 m log x i , t ! m , and variance σ H t 2 s 2 m = i = 1 m ( h i h ¯ ) 2 ( m 2 m ) ,   h i = log x i , t ! . Thus, we obtained m H t ~ N ( m μ H t , m 2 σ H t 2 ) . Next, based on the center limit theorem, O t ~ N ( μ O t , σ O t 2 ) , where the mean μ O t o ¯ = i = 1 m X i , t log θ i m ,   and variance σ O t 2 s 2 m = i = 1 m ( o i o ¯ ) 2 ( m 2 m ) , o i = x i , t log θ i . Thus, we have m O t ~ N ( m μ O t , m 2 σ O t 2 ) .
Let X 1 , X 2 , , X l be i.i.d. random samples, where X t ~ N ( μ X t , σ X t 2 ) ,   t = 1 , 2 , , l .   We can find distribution t = 1 l X t through moment generator function (Hogg et al. 2005; Walpole et al. 2012), so t = 1 l X t ~ N ( t = 1 l μ X t , t = 1 l σ X t 2 ) . Then, we get ( m H t + m O t ) ~ N ( m μ H t + m μ O t , m 2 σ H t 2 + m 2 σ O t 2 ) , so we have:
( m H t + m O t ) ~ N ( m ( μ O t μ H t ) , m 2 ( σ O t 2 σ H t 2 ) ) .
Next, for ( log n ! m H t + m O t ) ~ N ( m ( μ O t μ H t ) + log n ! , m 2 ( σ O t 2 u σ H t 2 ) ) , we find:
t = 1 l ( log n ! m H t + m O t ) ~ N ( μ ¨ , σ ¨ 2 ) ,
where
μ ¨ = t = 1 l ( m ( μ O t μ H t ) + log n ! ) = t = 1 l ( m ( i = 1 m X i , t log θ i m i = 1 m log x i , t ! m ) + log n ! ) = t = 1 l i = 1 m X i , t log θ i t = 1 l i = 1 m log x i , t ! + t = 1 l log n ! ,
and
σ ¨ 2 = t = 1 l m 2 ( σ O t 2 σ H t 2 ) = t = 1 l m 2 ( i = 1 m ( o i o ¯ ) 2 l ( m 1 ) i = 1 m ( h i h ¯ ) 2 l ( m 1 ) ) = t = 1 l m 2 ( i = 1 m ( x i , t log θ i i = 1 m x i , t log θ i m ) 2 ( m 2 m ) i = 1 m ( log x i , t ! i = 1 m log x i , t ! m ) 2 ( m 2 m ) ) σ ¨ 2 = m ( m 1 ) ( t = 1 l i = 1 m ( x i , t log θ i i = 1 m x i , t log θ i m ) 2 t = 1 l i = 1 m ( log x i , t ! i = 1 m log x i , t ! m ) 2 ) .  
To find the value of log n ! , we can approximate the Stirling formula in Equation (22):
n ! 2 π n ( n e ) n .
We can find Corollary 1 with a more equation from Theorem 3 with the same distribution.
Corollary 1.
Let L l ( θ ) = t = 1 l ( n ! i = 1 m x i , t ! i = 1 m θ i x i , t ) be a multinomial likelihood function, for n , m , l 1 , then random variable ( log L l ( θ ) t = 1 l log n ! ) has a normal distribution with mean:
μ = t = 1 l i = 1 m x i , t log θ i t = 1 l i = 1 m log x i , t ! ,
and variance σ 2 = σ ¨ 2 , Equation (19).
Proof of Corollary 1.
From  log L l ( θ ) we have:
log L l ( θ ) = t = 1 l ( log n ! i = 1 m log x i , t ! + i = 1 m x i , t log θ i ) ,
and thus,
log L l ( θ ) t = 1 l log n ! = t = 1 l ( i = 1 m log x i , t ! + i = 1 m x i , t log θ i ) .
Next, based on Equations (19) and (20), we have:
( m H t + m O t ) = t = 1 l ( i = 1 m log x i , t ! + i = 1 m x i , t log θ i ) .
Thus, we get:
t = 1 l ( i = 1 m log x i , t ! + i = 1 m x i , t log θ i ) ~ ( m ( μ O t μ H t ) , m 2 ( σ O t 2 σ H t 2 ) ) ,
suppose
μ = m ( μ O t μ H t ) = t = 1 l i = 1 m x i , t log θ i t = 1 l i = 1 m log x i , t !
and
σ 2 = m 2 ( σ O t 2 σ H t 2 ) = m ( m 1 ) ( t = 1 l i = 1 m ( x i , t log θ i i = 1 m x i , t log θ i m ) 2 t = 1 l i = 1 m ( log x i , t ! i = 1 m log x i , t ! m ) 2 ) .
As an illustration, the values of log L l ( θ ) where m = 38 ,   n = 2500 , l = 30 , for 𝓀 = 2000 times, are shown in Figure 5.
Next, as an illustration, the values of log L l ( θ ) , where m = 260 ,   n = 1 , 738 , 665 ,   l = 30 , for 𝓀 = 1000 , are shown in Figure 6.
The probability density function for histogram in Figure 5 is:
f ( x ) = 1 9435 2 π e 1 2 ( x 311 , 261 9435 ) 2 ,
and
f ( x ) = 1 4505 2 π e 1 2 ( x 308 , 256 4505 ) 2
for the histogram in Figure 6. Based on Figure 5 and Figure 6, the histogram of values of log L l ( θ ) are symmetrical. They are theoretically corresponding to Theorem 3. Corollary 1 state that ( log L l ( θ ) t = 1 l log n ! ) ~ N ( μ , σ 2 ) . It is the statistical form for multinomial goodness-of-fit test. It applies the following hypothesis test:
H0: 
( log L l + 1 ( θ ) t = 1 l + 1 log n ! ) follows a normal distribution with the mean  μ and variance  σ 2 ,
H1: 
( log L l + 1 ( θ ) t = 1 l + 1 log n ! ) does not follow a normal distribution with the mean  μ and variance  σ 2 .
If we look at the logarithmic form of the likelihood function for l + 1 observations, then Corollary 2 applies. This result illustrates that there is no need to store historical data in online testing, but it is enough to use the last calculation.
Corollary 2.
Value log L l + 1 ( θ ) t = 1 l + 1 log n ! , μ l + 1 and σ l + 1 2 ,   it can be obtained iteratively through  log L l ( θ ) t = 1 l log n ! ,   μ l and  σ l 2 ; thus:
log L l + 1 ( θ ) t = 1 l + 1 log n ! = ( log L l ( θ ) t = 1 l + 1 log n ! ) + ( i = 1 m log x i , ( l + 1 ) ! + i = 1 m x i , ( l + 1 ) log θ i ) , μ l + 1 = μ l + ( i = 1 m x i , ( l + 1 ) log θ i i = 1 m log x i , ( l + 1 ) ! ) ,
and
σ l + 1 2 = σ l 2 + m ( m 1 ) ( i = 1 m ( x i , ( l + 1 ) log θ i i = 1 m x i , ( l + 1 ) log θ i m ) 2 i = 1 m ( log x i , ( l + 1 ) ! i = 1 m log x i , ( l + 1 ) ! m ) 2 ) .
Proof of Corollary 2.
Consider the logarithm of the multinomial likelihood function for l observations as follows:
log L l ( θ ) = t = 1 l ( log n ! m i = 1 m log x i , t ! m + m i = 1 m x i , t log θ i m ) = log L l 1 ( θ ) + ( log n ! i = 1 m log x i , l ! + i = 1 m x i , l log θ i ) .
Iteratively, for l + 1 observations, based on the above form, the logarithm of the likelihood function can be written as:
log L l + 1 ( θ ) = log L l ( θ ) + ( log n ! i = 1 m log x i , ( l + 1 ) ! + i = 1 m x i , ( l + 1 ) log θ i ) .
Thus, based on the equation, we get:
log L l + 1 ( θ ) t = 1 l + 1 log n ! = ( log L l ( θ ) t = 1 l + 1 log n ! ) + ( i = 1 m log x i , ( l + 1 ) ! + i = 1 m x i , ( l + 1 ) log θ i ) .
The iterative process also applies to μ l + 1 and σ l + 1 2 , and is described as follows:
μ l = t = 1 l i = 1 m x i , t log θ i t = 1 l i = 1 m log x i , t ! = t = 1 l 1 i = 1 m x i , t log θ i + i = 1 m x i , l log θ i t = 1 l 1 i = 1 m log x i , t ! i = 1 m log x i , l ! = t = 1 l 1 i = 1 m x i , t log θ i t = 1 l 1 i = 1 m log x i , t ! + ( i = 1 m x i , l log θ i i = 1 m log x i , l ! ) = μ l 1 + ( i = 1 m x i , l log θ i i = 1 m log x i , l ! ) .
Based on the description above, we have:
μ l + 1 = μ l + ( i = 1 m x i , ( l + 1 ) log θ i i = 1 m log x i , ( l + 1 ) ! ) .
For variance, we have:
σ l 2 = m ( m 1 ) ( t = 1 l i = 1 m ( x i , t log θ i i = 1 m x i , t log θ i m ) 2 t = 1 l i = 1 m ( log x i , t ! i = 1 m log x i , t ! m ) 2 ) = m ( m 1 ) t = 1 l i = 1 m ( x i , t log θ i i = 1 m x i , t log θ i m ) 2 t = 1 l i = 1 m ( log x i , t ! i = 1 m log x i , t ! m ) 2 = m ( m 1 ) ( t = 1 l 1 i = 1 m ( x i , t log θ i i = 1 m x i , t log θ i m ) 2 + i = 1 m ( x i , l log θ i i = 1 m x i , l log θ i m ) 2 ( t = 1 l 1 i = 1 m ( log x i , t ! i = 1 m log x i , t ! m ) 2 + i = 1 m ( log x i , l ! i = 1 m log x i , l ! m ) 2 ) ) = m ( m 1 ) ( t = 1 l 1 i = 1 m ( x i , t log θ i i = 1 m x i , t log θ i m ) 2 t = 1 l 1 i = 1 m ( log x i , t ! i = 1 m log x i , t ! m ) 2 + i = 1 m ( x i , l log θ i i = 1 m x i , l log θ i m ) 2 i = 1 m ( log x i , l ! i = 1 m log x i , l ! m ) 2 ) σ l 2 = ( σ l 1 2 ) + m ( m 1 ) ( i = 1 m ( x i , l log θ i i = 1 m x i , l log θ i m ) 2 i = 1 m ( log x i , l ! i = 1 m log x i , l ! m ) 2 ) .
Thus, we find:
σ l + 1 2 = σ l 2 + m ( m 1 ) ( i = 1 m ( x i , ( l + 1 ) log θ i i = 1 m x i , ( l + 1 ) log θ i m ) 2 i = 1 m ( log x i , ( l + 1 ) ! i = 1 m log x i , ( l + 1 ) ! m ) 2 ) .
The hypothesis test used is the Z statistical test with the formula in Equation (26):
z s t a t i t i c a l = ( log L l + 1 ( θ ) t = 1 l + 1 log n ! ) μ σ ,
where log L l + 1 ( θ ) t = 1 l + 1 log n ! is random variable under the assumption H 0 . The decision-making criteria H 0 is not rejected when z α 2 z s t a t i t i c a l z α 2 , H 0 is rejected if z s t a t i t i c a l < z α 2 or z α 2 < z s t a t i t i c a l . The margin of error for the confidence interval ( 1 α ) in statistics on Equation (26) can be determined through Theorem 4. This theorem provides a formula for the error tolerance limit for the population average at the confidence interval ( 1 α ) .
Theorem 4.
Let  ( log L l + 1 ( θ ) t = 1 l + 1 log n ! ) ~ N ( μ l , σ l 2 ) be a random variable, where μ l = t = 1 l i = 1 m x i , t log θ i t = 1 l i = 1 m log x i , t !  and  σ l 2 = m ( m 1 ) t = 1 l i = 1 m ( x i , t log θ i i = 1 m x i , t log θ i m ) 2 t = 1 l i = 1 m ( log x i , t ! i = 1 m log x i , t ! m ) 2 ,  then margin of error ( ε )  for the confidence interval  ( 1 α )  is ε = z ( 1 α 2 ) σ l .
Proof of Theorem 4.
Let ε be a margin of error for the confidence interval   ( 1 α ) . Consider the following form:
P ( | ( log L l + 1 ( θ ) t = 1 l + 1 log n ! ) μ l | < ε ) = 1 α P ( ε < ( ( log L l + 1 ( θ ) t = 1 l + 1 log n ! ) μ l ) < ε ) = 1 α P ( ε σ l < ( ( log L l + 1 ( θ ) t = 1 l + 1 log n ! ) μ l ) σ l < ε σ l ) = 1 α P ( ε σ l < z s t a t < ε σ l ) = 1 α .
Based on the description, we have:
ε σ l = z ( 1 α 2 ) ε = z ( 1 α 2 ) σ l .
The steps for testing new data based on historical data are described in Algorithm 1.
Algorithm 1. Data testing.
  • Input  ( μ , σ 2 )
  • Find the estimated parameter  ( θ )  of the predictive distribution (based on historical data).
  • Enter the new observed value under the assumption H 0 .
  • Perform the Z statistical test at the significance level α.  H 0  is not rejected, it means that  ( l + 1   ) ~ N ( μ , σ 2 ) .
  • Update the parameters  ( θ )  and  ( μ , σ 2 )  if  H 0  is rejected.
  • Return to step 3.

4.2. Limitations

The smallest value or minimum limit of m , n , and l is determined based on the simulation of the values of log L ( θ ) so that Theorem 3 and Corol1ary 1 can be used. In the next step, the values of log L ( θ ) are tested for model fit using chi-squared statistics under the assumption H 0 , H 0 : data follow a normal distribution. The rejection criterion H 0   is based on a p-value, i.e., the probability value of obtaining a result is at least as extreme as the observed result of a statistical hypothesis test, assuming that the null hypothesis is true. The p-value is used as an alternative rejection point to provide a minor significance level with the null hypothesis being rejected. Obtaining a smaller p-value indicates more substantial evidence supporting the alternative hypothesis.
In the first simulation, for the value of m = 9 ,   l = 10 , and n = 115 , for 𝓀 = 500 , the histogram is obtained in Figure 7a. Based on the model fit test, a p-value < 0.0001 is obtained so that at a significance level of 0.05 , the value log L ( θ ) does not follows a normal distribution. Then, for the values of m = 10 ,   l = 8 , and n = 115 , for 𝓀 = 500 , the histogram is obtained in Figure 7b. A p-value = 0.018 was obtained based on the model fit test. At the 0.05 significance level, it can be stated that the log L ( θ ) value does not follow a normal distribution, similar to the previous simulation in Figure 7a.
The following simulation is carried out for the values of m = 10 ,   l = 9 , and n = 115 with 𝓀 = 500 ; the histogram is obtained in Figure 8a.
From these values, based on the model fit test, the p-value = 0.066 . This means that at a significance level of 0.05 , it can be concluded that the value of log L ( θ ) follows a normal distribution with mean μ ¨ = 2283.0220   and variance σ ¨ 2 = 11.877 .
Next, a simulation is carried out for the values of m = 10 ,   l = 9 , and n = 110 , for 𝓀 = 500 , to obtain the histogram in Figure 8b. Based on the model fit test, a p-value < 0.0001 was obtained with these values. Thus, at a significance level of 0.05 , the value of log L ( θ ) does not follow a normal distribution. After that, the last simulation was carried out for the values of m = 10 ,   l = 9 , and n = 111 , for 𝓀 = 500 ; the histogram is shown in Figure 9.
Based on the model fit test, the p-value = 0.363 at a significance level of 0.05 means that the value of log L ( θ ) follows a normal distribution with probability density function:
f ( x ) = 1 1.969 2 π e 1 2 ( x 2197.444 1.969 ) 2 .
This simulation concludes that in the case of multinomial distribution, Theorem 3 and Corollary 1 apply to the minimum values of m = 10 ,   l = 9 , and n = 111 .

4.3. Simulation

Let X t = [ X 1 , t   X 2 , t X m , t ] T be a random vector that stating the number of vehicles that might pass each road at tth day, where X i , t is a random variable stating the number of vehicles pass ith road at tth day. Then, assume X t follows a multinomial distribution with the parameter θ (probability). Furthermore, a total of 1 , 738 , 665 events are obtained with 260 event categories.
The network is constructed from the map of Bandung City-Indonesia, shown in Figure 1. We applied them to transportation data, namely traffic density. In this case, a map is presented in a dynamic network, and categories are represented as roads (Yudhanegara et al. 2020, 2021a, 2021b). The total number of events is the total number of vehicles passing on the road, see Figure 3 (Yudhanegara et al. 2021a).
Then, the data are generated 40 times under the assumption of a multinomial distribution with the probability mass function of X t ~ Mult   ( θ 1 , θ 2 , , θ 260 ; 1,738,665 ) , namely:
p ( x ; θ ) = 1,738,665 ! i = 1 260 x i , t ! i = 1 260 θ i x i , t ,
where i = 1 260 θ i = 1 and i = 1 260 x i , t = 1 , 738 , 665 . The parameter θ is generated randomly from the standard uniform distribution ( U ( 0 , 1 ) ) . We have 40 generated data, where 60 % ( 24 datas) was used as training data, and 40 % ( 16   datas) was used as testing data. The expected value and the estimated parameter of training data are obtained from the predictive distribution in Equation (6).
The training data that has been used as the network weight is presented in Figure 10. Figure 10 shows the network that has been clustered into four zones using the spectral bisection method. Each zone tends to change cluster members within its zone at different times. Furthermore, the data prediction simulation is presented in Table 2.
By using the estimated parameter ( θ ) of historical data, the value of Z s t a t i t i c a l is obtained, as shown in Table 2.
Next, to evaluate the prediction results of the predictive distribution model, we use MAPE . The results of the MAPE scores calculation is described in Table 3.
In general, based on the criteria for the MAPE scores in Table 1, the MAPE scores between the observed value and the expected value in Table 3 indicates that the model used is accurate because the MAPE scores are below 10%.

4.4. Optimization as Risk Management

The next goal of risk management is to provide information about potential sources of risk in the company (Lam 2016). Risk management is an effort to reduce risk in technical implementation and business decision making (Lam 2016). The purpose of risk management is to mitigate or track sources that have the potential to threaten business productivity and security. This tracking process can be carried out by research and procedural analysis of every company activity, from production to asset management. When risks are found and analyzed, it is necessary to make efforts so that risks do not occur and threaten business continuity.
Although it has extended and successive stages, the risk management process is one of the essential components of business management that can protect the company from many problems. In this study, particularly the item delivery strategy, optimization is one of the most critical components of risk management (Cooke and Pinter 1989; Better et al. 2008).
Next, we analyze the optimization of a dynamic network. Based on Table 2, for H 0 is rejected, if there is no change in the item delivery strategy (ignoring H 0 is rejected), then the optimization of delivery time in each zone will no longer be effective (Yudhanegara et al. 2022). This happens because the routes used are not correct. The result is significant delivery times on certain vehicles, while others are too small. The optimization of the time in the order of cases for H 0 is rejected as described in Table 4 and Table 5.

5. Conclusions

Let X t = [ X 1 , t   X 2 , t X m , t ] T be a random vector, X t ~ Mult   ( θ 1 , θ 2 , , θ m , n ) ,   t = 1 , 2 , , l , for n , m , l 1 , for the appropriate model fit test is used, namely the Z statistical test with log L l ( θ ) ~ N ( μ ¨ , σ ¨ 2 ) or ( log L l ( θ ) t = 1 l log n ! ) ~ N ( μ , σ 2 ) . The algorithm presented as the method of inferential statistics for online decisions. It is suitable for streaming data, some of which consist of traffic density data. This method can predict data with small error values, resulting in very accurate predictions. The method does not require large memory storage for historical data.
The case presented in this paper is suitable for controlling traffic jam in cities and traffic jam on holidays. The solution to the problems presented in this paper can be helpful for the transportation department to make decisions or provide information to the public. The impact on the item delivery company by having accurate information on traffic jams will enable them to formulate strategies for item deliveries.
For future research, we can simulate the data based on the predictive distribution of the multivariate Poisson cases. Moreover, a case study can be used with other discrete multivariate distributions. For example, there are cases where n is not constant, and n , the probability of the occurrence of the event being considered is θ 0 , and n . p = ω , where ω is the rate, so that further research is related to the multivariate Poisson distribution. Then, we can determine a new test method built based on the logarithm of the likelihood function for the multivariate Poisson distribution.
After obtaining the predictive distribution for the multivariate Poisson case, we will also apply it to education. Clustering for category data and decision-making methods will be needed for analyzing the risk of failing to graduate from university. Examples of data can be seen in a study conducted by Yudhanegara and Lestari (2019).

Author Contributions

Conceptualization, S.W.I. and M.R.Y.; methodology, S.W.I.; software, M.R.Y.; validation, S.W.I., K.N.S., and M.R.Y.; formal analysis, S.W.I.; investigation, K.N.S.; resources, M.R.Y.; data curation, K.N.S.; writing—original draft preparation, M.R.Y.; writing—review and editing, S.W.I.; visualization, K.N.S.; supervision, S.W.I.; project administration, S.W.I.; funding acquisition, S.W.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by RISET UNGGULAN LPPM ITB 2022 (Second Year), grant number NIP 197508041999031003 and the APC was funded by Sapto Wahyu Indratno.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are generated under the assumption of a multinomial distribution with our program. They are available on request from the corresponding author.

Acknowledgments

The authors would like to thank LPPM ITB for research funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Aranburu, Larraitz, Laureano F. Escudero, M. Araceli Garín, María Merino, and Gloria Pérez. 2016. Risk Management for Mathematical Optimization under Uncertainty. Documento de Trabajo BILTOKI 1: 1–18. [Google Scholar]
  2. Athreya, Krishna B., and Soumendra N. Lahiri. 2006. Measure Theory and Probability Theory. New York: Springer. [Google Scholar]
  3. Avetisyan, Marianna, and Jean-Paul Fox. 2012. The Dirichlet-multinomial model for multivariate randomized response data and small samples. Psicologica 33: 362–90. [Google Scholar]
  4. Azen, Razia, and Cindy M. Walker. 2010. Categorical Data Analysis for the Behavioral and Social Sciences. New York: Routledge Taylor & Francis Group. [Google Scholar]
  5. Bernardino, Raquel, and Ana Paias. 2018. Solving the family traveling salesman problem. European Journal of Operational Research 267: 453–66. [Google Scholar] [CrossRef]
  6. Better, Marco, Fred Glover, Gary Kochenberger, and Haibo Wang. 2008. Simulation optimization: Aplications in risk management. International Journal of Information Technology & Decision Making 7: 571–87. [Google Scholar] [CrossRef]
  7. Brock, Willian A., and Steven N. Durlauf. 2003. Multinomial Choice with Social Interactions. Technical Working Paper: National Bureau of Economic Research 288: 1–44. [Google Scholar]
  8. Casteigts, Arnaud, Paola Flocchini, Walter Quattrociocchi, and Nicola Santoro. 2011. Time-varying graphs and dynamic networks. International Conference on Ad-Hoc Networks and Wireless 27: 346–59. [Google Scholar]
  9. Cooke, Roger, and Janos Pinter. 1989. Optimization in risk management. Civil Engineering Systems 6: 122–28. [Google Scholar] [CrossRef]
  10. Degtereva, Viktoria, Maria Liubarskaia, Viktoria Merkusheva, and Alexey Artemiev. 2022. Increasing Importance of Risk Management in the Context of Solid Waste Sphere Reforming in Russian Regions. Risks 10: 79. [Google Scholar] [CrossRef]
  11. Dror, Moshe, and Pierre Trudeau. 1990. Split delivery routing. Naval Research Logistic 37: 383–402. [Google Scholar] [CrossRef]
  12. Elsner, Ulrich. 1997. Graph Partitioning. Chemnits: Technishe Universitat Chemnits. [Google Scholar]
  13. Hogg, Robert V., Joseph W. McKean, and Allen T. Craig. 2005. Introduction to Mathematical Statistics, 6th ed. Hoboken: Pearson. [Google Scholar]
  14. Johnson, Norman L., Samuel Kotz, and Narayanaswamy Balakrishnan. 1996. Discrete Multivariate Distributions. New York: John Wiley & Sons Inc. [Google Scholar]
  15. Klugman, Stuart A., Harry H. Panjer, and Gordon E. Willmot. 2012. Loss Models: From Data to Decisions, 4th ed. Hoboken: John Wiley & Sons Inc. [Google Scholar]
  16. Krokhmal, Pavlo, Michael Zabarankin, and Stan Uryasev. 2011. Modeling and optimization of risk. Surveys in Operations Research and Management Science 16: 49–66. [Google Scholar] [CrossRef]
  17. Lam, James. 2016. Strategic Risk Management: Optimizing the Risk Return Profile. New York: The Association of Accounts and Financial Professionals in Bussiness (IMA). [Google Scholar]
  18. Lewis, Colin D. 1982. Industrial and Business Forecasting Methods. London: Butterworths. [Google Scholar]
  19. Nelissen, Heleen E., Daniella Brals, Hafsat A. Ameen, Marijn van Der List, Berber Krammer, Tanimola M. Akande, Wendy Janssens, and Anja H. Van’t Hoog. 2020. The prominent roel of informal medicine vendors despite health insurance: A weekly diaries study in rural Nigeria. Health Policy and Planning 35: 54–363. [Google Scholar] [CrossRef] [PubMed]
  20. Newman, Mark E. J., and Michelle Girvan. 2004. Finding and evaluating community structure in networks. Physical Review E 69: 026113. [Google Scholar] [CrossRef] [PubMed]
  21. Nownes, Anthony J. 1992. Primaries, general elections, and voter turnout: A multinomial logit model of the decision to vote. American Politics Quarterly 20: 205–26. [Google Scholar] [CrossRef]
  22. Oosterhoff, John, and Willem R. Van Zwet. 1972. The likelihood ratio test for the multinomial distribution. Proceedings of the Sixth Berkeley Symposium on Mathematical Statistics and Probability 1: 31–49. [Google Scholar] [CrossRef]
  23. Owen, Art B. 2018. Introduction to Statistical Inference: Generalized Likelihood Ratios Test. Stanford: Stanford University, Available online: https://web.stanford.edu/class/archive/stats/stats200/stats200.1172/Lecture22.pdf (accessed on 23 March 2020).
  24. Reddy, Byreddy R., Ch. Deepika, G. Navya, and Gopal Raja. 2015. Factors effecting the usage of public transport multinomial logit model. International Journal of Computer Science and Information Technologies 6: 4672–675. [Google Scholar]
  25. Reinelt, Gerhard. 2012. TSPLIB: A traveling salesman problem library. ORSA Journal on Computing 3: 376–84. [Google Scholar] [CrossRef]
  26. Rith, Monorom, Fengqi Liu, Pai-Hsien Hung, Kento Yoh, Alexis M. Fillone, and Jose B. M. Biona. 2019. R programming language written for estimation of the integrated multinomial logit-linear regression model based on a copula approach: A technical article. Proceedings of the Eastern Asia Society for Transportation Studies 12: 1–11. [Google Scholar]
  27. Swamidas, Paul M. 2000. Encyclopedias of Production and Manufacturing Management. Boston: Kluwer Academic Publishing. [Google Scholar]
  28. Walpole, Ronald E., Raymon H. Myers, Sharon L. Myers, and Keying Ye. 2012. Probability and Statistics for Engineer and Scientists, 6th ed. Hoboken: Pearson. [Google Scholar]
  29. Yudhanegara, Mokhammad R., and Karunia E. Lestari. 2019. Clustering for multi-dimensional data set: A case study on educational. Journal of Physics: Conference Series 1280: 042025. [Google Scholar] [CrossRef]
  30. Yudhanegara, Mokhammad R., Sapto W. Indratno, and R. R. Kurnia N. Sari. 2020. Clustering for items distribution network. Journal of Physics: Conference Series 1496: 012019. [Google Scholar] [CrossRef]
  31. Yudhanegara, Mokhammad R., Sapto W. Indratno, and R. R. Kurnia N. Sari. 2021a. Dynamic items deivery network: Prediction and clustering. Heliyon 7: e06934. [Google Scholar] [CrossRef]
  32. Yudhanegara, Mokhammad R., Sapto W. Indratno, and R. R. Kurnia N. Sari. 2021b. Role of clustering method in items delivery optimization. Journal of Physics: Conference Series 2084: 012011. [Google Scholar] [CrossRef]
  33. Yudhanegara, Mokhammad R., Sapto W. Indratno, and R. R. Kurnia N. Sari. 2022. Prediction of traffic density and item delivery strategy through clustering. Sesiomadika 5: 27–46. [Google Scholar]
Figure 1. Maps of Bandung city as an item delivery network.
Figure 1. Maps of Bandung city as an item delivery network.
Risks 10 00122 g001
Figure 2. Item delivery network from the map.
Figure 2. Item delivery network from the map.
Risks 10 00122 g002
Figure 3. Network clustering with four zones/clusters.
Figure 3. Network clustering with four zones/clusters.
Risks 10 00122 g003
Figure 4. Item delivery strategy.
Figure 4. Item delivery strategy.
Risks 10 00122 g004
Figure 5. Histogram of values of log L l ( θ ) for 𝓀 = 2000 .
Figure 5. Histogram of values of log L l ( θ ) for 𝓀 = 2000 .
Risks 10 00122 g005
Figure 6. Histogram of values of log L l ( θ ) for 𝓀 = 1000 .
Figure 6. Histogram of values of log L l ( θ ) for 𝓀 = 1000 .
Risks 10 00122 g006
Figure 7. Histogram of values of log L l ( θ ) , (a) for 𝓀 = 500 ,   m = 9 ,   l = 10 ,   n = 115 ; and (b) for 𝓀 = 500 ,   m = 10 ,   l = 8 ,   n = 115 .
Figure 7. Histogram of values of log L l ( θ ) , (a) for 𝓀 = 500 ,   m = 9 ,   l = 10 ,   n = 115 ; and (b) for 𝓀 = 500 ,   m = 10 ,   l = 8 ,   n = 115 .
Risks 10 00122 g007
Figure 8. Histogram of values of log L l ( θ ) , (a) for 𝓀 = 500 ,   m = 10 ,   l = 9 ,   n = 115 ; and (b) for 𝓀 = 500 ,   m = 10 ,   l = 9 ,   n = 10 .
Figure 8. Histogram of values of log L l ( θ ) , (a) for 𝓀 = 500 ,   m = 10 ,   l = 9 ,   n = 115 ; and (b) for 𝓀 = 500 ,   m = 10 ,   l = 9 ,   n = 10 .
Risks 10 00122 g008
Figure 9. Histogram of values of log L l ( θ ) , for 𝓀 = 500 ,   m = 10 ,   l = 9 ,   n = 111 .
Figure 9. Histogram of values of log L l ( θ ) , for 𝓀 = 500 ,   m = 10 ,   l = 9 ,   n = 111 .
Risks 10 00122 g009
Figure 10. Network clustering at different times.
Figure 10. Network clustering at different times.
Risks 10 00122 g010
Table 1. Criteria for the MAPE score.
Table 1. Criteria for the MAPE score.
PercentageCriteria
MAPE < 10 % Very accurate
10 % MAPE < 20 % Good
20 % MAPE < 50 % Fair
MAPE 50 % Not accurate
Table 2. Summary of hypothesis testing with α = 0.05 , and z c r i t i c a l = 1.96   or z c r i t i c a l = 1.96 .
Table 2. Summary of hypothesis testing with α = 0.05 , and z c r i t i c a l = 1.96   or z c r i t i c a l = 1.96 .
Parameter ( θ ) and ( μ , σ 2 ) Data z s t a t i t i c a l DecisionItem Delivery Strategy Changes Parameter ( μ l , σ l 2 ) Error ( ε )
D1–24D1–24 vs. D25 0.154 H 0 is not rejectedNoD1–240.23912
D1–24D1–24 vs. D26 0.308 H 0 is not rejectedNoD1–250.22540
D1–24D1–24 vs. D27 0.462 H 0 is not rejectedNoD1–260.21168
D1–24D1–24 vs. D28 0.615 H 0 is not rejectedNoD1–270.23324
D1–24D1–24 vs. D29 0.769 H 0 is not rejectedNoD1–280.21168
D1–24D1–24 vs. D30 0.923 H 0 is not rejectedNoD1–290.24108
D1–24D1–24 vs. D31 1.003 H 0 is not rejectedNoD1–300.25872
D1–24D1–24 vs. D32 1.263 H 0 is not rejectedNoD1–310.24892
D1–24D1–24 vs. D33 1.465 H 0 is not rejectedNoD1–320.23912
D1–24D1–24 vs. D34 1.671 H 0 is not rejectedNoD1–330.23324
D1–24D1–24 vs. D35 1.865 H 0 is not rejectedNoD1–340.22932
D1–24D1–24 vs. D36 2.005 H 0 is rejectedYesD1–350.31752
D1–36D1–36 vs. D37 0.007 H 0 is not rejected,
by updating ( θ ) and ( μ , σ 2 )
NoD1–360.19404
D1–36D1–36 vs. D38 0.271 H 0 is not rejectedNoD1–370.2156
D1–36D1–36 vs. D39 0.429 H 0 is not rejectedNoD1–380.2410
D1–36D1–36 vs. D40 0.533 H 0 is not rejectedNoD1–390.2001
Note: D25 is the 25th data, and D1–24 is the 1st to 24th data.
Table 3. The results of the MAPE calculation and interpretation for the prediction.
Table 3. The results of the MAPE calculation and interpretation for the prediction.
Data MAPE   Score Interpretation Data MAPE   Score Interpretation
E1–24 vs. D250.9%very accurateE1–24 vs. D339.8%very accurate
E1–24 vs. D261.4%very accurateE1–24 vs. D348.8%very accurate
E1–24 vs. D271.0%very accurateE1–24 vs. D357.8%very accurate
E1–24 vs. D281.0%very accurateE1–36 vs. D374.3%very accurate
E1–24 vs. D291.9%very accurateE1–36 vs. D383.2%very accurate
E1–24 vs. D302.8%very accurateE1–36 vs. D392.2%very accurate
E1–24 vs. D314.3%very accurateE1–36 vs. D402.8%very accurate
E1–24 vs. D325.7%very accurate
Note: D25 is the 25th data, and E1–24 is the expected value of the 1st to 24th data.
Table 4. The risks when ignoring H 0 is rejected (for the yellow zone and gray zone).
Table 4. The risks when ignoring H 0 is rejected (for the yellow zone and gray zone).
Yellow ZoneGray Zone
Total time on initial strategyTotal time when ignoring H 0 is rejectedTotal time on initial strategyTotal time when ignoring H 0 is rejected
Vehicle 1Vehicle 2Vehicle 3Vehicle 1Vehicle 2Vehicle 3Vehicle 1Vehicle 2Vehicle 3Vehicle 1Vehicle 2Vehicle 3
86.42 min45.878 min58.751 min120.42 min170.3 min70.5 min73.448 min136.21 min152.18 min37.6 min50.9 min40.41 min
Table 5. The risks when ignoring H 0 is rejected (for the green zone and purple zone).
Table 5. The risks when ignoring H 0 is rejected (for the green zone and purple zone).
Green ZonePurple Zone
Total time on initial strategyTotal time when ignoring H 0 is rejectedTotal time on initial strategyTotal time when ignoring H 0 is rejected
Vehicle 1Vehicle 2Vehicle 3Vehicle 1Vehicle 2Vehicle 3Vehicle 1Vehicle 2Vehicle 3Vehicle 1Vehicle 2Vehicle 3
56.518 min53.063 min58.6235 min154.6 min40.9 min40.87 min70.03 min46.302 min50.13 min98.1 min140.3 min145.7 min
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Indratno, S.W.; Sari, K.N.; Yudhanegara, M.R. Optimization in Item Delivery as Risk Management: Multinomial Case Using the New Method of Statistical Inference for Online Decision. Risks 2022, 10, 122. https://doi.org/10.3390/risks10060122

AMA Style

Indratno SW, Sari KN, Yudhanegara MR. Optimization in Item Delivery as Risk Management: Multinomial Case Using the New Method of Statistical Inference for Online Decision. Risks. 2022; 10(6):122. https://doi.org/10.3390/risks10060122

Chicago/Turabian Style

Indratno, Sapto Wahyu, Kurnia Novita Sari, and Mokhammad Ridwan Yudhanegara. 2022. "Optimization in Item Delivery as Risk Management: Multinomial Case Using the New Method of Statistical Inference for Online Decision" Risks 10, no. 6: 122. https://doi.org/10.3390/risks10060122

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop