Next Article in Journal
Discrete Quantum Harmonic Oscillator
Next Article in Special Issue
Normal-G Class of Probability Distributions: Properties and Applications
Previous Article in Journal
A Robust and High Capacity Data Hiding Method for H.265/HEVC Compressed Videos with Block Roughness Measure and Error Correcting Techniques
Previous Article in Special Issue
Skewness of Maximum Likelihood Estimators in the Weibull Censored Data

Symmetry 2019, 11(11), 1361; https://doi.org/10.3390/sym11111361

Article
Generalized Truncation Positive Normal Distribution
by 1,*,
1
Departamento de Ciencias Matemáticas y Físicas, Facultad de Ingeniería, Universidad Católica de Temuco, Temuco 4780000, Chile
2
Departamento de Matemática, Facultad de Ingeniería, Universidad de Atacama, Copiapó 1530000, Chile
*
Author to whom correspondence should be addressed.
Received: 27 September 2019 / Accepted: 29 October 2019 / Published: 3 November 2019

## Abstract

:
In this article we study the properties, inference, and statistical applications to a parametric generalization of the truncation positive normal distribution, introducing a new parameter so as to increase the flexibility of the new model. For certain combinations of parameters, the model includes both symmetric and asymmetric shapes. We study the model’s basic properties, maximum likelihood estimators and Fisher information matrix. Finally, we apply it to two real data sets to show the model’s good performance compared to other models with positive support: the first, related to the height of the drum of the roller and the second, related to daily cholesterol consumption.
Keywords:
truncation; half-normal distribution; likelihood

## 1. Introduction

The half-normal (HN) distribution is a very important model in the statistical literature. Its density function has a closed-form and its cumulative distribution function (cdf) depends on the cdf of the standard normal model (or the error function), which is implemented in practically all mathematical and statistical software. Pewsey [1,2] provides the maximum likelihood (ML) estimation for the general location-scale HN distribution and its asymptotic properties. Wiper et al. [3] and Khan and Islam [4] perform analysis and applications for the HN model from a Bayesian framework. Moral et al. [5] also present the hnp R package, which produces half-normal plots with simulated envelopes using different diagnostics from a range of different fitted models. The HN model is also presented in the stochastic representation of the skew-normal distribution in Azzalini [6,7] and Henze [8]. In recent years this distribution has been used to model positive data, and it is becoming an important model in reliability theory despite the fact that it accommodates only decreasing hazard rates. Some of the generalizations of this distribution can be found in Cooray and Ananda [9], Olmos et al. [10], Cordeiro et al. [11], Gómez and Bolfarine [12], among others.
In particular, we focused on the extension of Cooray and Ananda [9]. The authors provided a motivation related to static fatigue life to consider the transformation $Z = σ Y 1 / α$, where $Y ∼ H N ( 1 )$. This model was named the generalized half-normal (GHN) distribution. An alternative way to extend the HN model was introduced by Gómez et al. [13] considering a normal distribution with mean and standard deviation $μ$ and $σ$, respectively, truncated to the interval $( 0 , + ∞ )$ and considering the reparametrization $λ = μ / σ$. This model was named the truncated positive normal (TPN) distribution with density function given by
$f ( z ; σ , λ ) = 1 σ Φ ( λ ) ϕ z σ − λ , z , σ > 0 , λ ∈ R ,$
where $ϕ ( · )$ and $Φ ( · )$ denote the density and cdf of the standard normal models, respectively. We use $TPN ( σ , λ )$ to refer to a random variable (r.v.) with density function as in Equation (1). Note that $TPN ( σ , 0 ) ≡ H N ( σ )$.
In this work we consider a similar idea to that used in Cooray and Ananda [9] to extend the TPN model including the transformation $Y = σ X 1 / α$, where $X ∼ T P N ( 1 , λ )$. We will refer to this distribution as the generalized truncation positive normal (GTPN).
The rest of the manuscript is organized as follows. Section 2 is devoted to study of some important properties of the model, as well as its moments, quantile and hazard functions and its entropy. In Section 3 we perform an inference and present the Fisher information matrix for the proposed model. Section 4 discusses the selection model in nested and non-nested models for the GTPN distribution. In Section 5 we carry out a simulation study in order to study properties of the ML estimators in finite samples for the proposed distribution. Section 6 presents two applications to real data-sets to illustrate that the proposed model is competitive versus other common models for positive data in the literature. Finally, in Section 7, we present some concluding remarks.

## 2. Model Properties

In this section we introduce the main properties of the GTPN model such as density, quantile and hazard functions, moments, among others.

#### 2.1. Stochastic Representation and Particular Cases

As mentioned previously, we say that a r.v. Z has $GTPN ( σ , λ , α )$ distribution if $Z = σ Y 1 / α$, where $Y ∼ T P N ( 1 , λ )$. By construction, the following models are particular cases for the GTPN distribution:
• $GTPN ( σ , λ , α = 1 ) ≡ TPN ( σ , λ )$.
• $GTPN ( σ , λ = 0 , α ) ≡ GHN ( σ , α )$.
• $GTPN ( σ , λ = 0 , α = 1 ) ≡ HN ( σ )$.
Figure 1 summarizes the relationships among the GTPN and its particular cases. We highlight that $λ = 0$ and $α = 1$ are within the parametric space (not on the boundary). Therefore, to decide between the GTPN versus the TPN, GHN, or HN distributions we can use classical hypothesis tests such as the likelihood ratio test (LRT), score test (ST), or gradient test (GT).

#### 2.2. Density, Cdf and Hazard Functions

Proposition 1.
For$Z ∼ G T P N ( σ , λ , α )$, the density function is given by
$f ( z ; σ , λ , α ) = α σ α Φ ( λ ) z α − 1 ϕ z σ α − λ , z ≥ 0 .$
where$σ , α > 0$and$λ ∈ R$.
Proof.
Considering the stochastic representation discussed in Section 2.1, we have that $z = g ( y ) = σ y 1 / α$ and
$f Z ( z ) = f Y ( g − 1 ( z ) ) d g − 1 ( z ) d z = 1 Φ ( λ ) ϕ ( g − 1 ( z ) − λ ) α z α − 1 σ α .$
Replacing $g − 1 ( z ) = ( z / σ ) α$ the result is obtained. □
Proposition 2.
For$Z ∼ G T P N ( σ , λ , α )$, the cdf and hazard function are given by
$F ( z ; σ , λ , α ) = Φ ( ( z σ ) α − λ ) + Φ ( λ ) − 1 Φ ( λ ) a n d h ( z ) = f ( z ) 1 − F ( z ) = α z α − 1 ϕ z σ α − λ σ α 1 − Φ z σ α − λ ,$
respectively, for all$z ≥ 0$.
Figure 2 shows the density and hazard functions for the $GTPN ( σ = 1 , λ , α )$ model, considering some combinations for $( λ , α )$. Note that the GTPN model can assume decreasing and unimodal shapes for the density function and decreasing or increasing shapes for the hazard function.

#### 2.3. Mode

Proposition 3.
The mode of the$G T P N ( σ , λ , α )$model is attained:
1.
at$z = σ 2 1 / α λ + λ 2 + 4 1 − 1 α 1 / α$whenever$α ≥ 1$or$0 < α < 1$and$λ > 0$,
2.
at$z = 0$in otherwise.
Proof.
Let $l = log ( f )$, where f is the density function defined in (2), of a direct computation we have
$∂ l ∂ z = l z = − 1 z α z σ 2 α − α λ z σ α − ( α − 1 ) .$
Note that $l z$ vanishes when $α z σ 2 α − α λ z σ α − ( α − 1 ) = 0$. Using the auxiliary variable $w = z σ α$ the last equation is rewritten as follows:
$α w 2 − α λ w − ( α − 1 ) = 0 .$
In the rest of the proof we use the discriminant of the quadratic equation in Equation (3), which is given by $Δ = α 2 λ 2 + 4 α ( α − 1 ) = α 2 [ λ 2 + 4 ( 1 − 1 α ) ]$ and its zeros are given by: $w = ( λ ± λ 2 + 4 ( 1 − 1 α ) ) / 2$.
If $α ≥ 1$ then $Δ ≥ λ 2$. In consequence, the mode is attained at $z 1 = σ 2 1 / α λ + λ 2 + 4 1 − 1 α 1 / α$.
If $0 < α < 1$ then $Δ < λ 2$. Here, two cases may occur. The first when $0 < Δ < λ 2$, in which case if $λ > 0$ its mode is attained at $z = σ 2 1 / α λ + λ 2 + 4 1 − 1 α 1 / α$, since $l z z < 0$, and if $λ < 0$, then the zeros of Equation (3) are negative, implying that function l is strictly decreasing. Its mode is therefore attained at zero. The other case is when $Δ < 0$, then we have that $α w 2 − α λ w − ( α − 1 ) > 0$ for all $w ≥ 0$, implying that $l z < 0$ for all $z ≥ 0$. Therefore, l is strictly decreasing and thus its mode is zero. □
Remark 1.
Note that$α ≥ 1$or$λ > 0$implies that the mode of the GTPN model is attached in a positive value.

#### 2.4. Quantiles

Proposition 4.
The quantile function for the$G T P N ( σ , λ , α )$is given by
$Q ( p ) = σ Φ − 1 ( 1 − ( 1 − p ) Φ ( λ ) ) + λ 1 α .$
Proof.
Follows from a direct computation, applying the definition of quantile function. □
Corollary 1.
The quartiles of the GTPN distribution are
1.
First quartile$σ [ Φ − 1 ( 1 − 3 4 Φ ( λ ) ) + λ ] 1 α$.
2.
Median$σ [ Φ − 1 ( 1 − Φ ( λ ) 2 ) + λ ] 1 α$.
3.
Third quartile$σ [ Φ − 1 ( 1 − 1 4 Φ ( λ ) ) + λ ] 1 α$.

#### 2.5. Central Moments

Proposition 5.
Let$Z ∼ G T P N ( σ , λ , α )$and$r = 1 , 2 , …$. The r-th non-central moment is given by
$μ r = E ( Z r ) = σ λ 1 / α r b r ( λ , α ) 2 Φ ( λ ) π ,$
where$b r ( λ , α ) = ∑ k = 0 ∞ r / α k λ 2 − k Γ k + 1 2 , λ 2 2$and$r / α k = 1 k ! ∏ n = 1 k − 1 ( r / α − n )$is the generalized binomial coefficient. When$r / α ∈ N$, the sum in$b r ( λ , α )$stops at$r / α$.
Proof.
Considering the stochastic representation of the GTPN model in Section 2.1, it is immediate that $E ( Z r ) = σ r E ( Y r α )$, where $Y ∼ T P N ( 1 , λ )$. This expected value can be computed using Proposition 2.2 in Gómez et al. [13]. □
Remark 2.
When$r / α ∈ N$, Closed Forms Can Be Obtained for$μ r$
Figure 3 illustrates the mean, variance, skewness, and kurtosis coefficients for the $GTPN ( σ = 1 , λ , α )$ model for some combinations of its parameters.

#### 2.6. Bonferroni and Lorenz Curves

In this subsection we present the Bonferroni and Lorenz curves (see Bonferroni [14]). These curves have applications not only in economics to study income and poverty, but also in medicine, reliability, etc. The Bonferroni curve is defined as
$B ( p ) = 1 p μ ∫ 0 q z f ( z ) d z , 0 ≤ p < 1 .$
where $μ = E ( Z )$, $q = F − 1 ( p )$. The Lorenz curve is obtained by the relation $L ( p ) = p B ( p )$. Particularly, it can be checked that for the GTPN model the Bonferroni curve is given by
$B ( p ) = 1 p μ E ( Z ) − σ Φ ( λ ) 2 π ∑ k = 0 ∞ 1 α k λ 1 α − k 2 k − 1 2 Γ ( k + 1 2 , ( q σ ) α − λ 2 2 ) .$
These curves serve as graphic methods for analysis and comparison, e.g., the inequality of non-negative distributions. See, for example, for a more detailed discussion [15].
Figure 4 shows the Bonferroni curve for the $GTPN ( σ = 1 , λ , α )$ model, considering different values for $λ$ and $α$.

#### 2.7. Shannon Entropy

Shannon entropy (see Shannon [16]) measures the amount of uncertainty for a random variable. It is defined as:
$S ( Z ) = − E log f ( Z ) .$
Therefore, it can be checked that the Shannon entropy for the GTPN model is
$S ( Z ) = log 2 π σ Φ ( λ ) α − ( α − 1 ) E log ( Z ) + μ 2 α 2 σ 2 α − λ μ α σ α + λ 2 2 ,$
Figure 5 shows the entropy curve for the $GTPN ( σ = 1.5 , λ , α )$ model, considering different values for $λ$ and $α$. We note that this function is increasing in $λ$ and $α$. where $E log ( Z ) = ∫ 0 ∞ log ( z ) f ( z ; σ , λ , α ) d z$. For $α = 1$, the Shannon entropy is reduced to
$S ( Z ) = log 2 π σ Φ ( λ ) + μ 2 2 σ 2 − λ μ σ + λ 2 2 ,$
which corresponds to the Shannon entropy for the TPN model; and for $α = 1$ and $λ = 0$, $S ( Z )$ is reduced to
$S ( Z ) = log π σ + μ 2 2 σ 2 ,$
which corresponds to the Shannon entropy for the HN distribution.

## 3. Inference

In this section we discuss the ML method for parameter estimation in the GTPN model.

#### 3.1. Maximum Likelihood Estimators

For a random sample $z 1 , … , z n$ from the $GTPN ( σ , λ , α )$ model, the log-likelihood function for $θ = ( σ , λ , α )$ is given by
$ℓ ( θ ) = n log ( α ) + ( α − 1 ) ∑ i = 1 n log ( z i ) − n α log ( σ ) − n log ( Φ ( λ ) ) − n 2 log ( 2 π ) − 1 2 ∑ i = 1 n z i σ α − λ 2 .$
Therefore, the score assumes the form $S ( θ ) = ( S σ ( θ ) , S λ ( θ ) , S α ( θ ) )$, where
$S σ ( θ ) = − n α σ + ∑ i = 1 n z i σ α − λ α z i σ α − 1 z i σ 2 , S λ ( θ ) = − n ξ ( λ ) + ∑ i = 1 n z i σ α − λ and S α ( θ ) = n α + ∑ i = 1 n log z i − n log ( σ ) − ∑ i = 1 n z i σ α − λ z i σ α log z i σ ,$
where $ξ ( λ ) = ϕ ( λ ) Φ ( λ )$ is the negative of the inverse Mills ratio. The ML estimators are obtained by solving the equation $S ( θ ) = 0 3$, where $0 p$ denotes a vector of zeros with dimension p. This equation has the following solution for $λ$
$λ ^ ( σ ^ , α ^ ) = ∑ i = 1 n z i σ ^ 2 α ^ − n ∑ i = 1 n z i σ ^ α ^ .$
Replacing Equation (4) in $S λ ( θ ) = 0$ and $S α ( θ ) = 0$, the problem is reduced to two equations. The solution of this problem needs to be solved by numerical methods such as Newton-Raphson. Below we discuss initial values for the vector $θ$ to initialize the algorithm.

#### 3.2. Initial Point to Obtain the Maximum Likelihood Estimators

In this subsection, we discuss the initial points for the iterative methods to find the ML estimators in the GTPN distribution.

#### 3.2.1. A Naive Point Based on the HN Model

In Section 2 we discuss that $GTPN ( σ , λ = 0 , α = 1 ) ≡ HN ( σ )$. Based on this fact, and considering that the ML estimator for $σ$ in the HN distribution has a closed-form, we can consider as an initial point $θ n a i v e = 1 n ∑ i = 1 n z i 2 , 0 , 1$.

#### 3.2.2. An Initial Point Based on Centiles

Let $q t$, $t = 1 , … , 99$, the t-th sample centile based on $z 1 , … , z n$. An initial point to $θ$ can be obtained by matching $q u , q 50$ and $q 100 − u$, with $u ∈ { 1 , 2 , … , 48 , 49 }$, with their respective theoretical counterparts. Defining $p = u / 100$, the equations obtained are
$q u = σ Φ − 1 ( 1 − ( 1 − p ) Φ ( λ ) ) + λ 1 α , q 50 = σ Φ − 1 ( 1 − 0.5 Φ ( λ ) ) + λ 1 α and q 100 − u = σ Φ − 1 ( 1 − p Φ ( λ ) ) + λ 1 α .$
The solutions for $σ$ and $α$ are
$σ ˜ = σ ˜ ( λ ˜ ) = q 100 − u Φ − 1 1 − p Φ ( λ ˜ ) + λ ˜ 1 / α ˜ ( λ ˜ ) and α ˜ = α ˜ ( λ ˜ ) = log Φ − 1 1 − ( 1 − p ) Φ ( λ ˜ ) + λ ˜ Φ − 1 1 − p Φ ( λ ˜ ) + λ ˜ log q u q 100 − u ,$
where $λ ˜$ is obtained from the non-linear equation
$σ ˜ ( λ ˜ ) Φ − 1 1 − 0.5 Φ ( λ ˜ ) + λ ˜ 1 / α ˜ ( λ ˜ ) = q 50 .$
Therefore, the initial point based on this method is given by $θ c e n t = σ ˜ , λ ˜ , α ˜$.

#### 3.3. An Initial Point Based on the Method of Moments

A more robust initial point can be obtained using the method of moments. The equations to solve are $μ r = z r ¯$, $r = 1 , 2 , 3$. The solution for $σ$ is
$σ ˜ ⋆ = σ ˜ ⋆ ( λ ˜ ⋆ , α ˜ ⋆ ) = 2 π z ¯ Φ ( λ ˜ ⋆ ) λ ˜ ⋆ 1 / α ˜ ⋆ b 1 ( λ ˜ ⋆ , α ˜ ⋆ ) .$
The solution for $λ$ and $α$ (say $λ ˜ ⋆$ and $α ˜ ⋆$, respectively) are obtained from the non-linear equations
$( σ ˜ ⋆ ( λ , α ) λ 1 / α ) 2 b 2 ( α , λ ) 2 Φ ( λ ) π = z 2 ¯ and ( σ ˜ ⋆ ( λ , α ) λ 1 / α ) 3 b 3 ( α , λ ) 2 Φ ( λ ) π = z 3 ¯ .$
Therefore, the initial point based on this method is given by $θ m o m = σ ˜ ⋆ , λ ˜ ⋆ , α ˜ ⋆$.

#### 3.4. Fisher Information Matrix

The Fisher information (FI) matrix for the GTPN distribution is given by $I F ( θ ) = I a b a , b ∈ { σ , λ , α }$. Consider the notation
$T j k = T j k ( λ ) = ∫ 0 ∞ [ log ( w ) ] k w j ϕ ( w − λ ) d w , j ∈ { 1 , 2 } , k ∈ { 0 , 1 , 2 } .$
Therefore,
$I σ σ = − α σ 2 + α ( 2 α + 1 ) σ 2 Φ ( λ ) T 20 − λ α ( α + 1 ) σ 2 Φ ( λ ) T 10 , I σ λ = α σ Φ ( λ ) T 10 , I σ α = 1 σ + 1 σ Φ ( λ ) [ λ ( T 10 + T 11 ) − T 20 − 2 T 21 ] , I λ λ = 1 − λ ξ ( λ ) − ξ 2 ( λ ) , I λ α = − 1 α Φ ( λ ) T 11 , I α α = 1 α 2 + 2 α 2 Φ ( λ ) T 22 − λ α 2 Φ ( λ ) T 12$
We observe that $T j k ( λ )$ is a continuous function and $lim | λ | → + ∞ T j k ( λ ) = 0$, then $T j k ( λ ) < + ∞$. Note that for $λ = 0$ and $α = 1$ this matrix is reduced to
$F I ( σ , 0 , 1 ) = 2 σ 2 1 σ 2 π − 4 σ T ^ 21 · 1 − 2 π − 2 T ^ 11 · · 1 + 4 T ^ 22 ,$
where $T ^ j k = T j k ( 0 )$. Additionally, we note that
$det ( F I ( σ , 0 , 1 ) ) = 2 σ 2 ( π − 3 ) π ( 1 + 4 T ^ 22 + 4 T ^ 21 ) − 4 ( T ^ 11 − 2 π T ^ 21 ) 2 ≈ 0.1021506 σ 2 > 0 , ∀ σ > 0 ,$
where $det ( · )$ denotes the determinant operator. Therefore, the FI matrix for the reduced model (HN) is invertible.

## 4. Model Discrimination

In this section we discuss some techniques to discriminate among the GTPN distribution and other models.

#### 4.1. GTPN Versus Submodels

An interesting problem to solve is the discrimination between GTPN and the three submodels represented in Figure 1. In other words, we are interested in testing the following hypotheses:
• $H 0 ( 1 ) : α = 1$ versus $H 1 ( 1 ) : α ≠ 1$ (TPN versus GTPN distribution).
• $H 0 ( 2 ) : λ = 0$ versus $H 1 ( 2 ) : λ ≠ 0$ (GHN versus GTPN distribution).
• $H 0 ( 3 ) : ( α , λ ) = ( 1 , 0 )$ versus $H 1 ( 3 ) : ( α , λ ) ≠ ( 1 , 0 )$ (HN versus GTPN distribution).
The three hypotheses can be tested considering the LRT, ST, and GT. Below we present the statistics for the three tests considered and for the three hypotheses of interest.

#### 4.1.1. Likelihood Ratio Test

The statistic for the LRT (say SLR) to tests $H 0 ( j )$, $j = 1 , 2 , 3$, is defined as
$SLR j = 2 ℓ σ ^ , λ ^ , α ^ − ℓ σ ^ 0 j , λ ^ 0 j , α ^ 0 j ,$
where $σ ^ 0 j , λ ^ 0 j$ and $α ^ 0 j$ denote the ML estimators for $σ$, $λ$ and $α$ restricted to $H 0 ( j )$, $j = 1 , 2 , 3$. Under $H 0 ( j )$, $j = 1 , 2$, $SLR j ∼ χ ( 1 ) 2$ and under $H 0 ( 3 )$, $SLR 3 ∼ χ ( 2 ) 2$, where $χ ( p ) 2$ denotes the Chi-squared distribution with p degrees of freedom. For $H 0 ( 3 )$, we obtain
$σ ^ 03 = 1 n ∑ i = 1 n z i 2 , λ ^ 03 = 0 and α ^ 03 = 1 ,$
whereas to test $H 0 ( 1 )$ and $H 0 ( 2 )$ the ML estimators under the null hypotheses need to be computed numerically. However, in both cases the problem is reduced to a unidimensional maximization. For details see Cooray and Ananda [9] and Gómez et al. [13], respectively.

#### 4.1.2. Score Test

The statistic for the ST (say SR) to test $H 0 ( j )$, $j = 1 , 2 , 3$, is defined as
$SR j = S θ ^ 0 j ⊤ I F θ ^ 0 j − 1 S θ ^ 0 j ,$
where $θ ^ j$ is the ML estimator under $H 0 ( j )$. Under $H 0 ( j )$, $j = 1 , 2$, $SR j ∼ χ ( 1 ) 2$ and under $H 0 ( 3 )$, $SR 3 ∼ χ ( 2 ) 2$.

The statistic for the GT (say ST) to tests $H 0 ( j )$, $j = 1 , 2 , 3$, is defined as
$ST j = S ( θ ^ j ) θ ^ − θ ^ j .$
Again, under $H 0 ( j )$, $j = 1 , 2$, $ST j ∼ χ ( 1 ) 2$ and under $H 0 ( 3 )$, $ST 3 ∼ χ ( 2 ) 2$. After some algebraic manipulations, we obtain that
$ST 1 = α ^ − 1 n 1 − log σ ^ 01 + ∑ i = 1 n log z i − z i σ ^ 01 − λ ^ 01 z i σ ^ 01 log z i σ ^ 01 . ST 2 = λ ^ − n 2 π + ∑ i = 1 n z i σ ^ 02 α ^ 02 . ST 3 = α ^ − 1 n 1 − log σ ^ 03 + ∑ i = 1 n log z i − z i σ ^ 03 2 log z i σ ^ 03 + λ ^ − n 2 π + ∑ i = 1 n z i σ ^ 03 .$

#### 4.2. Non-Nested Models

The comparison of non-nested models can be performed based on the AIC criteria (Akaike [17]), where the model with a lower AIC is preferred. However, in practice we can have a set of inappropriate models for a certain data set. For this reason, we also need to perform a goodness-of-fit validation. This can be performed, for instance, based on the quantile residuals (QR). For more details see Dunn and Smyth [18]. These residuals are defined as
$Q R i = G ( z i ; ψ ^ ) , i = 1 , … , n ,$
where $G ( ; ψ ^ )$ is the cdf of the specified distribution evaluated in the estimator for $ψ$. If the model is correctly specified, such residuals are a random sample from the standard normal distribution. This can be assessed using, for instance, the Anderson–Darling (AD), Cramer-Von-Mises (CVM) and Shapiro–Wilks (SW) tests. A discussion of these tests can be seen in Yazici and Yocalan [19].

## 5. Simulation

In this section we present a Monte Carlo (MC) simulation study in order to illustrate the behavior of the ML estimators. We consider three sample sizes: $n = 50 , 150$ and 300; two values for $σ$:1 and 2; two values for $λ$:3 and 4; and three values for $α$: $0.8 , 1$ and 2. For each combination of n, $σ$, $λ$ and $α$, we draw 10,000 samples of size n from the $GTPN ( σ , λ , α )$ model. To simulate a value from this distribution, we consider the following scheme:
• Simulate $X ∼ U n i f o r m ( 0 , 1 )$.
• Compute $Y = Φ − 1 ( 1 + ( X + 1 ) Φ ( λ ) ) + λ$.
• Compute $Z = σ Y 1 α$.
For each sample generated, ML estimators were computed numerically using the Newton-Raphson algorithm. Table 1 presents means and standard deviations for each parameter in each case. Notice that bias and standard deviations are reduced as the sample size increases, suggesting that the ML estimators are consistent.

## 6. Applications

In this section we present two real data applications to illustrate the better performance of the GTPN model over other well known models in the literature. For these comparisons we also consider the Weibull (WEI) and the Generalized Lindley (GL, Zakerzadeh [20]) models. The density function of the Weibull distribution is given by
$f ( x ; σ , λ ) = σ λ x σ − 1 e − x σ / λ ,$
with $x > 0$, $σ > 0$ and $λ > 0$, whereas for the GL model is given by
$f ( x ; α , θ , γ ) = θ 2 ( θ x ) α − 1 ( α + γ x ) e − θ x ( γ + θ ) Γ ( α + 1 ) ,$
with $x > 0$, $θ > 0$ $α > 0$ and $γ > 0$.

#### 6.1. Application 1

The data set was taken from Laslett [21], and consisted of $n = 115$ heights measured at 1 micron intervals along the drum of a roller (i.e., parallel to the axis of the roller). This was part of an extensive study of the surface roughness of the rollers. A statistical summary of the data set is presented in Table 2.
Initially, we calculate the estimators based on the centiles, naive, and moments of the GTPN distribution, which are $θ c e n t = ( 2.326 , 2.741 , 2.376 )$, $θ n a i v e = ( 3.557 , 0 , 1 )$, and $θ m o m = ( 3.075 , 1.572 , 3.203 )$. We used these estimations as initial values in computing the ML estimators for the GTPN model. Results are presented in Table 3. Note the high estimated standard error for the $γ$ parameter in the GL model. In addition, note that, based on the AIC criteria and BIC criteria [22], the GTPN model is preferred (among the fitted models) for this data set. Figure 6 shows the estimated density for each model in this data set, where the GTPN model appears to provide a better fit. Finally, Figure 7 also presents the qq-plot for the QR in the same models and the $p − v a l u e s$ for the three normality tests discussed in Section 4.2. Results suggest that the GTPN model is an appropriate model for this data set while the rest of models are not.

#### 6.2. Application 2

The data set to be investigated was taken from Nierenberg et al. [23], and is related to a study on plasma retinol and betacarotene levels from a sample of $n = 315$ subjects. More specifically, the response variable observed is grams of cholesterol consumed per day. Descriptive statistics for the variable are provided in Table 4.
Initially, we calculate the estimators based on the centiles, naive, and moments of the GTPN distribution, which are $θ c e n t = ( 2.326 , 2.741 , 2.376 )$, $θ n a i v e = ( 275.960 , 0 , 1 )$, and $θ m o m = ( 0.003 , − 86.024 , 1.854 )$. We used these estimations as initial values in computing the ML estimators for the GTPN model. Table 5 summarizes the fit for this data set. As in last application, we noted the high estimated standard error for the $γ$ parameter in the GL model. Again, based on the AIC criteria and BIC criteria, the preferred model is the GTPN. Figure 8 shows the estimated density for each model in the cholesterol data set. The GTPN model appears to provide a better fit. For this data set, we also present the hypothesis tests for the three particular models of the GTPN distribution discussed in Section 4.1. Specifically, for $H 0 ( 1 ) : α = 1$ versus $H 1 ( 1 ) : α ≠ 1$ we obtained $p − v a l u e s < 0.0001$, $0.0157$, and $< 0.0001$ for the SLR, SR, and ST tests, respectively. Therefore, with a significance of 5% we preferred the GTPN over the TPN model. For $H 0 ( 2 ) : λ = 0$ versus $H 1 ( 2 ) : λ ≠ 0$ we obtained $p − v a l u e s < 0.0001$, $0.0008$ and $< 0.0001$ for the SLR, SR and ST tests, respectively. Hence, with a significance of 5% we preferred the GTPN over the GHN model. Finally, $H 0 ( 3 ) : ( α , λ ) = ( 1 , 0 )$ versus $H 1 ( 3 ) : ( α , λ ) ≠ ( 1 , 0 )$ we obtained $p − v a l u e s < 0.0001$ for the three tests. Therefore, with a significance of 5% we preferred the GTPN over the HN model. Finally, Figure 9 also presents the qq-plot for the QR in the fitted models and the $p − v a l u e s$ for the three normality tests. Results suggest that the GTPN model is appropriate for this data set while the rest of the models are not.

## 7. Conclusions

In this work we introduce a new distribution for positive data named the GTPN model. This new distribution includes as particular cases three models well-known in the literature: the TPN, GHN, and HN models. The basic properties of the model and ML estimation were studied. We performed a simulation study in finite samples and two real data applications, showing the good performance of the model compared with other usual models in the literature.

## Author Contributions

All of the authors contributed significantly to this research article.

## Funding

The research of H.J.G. was supported by Proyecto de Investigación de Facultad de Ingeniería. Universidad Católica de Temuco. UCT-FDI012019. The research of O.V. was supported by Vicerrectoría de Investigación y Postgrado de la Universidad Católica de Temuco, Projecto FEQUIP 2019-INRN-03.

## Conflicts of Interest

The authors declare no conflict of interest.

## References

1. Pewsey, A. Large-sample inference for the general half-normal distribution. Commun. Stat. Theory Methods 2002, 31, 1045–1054. [Google Scholar] [CrossRef]
2. Pewsey, A. Improved likelihood based inference for the general half-normal distribution. Commun. Stat. Theory Methods 2004, 33, 197–204. [Google Scholar] [CrossRef]
3. Wiper, M.P.; Girón, F.J.; Pewsey, A. Objective Bayesian inference for the half-normal and half-t distributions. Commun. Stat. Theory Methods 2008, 37, 3165–3185. [Google Scholar] [CrossRef]
4. Khan, M.A.; Islam, H.N. Bayesian analysis of system availability with half-normal life time. Math. Sci. 2012, 9, 203–209. [Google Scholar] [CrossRef]
5. Moral, R.A.; Hinde, J.; Demétrio, C.G.B. Half-normal plots and overdispersed models in R: The hnp package. J. Stat. Softw. 2017, 81, 1–23. [Google Scholar] [CrossRef]
6. Azzalini, A. A class of distributions which includes the normal ones. Scand. J. Stat. 1985, 12, 171–178. [Google Scholar]
7. Azzalini, A. Further results on a class of distributions which includes the normal ones. Statistics 1986, 12, 199–208. [Google Scholar]
8. Henze, N. A probabilistic representation of the skew-normal distribution. Scand. J. Stat. 1986, 13, 271–275. [Google Scholar]
9. Cooray, K.; Ananda, M.M.A. A generalization of the half-normal distribution with applications to lifetime data. Commun. Stat. Theory Methods 2008, 10, 195–224. [Google Scholar] [CrossRef]
10. Olmos, N.M.; Varela, H.; Gómez, H.W.; Bolfarine, H. An extension of the half-normal distribution. Stat. Pap. 2012, 53, 875–886. [Google Scholar] [CrossRef]
11. Cordeiro, G.M.; Pescim, R.R.; Ortega, E.M.M. The Kumaraswamy generalized half-normal distribution for skewed positive data. J. Data Sci. 2012, 10, 195–224. [Google Scholar]
12. Gómez, Y.M.; Bolfarine, H. Likelihood-based inference for power half-normal distribution. J. Stat. Theory Appl. 2015, 14, 383–398. [Google Scholar] [CrossRef]
13. Gómez, H.J.; Olmos, N.M.; Varela, H.; Bolfarine, H. Inference for a truncated positive normal distribution. Appl. Math. J. Chin. Univ. 2015, 33, 163–176. [Google Scholar] [CrossRef]
14. Bonferroni, C.E. Elementi di Statistica Generale; Libreria Seber: Firenze, Italy, 1930. [Google Scholar]
15. Arcagnia, A.; Porrob, F. The Graphical Representation of Inequality. Rev. Colomb. Estad. 2014, 37, 419–436. [Google Scholar] [CrossRef]
16. Shannon, C.E. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
17. Akaike, H. A new look at the statistical model identification. IEEE Trans. Auto. Contr. 1974, 19, 716–723. [Google Scholar] [CrossRef]
18. Dunn, P.K.; Smyth, G.K. Randomized Quantile Residuals. J. Comput. Graph. Stat. 1996, 5, 236–244. [Google Scholar]
19. Yazici, B.; Yocalan, S. A comparison of various tests of normality. J. Stat. Comput. Simul. 2007, 77, 175–183. [Google Scholar] [CrossRef]
20. Zakerzadeh, H.; Dolati, A. Generalized Lindley Distribution. J. Math. Ext. 2009, 3, 1–17. [Google Scholar]
21. Laslett, G.M. Kriging and Splines: An Empirical Comparison of Their Predictive Performance in Some Applications. J. Am. Stat. Assoc. 1994, 89, 391–400. [Google Scholar] [CrossRef]
22. Schwarz, G. Estimating the dimension of a model. Ann. Stat. 1978, 6, 461–464. [Google Scholar] [CrossRef]
23. Nierenberg, D.W.; Stukel, T.A.; Baron, J.A.; Dain, B.J.; Greenberg, E.R. Determinants of plasma levels of beta-carotene and retinol. Am. J. Epidemiol. 1989, 130, 511–521. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Particular cases for the GTPN distribution.
Figure 1. Particular cases for the GTPN distribution.
Figure 2. Density and hazard functions for the $GTPN ( σ = 1 , λ , α )$ model with different combinations of $λ$ and $α$.
Figure 2. Density and hazard functions for the $GTPN ( σ = 1 , λ , α )$ model with different combinations of $λ$ and $α$.
Figure 3. Plots of the (a) expectation, (b) variance, (c) skewness and (d) kurtosis for $GTPN ( σ = 1 , λ , α )$ for $α ∈ { 0.75 , 1 , 1.5 }$ as a function of $λ$. In (d), the dashed line represents the kurtosis of the normal distribution.
Figure 3. Plots of the (a) expectation, (b) variance, (c) skewness and (d) kurtosis for $GTPN ( σ = 1 , λ , α )$ for $α ∈ { 0.75 , 1 , 1.5 }$ as a function of $λ$. In (d), the dashed line represents the kurtosis of the normal distribution.
Figure 4. Bonferroni curve for the generalized truncation positive normal (GTPN)$( σ , λ , α )$ model.
Figure 4. Bonferroni curve for the generalized truncation positive normal (GTPN)$( σ , λ , α )$ model.
Figure 5. Entropy for the $G T P N ( σ = 1 , α , λ )$ model.
Figure 5. Entropy for the $G T P N ( σ = 1 , α , λ )$ model.
Figure 6. Fit of the distributions for the Laslett data set.
Figure 6. Fit of the distributions for the Laslett data set.
Figure 7. QR for the fitted models in the Laslett data set. The $p − v a l u e s$ for the Anderson–Darling (AD), Cramer-Von-Mises (CVM) and Shapiro–Wilks (SW) normality tests are also presented to check if the RQ came from the standard normal distribution.
Figure 7. QR for the fitted models in the Laslett data set. The $p − v a l u e s$ for the Anderson–Darling (AD), Cramer-Von-Mises (CVM) and Shapiro–Wilks (SW) normality tests are also presented to check if the RQ came from the standard normal distribution.
Figure 8. Fit of the distributions for the cholesterol data set.
Figure 8. Fit of the distributions for the cholesterol data set.
Figure 9. RQ for the fitted models in the cholesterol data set. The $p − v a l u e s$ for the AD, CVM, and SW normality tests are also presented to check if the QR came from the standard normal distribution.
Figure 9. RQ for the fitted models in the cholesterol data set. The $p − v a l u e s$ for the AD, CVM, and SW normality tests are also presented to check if the QR came from the standard normal distribution.
Table 1. Monte Carlo (MC) simulation study for the maximum likelihood (ML) estimators in the $( GTPN σ , λ , α )$ model in 12 combinations of $σ$, $λ$ and $α$. The results summarize the mean and the standard deviation (sd) of the respective estimators obtained in the 10,000 replicates.
Table 1. Monte Carlo (MC) simulation study for the maximum likelihood (ML) estimators in the $( GTPN σ , λ , α )$ model in 12 combinations of $σ$, $λ$ and $α$. The results summarize the mean and the standard deviation (sd) of the respective estimators obtained in the 10,000 replicates.
n = 50n = 150n = 300n = 1000
True Valor$σ ^$$λ ^$$α ^$$σ ^$$λ ^$$α ^$$σ ^$$λ ^$$α ^$$σ ^$$λ ^$$α ^$
$σ$$λ$$α$meansdmeansdmeansdmeansdmeansdmeansdmeansdmeansdmeansdmeansdmeansdmeansd
130.81.5231.5733.0891.7970.9871.0291.2271.0293.0001.0520.8750.3501.0810.6143.01706340.8300.2021.0110.2333.0140.2670.8030.071
11.2861.1053.1181.8431.2270.7051.1110.7053.0271.0401.0830.4251.0440.4443.0200.6291.0310.2501.0060.1853.0120.2661.0040.089
21.0060.5103.2202.0362.4320.3091.0110.3093.0191.0552.1710.8501.0010.1963.0220.6142.0580.4810.9980.0933.0130.2662.0080.177
40.81.4891.7994.3912.0100.9710.8601.0870.8604.2801.1990.8170.2821.0170.5134.1640.7860.7980.1481.0030.3014.0490.4130.7990.082
11.2361.2194.4392.1821.2030.6311.0180.6314.2951.2221.0170.3430.9860.4184.1870.8260.9950.1910.9930.2304.0550.4040.9980.099
20.9620.5434.6422.4952.3600.3210.9570.3214.3231.3262.0330.6960.9740.2234.1680.8392.0000.3820.9930.1184.0450.4042.0000.198
230.83.0063.1313.1571.8760.9741.9962.4231.9963.0181.0450.8690.3402.1521.1853.0230.6240.8230.1942.0340.4723.0070.2690.8050.072
12.5122.1983.2151.9421.2071.4372.2371.4373.0281.0771.0860.4312.0940.8973.0170.6421.0330.2502.0090.3733.0150.2691.0040.089
21.9991.0393.3182.2232.4170.6312.0170.6313.0351.0962.1710.8622.0000.4053.0270.6382.0610.5041.9950.1863.0160.2672.0060.177
40.82.8803.5444.5622.2830.9701.7142.1611.7144.3261.2880.8130.2822.0241.0784.1940.8490.7960.1602.0030.5734.0530.4040.7990.079
12.4022.4344.6592.4191.1761.2832.0331.2834.3441.3461.0150.3551.9740.8354.1900.8440.9950.1911.9900.4654.0530.4080.9980.100
21.9131.1324.8022.8482.3780.6681.8980.6684.3931.4692.0240.7081.9380.4604.1980.9011.9930.3911.9840.2364.0470.4012.0000.198
Table 2. Descriptive statistics of the Laslett data set.
Table 2. Descriptive statistics of the Laslett data set.
Datasetn$X ¯$S2$b 1$$b 2$
Heights measured1153.480.52−1.246.30
Table 3. Estimation of the parameters and their standard errors (in parentheses) for the GTPN, TPN, WEI, and GL models for the data set. The AIC and BIC criteria are also included.
Table 3. Estimation of the parameters and their standard errors (in parentheses) for the GTPN, TPN, WEI, and GL models for the data set. The AIC and BIC criteria are also included.
EstimatedGTPNTPNWEIGL
$σ$2.354(0.379)0.720(0.047)5.855(0.435)-
$λ$2.512(0.495)4.842(0.333)3.744(0.062)4.093(0.539)
$α$2.227(0.401)--13.467(1.902)
$γ$---15.528(29.638)
AIC238.43254.63251.74308.67
BIC246.67260.12257.23316.90
Table 4. Descriptive statistics for the cholesterol data set.
Table 4. Descriptive statistics for the cholesterol data set.
Data Setn$X ¯$$S 2$$b 1$$b 2$
Heights measured315242.4617421.791.476.34
Table 5. Estimated parameters and their standard errors (in parentheses) for the GTPN, TPN, WEI, and GL models for the $c h o l e s t e r o l$ data set. The AIC and BIC criteria are also presented.
Table 5. Estimated parameters and their standard errors (in parentheses) for the GTPN, TPN, WEI, and GL models for the $c h o l e s t e r o l$ data set. The AIC and BIC criteria are also presented.
EstimatedGTPNTPNWEIGL
$σ$0.040(0.027)151.106(8.734)1.964(0.080)-
$λ$7.888(0.459)1.455(0.138)274.717(8.353)0.016(0.001)
$α$0.240(0.013)--2.831(0.293)
$γ$---2.067(6.720)
AIC3874.923943.253905.683879.07
BIC3886.173950.753913.183890.42