We start by dealing first with the extension of the ordinary normal bimodal distribution.
3.1. One-Parameter Bimodal Skew-Normal Distribution
Definition 3. If the random variable X is distributed according to the density functionthen we say that it is distributed according to the bimodal skew normal distribution with parameter λ which we denote by . Figure 1 depicts examples of the bimodal skew-normal (BSN) distribution given in Equation (
8) for different values of parameter
.
Proposition 4. Let , then .
Proof. Let
, then, for
, it follows that
Differentiating this last expression we conclude the proof by showing that
□
Remark 4. The density function , behaves like the density function of the bimodal normal model by the perturbation function . We note that the heights at the modes for a bimodal symmetric distribution are the same. However, for an asymmetric distribution the heights at the modes are not the same as shown in the next proposition.
Proposition 5. Let , . Moreover, let and () be the points at which the function reaches its the maximum value. Then,
- (a)
If then
- (b)
If then .
Proof. If (symmetric case) then with and .
Suppose now that ; it should then be required that and .
Hence, ∀
,
and, in particular,
. On the other hand, for the symmetric case, it is known that
and given that
, it follows that
The case
is proved similarly. □
Proposition 6. Let , and be the distribution functions of the random variables , respectively, and the density function of Y. Therefore,with . Proof. Let
the distribution function of the random variable
. Integrating by parts, with
and
, it follows that
and
, so that
□
Proposition 7. Let and with MGF and , respectively. Then,with and . Proof. Integrating by parts, with and , it follows that
and
so that
□
3.2. Two-Parameter Bimodal Skew-Normal Distribution
Definition 4. If the density function of a random variable X is such thatthen we say that X is distributed as the bimodal skew-normal distribution(see Elal-Oliviero et al. [21]) with parameters λ and α, which we denote by . Figure 2 depicts examples of the BSN distribution given in Equation (9) for different values of parameters λ and α. Remark 5. It is already known that changes in parameter λ lead to changes in values (heights) of the density of the model, that is, in . By incorporating the extra parameter α in the density , corresponding to the model, as α ranges in the interval the density changes from unimodal to bimodal, and vice versa, with great flexibility.
As we mentioned in the introduction, the BSN model was introduced by Elal-Olivero et al. [
21] and used for the Bayesian inference. Now we observe that we have constructed it based on a mixture of two distributions; below we study some of its properties, carry out a simulation study to see the behavior of the ML estimators and present an application to a real data set.
Proposition 8. Let , so that the following properties hold:
- (a)
If then
- (b)
If then
- (c)
If then
- (d)
If then
- (e)
Let , , and . Then,Proof. For , it follows thatwhich upon differentiation leads to: □
- (f)
Let , and . ThenProof.
We note that the density for the
model can be seen as a mixture between the
and
models. □
Proposition 9. Let , and and be the distribution function for the random variables W and Y respectively and the density function Y. Then, Proof. Let
be the distribution function of the random variable
X, where
. Then, the result follows by noticing that
□
Proposition 10. Let and , and let, and be the moment generating functions for the random variables W and Y, respectively; then,with , and Proof. The result follows by noticing that
□
Remark 6. The properties for the model presented next follow from general properties of the model presented in Azzalini [3] where is a symmetric (around zero) density function, G is a unidimensional (symmetric) distribution function such that exists and is an even function. - 1.
A stochastic representation for the model. Let and be independent random variables and defineThen . - 2.
Invariance perturbation property.
If and then are identically distributed.
Remark 7. The moments of the model can be computed easily, separating even from odd moments.
- 1.
Considering anti-perturbation invariance, the even moments of X and Z are the same, so that with .
- 2.
For the odd moments, considering that , it follows that:with . where the odd moments can be computed using derivations in Henze [2], leading to:with .
If we denote by
and
the asymmetry and kurtosis coefficients, respectively, then:
where
, with
are as given by
Representing the asymmetry for specific values of and by , then the following relation holds:
Table 2 and
Table 3 show asymmetry and kurtosis values for different values of
and
.
The parameter produces asymmetry in the symmetric model and, in particular when , the asymmetry is reflected in the change of height of the modes of the symmetric bimodal model.
The following proposition shows the identifiability of the model.
Proposition 11. The model is identifiable.
Proof. Let fixed. We will prove that if then , for all . Let us assume, for a contradiction, that , for all and fixed. Thus, after some algebraic procedures, we have that , for all . In consequence, is a contradiction of our assumption.
To prove that f is an injective function with respect to parameter , we use an analogous procedure, which concludes the proof. □
Below we show some properties involving conditional distributions and their relations with the model introduced in this paper.
Proposition 12. If and . Then .
Proof. The proof follows by noticing that
□
Corollary 1. If and , then .
Proposition 13. Let and . Then .
3.3. Location-Scale Extension
In practical scenarios, it is common to work with the location-scale transformation
, where
,
and
, with
and
. Therefore, the density function for the random variable
X, denoted as
is given by
Let us assume that
is a random sample of size
n from the distribution
. From (
10), the log-likelihood function is given as
where
, which is a continuous function of each parameter. Then, differentiating the log-likelihood function, we obtain that the elements of the score vector,
, where
, are given by
where
. Therefore, the ML estimator of
is the solution of the system
, which must be solved numerically.