Testing Symmetry of Unknown Densities via Smoothing with the Generalized Gamma Kernels

This paper improves a kernel-smoothed test of symmetry through combining it with a new class of asymmetric kernels called the generalized gamma kernels. It is demonstrated that the improved test statistic has a normal limit under the null of symmetry and is consistent under the alternative. A test-oriented smoothing parameter selection method is also proposed to implement the test. Monte Carlo simulations indicate superior finite-sample performance of the test statistic. It is worth emphasizing that the performance is grounded on the first-order normal limit and a small number of observations, despite a nonparametric convergence rate and a sample-splitting procedure of the test.


Introduction
Symmetry and conditional symmetry play a key role in numerous fields of economics and finance.Economists' focuses are often on asymmetry of price adjustments (Bacon [1]), innovations in asset markets (Campbell and Hentschel [2]) or policy shocks (Clarida and Gertler [3]).In addition, the mean-variance analysis in finance is consistent with investors' portfolio decision making if and only if asset returns are elliptically distributed (e.g., Chamberlain [4]; Owen and Rabinovitch [5]; Appendix B in Chapter 4 of Ingersoll [6]).Moreover, conditional symmetry in the distribution of the disturbance is often a key regularity condition for regression analysis.In particular, convergence properties of adaptive estimation and robust regression estimation are typically explored under this condition.For the former, Bickel [7] and Newey [8] demonstrate that conditional symmetry of the disturbance distribution in the contexts of linear regression and moment-condition models, respectively, suffices for adaptive estimators to attain their efficiency bounds.For the latter, Carroll and Welsh [9] warn invalidity in inference based on robust regression estimation when the regression disturbance is asymmetrically distributed.Indeed, symmetry of the disturbance distribution is often a key assumption for consistency of parameter estimators in certain versions of robust regression estimation (e.g., Lee [10,11]; Zinde-Walsh [12]; Bondell and Stefanski [13]).Based on their simulation studies, Baldauf and Santos Silva [14] also argue that lack of conditional symmetry in the disturbance distribution may lead to inconsistency of parameter estimates via robust regression estimation.
In view of the importance in the existence of symmetry, a number of tests for symmetry and conditional symmetry have been proposed.The tests can be classified into kernel and non-kernel methods.Examples for the former include Fan and Gencay [15], Ahmad and Li [16], Zheng [17], Diks and Tong [18], and Fan and Ullah [19].The latter falls into the tests based on: (i) sample moments (Randles et al. [20]; Godfrey and Orme [21]; Bai and Ng [22]; Premaratne and Bera [23]); (ii) regression percentile (Newey and Powell [24]); (iii) martingale transformation (Bai and Ng [25]); (iv) empirical processes (Delgado and Escanciano [26]; Chen and Tripathi [27]); and (v) Neyman's smooth test (Fang et al. [28]).Our focus is on the test by Fernandes, Mendes and Scaillet [29] (abbreviated as "FMS" hereafter).While this test can be viewed as the kernel-smoothed one, it has a unique feature.When a probability density function ("pdf") is symmetric about zero, its shapes on positive and negative sides must be mirror images each other.Then, after estimating pdfs on positive and negative sides separately using positive and absolute values of negative observations, respectively, FMS examine whether symmetry holds through gauging closeness between two density estimates.By this nature, we call the test the split-sample symmetry test ("SSST") hereafter.One of the features of the SSST is that it relies on asymmetric kernels with support on [0, ∞) such as the gamma ("G") kernel by Chen [30].Asymmetric kernel estimators are nonnegative and boundary bias-free, and achieve the optimal convergence rate (in the mean integrated squared error sense) within the class of nonnegative kernel estimators.It is also reported (e.g., p. 597 of Gospodinov and Hirukawa [31]; p. 651 of FMS) that asymmetric kernel-based estimation and inference possess nice finite-sample properties.The split-sample approach is expected to result in efficiency loss.However, it can attain the same convergence rate as the smoothed symmetry tests using symmetric kernels do.Furthermore, unlike these tests, the SSST does not require continuity of density derivatives at the origin.
The aim of this paper is to ameliorate the SSST further through combining it with the generalized gamma ("GG") kernels, a new class of asymmetric kernels with support on [0, ∞) that have been proposed recently by Hirukawa and Sakudo [32].Our particular focus is on two special cases of the GG kernels, namely, the modified gamma ("MG") and Nakagami-m ("NM") kernels.While superior finite-sample performance of the MG kernel has been reported in the literature, the NM kernel is also anticipated to have an advantage when applied to the SSST.It is known that finite-sample performance of a kernel density estimator depends on proximity in shape between the underlying density and the kernel chosen.As shown in Section 2, the NM kernel collapses to the half-normal pdf when smoothing is made at the origin, and the shape of the density is likely to be close to those on the positive side of single-peaked symmetric distributions.We also pay particular attention to the smoothing parameter selection.While existing articles on asymmetric kernel-smoothed tests (e.g., Fernandes and Grammig [33]; FMS) simply borrow the choice method based on optimality for density estimation, we tailor the idea of test-oriented smoothing parameter selection by Kulasekera and Wang [34,35] to the SSST.
The SSST with the GG kernels plugged in preserves all appealing properties documented in FMS.First, the SSST has a normal limit under the null of symmetry and it is also consistent under the alternative.Hence, unlike the tests by Delgado and Escanciano [26] and Chen and Tripathi [27], no simulated critical values are required.Second, Monte Carlo simulations indicate superior finite-sample performance of the SSST smoothed by the GG kernels.The performance is confirmed even when the entire sample size is 50, despite a nonparametric convergence rate and a sample-splitting procedure.Remarkably, the superior performance is based simply on first-order asymptotic results, and thus the assistance of bootstrapping appears to be unnecessary, unlike most of the smoothed tests employing fixed, symmetric kernels.This result complements previous findings on asymmetric kernel-smoothed tests by Fernandes and Grammig [33] and FMS.
The remainder of this paper is organized as follows.In Section 2 a brief review of a family of the GG kernels is provided.Section 3 proposes symmetry and conditional symmetry tests based on the GG kernels.Their limiting null distributions and power properties are also explored.As an important practical problem, Section 4 discusses the smoothing parameter selection.Our particular focus is on the choice method for power optimality.Section 5 conducts Monte Carlo simulations to investigate finite-sample properties of the test statistics.Section 6 summarizes the main results of the paper.Proofs are provided in the Appendix.
This paper adopts the following notational conventions: Γ (a) = ∞ 0 y a−1 exp (−y) dy (a > 0) is the gamma function; 1 {•} signifies an indicator function; • denotes the integer part; A = {tr (A A)} 1/2 is the Euclidian norm of matrix A; and c (> 0) denotes a generic constant, the quantity of which varies from statement to statement.The expression "X Lastly, in order to describe different asymptotic properties of an asymmetric kernel estimator across positions of the design point x (> 0) relative to the smoothing parameter b (> 0) that shrinks toward zero, we denote by "interior x" and "boundary x" a design point x that satisfies x/b → ∞ and x/b → κ for some 0 < κ < ∞ as b → 0, respectively.

Family of the GG Kernels: A Brief Review
Before proceeding, we provide a concise review on a family of the GG kernels.The family constitutes a new class of asymmetric kernels, and it consists of a specific functional form and a set of common conditions, as in Definition 1 below.The name "GG kernels" comes from the fact that the pdf of a GG distribution by Stacy [36] is chosen as the functional form.A major advantage of the family is that for each asymmetric kernel generated from this class, asymptotic properties of the kernel estimators (e.g., density and regression estimators) can be delivered by manipulating the conditions directly, as with symmetric kernels.

Definition 1. (Hirukawa and Sakudo
be a continuous function of the design point x and the smoothing parameter b.For such (α, β, γ), consider the pdf of GG (α, βΓ (α/γ) /Γ {(α + 1) /γ} , γ), i.e., This pdf is said to be a family of the GG kernels if it satisfies each of the following conditions: and the connection between x and ϕ b (x) Moreover, connections of α and γ at x = C 1 b, if any, are smooth.
, for some constant The family embraces the following two special cases 1 .Putting and γ = 1 in (1) generates the MG kernel It can be found that this is equivalent to the one proposed by Chen [30] by recognizing that α = ρ b (x) on p. 473 of Chen [30] and β/α = b.The same (α, β) and γ = 2 also yields the NM kernel The GG kernels are designed to inherit all appealing properties that the MG kernel possesses.We conclude this section by referring to the properties.Two properties below are basic ones.First, by construction, the GG kernels are free of boundary bias and always generate nonnegative density estimates everywhere.Second, the shape of each GG kernel varies according to the position at which smoothing is made; in other words, the amount of smoothing changes in a locally adaptive manner.To illustrate this property, Figure 1 plots the shapes of the MG and NM kernels at four different design points (x = 0.0, 0.5, 1.0, 2.0) at which the smoothing is performed.For reference, the G kernel is also drawn in each panel.When smoothing is made at the origin (Panel (A)), the NM kernel collapses to a half-normal pdf, whereas others reduce to exponential pdfs.As the design point moves away from the boundary (Panels (B-D)), the shape of each kernel becomes flatter and closer to symmetry.We should stress that Figure 1 is drawn with the value of the smoothing parameter fixed at b = 0.2.Unlike variable bandwidth methods for fixed, symmetric kernels (e.g., Abramson [37]), adaptive smoothing of these kernels can be achieved by a single smoothing parameter, which makes them much more appealing in empirical work.
The remaining three properties are on density estimates using the GG kernels.Third, when best implemented, each GG density estimator attains Stone's [38] optimal convergence rate in the mean integrated squared error within the class of nonnegative kernel density estimators.Fourth, the leading bias of each GG density estimator contains only the second-order derivative of the true density over the interior region, unlike many other asymmetric kernels including the G kernel.Fifth, the variance of the GG estimator tends to decrease as the design point moves away from the boundary.This property is particularly advantageous to estimating the distributions that have long tails with sparse data. 1 Hirukawa and Sakudo [32] present the Weibull kernel as yet another special case.However, it is not confirmed that this kernel satisfies Lemma 1 below, and thus the kernel is not investigated throughout.

SSST as a Special Case of Two-Sample Goodness-of-Fit Tests
This section proposes to combine the SSST with the GG kernels, explores asymptotic properties of the test statistic, and finally expands the scope of the test to testing the null of conditional symmetry.The SSST can be characterized as a special case of two-sample tests for equality of two unknown densities investigated by Anderson, Hall and Titterington [39].Suppose that we are interested in testing symmetry of the distribution of a random variable U ∈ R. Without loss of generality, we hypothesize that the distribution is symmetric about zero.If U has a pdf, then under the null, its shapes on positive and negative sides of the entire real line R must be mirror images each other.Let f and g be the pdfs to the right and left from the origin, respectively.Then, we would like to test the null hypothesis for almost all u ∈ R + against the alternative Accordingly, a natural test statistic should be built on the integrated squared error ("ISE") where F and G are cumulative distribution functions corresponding to f and g, respectively.
The name of the SSST comes from the way to construct a sample analog to , where N = n 1 + n 2 .Given the sub-samples, f and g can be estimated using a GG kernel with the smoothing parameter b as Similarly, (F, G) is replaced with their empirical measures (F n 1 , G n 2 ).In addition, because n 1 ∼ n 2 under H 0 , without loss of generality and for ease of exposition, we assume that N is even and that n : Although we could use Īn itself as the test statistic, the probability limit of I 1n plays a role in a non-vanishing center term of the asymptotic null distribution.Because the term is likely to cause size distortions in finite samples, we focus only on I n to construct the testing statistic.Now I n can be rewritten as is symmetric between Z i and Z j and that E Φ n Z i , Z j Z i = 0 almost surely under H 0 .It follows that I n is a degenerate U-statistic, and thus we may apply a martingale central limit theorem (e.g., Theorem 1 of Hall [40]; Theorem 4.7.3 of Koroljuk and Borovskich [41]).
Before describing the asymptotic properties of I n , we make two remarks.First, applying the idea of two-sample goodness-of-fit tests to the symmetry test is not new.Ahmad and Li [16] and Fan and Ullah [19] have also studied the symmetry test based on closeness of two density estimates measured by the ISE.They estimate densities using two samples, namely, the original entire sample {X i } N i=1 := {U i } N i=1 and the one obtained by flipping the sign of each observation {Y i } N i=1 := {−U i } N i=1 in our notations.Because each of X and Y has support on (−∞, ∞) by construction, a standard symmetric kernel is employed for density estimation unlike the SSST.Second, if X and Y are taken from two different distributions with support on [0, ∞), then I n can be viewed as a pure two-sample goodness-of-fit test.It can be immediately applied to the testing for equality of two unknown distributions of nonnegative economic and financial variables such as incomes, wages, short-term interest rates, and insurance claims.
To present the convergence properties of I n , we make the following assumptions.
Assumption 1. Two random samples {X i } n 1 i=1 and {Y i } n 2 i=1 are drawn independently from univariate distributions that have pdfs f and g with support on [0, ∞), respectively.Assumption 2. f and g are twice continuously differentiable on [0, ∞), and It is possible to use different asymmetric kernels and/or different smoothing parameters to estimate f and g.For convenience, however, we choose to employ the same asymmetric kernel function and a single smoothing parameter.Assumption 4. Let (X 1 , X 2 ) and (Y 1 , Y 2 ) be two independent copies of X and Y, respectively.Then, the followings hold: , where V I (2) is a kernel-specific constant given in Condition 5 of Definition 1.
Assumptions 1-3 are standard in the literature of asymmetric kernel smoothing.On the other hand, Assumption 4 has a different flavor.Convergence results on I n are built on several different moment approximations.While Definition 1 implies the statements in Lemma A2 in the Appendix, it is unclear whether the definition may even admit such approximations as in Assumption 4. The difficulty comes from the fact that unlike symmetric kernels, roles of design points and data points are nonexchangeable in asymmetric kernels.What makes the problem more complicated is that functional forms of (α, in the GG kernels are not fully specified in Definition 1. Considering that not all GG kernels may admit the moment approximations (a) and (b), we choose to make an extra assumption.Note that the MG and NM kernels fulfill Assumption 4, as documented in the next lemma.

Lemma 1. If Assumptions 1-3 hold, then each of the MG and NM kernels satisfies Assumption 4.
The theorem below delivers the convergence properties of I n and provides a consistent estimator of its asymptotic variance.

Theorem 1. Suppose that Assumptions 1-4 and n
We make a few remarks.First, it follows from Lemma 1 and Theorem 1 that the MG and NM kernels can be safely employed for the SSST, where values of V I (2) for these kernels are It also follows from Proposition 1 of FMS and Theorem 1 that limiting null distributions of nb 1/4 I n using the G and MG kernels coincide, as expected.Second, while a similar form to the asymptotic variance σ 2 can be found in Proposition 1 of FMS, σ 2 takes a more general form.Accordingly, the variance estimator σ2 is consistent under both H 0 and H 1 .Third, it can be inferred from Theorem 1 that the test statistic becomes T n := nb 1/4 I n / σ.As a consequence, the SSST is a one-sided test that rejects H 0 in favor of H 1 if T n > z α , where z α is the upper α-percentile of N (0, 1).The next proposition refers to consistency of the SSST.Observe that the power approaches one for local alternatives with convergence rates no faster than nb 1/4 , as well as for fixed alternatives.

SSST When Two Sub-Samples Have Unequal Sample Sizes
Convergence results in the previous section rely on the assumption that the sample sizes of two sub-samples {X i } n 1 i=1 and {Y i } n 2 i=1 are the same, i.e., so far n 1 = n 2 has been maintained.In reality, n 1 = n 2 is often the case, in particular, when the entire sample size N = n 1 + n 2 is odd or when H 1 is true.
Handling this case requires more tedious calculation.When n 1 = n 2 , I n can be rewritten as Following Fan and Ullah [19], we deliver convergence results under the assumption that two sample sizes n 1 and n 2 diverge at the same rate.The asymptotic variance of n 1 b 1/4 I n 1 ,n 2 and its consistent estimate are also provided.Because the essential arguments are the same as those for Theorem 1 and Proposition 1, we omit the proofs of Theorem 2 and Proposition 2 and simply state the results.Observe that when n 1 = n 2 = n, these results collapse to Theorem 1 and Proposition 1, respectively.
λ is given by The next corollary is a natural outcome from Theorem 2 and comes from the fact that under H 0 , n 1 ∼ n 2 or λ = 1 holds.Because N could be odd in this context, n should read n = N/2 .

Extension to a Test for Conditional Symmetry
So far we have maintained the assumption that the random variable U is observable and has a distribution that is symmetric about zero.However, often U is unobservable or the axis of symmetry is not zero.The former is typical when we are interested in symmetry of the distribution of the disturbance conditional on regressors in regression analysis.In this scenario, the test is conducted after U is replaced with the residual.For the latter, the test should be based on location-adjusted observations, i.e., transformed observations with an estimate of the axis of symmetry (e.g., the sample mean or the sample median) subtracted from U. These aspects motivate us to generalize the SSST to the testing for conditional symmetry.
Following FMS, we consider a testing for symmetry in the conditional distribution of the framework of a semiparametric context.Specifically, for a parameter space Θ 1 and a function ξ Observe that this is equivalent to test whether there is However, implementing this type of testing strategy requires to estimate the conditional pdf of V| V 2 nonparametrically.This is cumbersome, considering the curse of dimensionality in V 2 and another smoothing parameter choice.Instead, as in Zheng [17], Bai and Ng [25] and Delgado and Escanciano [26], we assume that there are a parameter space Θ 2 and a function ξ 2 : R d+1 × Θ 2 → R that can attain symmetry of the marginal distribution of Given the dependence of V (and thus U) on Θ 1 , we can finally rewrite our testing scheme as the one that tests, for a suitable parameter space Θ and a function ξ : R d+1 × Θ → R, symmetry of the marginal distribution of U = ξ (V 1 , V 2 ; θ 0 ) about zero for some θ 0 ∈ Θ.
Our remaining task is to demonstrate that there is no asymptotic cost in the test statistics with (ξ, θ 0 ) replaced by its estimator ξ, θ , as long as ξ, θ p → (ξ, θ 0 ) at a suitable rate of convergence.To control the convergence rate, we make Assumption 5 below.Observe that it allows for nonparametric rates of convergence; see Hansen [42], for instance, for uniform convergence rates of kernel estimators.

Smoothing Parameter Selection
How to choose the value of the smoothing parameter b is an important practical problem.Nonetheless, it appears that the issue has not been well addressed in the literature on testing problems using asymmetric kernels.While Fernandes and Grammig [33] adopt a method inspired by Silverman's [43] rule-of-thumb, FMS adjust the value chosen via cross validation.Both methods choose the smoothing parameter value from the viewpoint of optimality for density estimation.Such choices cannot be justified in theory or practice, because estimation-optimal values may not be equally optimal for testing purposes.In contrast, there are a few works on test-oriented smoothing parameter selection.For the test of equality in two unknown regression curves, Kulasekera and Wang [34,35], analytically explore the idea of choosing the smoothing parameter value that maximizes the power with the size preserved.Gao and Gijbels [44] combine this idea with the Edgeworth expansion for a bootstrap specification test of parametric regression models.
Below we tailor the procedure by Kulasekera and Wang [35] to the SSST.For a realistic setup, the case of n 1 = n 2 is exclusively considered.Their basic idea is from sub-sampling.Without loss of generality assume that {X i } n 1 i=1 and {Y i } n 2 i=1 are ordered samples.Then, the entire sample {X i } n 1 i=1 , {Y i } n 2 i=1 can be split into M sub-samples, where M = M n 1 is a non-stochastic sequence that satisfies 1/M + M/n 1 → 0 as n 1 → ∞.Given such M and (k 1 , k 2 ) := ( n 1 /M , n 2 /M ), the mth sub-sample is defined as , m = 1, . . ., M. This sub-sample yields the analogues to (3) and (4) as It follows that the test statistic using the mth sub-sample becomes Also denote the set of admissible values for b = b n 1 as for some prespecified exponent q ∈ (0, 1) and two constants 0 where c m (α) is the critical value for the size α test using the mth sub-sample.We pick the power-maximized bk 1 = Bk −q

and the smoothing parameter value bn 1 := Bn
−q 1 follows.The behavior of πM b k 1 can be examined by considering the local alternative , where c (α) is the critical value for the size α test using the entire sample.For such π (b n 1 ), Then, bn 1 is optimal in the sense of Proposition 3. The proof is omitted, because it is a minor modification of the one for Theorem 2.1 of Kulasekera and Wang [35]; indeed it can be established by recognizing that T n 1 ,n 2 d → N (I h /σ λ , 1) under H 1 , as in Proposition 3 of Fernandes and Grammig [33].
We conclude this section by stating how to obtain bn 1 in practice.Step 1 reflects that M should be divergent but smaller than both n 1 and n 2 in finite samples.Step 3 follows from the implementation methods in Kulasekera and Wang [34,35].Finally, Step 4 considers that there may be more than one maximizer of πM b k 1 .
Step 4: Set c m (α) ≡ z α and find bk 1 = inf arg max b k 1 ∈H k 1 πM b k 1 by a grid search.

Setup
It is widely recognized that asymptotic results on kernel-smoothed tests are not well transmitted to their finite-sample distributions, which reflects that omitted terms in the first-order asymptotics on the test statistics are highly sensitive to their smoothing parameter values in finite samples.On the other hand, Fernandes and Grammig [33] and FMS report superior finite-sample properties of asymmetric kernel-smoothed tests.To see which perspective dominates, this section investigates finite-sample performance of the test statistic for the SSST via Monte Carlo simulations.
To make a direct comparison with the results by FMS, we specialize in the conditional symmetry test using the same linear regression model y = β 0 + β 1 x + u as used in FMS.The data are generated in the following manner.First, the regressor x is drawn from N (0, 1).Second, the disturbance u, which is independent of x, is drawn from one of eight distributions with means of zero given in Table 1.Distributions with "S" (symmetric) and "A" (asymmetric) are used to investigate size and power properties of the test statistic, respectively.All the distributions are popularly chosen in the literature; the generalized lambda distribution ("GLD") by Ramberg and Schmeiser [45], in particular, is known to nest a wide variety of symmetric and asymmetric distributions 3 .Finally, the dependent variable y is generated by setting β 0 = β 1 = 1.Note: "S" and "A" stand for symmetric and asymmetric distributions, respectively."GLD" denotes the generalized lambda distribution by Ramberg and Schmeiser [45].The distribution is defined in terms of the inverse of the cumulative distribution function We are interested in testing symmetry of the conditional distribution of y given x.For this purpose the SSST is applied for the least-squares residual ûi := y i − β0 − β1 x i using the sample {(y i , x i )} N i=1 , where β0 , β1 are least-squares estimates of (β 0 , β 1 ).Finite-sample size and power properties of the test statistic T n 1 ,n 2 for two sub-samples with unequal sample sizes are examined against nominal 5% and 10% levels.The MG and NM kernels (denoted as "T n 1 ,n 2 -MG" and "T n 1 ,n 2 -NM", respectively) are employed as examples of the GG kernels.
Finite-sample properties of T n 1 ,n 2 -MG and T n 1 ,n 2 -NM are evaluated in comparison with other versions of the SSST.First, two versions of FMS's original test statistic built on an equivalence to our I n using the G kernel are considered."FMS-G-O" is FMS's truly original statistic, whereas "FMS-G-AltVar" is the one with the variance estimator replaced by σ2 λ given in Theorem 2. Second, T n 1 ,n 2 using the G kernel (denoted as "T n 1 ,n 2 -G") is also calculated.Notice that FMS-G-AltVar and T n 1 ,n 2 -G take exactly the same form.The only difference is the method of choosing the smoothing parameter b, which will be discussed shortly.Effects of changing the variance estimator, the method of choosing b, and the kernel choice can be examined by weighing FMS-G-O with FMS-G-AltVar, FMS-G-AltVar with T n 1 ,n 2 -G, and T n 1 ,n 2 -G with T n 1 ,n 2 -MG or T n 1 ,n 2 -NM, respectively.
The smoothing parameter b for FMS-G-O and FMS-G-AltVar is determined via making an adjustment for the value chosen by a cross-validation criterion; see p. 657 of FMS for details.On the other hand, the values of b for T n 1 ,n 2 -G, T n 1 ,n 2 -MG and T n 1 ,n 2 -NM are selected by the power-optimality criterion in the previous section.Implementation details are as follows: (i) all critical values in πM b k 1 are set at z 0.05 = 1.645; (ii) the shrinking rate of b is set at q = 4/9 because of N 1/2 -consistency of least-squares estimates and Theorem 3; (iii) three different values are considered for δ, namely, δ ∈ {0.3, 0.5, 0.7}; and (iv) the interval for b k 1 is set equal to H k 1 = [0.01, 0.64].The sample size is N ∈ {50, 100, 200}, and 1000 replications are drawn for each combination of the sample size N and the distribution of u.

3
Although the GLDs corresponding to A3 and A4 are used in Zheng [17] and FMS, they are found to have non-zero means.Therefore, we adjust the values of λ 1 and λ 2 with skewness and kurtosis maintained so that the resulting distributions have means of zero.

Simulation Results
Table 2 presents finite-sample rejection frequencies of each test statistic against nominal 5% and 10% levels across 1000 Monte Carlo samples.Critical values are simply based on the first-order normal limit, i.e., 1.645 and 1.280 correspond to the 5% and 10% levels, respectively.Panel (A) reports size properties.At first glance, we can find that the results of FMS-G-O are close to what is reported in Table 3 of FMS.It has the tendency of over-rejecting the null slightly against the nominal size.Comparing FMS-G-O with FMS-G-AltVar reveals that replacing the variance formula is likely to decrease the rejection frequencies.Changing the choice method of b further reduces the rejection frequencies, and T n 1 ,n 2 -G tends to result in mild under-rejection of the null.Effects of alternative kernel choices are mixed.While T n 1 ,n 2 -G and T n 1 ,n 2 -MG have similar size properties, T n 1 ,n 2 -NM looks more conservative in the sense that its rejection frequencies are slightly smaller.Impacts of varying δ are found to be minor at best.A concern is that all test statistics exhibit size distortions for S4.However, the distribution is platykurtic and has sharp boundaries at ±1.A platykurtic distribution is an exception rather than a rule in economics and finance, and a distribution with a compact support violates Assumption 1.In sum, all test statistics exhibit good size properties, although their convergence rates are nonparametric ones, effective sample sizes are (roughly) a half of the entire sample size N, and no size correction devices such as bootstrapping are used.
Panel (B) refers to power properties.We can immediately see that the rejection frequencies of each test statistic approach to one with the sample size N, which confirms consistency of the SSST.There is substantial improvement in power as the sample size increases from N = 50 to 100.Most of rejection frequencies become nearly one for as small as N = 200.After a closer look, we can find it hard to judge whether changing the variance formula from FMS-G-O to FMS-G-AltVar may affect power properties favorably or adversely.However, once the smoothing parameter value is chosen via the power-optimality criterion, power properties are improved in general.Power properties of T n 1 ,n 2 -G and T n 1 ,n 2 -MG again look alike, whereas T n 1 ,n 2 -NM appears to be more powerful than these two.Because the power tends to decrease with δ, it could be safe to choose δ = 0.3 from the viewpoint of power-maximization.Indeed, for N = 200 and δ = 0.3, each of T n 1 ,n 2 -G, T n 1 ,n 2 -MG and T n 1 ,n 2 -NM exhibits better power properties than FMS-G-O and FMS-G-AltVar.
For convenience, Panel (B) presents size-adjusted powers, where the best case scenario (i.e., δ = 0.3) is considered for T n 1 ,n 2 -G, T n 1 ,n 2 -MG and T n 1 ,n 2 -NM.These three test statistics again outperform FMS's original statistics in terms of size-adjusted powers, and T n 1 ,n 2 -NM appears to have the best power properties among three.All in all, Monte Carlo results indicate superior size and power properties of the SSST with the GG kernels plugged in.

Conclusions
The SSST developed by FMS is built on the idea of gauging the closeness between right and left sides of the axis of symmetry of an unknown pdf.To implement the test, we split the entire sample into two sub-samples and estimate both sides of the pdf nonparametrically using asymmetric kernels with support on [0, ∞).This paper has improved the SSST by combining it with the newly proposed GG kernels.The test statistic can be interpreted as a standardized version of a degenerate U-statistic.We deliver convergence properties of the test statistic and provide the asymptotic variance formulae for the cases of two sub-samples with equal and unequal sample sizes separately.It is demonstrated that the SSST smoothed by the GG kernels has a normal limit under the null of symmetry and is consistent under the alternative.As a part of the implementation method we also propose to select the smoothing parameter in a power-optimality criterion.Monte Carlo simulations indicate that the GG kernel-smoothed SSST with the power-maximized smoothing parameter value plugged in enjoys superior finite-sample properties.It should be stressed that the good performance of the SSST is grounded on the first-order normal limit and a small number of observations, despite its nonparametric convergence rate and sample-splitting procedure.
All the remaining statements can be shown in the same manner.To approximate the gamma function, we frequently refer to the following well-known formulae: 1. Stirling's formula ("SF"): as z → ∞.
2. Legendre's duplication formula ("LDF"): In addition, proofs of the above statements require the following lemma.Its proof is virtually the same as those for Lemmata A.1 and A.2 of Fernandes and Monteiro [46], and thus it is omitted.Lemma A1.For a constant D > 0 and two numbers x, y > 0, Proof of (A1).We apply the trimming argument as on p. 476 of Chen [30].For some ∈ (0, where L b (u; x, y) := K NM (u; x, b) K NM (u; y, b) for interior x, y.Then, the proof takes a multi-step approach including the following steps: Step 1: Step 1: Define By the property of GG random variables, (A5), (A9), and (A10), In the end, a first-order Taylor expansion of Step 2: For some t ∈ (0, 1), we split the interval for y into four subintervals as follows: Also denote h (x, y) := B b (x, y) f √ xy g (y).
Appendix A.2.1.Proof of Lemma A2 The variance approximation in Theorem 1 of Hirukawa and Sakudo [32] and the trimming argument on p.476 of Chen [30] yield (a). On the other hand, the bias approximation in Theorem 1 of Hirukawa and Sakudo [32] is applied to (b)-(d).As a consequence, (b) can be established by recognizing that E K X 2 (X 1 ) K Y 2 (Y 1 ) = E K X 2 (X 1 ) E K X 2 (X 1 ) , for instance.Moreover, (c) and (d) follow from the proofs for (d) and (f) in Lemma A1 of FMS.
With the assistance of Assumption 4 and Lemma A2, we can pick out the leading terms of E φ 2 n (Z 1 , Z 2 ) and E {φ n (Z 1 , Z 2 ) φ n (Z 2 , Z 1 )} as: The result immediately follows.
Appendix A.2.3.Proof of Lemma A4 It follows from Lemma A3 that Next, by Jensen's and C r -inequalities, Furthermore, applying C r -inequality repeatedly yields under H 0 .Essentially the same arguments as in the proofs of Lemmata 1 and A2 establish It follows from k < 3/2 that x 1−k f 3 (x) ≤ cx −1/2 in the neighborhood of the origin, and thus

Figure 1 .
Figure 1.Shapes of the GG Kernels When b = 0.2.

Table 1 .
Distributions of the Disturbance u in the Simulation Study.

Table 2 .
Size and Power of the SSST.