Next Article in Journal
Beyond Proteostasis: Lipid Metabolism as a New Player in ER Homeostasis
Next Article in Special Issue
Amino Acid Metabolism in Apicomplexan Parasites
Previous Article in Journal
L-Carnitine and Acylcarnitines: Mitochondrial Biomarkers for Precision Medicine
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comprehensive Comparative Analysis of Local False Discovery Rate Control Methods

Department of Mathematics and Statistics, Chonnam National University, Gwangju 61186, Korea
*
Author to whom correspondence should be addressed.
Metabolites 2021, 11(1), 53; https://doi.org/10.3390/metabo11010053
Submission received: 9 December 2020 / Revised: 11 January 2021 / Accepted: 11 January 2021 / Published: 14 January 2021
(This article belongs to the Special Issue Data Science in Metabolomics)

Abstract

:
Due to the advance in technology, the type of data is getting more complicated and large-scale. To analyze such complex data, more advanced technique is required. In case of omics data from two different groups, it is interesting to find significant biomarkers between two groups while controlling error rate such as false discovery rate (FDR). Over the last few decades, a lot of methods that control local false discovery rate have been developed, ranging from one-dimensional to k-dimensional FDR procedure. For comparison study, we select three of them, which have unique and significant properties: Efron’s approach, Ploner’s approach, and Kim’s approach in chronological order. The first approach is one-dimensional approach while the other two are two-dimensional ones. Furthermore, we consider two more variants of Ploner’s approach. We compare the performance of those methods on both simulated and real data.

1. Introduction

Due to the advent of advanced high-throughput technologies, a large amount of raw data have been produced and various methods that can appropriately preprocess such data have been developed. After various preprocessing steps, statistical methods are applied to the preprocessed, yet large-scale data. When the large-scale data consists of two different groups, an interesting question is whether there are some biomarkers showing significant difference between two groups. If it is the case, then it is crucial to find significant biomarkers while controlling error rate. For example, in case of classic test for single hypothesis, it is easy not only to find biomarkers but also to control type I error, which is the probability of false rejection under the true null. However, as the number of hypotheses to be tested increases (also known as multiple testing problem), it is getting more difficult to find some biomarkers while controlling type I error at the nominal level because very small amount of error is given to test for each hypothesis depending on the number of all hypotheses. To overcome such difficulty, various approaches had been tried [1,2,3,4,5,6,7,8,9]. As a solution, the control of false discovery rate was suggested as substitute for that of type I error. Formal definition of FDR was introduced in 1995. After that, new procedures that control false discovery rate have been developed [10,11,12,13,14,15]. More specifically, in 1995, Benjamini and Hochberg (BH) firstly introduced the formal definition of FDR and proposed a step-down procedure to control FDR [10]. Especially, the procedure considered the ordered p-values and controlled global FDR on average sense. Following the seminal paper, FDR issue has attracted a lot of researchers’ attention. After a while, focus moved on to the local FDR combined with model-based approach. Many methodologies, which control local false discovery rate in the mixture framework,
f ( z ) = π 0 f 0 ( z ) + ( 1 π 0 ) f 1 ( z ) ,
have been developed where f 0 ( z ) and f 1 ( z ) are null and alternative densities, respectively [11,12,13,15]. π 0 is the proportion of true null. Furthermore, f 1 could be one-sided or two-sided. In 2001, Efron and his colleagues (called Efron hereafter) provided a new procedure that controls one-dimensional local false discovery rate,
f d r 1 d ( z ) = π 0 f 0 ( z ) / f ( z )
where z is test statistic [11]. In the empirical Bayes’ perspective, they interpret the f d r 1 d ( z ) as a posteriori probability of false rejection when a gene with z belongs to the null. After that, Ploner and his coworkers (called Ploner hereafter) firstly introduced the two-dimensional local false discovery rate to improve on the Efron’s procedure:
f d r 2 d ( z 1 , z 2 ) = π 0 f 0 ( z 1 , z 2 ) f ( z 1 , z 2 ) ,
where ( z 1 , z 2 ) = ( t s t a t i s t i c , l o g ( s e ) ) are two-dimensional test statistic [15]. Dimension extension of test statistic caused the decrease in the number of false positives, leading to better performance. Kim and his colleagues (called Kim hereafter) also extended the dimension of test statistic, but in a different way that they apply Bonferroni correction to the combination of two marginal Efron’s local FDR [13]. Unlike Ploner, they explicitly represented two different types of composite null hypotheses: union null and intersection null (see Figure S1). Depending on the null type, they proposed different procedure that could control two-dimensional local false discovery rate.
For comparison, we here consider three different local FDR control methods: Efron, Ploner, and Kim. All methods consider mixture model as a density of test statistics. However, those methods can be differently categorized by their own properties. The Efron procedure can be used to control one-dimensional FDR while the other two (Ploner and Kim) can be applied to the control of two-dimensional FDR. Since Efron is the only method that controls one-dimensional local FDR, we here consider two more variants of Ploner. As mentioned in the literature, Ploner can be reduced into one dimensional approach through conditional expectation [15], which is called Ploner1d. Furthermore, classic two-sample t-statistic in Ploner1d can be replaced with Efron’s t, i.e., modified t-statistic. We call this Ploner1dE. The main difference between two 2-dimensional FDR procedures is whether the type of null matters or not. In case of Kim, the estimation of fdr2d severely depends on the type of null: union null or intersection null. However, the null type has no effect on the estimation of fdr2d in Ploner because they did not take null type into account when estimating fdr2d.
In this paper, we compare the performance of the all methods on both simulated and real data. In Section 2, we provide data analysis results which is followed by brief discussion in Section 3. We introduce the three methods in Section 4 and then provide brief conclusion in Section 5.

2. Results

2.1. Simulation Study

We consider three different scenarios: basic, mean shift, and scale change. In all scenarios, we consider the following mixture model
f ( z ) = π 0 f 0 ( z ) + π 1 f 1 ( z ) .
Throughout the simulation study, we consider π 0 = 0.8 and f 0 is the standard normal density. For comparison purpose, we generate data matrix, size of 3000 by 40, each group consisting of 20, respectively (see Figure S2). Furthermore, for each scenario, we consider two different types of true alternatives: one-sided and two-sided alternative. For one-sided alternative, right-tail alternative only is considered. In case of two-sided alternative, symmetric alternative density is considered. Even though we consider two different types of alternative, the results for two-sided alternative are more highlighted here because one-sided alternative can be considered as a subcase of two-sided alternative.
Efron is the only one-dimensional FDR control method to be considered here. However, Ploner also can be used for one-dimensional case. As mentioned in Ploner et al., f d r 1 d ( z 1 ) is defined as conditional expectation of f d r 2 d given Z 1 = z 1 , i.e., f d r 1 d ( z 1 ) = E [ f d r 2 d ( z 1 , Z 2 ) | z 1 ] . We call it Ploner1d. Furthermore, the classic t-statistic can be replaced with Efron’s t-statistic, i.e., t E . Then we get a new estimator of local false discovery rate coupled with Efron’s t, which we call Ploner1dE. Therefore, three methods for the estimation of fdr1d are considered here: Efron, Ploner1d, and Ploner1dE. Furthermore, as a reference, we provide true fdr1d. The calculation of the true fdr1d is given in Section 2.4 in Supplementary file. For the estimation of fdr2d, we consider two methods: Ploner2d and Kim.

2.1.1. Basic Scenario

In basic scenario, we set the problem easy to separate alternative from null. We consider two types of alternative densities, which are well-separated from the true null: one-sided or two-sided. We generate 3000 × 40 data matrix and then estimate FDR by using all methods. Once all estimates are given, we apply three different cutoffs ( 0.05 , 0.1 , 0.2 ) to the estimated false discovery rate. This process is repeated 100 times.
  • One-sided alternative In the mixture model, we set π 1 = 0.2 and about 600 rows have different group means. That is, random sample for group 1 (G1) come from N ( 2.5 , 1.5 2 ) and random sample for group 2 (G2) from N ( 0 , 1 ) . The estimated fdr1ds are provided in Figure 1: the estimated fdr1d by classic t (left) and modified t-statistic (right).
The estimated fdr1d by Ploner1d is given in left panel of Figure 1 and estimated fdr1d by Efron and Ploner1dE in the right, which are represented in different colors: Efron (green), Ploner1dE (blue). As a reference, true fdr1d is presented in red. Our focus is usually on the small value of FDR, i.e., right tail of estimated fdr1d in this case. Since the true alternative density is not much overlapped with the true null density, it is easy to find biomarkers belonging to alternatives (see Figure S3). In addition, as seen in the figure above, all methods show similar performance and control nominal FDR well.
For the estimation of fdr2d, we calculated two dimensional information such as t-stat and log se from the same data. Depending on the null type (union or intersection null), Kim’s method provides different results. However, Ploner’s approach provides the same results regardless of the null type. In essence, the estimation in Ploner is performed on the joint distribution of two-dimensional statistics.
The estimated fdr2ds for the union null are provided in the left panel of Figure 2. Top right part of it corresponds to the rejection region by Kim while about right half of the figure is the rejection region by Ploner. True null and true alternative are represented as dots in different colors: null (blue) and alternative (red). All red dots should have been rejected here. However, by Kim, only about half of them are correctly rejected. That is, many false negatives are observed. Unlike Kim, Ploner separates true alternative (red dots) from true null (blue dots) very well. The results for the intersection null are given in Figure 2b. All areas, except for the bottom left part, correspond to the rejection region by Kim. In this case, all red dots are correctly rejected. However, only a few true nulls (blue dots) are falsely rejected by Kim, producing some false positives. To be more precise, the number of false positives for each cutoff value are 1, 17, and 26, respectively. One more thing to mention is that Ploner’s rejection region is between two of Kim’s rejection region. That is, it includes Kim’s rejection region for union null while it is included in that for intersection null.
  • Two-sided alternative In the mixture model, we set π 1 f 1 = π 11 f 11 + π 12 f 12 where π 11 = π 12 = 0.1 . Again, about 600 rows have different group means. To be more specific, about 300 random sample for G1 come from N ( 2.5 , 1.5 2 ) and another 300 from N ( 2.5 , 1.5 2 ) , respectively. Just like the one-sided alternative, we follow the exactly same procedure with the new π 1 f 1 .
In Figure 3, the estimated fdr1ds are provided in different colors. The true fdr1d, as a reference, is presented in red. The vertical lines at the bottom of the figure represent true null (black) and true alternative (blue), respectively. As seen in the figure, true alternatives are well separated from true null at any nominal FDR level.
In Figure 4, we provide the estimated fdr2ds for union null (left) and intersection null (right). For union null, since we consider two-sided alternative here, two parts (top left and top right) correspond to the rejection region by Kim. Again, some red dots that should be rejected are not rejected, leading to false negatives. The number of false negatives by Kim for each cutoff value are 13, 0, and 0, respectively. However, for the intersection null, almost all true alternatives are correctly rejected.
For each method, mean and standard error of the estimated FDRs over 100 repetitions are summarized in Table 1. As seen in Table 1, all procedures control local FDR strictly.
Also, we selected one sample out of 100 and calculated some performance measures for each method when cutoff = 0.1 (Table 2). Not surprisingly, all methods in case of basic scenario show very good performance.

2.1.2. Mean Shift

We now make estimation situation more difficult than the basic scenario by moving the mean of alternative density close to that of true null density, i.e., smaller margin between two group means. We consider three different mean values of alternative density, μ = ( 1 , 1.5 , 2 ) . However, variance does not change here. We generate 3000 × 40 data matrix and then estimate local FDR by using all methods. Once all estimates are given, we apply the cutoff value of ( 0.05 , 0.1 , 0.2 ) to the estimated false discovery rate. This process is repeated 100 times. Again, we consider two types of alternative: one-sided or two-sided. As seen in the basic scenario, both one-sided and two-sided alternative show similar results and we here provide the results for two-sided alternative only. The results for one-sided alternative are provided in Supplementary file.
  • Two-sided alternative The scheme to generate random sample is similar to the basic scenario. Only difference is that 10 percent of r.s. for G1 come from N ( μ , 1.5 2 ) and another 10 percent from N ( μ , 1.5 2 ) . Three mean values μ = ( 1 , 1.5 , 2 ) are considered. Just like basic scenario, we follow the exactly same procedure with the new π 1 f 1 . When μ = 1.5 , Figure 5 and Figure 6 include the estimated fdr1d and the estimated fdr2d, respectively.
In fdr2d estimation when μ = 1.5 , Kim’s intersection null (Figure 6b) rejects correctly true alternatives, i.e., red dots.
When cutoff = 0.1, means and standard errors of the estimated FDRs for each method over 100 repetitions are summarized in Table 3. As summarized in Table 3, all procedures control local FDR strictly.
In addition, when cutoff = 0.1, box plots of the number of rejections by each method over 100 repetitions are summarized in Figure 7. As seen in the figure, the number of rejections decrease as μ decreases to keep nominal local FDR. Especially, when μ = 2 , i.e., case of easy separation, almost all true alternatives are correctly rejected.
In addition, we selected one sample out of 100 and calculated some performance measures for each method when μ = 1.5 and cutoff = 0.1 (Table 4).
ROC curve when μ = 1.5 is provided in Figure 8. For this, we considered various cutoff values ranging from 0 to 1 with steps of 0.05. Based on the plot, it is clear that all methods show satisfactory results.

2.1.3. Scale Change

Just like the mean-shift scenario, we make estimation situation more difficult, but in a different way by increasing the variance of alternative density. More specifically, we consider three different variances of alternative density, k σ 2 , i.e., k = ( 2 , 3 , 4 ) , σ 2 = 1.5 2 . We generate 3000 × 40 data matrix and then estimate FDR by using all methods. Once all estimates are given, we apply the cutoff value of ( 0.05 , 0.1 , 0.2 ) to the estimated false discovery rate. This process is repeated 100 times. Again, we here provide the results for two-sided alternative only. The results for one-sided alternative are provided in Supplementary file.
  • Two-sided alternative In the mixture model, we set π 1 f 1 = π 11 f 11 + π 12 f 12 where π 11 = π 12 = 0.1 . 10 percent of random sample come form N ( 2.5 , k × 1.5 2 ) and another 10 percent from N ( 2.5 , k × 1.5 2 ) . Three different k = ( 2 , 3 , 4 ) are considered. Just like basic scenario, we follow the exact same procedure with the new π 1 f 1 . When k = 4 , Figure 9 and Figure 10 include the estimated fdr1d and the estimated fdr2d, respectively.
When k = 4 , the estimated fdr2ds for union null (left) and intersection null (right) are provided in Figure 10. Different cutoff values by Kim are denoted by the dotted line in different colors: cutoff = 0.05 (green), cutoff = 0.1 (blue), and cutoff = 0.2 (red). In case of cutoff = 0.1, means and standard errors of estimated FDR over 100 repetitions are summarized in Table 5. As seen in Table 5, all procedures control local FDR strictly.
In addition, when cutoff = 0.1, the box plot of number of rejections by each method over 100 repetitions are summarized in Figure 11. As seen in the figure, the number of rejections decrease as k increases to keep nominal local FDR. Compared to mean shift scenario, the number of rejections does not change drastically as k varies.
In addition, we selected one sample and calculated some performance measures for each method when k = 4 and cutoff = 0.1 (Table 6). Clearly, all methods show satisfactory performance.
ROC curve when k = 4 is provided in Figure 12. For this, we considered various cutoff values ranging from 0 to 1 with steps of 0.05.

2.2. Real Data Analysis

We investigate the performance of all methods on two different real data set: omija data [13] and lymphoma data [15]. The first data set was used by Kim et al. while the second data set was used by Ploner et al. Experimental details about two sets of real data are provided in the original literature respectively [16,17].

2.2.1. Omija Data

The data matrix (3226 by 57) were obtained by using 57 Schisandra chinesis from two different countries: 27 species of China and 30 species of Korea. Three dimensional chromatograms were obtained by GC/MS and then converted into a data table by summing the intensities in the predetermined intervals from the instrument [13]. More details about the experiment are available in the original literature [17].
Two fdr1d estimates obtained by using modified t-statistic (left) and two fdr2d estimates (right) are provided in Figure 13.
The fdr2d estimates for union null are provided in the right panel of Figure 13. Each dot in the figure represents each retention time. Three different target FDRs are displayed in different colors: green (0.05), blue (0.1), and red (0.2). Surprisingly, we observe huge overlap at the top of two tornado plots. The dots rejected by Kim at the FDR level of 0.05 are presented in green. The Table 7 includes the number of rejection by each method.
In case of Efron, the numbers in Table 7 are the exactly same as the number of rejections given in original literature. However, for other two methods, we slightly modified the code to get the FDR estimation make more sense. Therefore, there are little difference between the numbers in Table 7 and the results in original literature.

2.2.2. Lymphoma Data

Lymphochip DNA micorarrays, which consist of 12,196 clones of complementary DNA, were used to get gene expression data from 240 patients with untreated diffuse large-B-cell lymphoma [16]. In addition, outcome information of two classes is available: 102 survivors and 138 death patients. The data were available and downloaded from the website (http://llmpp.nih.gov/DLBCL).
Two different fdr1d estimates obtained by using modified t-statistic (left) and two different fdr2d estimates(right) are provided in Figure 14.
The fdr2d estimates for union null are provided in the right panel of Figure 14. Each gene is represented as dot in the figure. Here we consider different set of cutoff values (0.1, 0.2, 0.3). The reason is that unlike the Omija data, the number of genes rejected by all methods except for Kim(Intersection null) are almost 0. Three different cutoff values of target FDR are displayed in different colors: green (0.1), blue (0.2), and red (0.3). As seen in Table 8, there are small number of rejections by Kim (union null). The green line is exactly overlaid on the blue line and blue line is not shown in Figure 14. In contrast to the tornado plots from Omija data, there is not much overlap between two tornado plots from lymphoma data. From this, we can imagine that the results in practice heavily depend on the nature of the data. The following table includes the number of rejection.

3. Discussion

In this paper, we study the performance of the methods that control one-dimensional and two-dimensional local false discovery rate. It is mentioned that two-dimensional local FDR has advantages over one-dimensional local FDR [15]. More specifically, for given target FDR (say 0.1), suppose that there are two different corresponding t-statistics, i.e., a < b , in case of one-dimensional local false discovery rate. Then, biomarkers with t-statistic smaller than a or larger than b are called significant. However, in two-dimensional case, the decision on the biomarkers with the same t-value varies depending on the additional information, i.e., standard error. Ploner et al. also mentioned this point. For example, as seen in Figure 1d in Ploner et al., there are many genes with t-statistic = 3.8 . Among them, the gene with log standard error = 1.0 is not rejected when the target FDR is 0.1. However, the gene with log standard error = 0.5 is rejected at the same level of FDR. Clearly, the decision on the genes with the same t-statistic could be different. It is also known that the small standard error inflates t-statistic and false positive may happen due to such small standard error. However, two-dimensional local FDR control method has a chance to figure out t-statistic inflated by small standard error. Hence, two-dimensional local FDR makes decision on the gene with small standard error conservatively compared to one-dimensional one. As a result, false positives could be reduced through dimension extension from 1D to 2D.
As we can imagine, the rejection region of Ploner2d is between two Kim’s rejection region. That is, we observed that the number of rejection by Ploner is smaller than Kim’s intersection null but bigger than Kim’s union null. From this, we got some insight on a new direction of future study. If we can find a new intermediate rejection region, which is the compromise of two rejection regions by Kim’s approach, new method with the intermediate rejection region would be better.

4. Materials and Methods

We consider methods controlling one-dimensional and two-dimensional local false discovery rate (fdr1d and fdr2d): Efron (2001) [11], Ploner (2006) [15], Kim (2018) [13]. Furthermore, there are two more variants of Ploner: Ploner1d and Ploner1dE. As mentioned earlier, Ploner1d is the method, which is reduced into one-dimensional method by conditional expectation. Ploner1dE employs the modified t-statistic instead of classic two-sample t-statistic.

4.1. Terminologies

Even though Soric coined the FDR terminology in 1989, Benjamini and Hochberg first defined the formal definition of the FDR ([9,10]). To introduce the FDR, the following confusion matrix is commonly used (Table 9).
It is assumed that the number of hypotheses m are known in advance and R is an observable random variable, meaning the number of hypotheses to be rejected. Then the FDR is defined as:
F D R = E V R ,
while the familywise error rate (FWER) is defined:
F W E R = P ( V 1 ) .
Benjamini and Hochberg showed that the FDR is smaller than or equal to the FWER [10]. The equivalence happens only when all hypotheses are true null, m 0 = m .
To define local false discovery rate, all methods to be considered here employ model-based approach. The following two-component mixture model is used and the local false discovery rate is defined as below:
f ( z ) = π 0 f 0 ( z ) + ( 1 π 0 ) f 1 ( z ) ,
f d r k d ( z ) = π 0 f 0 ( z ) f ( z ) ,
where z = ( z 1 , , z k ) t is k-dimensional statistic, f 0 and f 1 are null and alternative density functions, and f d r k d denotes the estimator of k-dimensional local false discovery rate. In this paper, we consider k = 1 , 2 and α is the target FDR level.

4.2. Efron’s Approach

Efron and his colleagues applied Bayes’s rule to the mixture model [11]. That is, they defined one-dimensional local false discovery rate as a posteriori probability of event given z, i.e.,
f d r 1 d ( z ) = P ( N o E v e n t | Z = z ) = π 0 f 0 ( z ) f ( z ) = π 0 ( z ) = 1 π 1 ( z ) ,
where f ( z ) = π 0 f 0 ( z ) + ( 1 π 0 ) f 1 ( z ) . Here, z is the modified t-statistic, which is defined as
t i E = D i ¯ S i + a 0 , i = 1 , , m ,
where D i ¯ is the average of paired differences, S i is their sample standard deviation, and a 0 is the 90th percentile of all S values. To find the best a 0 (90th percentile of S values), they searched for mappings that produce the least information loss during data reduction at the gene level, i.e., from a 320-vector in the experiment to a single number Z i . Such mapping also provides good separation between f ( Z ) and f 0 ( Z ) . In essence, among five numbers of a 0 = ( 0 , 5 th , 50 th , 90 th , ) , they selected the value maximizing π 1 ( Z ) / π 0 ( Z ) , which is equivalent to maximizing f ( Z ) / f 0 ( Z ) . Note that π 1 ( Z ) / π 0 ( Z ) is the exactly same as ( 1 / π 0 ) · f ( Z ) / f 0 ( Z ) 1 .
From an estimation perspective, there are three key things: π 0 , f 0 , and f. They first estimate the ratio f 0 / f by using logistic regression. To do so, they compute the null scores by repeatedly applying row-wise sign permutation to the difference data matrix. Since paired data was considered in the literature, they applied sign permutation. If independent sample is considered, then label permutation is typically used instead. Then the upper bound of π 0 is calculated:
π 0 m i n { f ( z ) / f 0 ( z ) } .
Combining two estimators together, f d r 1 d is estimated.
Efron et al. applied the method to the paired data. Analysis results show that if f d r 1 d is controlled at the level of 0.1, the estimation of π 0 has little effect on the decision of significance. More specifically, 106 genes are called significant when using π 0 = 1 while 127 genes are called significant when using π 0 = 0.811 . To find better upper bound for π 0 , they suggest new estimate, which is defined over an interval
π 0 A f ( z ) A f 0 ( z ) ,
where the interval A is near z = 0 ( A = [ 0.5 , 0.5 ] in simulation study).
Even though they applied the procedure to the paired data, transcriptional responses to ionizing radiation generated by Professor Gilbert Chu, the procedure could be generally applied to any two class problem including two sets of unpaired samples.
They also showed, from simulation study mimicking the radiation experiment, that the empirical Bayes approach is closely related to the frequentist version of FDR developed by Benjamini and Hochberg. The reason they use the terminology “empirical Bayes” is that the crucial ratio f 0 / f is estimated from the data rather than from a priori assumptions.

4.3. Ploner’s Approach

As seen in the Efron’s modified t-statistic, they tried to adjust the test statistic by adding some constant ( a 0 ) to the denominator in order to reduce false positives. Ploner and his colleagues also noticed such phenomenon that small standard error would inflate test statistic and increase false discovery rate [15]. To remedy such problem, they extended the dimension of test statistics and local false discovery rate in two-component mixture model. The two-dimensional local false discovery rate is defined:
f d r 2 d ( z 1 , z 2 ) = π 0 f 0 ( z 1 , z 2 ) f ( z 1 , z 2 ) ,
where z 1 and z 2 are t-statistic and log ( s e ) , respectively. Here classical t-statistic is used
t i P = X ¯ i 1 X ¯ i 2 s e i ,
with pooled standard error
s e i = ( n 1 1 ) s i 1 2 + ( n 2 1 ) s i 2 2 n 1 + n 2 2 1 / n 1 + 1 / n 2 ,
where X ¯ i j , s i j and n j are the gene-wise group mean, standard deviation and sample size for gene i and group j = 1, 2.
For the estimation of f d r 2 d , they consider a procedure requiring only a single smoothing of the ratio
f d r 2 d ( z ) = π 0 γ ( z ) B ( 1 γ ( z ) ) ,
where γ ( z ) = B f 0 ( z ) / [ f ( z ) + B f 0 ( z ) ] , and B ( = 100 ) is the number of permutations used to generate samples from null distribution, similar to the null score in Efron’s approach. However, group label is permuted instead of sign permutation in Efron’s paper because unpaired sample is used in this paper. Let Z be the m × k observation matrix from m genes. Each permutation of group labels generates a new dataset and statistic matrix Z * . After p permutations, we have a series of Z 1 * , … Z p * , which represent samples of Z under the null. Given the observed Z and p sets of null scores Z 1 * , … Z p * , γ ( z ) is estimated by using the mixed-model approach of Pawitan (2001, Section 18.10) [18]. More specifically, similar to Efron, they pre-bin the 2D space into small interval and then perform discrete smoothing of binomial data. The smoothed estimate of γ i j , which is computed by using the iteratively weighted least square algorithm, is the minimizer of the penalized log-likelihood
log L ( γ , λ ) = i j y i j log γ i j + ( N i j y i j ) log ( 1 γ i j ) + λ ( i , j ) ( k , l ) ( η i j η k l ) 2 ,
where ( i , j ) ( k , l ) means that ( i , j ) and ( k , l ) are primary neighbors in the 2D lattice. They assumed that the number of successes in ( i , j ) grid, y i j follows binomial distribution with success probability γ i j , i.e., y i j B i n ( N i j , γ i j ) . Here N i j is the total number of statistic in ( i , j ) grid. They also addressed bias correction issue, which are caused by boundary problem and data sparsity.
For the estimation of π 0 , they employ two different options. They use the same upper-bound estimate as used in Efron for simulation data while they use the mixture-based estimates for the real datasets, which is described in Pawitan et al. [19,20].
Theoretically, Ploner’s approach can be easily extended to the high-dimensional case (≥3). However, there exists practical difficulty of smoothing in higher dimension. Even in 2D case, they develop a procedure requiring single smoothing for γ ( z ) . However, smoothing problem still exists. That is, what is optimal for γ ( z ) is not necessarily optimal for f d r 2 d . It is also mentioned that the f d r 2 d performs as well as or better than the f d r 1 d , in dealing with misleadingly small standard errors. For f d r 1 d ( Z ) , they considered four different variants of t-statistic: standard t, Efron’s t, Smyth’s t, and SAM ([11,21,22]). Compared to the f d r 1 d coupled with four different t statistics, f d r 2 d has better overall performance in terms of empirical global FDR. It means that it is not necessary to modify t-statistic any longer. Rather, extension of the dimension of test statistic is more effective than the variation of test statistic.

4.4. Kim’s Approach

Ploner and his colleagues did not explicitly mention the type of null hypothesis. However, Kim and his coworkers presented two different bivariate composite null: intersection-type and union-type null (see Figure S1) [13]. For example, intersection-type composite null is represented:
H 0 = H 0 , 1 H 0 , 2 ,
where H 0 , 1 and H 0 , 2 represent each univariate marginal null. The first part H 0 , 1 : μ 1 = μ 2 represents equal mean and the second H 0 , 2 : σ 2 = 0 represents zero variance. It is clear that the form of alternative hypothesis is affected by the type of null, accordingly the rejection region should be constructed depending on the null hypothesis.
To test for composite null, they proposed a two stage procedure. In the first stage, they adopted Efron’s procedure to control one-dimensional local false discovery rate separately in each marginal space. In the second stage, to control 2d-fdr, they combined two marginal local false discovery rates by using the concept of Bonferroni’s correction. Regardless of the type of null, the first stage is the same. However, the second stage varies depending on the type of null.
Suppose that R 1 is rejection region for H 0 , 1 and R 2 is rejection region for H 0 , 2 , i.e., P ( R 1 ) = α 1 , P ( R 2 ) = α 2 , and α 1 + α 2 = α . Then rejection region for intersection null is R 1 R 2 . Thus, it is obvious that 2d-fdr is controlled because P ( R 1 R 2 ) P ( R 1 ) + P ( R 2 ) = α . However, in case of union null, it is more complicated to combine two marginal 1d-fdrs into one 2d-fdr. It is clear that P ( R 1 R 2 ) P ( R 1 R 2 ) P ( R 1 ) + P ( R 2 ) = α . However, in most cases, the difference between P ( R 1 ) + P ( R 2 ) and P ( R 1 R 2 ) is not negligible. When the difference is significant, it is obvious that decision rule would be too conservative. To avoid such problem, they tried to find optimal α 1 and α 2 such that
m a x ( α 1 , α 2 ) 1 + m i n n 1 n 12 , n 2 n 12 α ,
where n 1 , n 2 , n 12 are the number of hypotheses in R 1 , R 2 , R 1 R 2 , respectively. More details on the derivation of Equation (5) are provided in the original literature [13].
For the estimation of null part, π 0 and f 0 , they applied Gaussian mixture model to the data and selected the first mixture part by using “Mclust” function in the m c l u s t package because normality assumption was made on the null distribution [23,24]. In addition, the density f is estimated by kernel smoothing methods with optimal kernel. The kernel density estimator is obtained by using the “density” function in R.
Here we want to emphasize two main differences from Ploner’s method. Firstly, this procedure provides a closed-form and smooth decision rule on the bivariate test statistics, leading to easy decision for new observation. Secondly, the procedure can be applied to both type of composite null. In contrast, Ploner’s method did not clearly mention the type of null.

4.5. Estimation of f 0 ( z ) / f ( z )

The estimation of f 0 ( z ) / f ( z ) plays a key role in estimating local false discovery rate in the methods mentioned above. However, estimating numerator and denominator separately and then taking the ratio of them is not a good idea. Ploner and his colleagues commented on this issue: “optimal smoothing for the densities may not be optimal for the fdr because different amount of data are used.” Therefore, each method tries to estimate the ratio directly, yet in a different way.
In Efron’s approach, they calculate observed test statistic Z and generate B different sets of null statistic z. Each set of null statistic is generated by row-wise sign permutation ( B = 20 in the paper). Given Z and z, they consider Z and z as “successes” and “failures”, respectively. Then they estimate the probability
π ( z ) = f ( z ) f ( z ) + B f 0 ( z ) ,
with a natural spline with 7 degrees of freedom. In practice, they divided the range [−4, 4] into 139 equal intervals, and counted the number of values of Z ( n Z ) and z ( n z ) in each interval. The ratio of n Z to n z is used to estimate π ( z ) by logistic regression. Therefore, the estimator of f 0 / f is obtained by Equation (6)
f 0 f = 1 π ( z ) B π ( z ) f 0 f ^ = 1 π ^ ( z ) B π ^ ( z ) .
Similarly, Ploner and his colleagues try to estimate
γ ( z ) = 1 π ( z ) = B f 0 ( z ) f ( z ) + B f 0 ( z ) ,
and then estimate f d r 2 d
f d r 2 d ( z ) = π 0 γ ( z ) B ( 1 γ ( z ) ) .
Key difference is that Ploner et al. extended the dimension of test statistic to k dimension. Similar to Efron, the crucial part γ ( z ) is estimated by binning the data into small intervals. The discretized version of estimate r i j is the minimizer of the penalized log-likelihood function including single smoothing parameter.
Kim and his colleagues borrowed the idea of Efron to estimate the π ( z ) for marginal estimation of FDR. Furthermore, for dimension extension, they applied Bonferroni correction to the combination of two marginal FDR estimates. In addition, main difference from Ploner’s approach is that Kim et al. explicitly present two types of composite null: union and intersection null, and they propose a unified approach handling both types of null.

5. Conclusions

Ploner (2006) [15] commented on the issue of FDR control: “different types of t statistics have a little effect on the performance. However, dimension extension of statistics has a huge effect on the results”. In this paper, we assured the comments mentioned in Ploner et al. through comprehensive simulation study.
As an improvement of the Kim’s approach, we consider that combination of two rejection regions from Kim’s Union and Kim’s Intersection may produce a better rejection region. Therefore, we suggest a new rejection region (Figure 15), which improve on the control of FDR.
In case of Kim’s intersection null, some false positives were observed in the region (rectangle in red in Figure 15). In addition, false negatives were observed in the region (circle in green). Thus, we save some probability in two top corners and then get some space for the null (red rectangle area at the top). We plan to work on this issue as a direction of future study.

Supplementary Materials

The following are available online at https://www.mdpi.com/2218-1989/11/1/53/s1, Figure S1: Union null and Intersection null, Figure S2: Simulation setup: One-sided alternative and two-sided alternative, Figure S3: Basic scenario: true density, Figure S4: Mean shift scenario: True densities for μ = ( 1 , 1.5 , 2 ) , Figure S5: Mean shift scenario: Estimated fdr1d for μ = 1 , Figure S6: Mean shift scenario: Estimated fdr1d for μ = 1.5 , Figure S7: Mean shift scenario: Estimated fdr1d for μ = 2 , Figure S8: Mean shift scenario for μ = 1 : Estimated fdr2d for union null and intersection null, Figure S9: Mean shift scenario for μ = 1.5 : Estimated fdr2d for union null and intersection null, Figure S10: Mean shift scenario for μ = 2 : Estimated fdr2d for union null and intersection null, Figure S11: Mean shift scenario (two-sided): True densities for μ = ( 1 , 1.5 , 2 ) , Figure S12: Mean shift scenario (two-sided): Estimated fdr1d for μ = 1 , Figure S13: Mean shift scenario (two-sided): Estimated fdr1d for μ = 1.5 , Figure S14: Mean shift scenario (two-sided): Estimated fdr1d for μ = 2 , Figure S15: Mean shift scenario (two-sided) for μ = 1 : Estimated fdr2d for union and intersection null, Figure S16: Mean shift scenario (two-sided) for μ = 1.5 : Estimated fdr2d for union and intersection null, Figure S17: Mean shift scenario (two-sided) for μ = 2 : Estimated fdr2d for union and intersection null, Figure S18: Scale change scenario: True densities for k = ( 2 , 3 , 4 ) , Figure S19: Scale change scenario: Estimated fdr1d for k = 2 , Figure S20: Scale change scenario: Estimated fdr1d for k = 3 , Figure S21: Scale change scenario: Estimated fdr1d for k = 4 , Figure S22: Scale change scenario for k = 2 : Estimated fdr2d for union and intersection null, Figure S23: Scale change scenario for k = 3 : Estimated fdr2d for union and intersection null, Figure S24: Scale change scenario for k = 4 : Estimated fdr2d for union and intersection null, Figure S25: Scale change scenario (two-sided): True densities for k = ( 2 , 3 , 4 ) , Figure S26: Scale change scenario (two-sided): Estimated fdr1d for k = 2 , Figure S27: Scale change scenario (two-sided): Estimated fdr1d for k = 3 , Figure S28: Scale change scenario (two-sided): Estimated fdr1d for k = 4 , Figure S29: Scale change scenario for k = 2 : Estimated fdr2d for union and intersection null, Figure S30: Scale change scenario for k = 3 : Estimated fdr2d for union and intersection null, Figure S31: Scale change scenario for k = 4 : Estimated fdr2d for union and intersection null, Figure S32: Omija data: Estimated fdr2d for intersection null, Figure S33: Lymphoma data: Estimated fdr2d for intersection null, Table S1: Basic scenario (one-sided alternative): means and standard errors of estimated FDRs over 100 repetitions, Table S2: Basic scenario (one-sided alternative): Performance measure when cutoff = 0.1.

Author Contributions

Conceptualization, J.J.; methodology, J.J., S.J.K.; software, Y.O., S.J.K.; validation, J.J., S.J.K., and Y.O.; formal analysis, S.J.K.; investigation, J.J.; writing—original draft preparation, J.J.; writing—review and editing, J.J.; visualization, Y.O., S.J.K.; supervision, J.J.; funding acquisition, J.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2018R1D1A1B07042372 to JJ). This study was supported by the BK21 FOUR funded by the Ministry of Education (21A20132212094).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All samples used here are available at the github website https://github.com/jjs3098/Comparison-of-FDR-control-methods.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
BHBejamini and Hochberg
FDRFalse discovery rate
FWERFamilywise error rate
2d-fdrTwo-dimensional false discovery rate

References

  1. Alishahi, K.; Ehyaei, A.R.; Shojaie, A. A generalized Benjamini-Hochberg procedure for multivariate hypothesis testing. arXiv 2016, arXiv:1606.02386. [Google Scholar]
  2. Hochberg, Y. A sharper Bonferroni procedure for multiple tests of significance. Biometrika 1988, 75, 800–803. [Google Scholar] [CrossRef]
  3. Hochberg, Y.; Benjamini, Y. More powerful procedures for multiple significance testing. Stat. Med. 1990, 9, 811–818. [Google Scholar] [CrossRef] [PubMed]
  4. Holm, S. A simple sequentially rejective multiple test procedure. Scand. J. Stat. 1979, 6, 65–70. [Google Scholar]
  5. Hommel, G. A stagewise rejective multiple test procedure based on a modified Bonferroni test. Biometrika 1988, 75, 383–386. [Google Scholar] [CrossRef]
  6. Rom, D.M. A sequentially rejective test procedure based on a modified Bonferroni inequality. Biometrika 1990, 77, 663–665. [Google Scholar] [CrossRef]
  7. Saville, D.J. Multiple comparison procedures: The practical solution. Am. Stat. 1990, 44, 174–180. [Google Scholar]
  8. Simes, R.J. An improved Bonferroni procedure for multiple tests of significance. Biometrika 1986, 73, 751–754. [Google Scholar] [CrossRef]
  9. Soric, B. Statistical “discoveries” and effect-size estimation. J. Am. Stat. Assoc. 1989, 84, 608–610. [Google Scholar] [CrossRef]
  10. Benjamini, Y.; Hochberg, Y. Controlling the false discovery rate: A practical and powerful approach to multiple testing. J. R. Stat. Soc. B 1995, 57, 289–300. [Google Scholar] [CrossRef]
  11. Efron, B.; Tibshirani, R.; Storey, J.; Tusher, V. Empirical Bayes analysis of a microarray experiment. J. Am. Stat. Assoc. 2001, 96, 1151–1160. [Google Scholar] [CrossRef]
  12. Efron, B.; Tibshirani, R. Empirical Bayes Method and False Discovery Rates for Microarrays. Genet. Epidemiol. 2002, 23, 70–86. [Google Scholar] [CrossRef] [PubMed]
  13. Kim, Y.; Lim, J.; Lee, J.S.; Jeong, J. Controlling two-dimensional false discovery rates by combining two univariate multiple testing results with an application to mass spectral data. Chemom. Intell. Lab. Syst. 2018, 182, 149–157. [Google Scholar] [CrossRef]
  14. Kim, J.; Jeong, J. MetPC: Metabolite pipeline consisting of metabolite identification and biomarker discovery under the control of two-dimensional FDR. Metabolites 2019, 9, 103. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Ploner, A.; Calza, S.; Gusnanto, A.; Pawitan, Y. Multidimensional local false discovery rate for micorarray studies. Bioinformatics 2006, 22, 556–565. [Google Scholar] [CrossRef] [PubMed]
  16. Rosenwald, A.; Wright, G.; Chan, W.C.; Connors, J.M.; Campo, E.; Fisher, R.I.; Gascoyne, R.D.; Muller-Hermelink, H.K.; Smel, E.B.; Staudt, L.M. The use of molecular profiling to predict survival after chemotherapy for diffuse Large-B-cell Lymphoma. N. Engl. J. Med. 2002, 346, 1937–1947. [Google Scholar] [CrossRef] [PubMed]
  17. Lee, D.; Lee, S.J.; Kim, S.C.; Lim, J.; Kwon, S.W. Classification of spectral data using fused lasso logistic regression. Chemom. lntell. Lab. Syst. 2015, 142, 70–77. [Google Scholar]
  18. Pawitan, Y. Non-Gaussian smoothing. In All Likelihood: Statistical Modelling and Inference Using Likelihood; Oxford Science Publications: Oxford, UK, 2001; pp. 490–497. [Google Scholar]
  19. Pawitan, Y. False discovery rate, sensitivity and sample size for microarray studies. Bioinformatics 2005, 21, 3017–3024. [Google Scholar] [CrossRef]
  20. Pawitan, Y. Bias in the estimation of false discovery rate in micorarray studies. Bioinformatics 2005, 21, 3865–3872. [Google Scholar] [CrossRef] [Green Version]
  21. Smyth, G.K. Linear models and empirical Bayes methods for assessing differential expression in microarray experiments. Stat. Appl. Genet. Mol. Biol. 2004, 3, 3. [Google Scholar] [CrossRef]
  22. Tusher, V.G.; Tibshirani, R.; Chu, G. Significance analysis of microarrays applied to the ionizing radiation response. Proc. Natl. Acad. Sci. USA 2001, 98, 5116–5121. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Fraley, C.; Raftery, A.E. Model-based clustering, discriminant analysis, and density estimation. J. Am. Stat. Assoc. 2002, 97, 611–631. [Google Scholar] [CrossRef]
  24. Scrucca, L.; Fop, M.; Murphy, B.; Raftery, A.E. mclust5: Clustering, classification and density estimation using Gaussian finite mixture models. R J. 2016, 8, 289–317. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Basic scenario and one-sided alternative. Estimated fdr1ds by using classic t-statistic (a) and modified t (b). As a reference, true fdr1d is provided in red.
Figure 1. Basic scenario and one-sided alternative. Estimated fdr1ds by using classic t-statistic (a) and modified t (b). As a reference, true fdr1d is provided in red.
Metabolites 11 00053 g001
Figure 2. Basic scenario and one-sided alternative: estimated fdr2d for union null (a) and intersection null (b).
Figure 2. Basic scenario and one-sided alternative: estimated fdr2d for union null (a) and intersection null (b).
Metabolites 11 00053 g002
Figure 3. Basic scenario and two-sided. Estimated fdr1ds by using classic t-statistic (a) and modified t (b). As a reference, true fdr1d is provided in red.
Figure 3. Basic scenario and two-sided. Estimated fdr1ds by using classic t-statistic (a) and modified t (b). As a reference, true fdr1d is provided in red.
Metabolites 11 00053 g003
Figure 4. Basic scenario and two-sided. Estimated fdr2d for union null (a) and intersection null (b).
Figure 4. Basic scenario and two-sided. Estimated fdr2d for union null (a) and intersection null (b).
Metabolites 11 00053 g004
Figure 5. Mean shift scenario and two-sided. Estimated fdr1ds by (a) Classic t and Ploner1d (b) Efron and Ploner1dE when μ = 1.5 .
Figure 5. Mean shift scenario and two-sided. Estimated fdr1ds by (a) Classic t and Ploner1d (b) Efron and Ploner1dE when μ = 1.5 .
Metabolites 11 00053 g005
Figure 6. Mean shift scenario and two-sided. Estimated fdr2d when μ = 1.5 . (a) Union null (b) Intersection null.
Figure 6. Mean shift scenario and two-sided. Estimated fdr2d when μ = 1.5 . (a) Union null (b) Intersection null.
Metabolites 11 00053 g006
Figure 7. Mean shift scenario and two-sided. The number of rejections by each method over 100 repetitions when cutoff = 0.1. Box plots for three mean values by each method: (a) Efron (b) Ploner1d (c) Ploner1dE (d) Ploner2d (e) Kim (Union) (f) Kim (Intersection).
Figure 7. Mean shift scenario and two-sided. The number of rejections by each method over 100 repetitions when cutoff = 0.1. Box plots for three mean values by each method: (a) Efron (b) Ploner1d (c) Ploner1dE (d) Ploner2d (e) Kim (Union) (f) Kim (Intersection).
Metabolites 11 00053 g007
Figure 8. Mean shift scenario and two-sided. ROC curve when μ = 1.5 and cutoff values ranging from 0 to 1 with steps of 0.05.
Figure 8. Mean shift scenario and two-sided. ROC curve when μ = 1.5 and cutoff values ranging from 0 to 1 with steps of 0.05.
Metabolites 11 00053 g008
Figure 9. Scale change scenario and two-sided. Estimated fdr1ds by (a) Classic t and Ploner1d (b) Efron and Ploner1dE when k = 4 .
Figure 9. Scale change scenario and two-sided. Estimated fdr1ds by (a) Classic t and Ploner1d (b) Efron and Ploner1dE when k = 4 .
Metabolites 11 00053 g009
Figure 10. Scale change scenario and two-sided. Estimated fdr2d when k = 4 : (a) Union null (b) Intersection null.
Figure 10. Scale change scenario and two-sided. Estimated fdr2d when k = 4 : (a) Union null (b) Intersection null.
Metabolites 11 00053 g010
Figure 11. Scale change scenario and two-sided. The number of rejections by each method over 100 repetitions when cutoff = 0.1. Box plots for three mean values by each method: (a) Efron (b) Ploner1d (c) Ploner1dE (d) Ploner2d (e) Kim (Union) (f) Kim (Intersection).
Figure 11. Scale change scenario and two-sided. The number of rejections by each method over 100 repetitions when cutoff = 0.1. Box plots for three mean values by each method: (a) Efron (b) Ploner1d (c) Ploner1dE (d) Ploner2d (e) Kim (Union) (f) Kim (Intersection).
Metabolites 11 00053 g011
Figure 12. Scale change scenario and two-sided. ROC curve when k = 4 and cutoff values ranging from 0 to 1 with steps of 0.05.
Figure 12. Scale change scenario and two-sided. ROC curve when k = 4 and cutoff values ranging from 0 to 1 with steps of 0.05.
Metabolites 11 00053 g012
Figure 13. Omija data. (a): estimated fdr1d; (b): estimated fdr2d.
Figure 13. Omija data. (a): estimated fdr1d; (b): estimated fdr2d.
Metabolites 11 00053 g013
Figure 14. Lymphoma data. (a): estimated fdr1d; (b): estimated fdr2d.
Figure 14. Lymphoma data. (a): estimated fdr1d; (b): estimated fdr2d.
Metabolites 11 00053 g014
Figure 15. An example rejection region.
Figure 15. An example rejection region.
Metabolites 11 00053 g015
Table 1. Basic scenario (two-sided alternative): means and standard errors of estimated FDRs over 100 repetitions.
Table 1. Basic scenario (two-sided alternative): means and standard errors of estimated FDRs over 100 repetitions.
MethodsCutoff = 0.05Cutoff = 0.1Cutoff = 0.2
Efron0.0003 (0.00007)0.0004 (0.00009)0.0011 (0.00016)
Ploner1d0.0007 (0.00010)0.0014 (0.00015)0.0031 (0.00023)
Ploner1dE0.0001 (0.00004)0.0004 (0.00008)0.0013 (0.00016)
Ploner2d0.0005 (0.00010)0.0017 (0.00017)0.0051 (0.00032)
Kim(Intersection)0.0047 (0.00046)0.0133 (0.00097)0.0342 (0.00200)
Table 2. Basic scenario (two-sided alternative): Performance measures when cutoff = 0.1.
Table 2. Basic scenario (two-sided alternative): Performance measures when cutoff = 0.1.
SensitivitySpecificityAccuracyF1 Score
Efron0.99810.9990.998
Ploner1d0.99710.9990.997
Ploner1dE0.98210.9960.990
Ploner2d0.9920.9990.9980.994
Kim(Intersection)0.9980.9970.9980.994
Table 3. Mean shift scenario (two-sided alternative): means and standard errors of estimated FDRs over 100 repetitions (cutoff = 0.1).
Table 3. Mean shift scenario (two-sided alternative): means and standard errors of estimated FDRs over 100 repetitions (cutoff = 0.1).
Methods μ = 1 μ = 1.5 μ = 2
Efron0.0114 (0.00085)0.0086 (0.00051)0.0026 (0.00021)
Ploner1d0.0398 (0.00140)0.0177 (0.00065)0.0059 (0.00034)
Ploner1dE0.0236 (0.00096)0.0101 (0.00047)0.0027 (0.00020)
Ploner2d0.0110 (0.00065)0.0052 (0.00035)0.0028 (0.00022)
Kim(Intersection)0.0316 (0.00164)0.0220 (0.00140)0.0202 (0.00217)
Table 4. Mean shift scenario (two-sided alternative): Performance measures when μ = 1.5 and cutoff = 0.1.
Table 4. Mean shift scenario (two-sided alternative): Performance measures when μ = 1.5 and cutoff = 0.1.
SensitivitySpecificityAccuracyF1 Score
Efron0.7880.9990.9570.880
Ploner1d0.8000.9990.9590.887
Ploner1dE0.6830.9980.9350.808
Ploner2d0.7970.9980.9580.883
Kim(Intersection)0.7680.9990.9530.867
Table 5. Scale change scenario (two-sided alternative): means and standard errors of estimated FDRs over 100 repetitions (cutoff = 0.1).
Table 5. Scale change scenario (two-sided alternative): means and standard errors of estimated FDRs over 100 repetitions (cutoff = 0.1).
Methods k = 2 k = 3 k = 4
Efron0.0014 (0.00018)0.0026 (0.00025)0.0032 (0.00027)
Ploner1d0.0063 (0.00030)0.0137 (0.00055)0.0209 (0.00062)
Ploner1dE0.0004 (0.00008)0.0003 (0.00008)0.0002 (0.00007)
Ploner2d0.0015 (0.00017)0.0025 (0.00023)0.0034 (0.00030)
Kim(Intersection)0.0050 (0.00033)0.0070 (0.00036)0.0085 (0.00044)
Table 6. Scale change scenario (two-sided alternative): Performance measures when k = 4 and cutoff = 0.1.
Table 6. Scale change scenario (two-sided alternative): Performance measures when k = 4 and cutoff = 0.1.
SensitivitySpecificityAccuracyF1
Efron0.9100.9990.9810.950
Ploner1d0.81710.9630.899
Ploner1dE0.6280.9950.9210.762
Ploner2d0.7930.9970.9560.879
Kim(Intersection)10.9970.9970.993
Table 7. Omija data: The number of rejection by each method.
Table 7. Omija data: The number of rejection by each method.
1D 2D
CutoffEfronPloner1dPloner1dEPloner2dKim (Union)Kim (Intersection)
0.05131425129011096362689
0.1491638169513557962797
0.21271906207018409632902
Table 8. Lymphoma data: The number of genes rejected by each method.
Table 8. Lymphoma data: The number of genes rejected by each method.
1D 2D
CutoffEfronPloner1dPloner1dEPloner2dKim (Union)Kim (Intersection)
0.11771610142667
0.2129815437043185
0.3155850669053569
Table 9. Confusion matrix when testing m hypotheses simultaneously.
Table 9. Confusion matrix when testing m hypotheses simultaneously.
Decision
NullAlternativeTotal
True null m 0 V V m 0
True alternativeT R V m m 0
m R Rm
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kim, S.J.; Oh, Y.; Jeong, J. Comprehensive Comparative Analysis of Local False Discovery Rate Control Methods. Metabolites 2021, 11, 53. https://doi.org/10.3390/metabo11010053

AMA Style

Kim SJ, Oh Y, Jeong J. Comprehensive Comparative Analysis of Local False Discovery Rate Control Methods. Metabolites. 2021; 11(1):53. https://doi.org/10.3390/metabo11010053

Chicago/Turabian Style

Kim, Shin June, Youngjae Oh, and Jaesik Jeong. 2021. "Comprehensive Comparative Analysis of Local False Discovery Rate Control Methods" Metabolites 11, no. 1: 53. https://doi.org/10.3390/metabo11010053

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop