Next Article in Journal
The Well-Posedness and Ergodicity of a CIR Equation Driven by Pure Jump Noise
Previous Article in Journal
Application of Mathematical Modeling and Numerical Simulation of Blood Biomarker Transport in Paper-Based Microdevices
Previous Article in Special Issue
A New Measure for an Acceptable Level of Homogeneity in Meta-Informatics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Online Multi-Layer FDR Control

Department of Biostatistics, University of Nebraska Medical Center, Omaha, NE 68198, USA
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(12), 1937; https://doi.org/10.3390/math13121937
Submission received: 19 April 2025 / Revised: 3 June 2025 / Accepted: 4 June 2025 / Published: 11 June 2025
(This article belongs to the Special Issue Statistical Methods in Bioinformatics and Health Informatics)

Abstract

:
When hypotheses are tested in a stream and real-time decision-making is needed, online sequential hypothesis testing procedures are needed. Furthermore, these hypotheses are commonly partitioned into groups by their nature. For example, RNA nanocapsules can be partitioned based on the therapeutic nucleic acids (siRNAs) being used, as well as the delivery nanocapsules. When selecting effective RNA nanocapsules, simultaneous false discovery rate control at multiple partition levels is needed. In this paper, we develop hypothesis testing procedures which control the false discovery rate (FDR) simultaneously for multiple partitions of hypotheses in an online fashion. We provide rigorous proofs for their FDR or modified FDR (mFDR) control properties and use extensive simulations to demonstrate their performances.

1. Introduction

In early-stage drug development research, lead compounds are often selected in an online fashion for the next-stage experiment (i.e., from discovery to an in vitro study or from in vitro to in vivo studies) [1]. Because the data enters in a stream or in batches, and real-time decision-making is desirable, online sequential hypothesis testing procedures need to be adopted. Furthermore, hypotheses in drug discovery are commonly partitioned into groups by their nature, and multiple partitioning methods are simultaneously of interest. For example, in developing innovative RNA nanocapsules to treat cancer metastasis, therapeutic nucleic acids (siRNAs) as well as delivery nanocapsules are designed [2,3]. At each step of the screening, a decision needs to be made whether to follow up on a particular combination of an siRNA and nanocapsule. In this case, these combinations are naturally partitioned according to their siRNA and nanocapsule formulations. Ideally, to select the candidate lead compounds, we are interested in controlling the false positives in terms of both the siRNA and nanocapsule formulations. We are interested in other potential partitions, for example, based on experiments in different cancer types [4].
Beyond drug development, the need for multi-layer false discovery rate (FDR) control also arises in many modern applications. In real-time A/B testing on online platforms, experiments are often run concurrently across user segments, device types, or product categories, and the results arrive continuously [5,6]. Similar challenges also exist in genomics, where hypotheses (e.g., gene expressions or mutations) may be partitioned by biological pathways or genomic regions and tested sequentially in large-scale screening studies [7]. These situations require statistical methods that can make decisions in an online fashion while still simultaneously controlling the FDR across multiple layers of structures. Motivated by these examples, we aimed to develop a hypothesis testing procedure that controls the FDR simultaneously for multiple partitions of hypotheses in an online fashion.
Mathematically, at time t N , we observe a stream of null hypotheses, H i ( t ) j ( t ) , where i ( t ) N denotes the index for the siRNA being tested at time t, and j ( t ) N denotes the index for the nanocapsule being tested at time t. H i ( t ) j ( t ) indicates a hypothesis with respect to the siRNA i ( t ) and nanocapsule j ( t ) .
We define two partition-level null hypotheses, H i ( t ) 1 and H j ( t ) 2 , such that
H i ( t ) 1 is true H i ( t ) j is true for j u = 1 t { j ( u ) } ,
H j ( t ) 2 is true H i j ( t ) is true for i u = 1 t { i ( u ) } ,
where H i ( t ) 1 is the partition-level hypothesis with respect to the siRNAs, and H i ( t ) 2 is the partition-level hypothesis with respect to the nanocapsules. H i ( t ) j is the hypothesis where the siRNA index i ( t ) is fixed at time t and varies over the nanocapsule indices j, and H i j ( t ) is a hypothesis where the nanocapsule index j ( t ) is fixed at time t and varies over the siRNA indices i. At time t, we make a decision on whether to reject H i ( t ) j ( t ) based on its corresponding test statistic and p-value, as well as the history information on the previous hypotheses (statistics, p-values, and decisions); i.e., we let δ t = 1 if we reject H i ( t ) j ( t ) , and δ t = 0 if we accept H i ( t ) j ( t ) . Our goal is to control the group-level FDR with both partitions according to the siRNA and according to the nanocapsules, which we denote as FDR 1 ( t ) and FDR 2 ( t ) . Let
S ^ 1 ( t ) = { i ( u ) : δ u = 1 } , S ^ 2 ( t ) = { j ( u ) : δ u = 1 } ,
FDR 1 ( t ) = E | S ^ 1 ( t ) H i ( t ) 1 | | S ^ 1 ( t ) | 1 , FDR 2 ( t ) = E | S ^ 2 ( t ) H j ( t ) 2 | | S ^ 2 ( t ) | 1 ,
where S ^ 1 ( t ) and S ^ 2 ( t ) denote the sets of rejected hypotheses at time t grouped by the siRNA and nanocapsule partitions, respectively; H i ( t ) 1 and H j ( t ) 2 denote the corresponding sets of true null hypotheses for the siRNA and nanocapsule partitions, respectively. ∨ is the maximum operator: | S ^ 1 ( t ) | 1 = max ( | S ^ 1 ( t ) | , 1 ) .
To control the FDR for individual online hypotheses, Foster and Stine (2008) [8] introduced the α -investing framework, a foundational method for online FDR control. This approach treats significance thresholds as a budget, namely α -wealth, which is spent to test hypotheses and “spent” upon discoveries (i.e., rejected null hypotheses). This method controls the modified FDR (mFDR) and is sensitive to test ordering. The modified false discovery rate (mFDR) is an alternative measure defined by the ratio between the expected number of false discoveries and the expected number of discoveries. To address these limitations, Javanmard and Montanari (2018) [9] proposed two alternative procedures, LOND (Levels based On Number of Discoveries) and LORD (Levels based On Recent Discoveries), to provide finite-sample FDR control under dependency structures. In LOND, the significance level at each step is scaled according to the number of previous rejections. The LORD procedure, on the other hand, updates the significance thresholds after each discovery, allowing for greater adaptability. Both methods depends on independent p-values. More recently, these methods have been further extended for improved power and relaxed dependence assumptions for the p-values [10,11,12].
For FDR control in grouped or structured hypotheses, such as genes grouped by biological pathways or drug compounds tested across cancer types, Ramdas and Barber (2019) [13] developed the p-filter, a method for multi-partitioned data settings where hypothesis testing occurs across distributed or federated environments. The p-filter maintains a dynamic filtration level that propagates across data partitions, enabling asynchronous hypothesis testing while preserving FDR guarantees. Ramdas et al. (2019) [14] developed the DAGGER method, which controls the FDR in hypothesis testing structured using directed acyclic graphs, where nodes represent hypotheses with parent–child dependencies. DAGGER processes hypotheses top-down, ensuring that a node is rejected only if its parents are rejected, and guarantees FDR control under the independence or positive or arbitrary dependence of p-values.
In many practical applications related to drug discovery, as we described in the Introduction, there is a need to control the FDR at multiple partition levels simultaneously in an online fashion; there is no existing method for this kind of setting. We aimed to address this gap by developing an FDR control method to control the group-level FDR defined based on multiple structures or partitions (multiple layers) in an online fashion. In our setting, not only do the hypotheses enter sequentially, but the group membership of the hypothesis entering each layer is also not predefined. We developed a novel framework that integrates online FDR control with multi-layer structured testing capabilities to provide theoretically guaranteed FDR control in multi-layer online testing environments. The main contributions of this paper are summarized below:
  • We present multi-layer α -investing, LOND, and LORD algorithms for controlling the FDR in online multiple-layer problems.
  • We provide theoretical guarantees for the multi-layer α -investing, LOND, and LORD methods.
  • We demonstrate the numerical FDR control’s performance and the power of our methods over an extensive range of simulation settings.
The rest of this paper is structured as follow. In Section 2, we provide a mathematical formulation of the online multi-layer hypothesis testing problem. In Section 3, we illustrate the multi-layer α -investing, LOND, and LORD algorithms. In Section 4, we show our main theoretical results, that the multi-layer α -investing method simultaneously controls the mFDR in different layers and the multi-layer LOND and LORD simultaneously control the FDR and mFDR in different layers. In Section 5, we demonstrate the performance of our methods with the results from extensive simulation studies.

2. Model Framework

Now we formulate the problem for more general scenarios. From now on, we use the following notations. Suppose we have a series of (unlimited) null hypotheses, H i 0 s, i N , and M partitions of the space N , denoted as A 1 m , , A G m m N for m = 1 , , M . We call these M different partitions “layers” of hypotheses. We can define the group-level null hypotheses in layer m at time t as
H A g m 0 ( t ) = j A g m [ t ] H j 0 , where g = 1 , , G m ,
where [t] = { 1 , 2 , , t } . We also define a group index for the ith hypothesis in the mth layer as g i m , i.e., i A g i m m .
In the online setting, H i 0 s come in sequentially, and the corresponding test statistics z i s and associated p-values p i s are also observed sequentially. For each H i 0 we need to make the decision δ i whether to reject it based on z i , p i , and the available history information; i.e., δ i = 1 if we reject H i 0 , and δ i = 0 if we accept H i 0 . At time t (t is discretized), we have three different types of history information:
  • All information, F t a = { z i , p i , δ i , i = 1 , , t } .
  • All decisions, F t d = { δ i , i = 1 , , t } .
  • A summary of decisions, for example, D t = i = 1 t δ i .
Now, we define decisions at the group level using decisions at the individual level. At the individual level, we define the rejection set at time t as S ^ ( t ) = { i : δ i = 1 } [ t ] . Since the decision whether to reject a group will change over time, we use δ g m ( t ) to denote the decision for group g ( g { 1 , , G m } ) in layer m at time t, where
δ g m ( t ) = 1 if S ^ ( t ) A g m and δ g m ( t ) = 0 otherwise .
So for the mth partition, the group-level selected sets at time t are
S ^ m ( t ) = { g { 1 , , G m } : δ g m ( t ) = 1 } .
Similarly, we denote the truth about H i at the individual level by θ i ; θ i = 0 when H i 0 should be accepted, while θ i = 1 when H i 0 should be rejected. We can obtain the true set S ( t ) = { i : θ i = 1 } [ t ] and
θ g m ( t ) = 1 if S ( t ) A g m and θ g m ( t ) = 0 otherwise .
For the mth partition, the group-level true null set at time t is S m ( t ) = { g { 1 , , G m } : θ g m ( t ) = 1 } .
Now we can define the time-dependent false discovery rate (FDR) and modified false discovery rate (mFDR) for each layer separately. For the mth layer, we have
FDR m ( t ) = E [ V m ( t ) R m ( t ) 1 ] and mFDR η m ( t ) = E V m ( t ) E R m ( t ) + η ,
where V m ( t ) = | S ^ m ( t ) ( S m ( t ) ) c | and R m = | S ^ m ( t ) | .

3. Algorithms

3.1. Multi-Layer α -Investing

In this section, we illustrate the generalized α -investing algorithm for the multi-layer setting.
A multi-layer α -investing procedure is an adaptive sequential hypothesis testing procedure. Suppose that the overall target FDR is α for every layer. For each time point, t ( t = 1 , , N where we allow N ), and partition layer, m ( m = 1 , , M ), we define the significance level α t m , the spending function ϕ t m , and the rewarding function ψ t m that satisfy the assumptions in Theorem 1 and Lemma 1 in Section 4 (a simple choice is an alpha-investing [8] procedure with α t m = α , ϕ t m = α 1 α , and ψ t m = ϕ t m + α ). Each hypothesis, H i 0 , will be rejected if its p-values, p i m s, pass the corresponding thresholds α i m s for all those ms for which a decision is still pending (that is, m s s.t. δ g i m m ( i 1 ) = 0 ). The wealth function W m ( t ) is initialized at α η and then updated under the following rule: at time t = i , if the hypothesis H i 0 is rejected, then we gain a reward, ψ i m , for all the layers tested in this round. In the meantime, no matter whether H i 0 is rejected, we spend ϕ i m for those layers that are tested in this round. In other words, we can write W m ( i ) as
W m ( i ) = W m ( i 1 ) + ψ i m ϕ i m , δ g i m m ( i 1 ) = 0 , δ g i m m ( i ) = 1 W m ( i 1 ) ϕ i m , δ g i m m ( i 1 ) = 0 , δ g i m m ( i ) = 0
This procedure simultaneously controls the mFDR in all the layers. We state the multi-layer α -investing algorithm in Algorithm 1.
Algorithm 1 Multi-layer α -investing.
  1:
Initialize δ g m = 0 and W m ( 0 ) = α η , for m = 1 , , M and all g { 1 , , G m } .
  2:
while min m W m ( i ) > 0 and i N  do
  3:
    δ i = 1
  4:
   for  m = 1 , , M  do
  5:
     if  δ g i m m = 0  then
  6:
        if  p i m α i m  then
  7:
           δ i = 0 , break
  8:
        end if
  9:
     end if
10:
   end for
11:
   if  δ i = 1  then
12:
     for  m = 1 , , M  do
13:
        if  δ g i m m = 0  then
14:
           δ g i m m = 1 and W m ( i ) = W m ( i 1 ) + ψ i m ϕ i m
15:
        end if
16:
     end for
17:
   else
18:
     for  m = 1 , , M  do
19:
        if  δ g i m m = 0  then
20:
           W m ( i ) = W m ( i 1 ) ϕ i m
21:
        end if
22:
     end for
23:
   end if
24:
    i = i + 1
25:
end while
Remark: The statistics p i m s are also allowed to just be individual p i s for all the layers or a combined score which is a function of all the p j s s.t. j i and g j m = g i m .

3.2. Multi-Layer LOND and LORD

In this section, we generalize the LOND and LORD algorithms to a multi-layer setting.
To extend the LOND algorithm to a multi-layer setting, we set the significance level for the hypothesis at time i in layer m to be β i m ( R m ( i 1 ) + 1 ) , where
β i m : i = 1 β i m = α for m = 1 , , M
are M predefined infinite sequences that sum to 1, and R m ( i 1 ) = | S ^ m ( i 1 ) | . Each hypothesis, H i 0 , will be rejected if its p-value, p i m , passes the threshold β i m ( R m ( i 1 ) + 1 ) for all those ms for which a decision is still pending ( m s s.t. δ g i m m ( i 1 ) = 0 ). We state the multi-layer LOND algorithm in Algorithm 2, and we will show that the multi-layer LOND algorithm can simultaneously control the multi-layer FDR and mFDR at level α in the next section.
Remark: Since in layer m, we did not need to test a hypothesis at every time point, t = 1 , , i , in fact, we only performed κ i m tests, where κ i m = i j = 1 i δ g j m m ( i ) + j = 1 G m δ j m ( i ) ; we can improve the method by setting the significance level for the hypothesis at time i in layer m to be β κ i m m ( R m + 1 ) . By doing this, the power is increased and the multi-layer FDRs are still controlled. We will show this improvement through a simulation in Section 4 but omit the theoretical proof of it.
To extend the LORD algorithm to the multi-layer setting, we set M predefined sequences, { β j m : j = 1 β j m = α } . For the hypothesis at time i in layer m, we use the significance level α i m = β ζ i m m , where ζ i m = κ i m κ τ i m m + 1 , τ i m = max { j < i : δ g j m m ( j 1 ) = 0 , δ g j m m ( j ) = 1 } , and κ i m = i j = 1 i δ g j m m ( i ) + j = 1 G m δ j m ( i ) . Define ζ i m = κ i m κ τ i m m + 1 . We can show that the modified LORD algorithm simultaneously controls the multi-layer FDR and mFDR at level α . We state the multi-layer LORD algorithm in Algorithm 3.
Algorithm 2 Multi-layer LOND.
  1:
Select sequences j = 1 β j m = α and initial R m = 0 , and δ g m = 0 , for m = 1 , , M and all g { 1 , , G m } .
  2:
while  i N  do
  3:
    δ i = 1
  4:
   for  m = 1 , , M  do
  5:
     if  δ g i m m = 0  then
  6:
        if  p i m β i m ( R m + 1 )  then
  7:
           δ i = 0 , break
  8:
        end if
  9:
     end if
10:
   end for
11:
   if  δ i = 1  then
12:
     for  m = 1 , , M  do
13:
        if  δ g i m m = 0  then
14:
           R m = R m + 1 , δ g i m m = 1
15:
        end if
16:
     end for
17:
   end if
18:
    i = i + 1
19:
end while
Algorithm 3 Multi-layer LORD.
  1:
Select sequences g = 1 β g m = α and initial ζ m = 1
  2:
while  i N  do
  3:
    δ i = 1
  4:
   for  m = 1 , , M  do
  5:
     if  δ g i m m = 0  then
  6:
        if  p i m β ζ m m  then
  7:
           δ i = 0 , break
  8:
        end if
  9:
     end if
10:
   end for
11:
   if  δ i = 1  then
12:
     for  m = 1 , , M  do
13:
        if  δ g i m m = 0  then
14:
           ζ m = 1 and δ g i m m = 1
15:
        end if
16:
     end for
17:
   else
18:
     for  m = 1 , , M  do
19:
        if  δ g i m m = 0  then
20:
           ζ m = ζ m + 1
21:
        end if
22:
     end for
23:
   end if
24:
    i = i + 1
25:
end while

4. Theoretical Results

We will show that multi-layer α -investing method simultaneously controls the mFDR in different layers and the multi-layer LOND and LORD simultaneously control both the FDR and mFDR in different layers. First we introduce one preliminary assumption; this assumption is weaker than the independence assumption across the tests.
Assumption 1.
m { 1 , , M } , and for all the layers that need to be tested at time j,
θ j = 0 : P θ ( δ g j m m ( j ) = 1 | F j 1 ) α j m , θ j = 1 : P θ ( δ g j m m ( j ) = 1 | F j 1 ) ρ j m .
Remark: This assumption is a simple extension of Assumption 2.2 in the paper of Aharoni and Rosset (2014) [15] regarding the generalized α -investing method. It requires that the level and power traits of a test hold even given the history of rejections. This is somewhat weaker than requiring independence. In particular, ρ j m 1 , m { 1 , , M } is the maximal probability of rejection for any parameter value in the alternative hypothesis, which quantifies the “best power”.
Next we state the theoretical m F D R control guarantee for the multi-layer generalized α -investing procedure.
Theorem 1.
Under Assumption 1, if 0 ψ j m min ( ϕ j m ρ j m + α , ϕ j m α j m + α + 1 ) , then a multi-layer generalized α-investing procedure simultaneously controls the m F D R η m for all the layers, m = 1 , , M .
The proof of Theorem 1 is postponed to Appendix A. One key step is to establish a submartingale, as stated in the following lemma.
Lemma 1.
Under Assumption 1 and the multi-layer α-investing procedure, the stochastic processes A m ( j ) = α R m ( j ) V m ( j ) + α η W m ( j ) , m = 1 , , M are submartingales with respect to F j .
Next, we state the F D R and m F D R control guarantees for the multi-layer LOND and LORD methods.
Theorem 2.
Suppose that conditional on R m ( i 1 ) = | S ^ m ( i 1 ) | , m = 1 , , M , we have
θ i Θ , m = 1 , , M ,
P θ i = 0 δ g i m m ( i ) = 1 | R m ( i 1 ) , δ g i m m ( i 1 ) = 0 E α i m | R m ( i 1 ) , δ g i m m ( i 1 ) = 0 ,
where α i m = β i m ( R m ( i 1 ) + 1 ) and δ i m are given by the LOND algorithm. Then this rule simultaneously controls the multi-layer FDRs and mFDRs at levels less than or equal to α; i.e.,
n 1 , m = 1 , , M , FDR m ( n ) α and mFDR m ( n ) α .
Theorem 3.
Suppose that conditional on the history of decisions in layer m, F t m = σ δ g i m m ( i ) , i = 1 , , t , we have θ Θ , m = 1 , , M that need to be tested at time i:
P θ i = 0 δ g i m m ( i ) = 1 F t m E α i m F t m .
Then, the LORD rule simultaneously controls the multi-layer mFDR to ensure it is less than or equal to α; i.e., mFDR 1 m ( n ) α for all n 1 . Further, it simultaneously controls the multi-layer FDR at every discovery. More specifically, letting d k m be the time of the kth discovery in layer m, we have the following for all k 1 :
sup θ Θ E θ FDP m ( d k m ) · I ( d k m < ) α , m = 1 , , M .
All the proofs of Lemma 1 and Theorems 1–3 are in Web Appendix A.

5. Numerical Studies

We conducted extensive simulations to evaluate the performance of our proposed multi-layer methods in various scenarios. We simulated data from a two-layer model with different settings to evaluate the proposed methods. For each setting, we compared the FDR and mFDR control and power at both the group and individual levels. For all simulations, we set the target FDR level at α = 0.1 and ran 100 repeats.

5.1. Simulation of Group Structures

We considered three types of group structures: the first two were balanced, while the third was unbalanced. For the balanced structure, we denoted G as the number of groups and the number of hypotheses per group as n. Therefore, the total number of features was N = n G . The three types of group structures are described below:
  • Balanced Block Structure: The features came in group by group; i.e., all the features from the first groups came in, then all the features from the second group followed, and so on.
  • Balanced Interleaved Structure: The first feature was from group 1, the second feature was from group 2,⋯, the nth feature was from group n, the ( n + 1 ) th feature was from group 1, and so on.
  • Unbalanced Structure: The features were randomly assigned to be in either the current group or a new group by some random mechanism. To be more specific, we fixed the total number of features as N and denoted the membership of the tth feature as G ( t ) , letting
    P ( G ( t ) = g | G ( t 1 ) = g ) = 1 p 1 , P ( G ( t ) = g | G ( t 1 ) = g ) = p 1 G 1 ,
    where g g { 1 , , G } .

5.2. Simulation of True Signals

We considered three signal structures:
  • Fixed Pattern: The first s % of groups were true groups. Within the true groups, the first k % of features were the true features.
  • Random Pattern: The true groups and true signals were randomly sampled. We first randomly sampled s % of the total groups to be true groups. Then within each true group, we randomly sampled k % of the features as the true features.
  • Markov Pattern: The signals were from a hidden Markov model. We considered a Markov model with two hidden states: a stationary state and an eruption state. In the stationary state, θ i B i n o m i a l ( N , 0.5 ) ; in the erruption state, P ( θ i = 0 | θ i 1 = 0 ) = P ( θ i = 1 | θ i 1 = 1 ) = 0.9 . At each step, the probability that the signal remained in the same state as the previous step was 0.9, and it had a probability of 0.1 of changing to the other state.

5.3. Simulation of Signal Strength

We considered three types of signal strength settings:
  • Constant Strength: The signal strengths were equal; the p-values for the true signals were from a two-sided test with z-statistics from N ( 1.5 β , 1 ) .
  • Increasing Strength: The signal strengths were increasing; the p-values were from a two-sided test with z-statistics from N ( β t , 1 ) , where β t = β ( 1 + t / t T O L ) . Here t is the number of true signals until that moment, and t T O L is the total number of true signals in the simulation.
  • Decreasing Strength: The signal strengths were decreasing; the p-values were from a two-sided test with z-statistics from N ( β t , 1 ) , where β t = β ( 2 t / t T O L ) , with t and t T O L defined the same as in (2).
All the simulation designs satisfied Assumption 1 by construction. Specifically, under the null hypothesis ( θ i = 0 ), p-values were generated independently from the Uniform ( 0 , 1 ) distribution. This guaranteed that the probability of rejection at each step was less than or equal to the target significance level α i m (i.e., P ( p j α j m ) = α j m ), satisfying the first condition of Assumption 1. Under the alternative hypothesis ( θ i = 1 ), p-values were generated from two-sided tests using normally distributed test statistics with non-zero means. These distributions were stochastically smaller than the null distribution, which led to a higher probability of rejection. However, this probability was still finite and bounded above by one, ensuring that the second condition of Assumption 1 was satisfied with a constant ρ i m 1 . Moreover, our simulation generated data sequentially and independently, making the conditional probability bounds in Assumption 1 valid under the filtration F i 1 . Therefore, both conditions of Assumption 1 were met in all the simulation settings, ensuring that the theoretical guarantees in Theorems 1–3 were applicable.
We compared seven methods for each setting: the original α -investing (GAI), original LORD (LORD), original LOND (LOND), multi-layer α -investing (ml-GAI), multi-layer LORD (ml-LORD), multi-layer LOND (ml-LOND), and modified multi-layer LOND (ml-LOND_m) methods.

5.4. Simulation Results

Our baseline configuration used the following parameters: 20% true groups ( s = 20 ), 50% true signals within true groups ( k = 100 ), 20 groups in total ( G = 20 ), 10 features per group ( n = 10 ), a Balanced Block Structure, Fixed Pattern, and a Constant Strength. We varied the signal strength parameter β to examine how the methods performed across different effect sizes.
Figure 1 shows the power, FDR, and mFDR for both the individual and group layers under different group structures. For the Balanced Block Structure, all the methods successfully maintained FDR and mFDR control at both the individual and group levels except GAI. For GAI, both the group-level FDR and mFDR exceeded 0.4 when β exceeded 2.0. The ml-LORD and LORD methods exhibited the highest power at both levels. As β increased beyond 4.5, the power of all the methods approached 1.0. For the Balanced Interleaved Structure, we observed that the LORD methods (both LORD and ml-LORD) significantly outperformed all other methods in terms of power at both the individual and group levels. Importantly, all the multi-layer methods maintained FDR and mFDR control at both levels. The Unbalanced Structure demonstrated results largely consistent with those of the Balanced Interleaved Structure. However, a slight exceedance of the 0.1 threshold was observed for the LORD procedure when β reached 0.5.
Figure 2 illustrates the performance metrics across different true signal patterns. All the methods controlled both the individual and group FDR and mFDR in the Fixed Pattern and Markov Pattern. However, LORD failed to control both the group-level mFDR and FDR for a Random Pattern when β exceeded 3. In terms of power, both LORD and ml-LORD obviously had better performance than the other methods in these three patterns.
Figure 3 shows the impact of different signal strength patterns on the power and FDR control. The GAI method failed to control both the group-level mFDR and FDR across the three patterns. In contrast, all the multi-layer methods successfully controlled both the mFDR and FDR. Regarding power, all the versions of the LOND procedure (LOND, ml-LOND, and ml-LOND_m) exhibited lower power compared to the α -investing (GAI, ml-GAI) and LORD (LORD, ml-LORD) methods. Among the three patterns, the ml-LORD method demonstrated the best overall performance, as it consistently achieved the highest power while maintaining FDR control. Furthermore, Figure 4 shows the results with a reduced number of true signals ( k = 50 ). Most methods continued to control both the FDR and mFDR. However, a slight exceedance of the 0.1 threshold was observed for the LORD method when the effect size β exceeded 3.0.
Our extensive simulation studies highlighted the advantages of using multi-layer FDR control methods over the original ones. Under a Balanced Block Structure, the GAI procedure failed to maintain group-level FDR control. In more complex scenarios, such as the Random Pattern setting, the LORD method also failed to adequately control the group-level FDR. In contrast, all the multi-layer methods consistently maintained valid control of both the FDR and mFDR across all the settings. In terms of power, the modified multi-layer LORD method slightly outperformed the original LORD approach while maintaining comparable FDR control.
All the simulation results are consistent with theoretical results showing that the multi-layer α -investing method simultaneously controls the mFDR in multiple layers (Theorem 1) and the multi-layer LOND and LORD procedures achieve the simultaneous control of both the FDR and mFDR across layers (Theorems 2 and 3). These findings demonstrate that multi-layer methods not only ensure more reliable FDR/mFDR control, particularly at the group level, but also deliver better statistical power compared to the original methods.

6. Discussion

In summary, we propose three methods that simultaneously control the mFDR or FDR for multiple layers of groups (partitions) of hypotheses in an online manner. The mFDR or FDR controls were theoretically guaranteed and demonstrated with simulations. These methods fill an important gap in drug discovery when we need to control the FDR in drug screening for multiple layers of features (i.e., siRNAs, nanocapsules, target cell lines, etc.). For example, when screening siRNA–nanocapsule combinations, our methods allow researchers to make statements about both the efficacy of specific combinations and the general performance of specific siRNAs or nanocapsules across all the combinations tested, which can accelerate the identification of promising candidates and reduce false leads. In addition, the proposed algorithms for the methods are also computationally efficient. The computational overhead of our multi-layer methods is minimal compared to that of the original methods. This makes our methods highly scalable, even for large-scale screening applications involving thousands of hypotheses across multiple layers.
A challenge we face is that the power of these methods reduces as the number of layers increases. We can potentially increase the power by adding more assumptions and estimating the pattern of the true signals and the correlation between the p-values more accurately. The modified multi-layer LOND method ( ml - LOND m ) shows an effective example. It had better power performance than the multi-layer LOND method in the simulation; developing such methods with theoretical guarantees will be the subject of our future work.
Another limitation lies in the assumption of independence or weak dependence among p-values across different layers. In real-world settings, particularly in high-throughput biological studies or multi-modal drug screenings, their characteristics are often correlated. Our multi-layer online frameworks do not assume independence across multiple characteristics because the key assumptions are based on the conditional distribution conditioned on the history information. The construction of statistics for such scenarios is a practical challenge. Some extensions of the conditional permutation test [16] can be explored as a future direction.

Author Contributions

Conceptualization, R.D.; methodology, R.D.; software, R.W.; validation, R.W.; formal analysis, R.W.; investigation, R.W.; resources, R.D.; data curation, R.W.; writing—original draft preparation, R.D. and R.W.; writing—review and editing, R.D. and R.W.; visualization, R.W.; supervision, R.D.; project administration, R.D.; funding acquisition, R.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Institute of General Medical Sciences under grant U54 GM115458.

Data Availability Statement

The R code and simulation data are available upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Technical Proof

Proof of Lemma 1.
Denote E θ j 1 = E θ ( · | F j 1 ) . The key here is that ϕ j m , ψ j m , and α j m are measurable with respect to F j 1 . For a layer, m, that we need to test at time j, we can deduce that δ g j m m ( j 1 ) = 0 . If θ j = 0 and θ g j m m ( j 1 ) = 0 , then θ g j m m ( j ) = 0 and V m ( j ) V m ( j 1 ) = δ g j m m ( j ) , and we have
E θ j 1 ( A m ( j ) A m ( j 1 ) ) = ( 1 α + ψ j m ) E θ j 1 ( δ g j m m ( j ) ) + ϕ j m ( 1 α + ψ j m ) α j m + ϕ j m ( 1 α + ϕ j m / α j m + α 1 ) α j m + ϕ j m = 0
If θ j = 0 and θ g j m m ( j 1 ) = 1 , then θ g j m m ( j ) = 1 and V m ( j ) V m ( j 1 ) = 0 , and we have
E θ j 1 ( A m ( j ) A m ( j 1 ) ) = ( α + ψ j m ) E θ j 1 ( δ g j m m ( j ) ) + ϕ j m ( α + ψ j m ) α j m + ϕ j m ( α + ϕ j m / α j m + α 1 ) α j m + ϕ j m = α j m 0
If θ j = 1 , then θ g j m m ( j ) = 1 and V m ( j ) V m ( j 1 ) = 0 , and we have
E θ j 1 ( A m ( j ) A m ( j 1 ) ) = ( α + ψ j m ) E θ j 1 ( δ g j m m ( j ) ) + ϕ j m ( α + ψ j m ) ρ j m + ϕ j m ( α + ϕ j m / ρ j m + α ) ρ j m + ϕ j m = 0
This finishes the proof that A m ( j ) is a submartingale. □
Proof of Theorem 1.
Since R m ( 0 ) = 0 , V m ( 0 ) = 0 , and W m ( 0 ) = α η , we have A m ( 0 ) = 0 . Based on Lemma 1, we have E θ ( A m ( t ) ) = α E θ R m ( t ) E θ V m ( t ) + α η E θ W m ( t ) 0 . Since W m ( t ) 0 , E θ ( A m ( t ) ) = α E θ R m ( t ) E θ V m ( t ) + α η 0 , which leads to m F D R η m = E θ V m ( t ) E θ R m ( t ) + η α as long as t is a stopping time. □
Proof of Theorem 2.
We first show that F D R m ( n ) α . Write
F D R m ( n ) = E θ V m ( n ) R m ( n ) 1 E θ j = 1 n δ g j m m ( j ) R m ( n ) 1 = E θ j = 1 n δ g j m m ( j ) α j m · α j m R m ( n ) 1
Since R m ( n ) 1 R m ( j ) 1 , we have
α j m R m ( n ) 1 α j m R m ( j 1 ) + 1 = β j m
and by assumption, we know that θ Θ and m = 1 , , M , and we have
E θ ( δ g j m m ( j ) α j m | R m ( j 1 ) , δ g i m m ( i 1 ) = 0 ) 0 .
So we have
E θ δ g j m m ( j ) α j m E θ δ g j m m ( j ) α j m | R m ( j 1 ) , δ g i m m ( i 1 ) = 0 E ( 1 ) = 1
since α j m is measurable with respect to R m ( j 1 ) . Summing over n, we obtain F D R m ( n ) j = 1 β j m = α .
Next we prove that m F D R m ( n ) α . Since
E θ ( δ g i m m α j m ) = E [ E θ ( δ g i m m α j m | R m ( j 1 ) , δ g i m m ( i 1 ) = 0 ) ] 0 .
Summing over n, we have
E θ ( V m ( n ) ) i = 1 n E θ α i m = i = 1 n E θ ( β i m R m ( i 1 ) + 1 ) = i = 1 n E θ β i m j = 1 i 1 δ j m + i = 1 n β i m = j = 1 n 1 i = j + 1 n β i m E θ ( δ i m ) + i = 1 n β i m i = 1 n β i m ( E θ ( R m ( n ) ) + 1 )
So m F D R m ( n ) i = 1 n β i m α . □
Proof of Theorem 3.
Because the sum of significance level α i m between two consecutive discoveries is bounded by α ,
E θ ( V m ( n ) ) i = 1 n E θ α i m α E θ ( R m ( n ) + 1 )
This leads to the result that m F D R 1 m ( n ) α .
Next, we prove the result for the FDR. Denote X i m = 1 if d i m < and the ith discovery is false, X i m = 0 if d i m < and the ith discovery is true, and X i m = 2 otherwise. First we have
E θ ( X i m I ( d i m < ) | F d i 1 m ) l = d i 1 m + 1 P θ l = 0 ( δ g l m m ( l ) = 1 , d i m = l | F d i 1 m )
Given d i 1 m , we know that ζ i m = κ i m κ d i 1 m m + 1 . This implies that for any l = d i 1 m + 1 , , , ζ l m = ζ l 1 m is equivalent to κ l m = κ l 1 m , which means that l will not be tested in the layer m and it is impossible for d i m = l . So we have
l = d i 1 m + 1 P θ l = 0 ( δ g l m m ( l ) = 1 , d i m = l | F d i 1 m m ) = l = d i 1 m + 1 P θ l = 0 ( δ g l m m ( l ) = 1 , d i m = l | F d i 1 m m ) I ( ζ i m ζ i 1 m ) l = d i 1 m + 1 α i m I ( ζ i m ζ i 1 m ) r = 1 β r m = α
The last part holds since α i m = β ζ i m m .
E θ ( F D P ( d k m ) I ( d i m < ) ) = E V m ( d k m ) k I ( d i m < ) = 1 k i = 1 k E ( X i m I ( d i m < ) ) 1 k i = 1 k E [ E ( X i m I ( d i m < ) ) | F d i 1 m ] α

References

  1. Deore, A.B.; Dhumane, J.R.; Wagh, R.; Sonawane, R. The stages of drug discovery and development process. Asian J. Pharm. Res. Dev. 2019, 7, 62–67. [Google Scholar] [CrossRef]
  2. Lee, J.M.; Yoon, T.J.; Cho, Y.S. Recent developments in nanoparticle-based siRNA delivery for cancer therapy. BioMed Res. Int. 2013, 2013, 782041. [Google Scholar] [CrossRef] [PubMed]
  3. Ebrahimi, N.; Manavi, M.S.; Nazari, A.; Momayezi, A.; Faghihkhorasani, F.; Abdulwahid, A.H.R.R.; Rezaei-Tazangi, F.; Kavei, M.; Rezaei, R.; Mobarak, H.; et al. Nano-scale delivery systems for siRNA delivery in cancer therapy: New era of gene therapy empowered by nanotechnology. Environ. Res. 2023, 239, 117263. [Google Scholar] [CrossRef] [PubMed]
  4. Zhang, X.D.; Kuan, P.F.; Ferrer, M.; Shu, X.; Liu, Y.C.; Gates, A.T.; Kunapuli, P.; Stec, E.M.; Xu, M.; Marine, S.D.; et al. Hit selection with false discovery rate control in genome-scale RNAi screens. Nucleic Acids Res. 2008, 36, 4667–4679. [Google Scholar] [CrossRef] [PubMed]
  5. Taddy, M.; Gardner, M.; Chen, L.; Draper, D. Heterogeneous treatment effects in digital experimentation. arXiv 2015, arXiv:1412.8563. [Google Scholar]
  6. Xie, Y.; Chen, N.; Shi, X. False discovery rate controlled heterogeneous treatment effect detection for online controlled experiments. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 876–885. [Google Scholar]
  7. Goeman, J.J.; Bühlmann, P. Analyzing gene expression data in terms of gene sets: Methodological issues. Bioinformatics 2007, 23, 980–987. [Google Scholar] [CrossRef] [PubMed]
  8. Foster, D.P.; Stine, R.A. α-investing: A procedure for sequential control of expected false discoveries. J. R. Stat. Soc. Ser. B Stat. Methodol. 2008, 70, 429–444. [Google Scholar] [CrossRef]
  9. Javanmard, A.; Montanari, A. Online rules for control of false discovery rate and false discovery exceedance. Ann. Stat. 2018, 46, 526–554. [Google Scholar] [CrossRef]
  10. Ramdas, A.; Zrnic, T.; Wainwright, M.; Jordan, M. SAFFRON: An adaptive algorithm for online control of the false discovery rate. In Proceedings of the International Conference on Machine Learning; PMLR: Stockholm, Sweden, 2018; pp. 4286–4294. [Google Scholar]
  11. Tian, J.; Ramdas, A. Online control of the familywise error rate. Stat. Methods Med. Res. 2021, 30, 976–993. [Google Scholar] [CrossRef] [PubMed]
  12. Zehetmayer, S.; Posch, M.; Koenig, F. Online control of the False Discovery Rate in group-sequential platform trials. Stat. Methods Med. Res. 2022, 31, 2470–2485. [Google Scholar] [CrossRef] [PubMed]
  13. Ramdas, A.K.; Barber, R.F.; Wainwright, M.J.; Jordan, M.I. A unified treatment of multiple testing with prior knowledge using the p-filter. Ann. Stat. 2019, 47, 2790–2821. [Google Scholar] [CrossRef]
  14. Ramdas, A.; Chen, J.; Wainwright, M.J.; Jordan, M.I. A sequential algorithm for false discovery rate control on directed acyclic graphs. Biometrika 2019, 106, 69–86. [Google Scholar] [CrossRef]
  15. Aharoni, E.; Rosset, S. Generalized α-investing: Definitions, optimality results and application to public databases. J. R. Stat. Soc. Ser. B Stat. Methodol. 2014, 76, 771–794. [Google Scholar] [CrossRef]
  16. Berrett, T.B.; Wang, Y.; Barber, R.F.; Samworth, R.J. The conditional permutation test for independence while controlling for confounders. J. R. Stat. Soc. Ser. B Stat. Methodol. 2020, 82, 175–197. [Google Scholar] [CrossRef]
Figure 1. Power, FDR, and mFDR for individual (upper panel) and group (lower panel) layers with s = 20, k = 100, n = 10, G = 20, Fixed Pattern, and Constant Strength for different group structures: Balanced Block Structure (Left), Balanced Interleaved Structure (Middle), and Unbalanced Structure (Right).
Figure 1. Power, FDR, and mFDR for individual (upper panel) and group (lower panel) layers with s = 20, k = 100, n = 10, G = 20, Fixed Pattern, and Constant Strength for different group structures: Balanced Block Structure (Left), Balanced Interleaved Structure (Middle), and Unbalanced Structure (Right).
Mathematics 13 01937 g001
Figure 2. Power, FDR, and mFDR for individual (upper panel) and group (lower panel) layers with s = 20, k = 100, n = 10, G = 20, Balanced Interleaved Structure, and Constant Strength for different signal structures: Fixed Pattern (Left), Random Pattern (Middle), and Markov pattern (Right).
Figure 2. Power, FDR, and mFDR for individual (upper panel) and group (lower panel) layers with s = 20, k = 100, n = 10, G = 20, Balanced Interleaved Structure, and Constant Strength for different signal structures: Fixed Pattern (Left), Random Pattern (Middle), and Markov pattern (Right).
Mathematics 13 01937 g002
Figure 3. Power, FDR, and mFDR for individual (upper panel) and group (lower panel) layers with s = 20, k = 100, n = 10, G = 20, Balanced Block Structure, and Fixed Pattern for different signal strengths: Constant Strength (Left), Increasing Strength (Middle), and Decreasing Strength (Right).
Figure 3. Power, FDR, and mFDR for individual (upper panel) and group (lower panel) layers with s = 20, k = 100, n = 10, G = 20, Balanced Block Structure, and Fixed Pattern for different signal strengths: Constant Strength (Left), Increasing Strength (Middle), and Decreasing Strength (Right).
Mathematics 13 01937 g003
Figure 4. Power, FDR, and mFDR for individual (upper panel) and group (lower panel) layers with s = 20, k = 50, n = 10, G = 20, Bbalanced Interleaved Structure, and Random Pattern for different signal strengths: Constant Strength (Left), Increasing Strength (Middle), and Decreasing Strength (Right).
Figure 4. Power, FDR, and mFDR for individual (upper panel) and group (lower panel) layers with s = 20, k = 50, n = 10, G = 20, Bbalanced Interleaved Structure, and Random Pattern for different signal strengths: Constant Strength (Left), Increasing Strength (Middle), and Decreasing Strength (Right).
Mathematics 13 01937 g004
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, R.; Dai, R. Online Multi-Layer FDR Control. Mathematics 2025, 13, 1937. https://doi.org/10.3390/math13121937

AMA Style

Wang R, Dai R. Online Multi-Layer FDR Control. Mathematics. 2025; 13(12):1937. https://doi.org/10.3390/math13121937

Chicago/Turabian Style

Wang, Runqiu, and Ran Dai. 2025. "Online Multi-Layer FDR Control" Mathematics 13, no. 12: 1937. https://doi.org/10.3390/math13121937

APA Style

Wang, R., & Dai, R. (2025). Online Multi-Layer FDR Control. Mathematics, 13(12), 1937. https://doi.org/10.3390/math13121937

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop