Next Article in Journal
Conscious Neighborhood-Based Jellyfish Search Optimizer for Solving Optimal Power Flow Problems
Previous Article in Journal
Hybrid Adaptive MPC with Edge AI for 6-DoF Industrial Robotic Manipulators
Previous Article in Special Issue
Inflation of Familywise Error Rate in Treatment Efficacy Testing Due to the Reallocation of Significance Levels Based on Safety Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Subset Selection with Curtailment Among Treatments with Two Binary Endpoints in Comparison with a Control

1
Department of Mathematics, Syracuse University, Syracuse, NY 13244, USA
2
Department of Mathematics and Statistics, University of North Florida, Jacksonville, FL 32224, USA
3
Department of Mathematics, Le Moyne College, Syracuse, NY 13214, USA
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(19), 3067; https://doi.org/10.3390/math13193067
Submission received: 30 June 2025 / Revised: 23 August 2025 / Accepted: 3 September 2025 / Published: 24 September 2025
(This article belongs to the Special Issue Sequential Sampling Methods for Statistical Inference)

Abstract

We propose a sequential procedure with a closed and adaptive structure. It selects a subset of size t ( > 0 ) from k ( t ) treatments in such a way that any treatment superior to the control is guaranteed to be included. All the experimental treatments and the control are assumed to produce two binary endpoints, and the procedure is based on those two binary endpoints. A treatment is considered superior if both its endpoints are larger than those of the control. While responses across treatments are assumed to be independent, dependence between endpoints within each treatment is allowed and modeled via an odds ratio. The proposed procedure comprises explicit sampling, stopping, and decision rules. We demonstrate that, for any sample size n and parameter configuration, the probability of correct selection remains unchanged when switching from the fixed-sample-size procedure to the sequential one. We use the bivariate binomial and multinomial distributions in the computation and derive design parameters under three scenarios: (i) independent endpoints, (ii) dependent endpoints with known association, and (iii) dependent endpoints with unknown association. We provide tables with the sample size savings achieved by the proposed procedure compared to its fixed-sample-size counterpart. Examples are given to illustrate the procedure.

1. Background, Introduction, and Motivation

As both efficacy and safety are crucial in Phase II evaluations, we propose a closed adaptive sequential procedure to compare k ( > 1 ) experimental treatments with a control group. Each treatment, including the control, is assessed using two binary endpoints: one for efficacy and one for safety. An experimental treatment is considered superior to the control if it demonstrates higher success probabilities in both endpoints. Let the control treatment be denoted by π 0 , and the k experimental treatments by π 1 , π 2 , , π k . The outcomes for each treatment consist of two binary endpoints that are modeled marginally as Bernoulli random variables with unknown success probabilities. Specifically, for the control, the success probabilities are denoted by ( p 0 , 1 , p 0 , 2 ) , and for each experimental treatment π i ( i = 1 , , k ) , the corresponding success probabilities are ( p i , 1 , p i , 2 ) . Comparisons are made between each experimental treatment and the control by evaluating whether p i , 1 > p 0 , 1 and p i , 2 > p 0 , 2 . The proposed procedure incorporates curtailment, allowing for early termination of sampling when sufficient evidence has been gathered to make a decision, thus potentially stopping before reaching the maximum sample size N. The term closed highlights the existence of an upper bound N, whereas adaptive indicates that continuation decisions are guided by interim outcomes.
In medical and pharmaceutical research, a common statistical problem involves selecting the best among k ( k > 1 ) treatments, each yielding Bernoulli outcomes. Classical methods addressing this issue include the fixed-sample Indifference Zone approach of [1], and the Subset Selection strategy of [2], which aims to identify a group likely to contain the best treatment. Later studies by [3] and [4] further explored this framework for comparisons involving a control. A notable limitation of these approaches is their exclusive focus on single binary endpoints under non-adaptive, fixed-sample designs.
In contrast to earlier fixed-sample approaches, we consider a curtailed selection procedure involving two Bernoulli endpoints. Under this design, each treatment is sampled sequentially, up to a pre-specified maximum of N observations. Sampling for a treatment is terminated early if there is sufficient evidence that it is no longer a viable candidate, or once the sample limit N is reached. The formal definition of a “contending treatment” is given in Procedure R (Section 3). This early stopping strategy can substantially reduce the total number of observations required for treatment selection.
Curtailment has previously been applied in clinical trials with Bernoulli outcomes, particularly in the context of hypothesis testing ([5]) and treatment selection ([6,7,8]). However, these studies have largely focused on a single Bernoulli endpoint. In this work, we extend the notion of curtailment to the selection problem involving two Bernoulli endpoints. Related investigations include [9] for normally distributed outcomes, and [10,11,12] for binary endpoint designs.
Recent developments in ranking and selection (R&S) have highlighted the importance of incorporating covariates into selection procedures to support personalized decision-making. Ref. [13] introduced a two-stage R&S procedure with covariates (R&S-C) based on a linear regression framework, addressing scenarios where the mean performance of treatments varies systematically with patient-specific covariates. Their approach ensures a predefined average probability of correct selection across the covariate space, enabling tailored treatment recommendations in heterogeneous populations. Although their framework differs from ours by considering a single continuous endpoint and not incorporating curtailment, their emphasis on individual-level heterogeneity underscores the broader need for flexible experimental designs that adapt to diverse patient characteristics.
In addition, Bayesian sequential approaches such as the Knowledge Gradient (KG) method have also been proposed to incorporate covariates into ranking and selection problems ([14]). These methods provide theoretical consistency guarantees and computationally efficient decision-making frameworks, which are particularly valuable when dealing with large covariate spaces or high-dimensional decision scenarios. Nevertheless, Bayesian methods often require stronger assumptions and more complex computations than frequentist procedures. Our study, while not directly modeling covariates, provides a simpler, frequentist alternative designed specifically for dual binary endpoints that facilitates easier implementation in clinical trial settings.
More recently, ref. [15] discussed a curtailed procedure for subset selection involving two binary Bernoulli endpoints. However, their approach compares each experimental treatment to a well-established standard treatment. This design is most appropriate when a widely accepted reference treatment exists. In contrast, our procedure compares new treatments against a control treatment, which may or may not be a recognized standard.
There are many situations in which our approach is more applicable. For example, in the absence of a universally accepted standard treatment—such as when placebo is the only baseline option—it becomes necessary to evaluate new treatments in relation to a control. Similarly, even when a standard treatment exists, it may have been validated only in limited populations (e.g., specific age groups, races, or genders). In such cases, it is important to assess whether the standard treatment continues to perform well in broader or different patient populations. Our design, which explicitly includes the control treatment in the experiment, enables such comparisons and provides a flexible and inclusive framework for decision-making.
This study tackles the selection problem involving two binary endpoints through a subset selection framework. We propose a curtailed, closed sequential sampling design, where the total number of observations obtained from each of the k + 1 candidate populations (including the control) is a bounded random variable. We assume that treatment effects manifest quickly relative to the overall trial duration. The proposed procedure is constructed with reference to a fixed-sample-size method, which will be detailed in later sections. Our results show that the sequential design achieves the same probability of correctly identifying superior treatments as the fixed-sample method, while requiring fewer samples from inferior alternatives. Section 2 sets forth the modeling assumptions, trial goals, and statistical criteria for studies involving dual binary endpoints. Section 3 introduces the fixed-sample approach used as the basis for comparison in evaluating the curtailed procedure.
In Section 4, we introduce a curtailed sequential selection method that fulfills the objective of this study. It is shown that the proposed design achieves the same probability of correct selection as the corresponding fixed-sample-size method across the entire parameter space. Section 5 compares the expected sample sizes between the curtailed and non-curtailed versions of the procedure. In Section 6, we present two numerical examples that demonstrate the implementation of the proposed approach. Lastly, Section 7 concludes the paper with final remarks.

2. Assumptions, Goal, and Probability Requirements

Assumptions: Assume that n independent subjects receive a treatment, and two binary outcomes—typically interpreted as therapeutic efficacy (“response”) and safety (“nontoxicity”)—are recorded for each individual. Following the notation introduced by [11], let X i j denote the number of subjects experiencing outcome i on the first endpoint and outcome j on the second, where i , j { 1 , 2 } , with 1 indicating “success” and 2 indicating “failure.” The observed data can be organized into a 2 × 2 contingency table (see Table 1).
We assume that the random vector X = ( X 11 , X 12 , X 21 , X 22 ) follows a multinomial distribution with probabilities P = ( p 11 , p 12 , p 21 , p 22 ) , where:
  • p 11 is the probability of success on both endpoints,
  • p 12 is the probability of success on endpoint 1 and failure on endpoint 2,
  • p 21 is the probability of failure on endpoint 1 and success on endpoint 2,
  • p 22 is the probability of failure on both endpoints.
The use of a multinomial distribution to model two dependent binary endpoints is standard in clinical trial design literature. In particular, both [11] and [12] adopt the multinomial formulation in their Phase II design studies. We follow the same convention in this work, leveraging its flexibility in capturing both marginal probabilities and the association structure between the two endpoints.
All developments in this paper are carried out under a joint multinomial (equivalently, bivariate–binomial) model for two dependent binary endpoints, a modeling choice that is widely used in Phase II clinical trial design. The real-data examples are intended to illustrate implementation rather than to validate distributional assumptions. In applied use, investigators should justify or check the adequacy of the multinomial model for their specific data before adopting the procedure.
Let X 1 = X 11 + X 12 and X 2 = X 11 + X 21 represent the number of successes on endpoints 1 and 2, respectively. The marginal probabilities of success are given by p 1 = p 11 + p 12 and p 2 = p 11 + p 21 , respectively. Consequently, X 1 Binomial ( n , p 1 ) and X 2 Binomial ( n , p 2 ) . We denote the binomial probability mass function with parameters n and p by b ( n , p , · ) .
The joint distribution of ( X 1 , X 2 ) depends not only on p 1 and p 2 , but also on the association between the two endpoints. To quantify this association, we use the odds ratio ϕ = p 11 p 22 p 21 p 12 , which is a natural and widely used measure in 2 × 2 tables. Notably, ϕ is independent of the marginal probabilities p 1 and p 2 . When ϕ = 1 , the two endpoints are independent; ϕ > 1 indicates a positive association, and  ϕ < 1 indicates a negative association.
This study investigates two binary outcomes—efficacy and safety—for each experimental treatment, comparing them against those of a control. The success probabilities associated with the control treatment on these two endpoints are denoted by p 0 , 1 and p 0 , 2 , respectively. Let π i , for  i = 1 , , k , denote the k experimental treatments being evaluated, with  π 0 representing the control group. Each treatment is associated with two binary outcomes. To distinguish between the two endpoints within each treatment, we use a second subscript j in the notation, where j = 1 corresponds to the efficacy endpoint and j = 2 to the safety endpoint. Thus, the success probabilities for treatment π i are denoted by p i , 1 and p i , 2 for the efficacy and safety endpoints, respectively, for  i = 0 , 1 , , k . We assume that the k + 1 treatment groups are statistically independent, indicating that responses from different treatment arms do not influence one another. In contrast, within any single treatment arm, dependence between the two endpoints may exist. To classify treatments based on their performance, we partition the parameter space { ( p 1 , 1 , p 1 , 2 ) , , ( p k , 1 , p k , 2 ) 0 < p i , j < 1 for i = 1 , , k , j = 1 , 2 } using four prespecified constants: δ 0 , 1 , δ 0 , 2 , δ 1 , 1 , and δ 1 , 2 . These constants satisfy the conditions < δ 0 , 1 < δ 1 , 1 with δ 1 , 1 > 0 , and  < δ 0 , 2 < δ 1 , 2 with δ 1 , 2 > 0 . In this framework, a treatment π i is considered ineffective if p i , 1 p 0 , 1 + δ 0 , 1 or p i , 2 p 0 , 2 + δ 0 , 2 , and considered effective if p i , 1 p 0 , 1 + δ 1 , 1 and p i , 2 p 0 , 2 + δ 1 , 2 , where p 0 , 1 and p 0 , 2 are success probabilities of the control treatment, and we assume that these two probabilities are known prior to conducting the selection procedure. Our objective is to classify the k experimental treatments into two groups: those that are effective and those that are ineffective. We now describe the formal selection goal.
The threshold parameters δ 0 , 1 , δ 0 , 2 , δ 1 , 1 , δ 1 , 2 are pre-specified quantities that reflect the minimal clinically meaningful differences in efficacy and safety between an experimental treatment and the control. In practice, these values are typically determined by subject-matter experts based on prior clinical knowledge, historical studies, or the expected magnitude of benefit that would justify advancing a treatment to further development stages. From a statistical perspective, the role of these thresholds is analogous to the effect size specified in power calculations for hypothesis testing. Just as the effect size quantifies the magnitude of difference that a study aims to detect with sufficient power, the  δ values define the selection region where a treatment is considered promising in terms of both endpoints. Importantly, the framework allows these thresholds to be flexibly tuned to reflect the specific objectives or priorities of a trial. For example, if the primary goal is to achieve a substantial improvement in efficacy while maintaining comparable safety, one may set δ 1 , 1 to a larger value and δ 1 , 2 close to zero. Conversely, if safety improvement is the main concern—such as in cases where toxicity is a known issue—one may choose a small δ 1 , 1 and a larger δ 1 , 2 . A detailed illustration of how these thresholds can be specified based on the desired benefit-risk trade-off is provided in Example 2 (Chemotherapy of Acute Leukemia), where the clinical objective is to improve safety while maintaining efficacy.
Goal: Select a subset consisting of those treatments π i for which p i , 1 > p 0 , 1 and p i , 2 > p 0 , 2 ; that is, include all experimental treatments that demonstrate superiority over the control treatment with respect to both efficacy and safety. If no such treatment exists—i.e., if no π i satisfies both p i , 1 > p 0 , 1 and p i , 2 > p 0 , 2 —then none of the k experimental treatments should be selected.
Probability requirements: Let P 0 and P 1 be pre-specified probability constants satisfying 2 k < P 0 < 1 and ( 1 2 k ) / k < P 1 < 1 . The probability requirements for the selection procedure are defined as follows:
P ( All π i with p i , 1 p 0 , 1 + δ 1 , 1 and p i , 2 p 0 , 2 + δ 1 , 2 are included in the selected subset , for i = 1 , , k ) P 1 ,
and
P ( No π i is selected whenever p i , 1 p 0 , 1 + δ 0 , 1 or p i , 2 p 0 , 2 + δ 0 , 2 for all i = 1 , , k ) P 0 .
Let C S 1 denote the event that the selected subset correctly includes all effective treatments, provided such treatments exist. Specifically, C S 1 occurs when every treatment π i satisfying p i , 1 p 0 , 1 + δ 1 , 1 and p i , 2 p 0 , 2 + δ 1 , 2 is included in the selected subset. Similarly, let C S 0 denote the event that no treatment is selected when none are truly effective. That is, C S 0 occurs if p i , 1 p 0 , 1 + δ 0 , 1 or p i , 2 p 0 , 2 + δ 0 , 2 holds for all i = 1 , , k . Under this framework, the selection procedure is required to satisfy the following probability criteria:
P ( C S 1 ) P 1 , P ( C S 0 ) P 0 ,
where P 1 and P 0 are prespecified thresholds that represent the minimum acceptable probabilities for correctly identifying effective treatments and correctly excluding ineffective treatments, respectively.
Remark 1.
A selection is considered correct when at least one effective experimental treatment exists and the chosen subset successfully includes all such effective options. The rationale for selecting a subset—rather than identifying a single best treatment—is that no natural ordering can be established among the pairs of success probabilities ( p 1 , 1 , p 1 , 2 ) , , ( p k , 1 , p k , 2 ) unless one endpoint is explicitly prioritized over the other. Since this paper does not assume any preference between the two endpoints, we adopt a subset selection approach.

3. Fixed Sample Size Procedure

This section introduces the fixed-sample-size selection method, which serves as a benchmark for evaluating the curtailed procedure outlined in Section 4. This fixed-sample-size procedure was derived by [16]. We also present results concerning the determination of design parameters that guarantee the fixed-sample-size procedure meets the probability criteria specified in Probability requirements (1) and (2).
For prespecified design parameters n , c 1 , c 2 , the selection procedure is defined as follows:
Procedure H:
Suppose we obtain n observations from each of the k Bernoulli experimental treatments and the control treatment. Let X i , 1 and X i , 2 denote the number of observed successes for the first and second binary endpoints of treatment i, where i = 0 , 1 , 2 , , k . For fixed positive constants c 1 and c 2 , Procedure H proceeds as follows:
(1)
Include in the selected subset all treatments π i such that X i , 1 X 0 , 1 c 1 and X i , 2 X 0 , 2 c 2 ;
(2)
If no treatment π i satisfies both conditions, select the control treatment π 0 .
In ranking and selection problems, it is common practice to derive an explicit formula for the probability of a correct selection, denoted by P ( C S ) , and then identify the least favorable configuration (LFC), under which P ( C S ) reaches its minimum. Design parameters are then determined such that P ( C S | LFC ) exceeds a pre-specified threshold P .
However, in this particular subset selection problem, no closed-form expression for P ( C S ) could be obtained. Instead, a lower bound was developed, together with the configuration that minimizes it. If this minimum bound exceeds P , then P ( C S ) will necessarily be no less than P across all parameter settings.
We denote by C F G 1 the configuration where p i , 1 = p 0 , 1 + δ 1 , 1 , p i , 2 = p 0 , 2 + δ 1 , 2 , i = 1 , , k and by C F G 0 the parameter configuration where p i , 1 = p 0 , 1 + δ 0 , 1 ,   p i , 2 = p 0 , 2 + δ 0 , 2 , i = 1 , , k . Then C F G 1 and C F G 0 are the configurations under which the lower bounds P L ( C S 1 ) and P L ( C S 0 ) of the probabilities of correct selections P ( C S 1 ) and P ( C S 0 ) , respectively, were computed. P L ( C S 1 ) also depends on the odds ratios between the two endpoints of each of the k treatments, while P L ( C S 0 ) does not. It is assumed that the degree of association between the two endpoints is identical across all k treatments. Three scenarios are considered: (1) the endpoints are independent; (2) the endpoints are dependent with a known association structure; and (3) the endpoints are dependent but the association is unknown. When the association is not known, it was shown that the minimum of P L ( C S 1 ) is attained when the odds ratio is zero. However, numerical computations showed that the sample size varies very little with the odds ratio. We now present the theorems establishing the lower bounds P L ( C S 0 ) and P L ( C S 1 ) for the correct selection probabilities P ( C S 0 ) and P ( C S 1 ) , respectively. The corresponding proofs are provided in [16] and also included in Appendix A for completeness.
Case 1: ϕ i = 1 , i = 0 , 1 , , k . We begin with the scenario where the two endpoints are independent. Specifically, we assume that ϕ i = 1 for all i = 0 , 1 , , k . Under this assumption, X i , 1 and X i , 2 are mutually independent binomial random variables with parameters ( n , p i , 1 ) and ( n , p i , 2 ) , respectively, for each i = 0 , 1 , , k .
Theorem 1
(Adapted from [16]). For fixed k , p 0 , 1 , p 0 , 2 , δ 0 , 1 , δ 0 , 2 , δ 1 , 1 , δ 1 , 2 , the probability requirements are satisfied by choosing values of n , c 1 , c 2 that simultaneously satisfy   
( x 0 , 1 = 0 n [ x i , 1 = c 1 + x 0 , 1 n b ( n , p 0 , 1 + δ 1 , 1 , x i , 1 ) ] k b ( n , p 0 , 1 , x 0 , 1 ) ) × ( x 0 , 2 = 0 n [ x i , 2 = c 2 + x 0 , 2 n b ( n , p 0 , 2 + δ 1 , 2 , x i , 2 ) ] k b ( n , p 0 , 2 , x 0 , 2 ) ) P 1  and
x 0 , 1 = 0 n x 0 , 2 n ( 1 max [ P ( X i , 1 c 1 + x 0 , 1 | p i , 1 = p 0 , 1 + δ 0 , 1 ) , P ( X i , 2 c 2 + x 0 , 2 | p i , 2 = p 0 , 2 + δ 0 , 2 ) ] ) k × b ( n , p 0 , 1 , x 0 , 1 ) b ( n , p 0 , 2 , x 0 , 2 ) P 0 .
Case 2: ϕ i 1 is specified for i = 0 , 1 , , k . We next consider the scenario where the two endpoints within each treatment arm exhibit dependence, with the association parameter ϕ i known in advance.
Theorem 2
(Adapted from [16]). For fixed values of k , p 0 , 1 , p 0 , 2 , δ 0 , 1 , δ 0 , 2 , δ 1 , 1 , δ 1 , 2 , ϕ i , i = 0 , 1 , , k , the probability requirements are satisfied by choosing values of n , c 1 , c 2 that simultaneously satisfy
x 0 , 1 = 0 n x 0 , 2 = 0 n [ Π i = 1 k ( x i , 1 = c 1 + x 0 , 1 n x i , 2 = c 2 + x 0 , 2 n P ( X i , 1 = x i , 1 , X i , 2 = x i , 2 | p i , 1 = p 0 , 1 + δ 1 , 1 , p i , 2 = p 0 , 2 + δ 1 , 2 , ϕ i ) ) ] × P ( X 0 , 1 = x 0 , 1 , X 0 , 2 = X 0 , 2 | p 0 , 1 , p 0 , 2 , ϕ 0 ) P 1  and
x 0 , 1 = 0 n x 0 , 2 n ( 1 max [ P ( X i , 1 c 1 + x 0 , 1 | p i , 1 = p 0 , 1 + δ 0 , 1 ) , P ( X i , 2 c 2 + x 0 , 2 | p i , 2 = p 0 , 2 + δ 0 , 2 ) ] ) k × P ( X 0 , 1 = x 0 , 1 , X 0 , 2 = X 0 , 2 | p 0 , 1 , p 0 , 2 , ϕ 0 ) P 0 . where p i , 11 = a i a i 2 + b i 2 ( ϕ i 1 ) , a i = 1 + ( ϕ i 1 ) ( p 0 , 1 + δ 1 , 1 + p 0 , 2 + δ 1 , 2 ) , b i = 4 ϕ i ( ϕ i 1 ) ( p 0 , 1 + δ 1 , 1 ) ( p 0 , 2 + δ 1 , 2 ) , p i , 12 = p 0 , 1 + δ 1 , 1 p i , 11 .
Case 3: ϕ i unspecified for i = 1 , 2 , , k , ϕ 0 specified. We now examine the setting in which the dependence structure between the two endpoints is unknown for the experimental treatments, while the association for the control treatment remains specified.
Theorem 3
(Adapted from [16]). For fixed k , p 0 , 1 , p 0 , 2 , δ 0 , 1 , δ 0 , 2 , δ 1 , 1 , δ 1 , 2 , the probability requirements are satisfied by choosing values of n , c 1 , c 2 that simultaneously satisfy
x 0 , 1 = 0 n x 0 , 2 = 0 n [ x i , 1 = c 1 + x 0 , 1 n x i , 2 = c 2 + x 0 , 2 n b ( n , p 0 , 2 + δ 1 , 2 , x i , 2 ) f ( x i , 1 , x i , 2 ) ] k × P ( X 0 , 1 = x 0 , 1 , X 0 , 2 = x 0 , 2 | p 0 , 1 , p 0 , 2 , ϕ 0 ) P 1 and
( x 0 , 1 = 0 n x 0 , 2 n ( 1 max [ P ( X i , 1 c 1 + x 0 , 1 | p i , 1 = p 0 , 1 + δ 0 , 1 ) , P ( X i , 2 c 2 + x 0 , 2 | p i , 2 = p 0 , 2 + δ 0 , 2 ) ] ) k × P ( X 0 , 1 = x 0 , 1 , X 0 , 2 = X 0 , 2 | p 0 , 1 , p 0 , 2 , ϕ 0 ) P 0 where
f ( x i , 1 , x i , 2 ) = b n x i , 2 , p 0 , 1 + δ 1 , 1 1 ( p 0 , 2 + δ 1 , 2 ) , x i , 1 if p 0 , 1 + δ 1 , 1 < 1 ( p 0 , 2 + δ 1 , 2 ) b x i , 2 , p 0 , 1 + δ 1 , 1 + p 0 , 2 + δ 1 , 2 1 p 0 , 2 + δ 1 , 2 , x i , 1 + x i , 2 n if p 0 , 1 + δ 1 , 1 > 1 ( p 0 , 2 + δ 1 , 2 ) b n , p 0 , 2 + δ 1 , 2 , x i , 2 1 x i , 1 + x i , 2 = n if p 0 , 1 + δ 1 , 1 = 1 ( p 0 , 2 + δ 1 , 2 )
,
1 x i , 1 + x i , 2 = n = 1 x i , 1 + x i , 2 = n 0 x i , 1 + x i , 2 n
and
ϕ i = 0 , i = 1 , 2 , , k
Remark 2.
The lower bound on P ( C S 0 ) depends only on the odds ratio for the control treatment.
Remark 3.
Ref. [16] demonstrated that P ( C S 1 ) increases with the odds ratios of the experimental treatments π i , for  i = 1 , 2 , , k . Therefore, the minimum value of P ( C S 1 ) is achieved when the odds ratios of all experimental treatments are zero. To obtain a lower bound for P ( C S 1 ) , we evaluated it under the assumption that all treatments tested have odds ratios equal to zero. As a result, the scenario with unspecified odds ratios is effectively handled by considering the scenario where the odds ratios of the tested treatments are zeros.
Remark 4.
The theoretical results presented here are obtained under the assumption that all k treatments share the same type of dependence between the two endpoints. For instance, the endpoints may be mutually independent across all treatments, or they may follow a common, known dependence structure. Nevertheless, due to the generality of our derivations, the framework can also accommodate settings in which different treatments exhibit heterogeneous dependence types—such as independence for some and unspecified association for others.
Remark 5.
Under configuration C F G 1 , where p i , 1 = p 0 , 1 + δ 1 , 1 and p i , 2 = p 0 , 2 + δ 1 , 2 for all i = 1 , , k , the quantity P L ( C S 1 ) represents a lower bound on the probability of correctly selecting π 1 , , π k when the experimental treatments are truly superior to the control. In this sense, P L ( C S 1 ) is analogous to a lower bound on the power in classical hypothesis testing. On the other hand, under configuration C F G 0 , where p i , 1 = p 0 , 1 + δ 0 , 1 and p i , 2 = p 0 , 2 + δ 0 , 2 for all i = 1 , , k , the quantity P L ( C S 0 ) denotes the probability of correctly selecting π 0 , reflecting the scenario in which none of the experimental treatments are truly superior to the control. In this case, 1 P L ( C S 0 ) is conceptually analogous to the upper bound of the family-wise error rate (FWER) in multiple hypothesis testing, as it represents the probability of erroneously selecting at least one ineffective treatment.

4. Proposed Curtailment Procedure

To address the objective described in Section 2, we introduce a curtailed procedure, denoted by R. This sequential approach utilizes curtailment to reduce the required sample size for treatments that are evidently ineffective or demonstrate sufficient efficacy. Let n represent the maximum number of observations allowed per treatment.
Curtailment Procedure R:
A treatment is considered contending if it has not yet been eliminated from the study. The procedure begins with all k + 1 populations classified as contending. Sampling follows a vector-at-a-time approach. At “Step M”, where 1 M n , a total of M vectors have been drawn. Let Y ( i , 1 ) , M and Y ( i , 2 ) , M denote the cumulative numbers of successes observed for endpoints 1 and 2 of treatment π i through Step M.
Sampling Rule. A vector-at-a-time sampling strategy is adopted, subject to the following constraints:
(a) A maximum of n observations may be collected from each of the ( k + 1 ) populations. Sampling proceeds sequentially from each contending treatment, one observation at a time, until either the treatment accumulates n observations or is eliminated based on criteria (b) or (c) below.
(b) At any step M, if the number of successes for the two endpoints, Y ( i , 1 ) , M and Y ( i , 2 ) , M , of treatment π i satisfy
Y ( i , 1 ) , M + n M < c 1 + Y ( 0 , 1 ) , M or Y ( i , 2 ) , M + n M < c 2 + Y ( 0 , 2 ) , M ,
then eliminate treatment π i and stop sampling from it.
(c) At any step M, if the number of successes for the two endpoints, Y ( i , 1 ) , M and Y ( i , 2 ) , M , of treatment π i satisfy
Y ( i , 1 ) , M c 1 + Y ( 0 , 1 ) , M + n M and Y ( i , 2 ) , M c 2 + Y ( 0 , 2 ) , M + n M ,
then terminate sampling for treatment π i .
Stopping Rule:
The experiment is terminated at the first step M at which any one of the following three conditions is met:
(i)
There exists a partition A , B of the set { 1 , 2 , , k } such that:
Y ( i , 1 ) , M c 1 + Y ( 0 , 1 ) , M + n M and Y ( i , 2 ) , M c 2 + Y ( 0 , 2 ) , M + n M , for all i A , Y ( j , 1 ) , M + n M < c 1 + Y ( 0 , 1 ) , M or Y ( j , 2 ) , M + n M < c 2 + Y ( 0 , 2 ) , M , for all j B .
(ii)
For all i { 1 , 2 , , k } ,
Y ( i , 1 ) , M + n M < c 1 + Y ( 0 , 1 ) , M or Y ( i , 2 ) , M + n M < c 2 + Y ( 0 , 2 ) , M .
(iii)
M = n .
Decision Rule:
(a) If sampling terminates under condition (i) of the above Stopping Rule, then all treatments π i contained in set A are included in the selected subset.
(b) If sampling terminates under condition (ii) of the above Stopping Rule, we conclude that none of the experimental treatments demonstrates a statistically significant improvement over the control treatment π 0 .
(c) If sampling terminates under condition (iii) of the above Stopping Rule, we include in the selected subset all treatments π i satisfying Y ( i , 1 ) , n c 1 + Y ( 0 , 1 ) , n and Y ( i , 2 ) , n c 2 + Y ( 0 , 2 ) , n . If no such treatments exist—i.e., the selected subset is empty—we conclude that none of the experimental treatments is significantly better than the control π 0 .
Theorem 4.
Given fixed k and n, procedures H and R produce the same subset of treatments  { π 0 , π 1 , , π k } if they are applied with identical threshold parameters c 1 and c 2 . This equivalence holds uniformly over the parameter space { ( p 0 , 1 , p 0 , 2 ) , ( p 1 , 1 , p 1 , 2 ) , , ( p k , 1 , p k , 2 ) } .
Proof. 
Decision Rule (c) of Procedure R, which applies exclusively when sampling terminates under Stopping Rule (iii), coincides with the decision rule of Procedure H in the case where m = n . Consequently, for any sampling outcome in which each of the k treatments reaches the maximum of n observations, procedures R and H will yield the same selected subset. It therefore suffices to examine only those cases in which Procedure R makes a decision under Rule (a) or (b).
Decision Rules (a) and (b) are applied exclusively when sampling terminates under Stopping Rules (i) and (ii), respectively. Observe that sampling under either of these rules always occurs before the maximum number of observations is reached; that is, M < n , and hence n M > 0 .
If sampling terminates under Stopping Rule (i), then under Procedure R, Decision Rule (a) selects all treatments π i that satisfy the condition:
Y ( i , 1 ) , M c 1 + Y ( 0 , 1 ) , M + n M and Y ( i , 2 ) , M c 2 + Y ( 0 , 2 ) , M + n M .
Now suppose the experiment were to continue as it would under Procedure H. Let Y ( i , 1 ) , n and Y ( i , 2 ) , n denote the total number of successes for treatment π i at endpoints 1 and 2, respectively, after n observations. Then, by Rule (2) of Procedure H, treatment π i would be selected if
Y ( i , 1 ) , n c 1 Y ( 0 , 1 ) , n and Y ( i , 2 ) , n c 2 Y ( 0 , 2 ) , n .
Observe that:
Y ( i , 1 ) , n c 1 Y ( i , 1 ) , M c 1 Y ( 0 , 1 ) , M + n M Y ( 0 , 1 ) , n , Y ( i , 2 ) , n c 2 Y ( i , 2 ) , M c 2 Y ( 0 , 2 ) , M + n M Y ( 0 , 2 ) , n .
Hence, the same subset of treatments would be selected by Procedure H.
Similarly, if sampling terminates under Stopping Rule (ii), then Decision Rule (b) of Procedure R results in the selection of no experimental treatments. This outcome coincides precisely with the decision specified by Rule (2) of Procedure H.
This concludes the proof of the theorem.    □

5. Tables

In this section, we assess the performance of the curtailment procedure by examining its potential for sample size reduction compared to the corresponding non-curtailment procedure H. Throughout the analysis, we assume a common association structure between the two endpoints across all ( k + 1 ) treatments. Nevertheless, our results indicate that the parameter derivations for the curtailment procedure are also applicable to settings in which the association structures vary across treatments.
We examine the following configurations: k = 2 , 3 ; p 0 , 1 = 0.40 , 0.50 , 0.60 ; p 0 , 2 = 0.60 ; δ 0 , 1 = δ 0 , 2 = 0.01 ; δ 1 , 1 = 0.30 , δ 1 , 2 = 0.25 ; P 0 = 0.90 ; P 1 = 0.80 , 0.85 , 0.90 ; and ϕ = 0 , 0.01 , 0.1 , 1 , 2 , 4 , 8 , 100 . These parameter settings are consistent with those adopted in [16] for generating Table 2, Table 3, Table 4, Table 5, Table 6 and Table 7 under the fixed-sample-size procedure H. For each configuration, the minimum required sample size per treatment, n, along with the corresponding thresholds ( c 1 , c 2 ) , was determined to ensure that the specified probability constraints were satisfied. If multiple ( c 1 , c 2 ) combinations met these constraints, the design yielding the highest probability of selecting a single effective treatment was chosen. This n is then used as the maximum number of observations per treatment under the curtailment procedure R. As established in Theorem 4, selecting the same values of n and ( c 1 , c 2 ) ensures that the curtailment procedure meets the same probability requirements as Procedure H. This result ensures that the theoretical properties derived under the fixed-sample framework—such as those in Theorems 1–3—can be validly transferred to the curtailed procedure without requiring separate proofs. In this sense, Theorem 4 provides the theoretical foundation for leveraging fixed-sample thresholds within a sequential, curtailed sampling design.
We define N as the total number of observations required by the fixed-sample-size procedure to meet the prescribed probability constraints, with  N = ( k + 1 ) × n . This value of N also serves as an upper limit on the total number of observations under the curtailment procedure. Let E 1 ( N | R ) and E 0 ( N | R ) represent the expected sample sizes of Procedure R under the configurations C F G 1 and C F G 0 , respectively, as described in Section 3. These configurations were also used to compute the lower bounds P L ( C S 1 ) and P L ( C S 0 ) for selecting a correct subset under the alternative and null hypotheses, and to determine the design parameters for Procedure H under the fixed-sample-size setting.
We define the average expected sample size under the curtailment procedure as E ( N R ) = 1 2 E 1 ( N R ) + E 0 ( N R ) , following the approach of Thall, Simon, and Ellenberg, to account for performance under both configurations. The quantities E 1 ( N R ) and E 0 ( N R ) are estimated via simulation (10,000 repetitions), implemented in R. To generate bivariate binary data with marginal probabilities p 1 , p 2 and odds ratio ϕ 1 , we compute:
p 11 = a a 2 + b 2 ( ϕ 1 ) , a = 1 + ( ϕ 1 ) ( p 1 + p 2 ) , b = 4 ϕ ( ϕ 1 ) p 1 p 2 , p 12 = p 1 p 11 , p 21 = p 2 p 11 , p 22 = 1 ( p 11 + p 12 + p 21 ) ,
and then simulate binary outcomes from a 2 × 2 table with cell probabilities p i j .
Table 2, Table 3, Table 4, Table 5, Table 6 and Table 7 summarize, for each parameter setting, the total sample size N required by the fixed-sample-size procedure, the expected sample sizes E 1 ( N | R ) and E 0 ( N | R ) under the curtailment procedure, and the percentage reduction in observations achieved via curtailment, computed as R S ( % ) = N E ( N | R ) N × 100 . These results clearly demonstrate that Procedure R requires significantly fewer observations than Procedure H to achieve equivalent performance guarantees.
Table 2, Table 3, Table 4, Table 5, Table 6 and Table 7 also demonstrate that the odds ratio ϕ has little impact on the expected sample size when ϕ values are relatively close, under both C F G 1 and C F G 0 , for the curtailment procedure. Greater variability is observed under C F G 0 . This pattern is consistent with the findings of [12], who considered only moderate odds ratios ( ϕ = 2 , 4 , 8 ) in the context of hypothesis testing, and observed minimal sample size variation under curtailment, with larger variability under the null hypothesis. In our procedure, however, when ϕ varies substantially—for example, from 1 to 100—we observe a marked decrease in expected sample size under both C F G 1 and C F G 0 . Ref. [12] did not report average expected sample sizes, but instead presented results under both the null and alternative hypotheses and calculated percentage savings. Their results showed modest sample size savings under the alternative hypothesis and more pronounced savings under the null, consistent with our own findings.

6. Examples

6.1. Immunotherapy in Elderly Patients with Non-Small Cell Lung Cancer

To demonstrate the practical utility of our method and illustrate how to use the simulation-based reference tables, we construct the following example. The example is intended to illustrate how to implement the proposed design and to use the simulation-based reference tables; it is not intended to validate the underlying distributional assumptions on this dataset.
This example considers an experimental trial involving two immunotherapy-based treatments for elderly patients (≥75 years old) diagnosed with advanced non-small cell lung cancer (NSCLC). The trial compares two immunotherapy strategies—PD1-A (anti-PD-1 monotherapy) and PD1-B (anti-PD-1 combined with low-dose chemotherapy)—against the standard chemotherapy regimen consisting of carboplatin and pemetrexed, which serves as the control treatment.
While carboplatin plus pemetrexed is considered the standard of care in general NSCLC populations, this regimen has not been adequately studied in patients aged 75 and above. Historically, in younger NSCLC populations, this standard chemotherapy yields approximately 40 % objective response rate (ORR), and around 40 % of patients experience grade 3 or higher treatment-related adverse events. These outcomes establish the benchmark efficacy and safety rates for the control treatment as p 0 , 1 = 0.40 and p 0 , 2 = 0.60 , respectively. Prior analyses in younger patients suggest an odds ratio of approximately 2 between efficacy and safety, indicating that patients who do not experience toxicity are more likely to respond to treatment.
This study aims to assess whether PD1-A or PD1-B outperforms the control treatment with respect to efficacy and safety. Specifically, the experimenter seeks an increase in the response rate of at least 0.30 and a reduction in high-grade toxicity of at least 0.25 , corresponding to δ 1 , 1 = 0.30 and δ 1 , 2 = 0.25 . If both experimental treatments fail to demonstrate improvements over the control, the standard chemotherapy will be selected, with thresholds δ 0 , 1 = δ 0 , 2 = 0.01 . This means that whenever a tested treatment exceeds the control by at least 0.30 in efficacy and 0.25 in safety, the procedure guarantees a probability of at least P 1 of selecting all such treatments. Conversely, if none of the experimental treatments surpass the control by more than 0.01 in either endpoint, the procedure ensures a probability of at least P 0 of correctly selecting only the control treatment.
When P 1 = 0.85 and P 0 = 0.90 , Table 3 shows that the fixed sample size procedure requires n = 81 observations per treatment, with corresponding critical values c 1 = 14 and c 2 = 12 . Therefore, a total of 3 × 81 = 243 observations is needed under the fixed sample size procedure to meet the specified probability constraints. In contrast, the curtailment procedure, while also using at most n = 81 observations per treatment and the same critical values c 1 = 14 , c 2 = 12 , is expected to achieve the same probability guarantees with fewer observations on average. According to Table 3, the expected relative sample size saving from using the curtailment procedure is approximately 11.2767 % .

6.2. Chemotherapy of Acute Leukemia

This example is adapted from a clinical trial scenario described in [17], with minor modifications to align with our design framework. This example likewise serves to demonstrate implementation rather than to validate the distributional assumptions for this dataset.
This example involves an experimental trial comparing two different combinations of Gemcitabine and Cyclophosphamide—denoted as GemCy1 and GemCy2—each with varying dosage proportions, against the standard Ara-C regimen for treating patients with good-prognosis acute myelogenous leukemia (AML) or myelodysplastic syndrome. Historically, the standard treatment Ara-C yields approximately 70 % of patients achieving complete remission (CR), while around 38 % either die or experience severe myelosuppression within the first five weeks. These historical outcomes establish the efficacy and safety rates for the control treatment as p 0 , 1 = 0.7 and p 0 , 2 = 0.62 , respectively. Additionally, the odds ratio between efficacy and safety is estimated to be 3.05 , indicating that patients who do not experience toxicity are more likely to achieve complete remission.
The goal of the trial is to determine whether either GemCy1 or GemCy2 surpasses the control treatment in terms of benefit-risk profile. Since the toxicity associated with Ara-C is relatively severe, our design prioritizes safety improvement while maintaining efficacy. Specifically, we consider a benefit-risk trade-off scenario in which the efficacy threshold is relaxed—i.e., we set δ 1 , 1 = 0 , requiring only that the experimental treatments match the control in efficacy—while demanding a safety improvement of at least δ 1 , 2 = 0.35 . This reflects the objective of identifying a new treatment that is comparably effective but significantly safer. If neither experimental treatment demonstrates superiority in safety, the control treatment is selected. To enforce this, we use δ 0 , 1 = 0.2 and δ 0 , 2 = 0.01 . The negative value of δ 0 , 1 indicates that an experimental treatment will be excluded if its efficacy falls more than 0.2 below that of the control (i.e., if p i , 1 < p 0 , 1 0.2 ). This ensures that a treatment cannot be selected solely on the basis of improved safety if it sacrifices too much efficacy.
When P 1 = 0.85 and P 0 = 0.85 , the grid search results show that the fixed sample size procedure requires n = 86 observations per treatment, with corresponding critical values c 1 = 8 and c 2 = 17 . Consequently, implementing the fixed sample size procedure requires a total of 3 × 86 = 258 observations to meet the required probabilistic conditions. In contrast, the curtailment procedure, while also using at most n = 86 observations per treatment and the same critical values c 1 = 8 , c 2 = 17 , is expected to achieve the same probability guarantees with fewer observations on average. According to the simulation program derived by the curtailed procedure, the expected relative sample size saving from using the curtailment procedure is approximately 12.13 % .
Remark 6.
The thresholds δ 0 , 1 , δ 0 , 2 , δ 1 , 1 , δ 1 , 2 are design parameters specified by the investigator to reflect the clinical priorities of the trial. If the goal is to prioritize efficacy, one may choose a relatively large value for δ 1 , 1 (requiring a strong efficacy improvement) and a small or even zero value for δ 1 , 2 (allowing safety to remain comparable to control). Conversely, higher values of δ 1 , 2 can be used to enforce stricter safety improvements. This flexibility allows the proposed procedure to accommodate various benefit-risk trade-offs tailored to specific therapeutic contexts. In this framework, the probability of correctly selecting a truly effective treatment—i.e., one that exceeds both thresholds δ 1 , 1 and δ 1 , 2 —is guaranteed to exceed P 1 , which serves a role analogous to the power in hypothesis testing.
Remark 7.
The thresholds δ 0 , 1 and δ 0 , 2 play a complementary role by specifying minimal margins below which a treatment is considered insufficiently effective or safe. Treatments that fail to exceed either p 0 , 1 + δ 0 , 1 or p 0 , 2 + δ 0 , 2 are excluded from selection, ensuring that only those with meaningful improvements over the control are considered. This design component helps control the probability of incorrectly selecting inferior treatments, serving a role analogous to controlling the type I error rate in hypothesis testing.
Remark 8.
The values of p 0 , 1 and p 0 , 2 for the control group are typically known or can be reliably estimated from historical data, since the control treatment usually corresponds to an existing standard-of-care therapy with well-documented efficacy and safety profiles.In situations where the exact values of p 0 , 1 and p 0 , 2 are uncertain, a conservative design strategy can be employed. Specifically, one may perform a grid search over a plausible range of ( p 0 , 1 , p 0 , 2 ) pairs and identify the worst-case configuration—i.e., the combination that yields the largest required sample size n. The study can then be designed based on this worst-case scenario to ensure robust performance across all reasonable control assumptions.

7. Conclusions

This study introduces a curtailment procedure for selecting a random-sized subset that includes the best treatment, provided it demonstrates a significant improvement over the control. The comparison is based on two Bernoulli endpoints observed for each of the k experimental treatments and the control group. The proposed method builds upon the fixed sample size approach outlined by [16]. While meeting the same probability constraints for successful selection as the original method, the curtailment approach requires fewer observations from the experimental arms. Simulation results show that this method can yield a relative total sample size reduction of approximately 10– 12 % compared to the fixed sample size procedure. The adaptive nature of curtailment is especially advantageous, as it not only decreases the overall sample size but also avoids unnecessary sampling from less promising treatments. However, to ensure its practical feasibility, the time between treatment administration and response observation should be short relative to the experiment duration.
Furthermore, simulation results suggest that variations in the odds ratio have negligible influence on the required sample size. The insensitivity to odds ratios contributes to the robustness of the proposed method, particularly when the assumption of independence between endpoints is violated. In our study, we assumed a consistent pattern of association between the two endpoints across all treatment arms. Nevertheless, this assumption can be readily relaxed to allow for differing association structures among treatments.
As a potential direction for future research, when the true success probabilities p i j are unknown at the design stage, one may estimate the association parameter ϕ from historical or pilot data and obtain a confidence set I ϕ . Rather than reporting ranges for ( n , c 1 , c 2 ) , a robust, worst-case calibration could be performed to choose a single triplet ( n , c 1 , c 2 ) meeting the selection-probability guarantees for all ϕ I ϕ . Our design tables suggest that the resulting robust triplet often coincides with the design at one endpoint of I ϕ , reflecting the empirical stability of ( c 1 , c 2 ) across nearby ϕ values. A systematic development and comparison of such interval-based robust calibration is left for future study.
In applied use, practitioners should justify or check the adequacy of the joint multinomial (equivalently, bivariate–binomial) assumption for their specific datasets prior to adopting the procedure. The present work focuses on the methodological framework and guarantees under this assumption. As another avenue for future work, it would be valuable to investigate the empirical validity of the underlying distributional assumptions—such as the bivariate binomial or multinomial model—when applying the proposed procedure to real-world clinical datasets.
Additionally, a reader may wonder about selecting the single most efficacious treatment from the identified subset. That task falls under the select-the-best framework, which has distinct objectives and methodological requirements compared to the subset selection problem studied here. Both approaches have distinct objectives and design considerations, and each has its own advantages and limitations. The two approaches are referred to as the Subset Selection Approach and the Indifference Zone Approach, respectively. Please see [18]. We view these two directions as complementary rather than contradictory, and believe they are best addressed in separate articles.

Author Contributions

Conceptualization, P.C. and L.H.; methodology, E.M.B. and C.Y.; software, C.Y.; validation, E.M.B., P.C. and L.H.; formal analysis, C.Y.; investigation, E.M.B. and C.Y.; resources, C.Y.; data curation, E.M.B. and C.Y.; writing—original draft preparation, C.Y.; writing—review and editing, E.M.B., P.C. and L.H.; visualization, C.Y.; supervision, P.C.; project administration, L.H.; funding acquisition, P.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

The authors thank the anonymous reviewers for their valuable comments and suggestions.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Appendix A. Proofs of Theorems 1–3

Theorems 1–3 and their corresponding proofs are adapted from our submitted manuscript ([16]). They are included here for the reader’s convenience and to ensure the completeness of the current work.
To prove Theorems 1–3, we introduced the following monotonicity properties of the joint probability
P ( X 1 c 1 , X 2 c 2 p 1 , p 2 , ϕ ) ,
which were introduced by [19].
Theorem A1.
Let n 1 , 1 c 1 n , 1 c 2 n be fixed constants. a. For fixed ϕ and p 2 , P ( X 1 c 1 , X 2 c 2 | p 1 , p 2 , ϕ ) is increasing in p 1 . b. For fixed ϕ and p 1 , P ( X 1 c 1 , X 2 c 2 | p 1 , p 2 , ϕ ) is increasing in p 2 . c. For fixed p 1 and p 2 , P ( X 1 c 1 , X 2 c 2 | p 1 , p 2 , ϕ ) is increasing in ϕ.
We now provide the proof of Theorem 1 from Section 3.
Proof. 
Suppose that there are k 1 k treatments superior to the control. Without loss of generality, assume that treatments π i , i = 1 , 2 , , k 1 are better than the control. Since ϕ = 1 , efficacy and safety are independent, and each endpoint follows a binomial distribution with success probabilities p i , 1 and p i , 2 , respectively. Let b ( n , p , x ) denote the probability mass function of a binomial distribution, i.e.,
b ( n , p , x ) = n x p x ( 1 p ) n x , x = 0 , 1 , , n .
Then by the definition of P ( C S 1 ) , we have:
P ( C S 1 ) = P ( X i , 1 X 0 , 1 + c 1 , X i , 2 X 0 , 2 + c 2 , i = 1 , 2 , , k 1 | p i , 1 p 0 , 1 + δ 1 , 1 , p i , 2 p 0 , 2 + δ 1 , 2 , i = 1 , 2 , , k 1 ) P ( X i , 1 X 0 , 1 + c 1 , X i , 2 X 0 , 2 + c 2 , i = 1 , 2 , , k | p i , 1 p 0 , 1 + δ 1 , 1 , p i , 2 p 0 , 2 + δ 1 , 2 , i = 1 , 2 , , k ) x 0 , 1 = 0 n i = 1 k P ( X i , 1 x 0 , 1 + c 1 | p i , 1 p 0 , 1 + δ 1 , 1 ) b ( n , p 0 , 1 , x 0 , 1 ) × x 0 , 2 = 0 n i = 1 k P ( X i , 2 x 0 , 2 + c 2 | p i , 2 p 0 , 2 + δ 1 , 2 ) b ( n , p 0 , 2 , x 0 , 2 ) ( x 0 , 1 = 0 n [ x 1 , 1 = x 0 , 1 + c 1 n b ( n , p 0 , 1 + δ 1 , 1 , x 1 , 1 ) ] k b ( n , p 0 , 1 , x 0 , 1 ) ) × ( x 0 , 2 = 0 n [ x 1 , 2 = x 0 , 2 + c 2 n b ( n , p 0 , 2 + δ 1 , 2 , x 1 , 2 ) ] k b ( n , p 0 , 2 , x 0 , 2 ) )
Then, if
( x 0 , 1 = 0 n [ x 1 , 1 = x 0 , 1 + c 1 n b ( n , p 0 , 1 + δ 1 , 1 , x 1 , 1 ) ] k b ( n , p 0 , 1 , x 0 , 1 ) ) × ( x 0 , 2 = 0 n [ x 1 , 2 = x 0 , 2 + c 2 n b ( n , p 0 , 2 + δ 1 , 2 , x 1 , 2 ) ] k b ( n , p 0 , 2 , x 0 , 2 ) ) P 1 , we obtain that P ( C S 1 ) P 1
Now we show that x 0 , 1 = 0 n x 0 , 2 n ( 1 max [ P ( X i , 1 c 1 + x 0 , 1 | p i , 1 = p 0 , 1 + δ 0 , 1 ) , P ( X i , 2 c 2 + x 0 , 2 | p i , 2 = p 0 , 2 + δ 0 , 2 ) ] ) k × b ( n , p 0 , 1 , x 0 , 1 ) b ( n , p 0 , 2 , x 0 , 2 ) P 0 .
By the definition of P ( C S 0 ) , we have:
P ( C S 0 ) = P ( X i , 1 < X 0 , 1 + c 1 or X i , 2 < X 0 , 2 + c 2 , i = 1 , , k | p i , 1 p 0 , 1 + δ 0 , 1 or p i , 2 p 0 , 2 + δ 0 , 2 , i = 1 , , k ) = i = 1 k P ( X i , 1 < X 0 , 1 + c 1 or X i , 2 < X 0 , 2 + c 2 | p i , 1 p 0 , 1 + δ 0 , 1 or p i , 2 p 0 , 2 + δ 0 , 2 ) = i = 1 k ( 1 P ( X i , 1 X 0 , 1 + c 1 , X i , 2 X 0 , 2 + c 2 | p i , 1 p 0 , 1 + δ 0 , 1 or p i , 2 p 0 , 2 + δ 0 , 2 ) )
Now consider a typical term in the product of the above expression:
P ( X i , 1 X 0 , 1 + c 1 , X i , 2 X 0 , 2 + c 2 | p i , 1 p 0 , 1 + δ 0 , 1 or p i , 2 p 0 , 2 + δ 0 , 2 ) max ( P ( X i , 1 X 0 , 1 + c 1 , X i , 2 X 0 , 2 + c 2 | p i , 1 p 0 , 1 + δ 0 , 1 , p i , 2 p 0 , 2 + δ 0 , 2 ) , P ( X i , 1 X 0 , 1 + c 1 , X i , 2 X 0 , 2 + c 2 | p i , 1 p 0 , 1 + δ 0 , 1 , p i , 2 p 0 , 2 + δ 0 , 2 ) , P ( X i , 1 X 0 , 1 + c 1 , X i , 2 X 0 , 2 + c 2 | p i , 1 p 0 , 1 + δ 0 , 1 , p i , 2 p 0 , 2 + δ 0 , 2 ) ) = max ( P ( X i , 1 X 0 , 1 + c 1 | p i , 1 p 0 , 1 + δ 0 , 1 ) P ( X i , 2 X 0 , 2 + c 2 | p i , 2 p 0 , 2 + δ 0 , 2 ) , P ( X i , 1 X 0 , 1 + c 1 | p i , 1 p 0 , 1 + δ 0 , 1 ) P ( X i , 2 X 0 , 2 + c 2 | p i , 2 p 0 , 2 + δ 0 , 2 ) , P ( X i , 1 X 0 , 1 + c 1 | p i , 1 p 0 , 1 + δ 0 , 1 ) P ( X i , 2 X 0 , 2 + c 2 | p i , 2 p 0 , 2 + δ 0 , 2 ) ) max ( P ( X i , 1 X 0 , 1 + c 1 | p i , 1 = p 0 , 1 + δ 0 , 1 ) P ( X i , 2 X 0 , 2 + c 2 | p i , 2 = p 0 , 2 + δ 0 , 2 ) , P ( X i , 1 X 0 , 1 + c 1 | p i , 1 = p 0 , 1 + δ 0 , 1 ) P ( X i , 2 X 0 , 2 + c 2 | p i , 2 = 1 ) , P ( X i , 1 X 0 , 1 + c 1 | p i , 1 = 1 ) P ( X i , 2 X 0 , 2 + c 2 | p i , 2 = p 0 , 2 + δ 0 , 2 ) ) = max ( P ( X i , 1 X 0 , 1 + c 1 | p i , 1 = p 0 , 1 + δ 0 , 1 ) , P ( X i , 2 X 0 , 2 + c 2 | p i , 2 = p 0 , 2 + δ 0 , 2 ) )
Therefore,
P ( C S 0 ) i = 1 k ( 1 max ( P ( X i , 1 X 0 , 1 + c 1 | p i , 1 = p 0 , 1 + δ 0 , 1 ) , P ( X i , 2 X 0 , 2 + c 2 | p i , 2 = p 0 , 2 + δ 0 , 2 ) ) ) = x 0 , 1 = 0 n x 0 , 2 = 0 n i = 1 k [ 1 max ( P ( X i , 1 x 0 , 1 + c 1 | p i , 1 = p 0 , 1 + δ 0 , 1 ) , P ( x i , 2 X 0 , 2 + c 2 | p i , 2 = p 0 , 2 + δ 0 , 2 ) ) ] × b ( n , p 0 , 1 , x 0 , 1 ) b ( n , p 0 , 2 , x 0 , 2 ) = x 0 , 1 = 0 n x 0 , 2 = 0 n [ 1 max ( P ( X i , 1 x 0 , 1 + c 1 | p i , 1 = p 0 , 1 + δ 0 , 1 ) , P ( x i , 2 X 0 , 2 + c 2 | p i , 2 = p 0 , 2 + δ 0 , 2 ) ) ] k × b ( n , p 0 , 1 , x 0 , 1 ) b ( n , p 0 , 2 , x 0 , 2 )
and so if
x 0 , 1 = 0 n x 0 , 2 = 0 n [ 1 max ( P ( X i , 1 x 0 , 1 + c 1 | p i , 1 = p 0 , 1 + δ 0 , 1 ) , P ( x i , 2 X 0 , 2 + c 2 | p i , 2 = p 0 , 2 + δ 0 , 2 ) ) ] k × b ( n , p 0 , 1 , x 0 , 1 ) b ( n , p 0 , 2 , x 0 , 2 ) P 0 ,
we obtain that P ( C S 0 ) P 0 .
We now provide the proof of Theorem 2 from Section 3.
Proof. 
Suppose that there are k 1 k , treatments better than the control treatment. Without loss of generality we can assume that treatments π i , i = 1 , 2 , , k 1 are better than the control.
The joint probability distribution of X 1 and X 2 can be written as:
P ( X 1 = x 1 , X 2 = x 2 | p 1 , p 2 , ϕ )
= i = m a x ( 0 , X 1 + X 2 n ) m i n ( X 1 , X 2 ) n ! i ! ( X 1 i ) ! ( X 2 i ) ! ( n X 1 X 2 + i ) ! p 11 i p 12 X 1 i p 21 X 2 i p 22 n X 1 X 2 + i
where
p 11 = a a 2 + b 2 ( ϕ 1 ) , a = 1 + ( ϕ 1 ) ( p e + p s ) , b = 4 ϕ ( ϕ 1 ) p e p s
p 12 = p e p 11
p 21 = p s p 11
p 22 = 1 p 11 p 12 p 21
By the monotonicity properties of the joint probability introduced by [19], we have:
P ( C S 1 ) = P ( X i , 1 X 0 , 1 + c 1 , X i , 2 X 0 , 2 + c 2 , i = 1 , 2 , , k 1 | p i , 1 p 0 , 1 + δ 1 , 1 , p i , 2 p 0 , 2 + δ 1 , 2 , ϕ i , i = 1 , 2 , , k 1 ) P ( X i , 1 X 0 , 1 + c 1 , X i , 2 X 0 , 2 + c 2 , i = 1 , 2 , , k | p i , 1 p 0 , 1 + δ 1 , 1 , p i , 2 p 0 , 2 + δ 1 , 2 , ϕ i , i = 1 , 2 , , k ) = x 0 , 1 = 0 n x 0 , 2 = 0 n [ i = 1 k P ( X i , 1 x 0 , 1 + c 1 , X i , 2 x 0 , 2 + c 2 | p i , 1 p 0 , 1 + δ 1 , 1 , p i , 2 p 0 , 2 + δ 1 , 2 , ϕ i ) ] × P ( X 0 , 1 = x 0 , 1 , X 0 , 2 = x 0 , 2 | p 0 , 1 , p 0 , 2 , ϕ 0 ) x 0 , 1 = 0 n x 0 , 2 = 0 n [ i = 1 k P ( X i , 1 x 0 , 1 + c 1 , X i , 2 x 0 , 2 + c 2 | p i , 1 = p 0 , 1 + δ 1 , 1 , p i , 2 = p 0 , 2 + δ 1 , 2 , ϕ i ) ] × P ( X 0 , 1 = x 0 , 1 , X 0 , 2 = x 0 , 2 | p 0 , 1 , p 0 , 2 , ϕ 0 ) = x 0 , 1 = 0 n x 0 , 2 = 0 n [ Π i = 1 k ( x i , 1 = c 1 + x 0 , 1 n x i , 2 = c 2 + x 0 , 2 n P ( X i , 1 = x i , 1 , X i , 2 = x i , 2 | p i , 1 = p 0 , 1 + δ 1 , 1 , p i , 2 = p 0 , 2 + δ 1 , 2 , ϕ i ) ) ] × P ( X 0 , 1 = x 0 , 1 , X 0 , 2 = X 0 , 2 | p 0 , 1 , p 0 , 2 , ϕ 0 ) P 1
P ( C S 0 ) = P ( X i , 1 < X 0 , 1 + c 1 or X i , 2 < X 0 , 2 + c 2 , i = 1 , , k | p i , 1 p 0 , 1 + δ 0 , 1 or p i , 2 p 0 , 2 + δ 0 , 2 , i = 1 , , k ) = i = 1 k P ( X i , 1 < X 0 , 1 + c 1 or X i , 2 < X 0 , 2 + c 2 | p i , 1 p 0 , 1 + δ 0 , 1 or p i , 2 p 0 , 2 + δ 0 , 2 ) = i = 1 k ( 1 P ( X i , 1 X 0 , 1 + c 1 , X i , 2 X 0 , 2 + c 2 | p i , 1 p 0 , 1 + δ 0 , 1 or p i , 2 p 0 , 2 + δ 0 , 2 ) )
Since
P ( X i , 1 c 1 , X i , 2 c 2 | p i , 1 p 0 , 1 + δ 0 , 1 o r p i , 2 p 0 , 2 + δ 0 , 2 ) m a x ( P ( X i , 1 c 1 , X i , 2 c 2 | p i , 1 p 0 , 1 + δ 0 , 1 , p i , 2 p 0 , 2 + δ 0 , 2 ) , P ( X i , 1 c 1 , X i , 2 c 2 | p i , 1 p 0 , 1 + δ 0 , 1 , p i , 2 p 0 , 2 + δ 0 , 2 ) , P ( X i , 1 c 1 , X i , 2 c 2 | p i , 1 p 0 , 1 + δ 0 , 1 , p i , 2 p 0 , 2 + δ 0 , 2 ) ) m a x ( P ( X i , 1 c 1 , X i , 2 c 2 | p i , 1 = p 0 , 1 + δ 0 , 1 , p i , 2 = p 0 , 2 + δ 0 , 2 ) P ( X i , 1 c 1 , X i , 2 c 2 | p i , 1 = p 0 , 1 + δ 0 , 1 , p i , 2 = 1 ) , P ( X i , 1 c 1 , X i , 2 c 2 | p i , 1 = 1 , p i , 2 = p 0 , 2 + δ 0 , 2 ) ) = m a x ( P ( X i , 1 c 1 | p i , 1 = p 0 , 1 + δ 0 , 1 ) , P ( X i , 2 c 2 | p i , 2 = p 0 , 2 + δ 0 , 2 ) )
We obtain that
i = 1 k ( 1 P ( X i , 1 X 0 , 1 + c 1 , X i , 2 X 0 , 2 + c 2 | p i , 1 p 0 , 1 + δ 0 , 1 or p i , 2 p 0 , 2 + δ 0 , 2 ) ) x 0 , 1 = 0 n x 0 , 2 = 0 n i = 1 k [ 1 m a x ( P ( X i , 1 x 0 , 1 + c 1 | p i , 1 = p 0 , 1 + δ 0 , 1 ) , P ( X i , 2 x 0 , 2 + c 2 | p i , 2 = p 0 , 2 + δ 0 , 2 ) ) ] × P ( X 0 , 1 = x 0 , 1 , X 0 , 2 = x 0 , 2 ) = x 0 , 1 = 0 n x 0 , 2 = 0 n [ 1 m a x ( P ( X i , 1 x 0 , 1 + c 1 | p i , 1 = p 0 , 1 + δ 0 , 1 ) , P ( X i , 2 x 0 , 2 + c 2 | p i , 2 = p 0 , 2 + δ 0 , 2 ) ) ] k × P ( X 0 , 1 = x 0 , 1 , X 0 , 2 = x 0 , 2 )
So P ( C S 0 ) x 0 , 1 = 0 n x 0 , 2 = 0 n [ 1 m a x ( P ( X i , 1 x 0 , 1 + c 1 | p i , 1 = p 0 , 1 + δ 0 , 1 ) , P ( X i , 2 x 0 , 2 + c 2 | p i , 2 = p 0 , 2 + δ 0 , 2 ) ) ] k × P ( X 0 , 1 = x 0 , 1 , X 0 , 2 = x 0 , 2 )
We now provide the proof of Theorem 3 from Section 3.
Proof. 
Suppose that there are k 1 k , treatments better than the control treatment. Without loss of generality we can assume that treatments π i , i = 1 , 2 , , k 1 are better than the control.
By the monotonicity properties of the joint probability introduced by [19], we have:
P ( C S 1 ) = P ( X i , 1 X 0 , 1 + c 1 , X i , 2 X 0 , 2 + c 2 , i = 1 , 2 , , k 1 | p i , 1 p 0 , 1 + δ 1 , 1 , p i , 2 p 0 , 2 + δ 1 , 2 , ϕ i , i = 1 , 2 , , k 1 ) P ( X i , 1 X 0 , 1 + c 1 , X i , 2 X 0 , 2 + c 2 , i = 1 , 2 , , k | p i , 1 p 0 , 1 + δ 1 , 1 , p i , 2 p 0 , 2 + δ 1 , 2 , ϕ i , i = 1 , 2 , , k ) = x 0 , 1 = 0 n x 0 , 2 = 0 n [ i = 1 k P ( X i , 1 x 0 , 1 + c 1 , X i , 2 x 0 , 2 + c 2 | p i , 1 p 0 , 1 + δ 1 , 1 , p i , 2 p 0 , 2 + δ 1 , 2 , ϕ i ) ] × P ( X 0 , 1 = x 0 , 1 , X 0 , 2 = x 0 , 2 | p 0 , 1 , p 0 , 2 , ϕ 0 ) x 0 , 1 = 0 n x 0 , 2 = 0 n [ i = 1 k P ( X i , 1 x 0 , 1 + c 1 , X i , 2 x 0 , 2 + c 2 | p i , 1 = p 0 , 1 + δ 1 , 1 , p i , 2 = p 0 , 2 + δ 1 , 2 , ϕ i = 0 ) ] × P ( X 0 , 1 = x 0 , 1 , X 0 , 2 = x 0 , 2 | p 0 , 1 , p 0 , 2 , ϕ 0 ) = x 0 , 1 = 0 n x 0 , 2 = 0 n [ x i , 1 = c 1 + x 0 , 1 n x i , 2 = c 2 + x 0 , 2 n P ( X i , 1 = x i , 1 , X i , 2 = x i , 2 | p i , 1 = p 0 , 1 + δ 1 , 1 , p i , 2 = p 0 , 2 + δ 1 , 2 , ϕ i = 0 ) ] k × P ( X 0 , 1 = x 0 , 1 , X 0 , 2 = X 0 , 2 | p 0 , 1 , p 0 , 2 , ϕ 0 ) P 1
The first inequality now follows if we use the expression of the joint probability distribution of X 1 , 1 and X 1 , 2 when ϕ = 0, described in Section 3.
The proof of second inequality follows similarly to the proof of the second inequality in the previous theorem. □

References

  1. Sobel, M.; Huyett, M.J. Selecting the Best One of Several Binomial Populations. Bell Syst. Tech. J. 1957, 36, 537–576. [Google Scholar] [CrossRef]
  2. Gupta, S.S.; Sobel, M. On Selecting a Subset which Contains all Populations Better than a Standard. Ann. Math. Stat. 1958, 29, 235–244. [Google Scholar] [CrossRef]
  3. Dunnett, C.M. Selection of the Best Treatment in Comparison to a Control with an Application to a Medical Trial. In Design of Experiments, Ranking, and Selection; Santner, T.J., Tamhane, A.C., Eds.; Marcel Dekker: New York, NY, USA, 1984; pp. 47–66. [Google Scholar]
  4. Thall, P.F.; Simon, R.; Ellenberg, S.S. Two-stage designs for comparative clinical trials. Biometrika 1988, 75, 303–310. [Google Scholar] [CrossRef]
  5. Carsten, C.; Chen, P. Curtailed Two-Stage Matched Pairs Design in Double-Arm Phase II Clinical Trials. J. Biopharm. Stat. 2016, 26, 816–822. [Google Scholar] [CrossRef] [PubMed]
  6. Bechhofer, R.E.; Kulkarni, R.V. Closed Adaptive Sequential Procedures for Selecting the Best of k ≥ 2 Bernoulli Populations. No. TR510. 1981. Available online: https://apps.dtic.mil/sti/html/tr/ADA115653/ (accessed on 2 September 2025).
  7. Buzaianu, E.M.; Chen, P. Curtailment Procedure for Selecting Among Bernoulli Populations. Commun. Stat. Theory Methods 2008, 37, 1085–1102. [Google Scholar] [CrossRef]
  8. Jennison, C. Equal Probability of Correct Selection for Bernoulli Selection Procedures. Commun. Stat. Theory Methods 1983, 12, 2887–2896. [Google Scholar] [CrossRef]
  9. Jennison, C.; Turnbull, B.W. Group sequential tests for bivariate response: Interim analysis of clinical trials with both efficacy and safety endpoints. Biometrics 1993, 49, 741–752. [Google Scholar] [CrossRef] [PubMed]
  10. Bryant, J.; Day, R. Incorporating Toxicity Considerations into the Design of Two-Stage Phase II Clinical Trials. Biometrics 1995, 51, 1372–1383. [Google Scholar] [CrossRef] [PubMed]
  11. Conway, M.R.; Petroni, G.R. Bivariate Sequential Designs for Phase II Trials. Biometrics 1995, 51, 656–664. [Google Scholar] [CrossRef]
  12. Chen, C.M.; Chi, Y. Curtailed two-stage designs with two dependent binary endpoints. Pharm. Stat. 2012, 11, 57–62. [Google Scholar] [CrossRef] [PubMed]
  13. Shen, H.; Hong, L.J.; Zhang, X. Ranking and selection with covariates for personalized decision making. INFORMS J. Comput. 2021, 33, 1500–1519. [Google Scholar] [CrossRef]
  14. Ding, L.; Hong, L.J.; Shen, H.; Zhang, X. Knowledge Gradient for Selection with Covariates: Consistency and Computation. In Proceedings of the 2019 Winter Simulation Conference, National Harbor, MD, USA, 8–12 December 2019; pp. 3425–3436. [Google Scholar]
  15. Buzaianu, E.M.; Chen, P.; Hsu, L. A Curtailed Procedure for Selecting Among Treatments With Two Bernoulli Endpoints. Sankhya B 2022, 84, 320–339. [Google Scholar] [CrossRef]
  16. Yin, C.; Buzaianu, E.M.; Chen, P.; Hsu, L. A Design for Selecting Among Treatments with Two Binary Endpoints. In Manuscript Submitted to Communications in Statistics—Theory and Methods. Status: Under Review; Taylor & Francis Inc.: Philadelphia, PA, USA, 1976. [Google Scholar]
  17. Thall, P.F.; Cheng, S.C. Treatment comparisons based on two-dimensional safety and efficacy alternatives in oncology trials. Biometrics 1999, 55, 746–753. [Google Scholar] [CrossRef] [PubMed]
  18. Gupta, M.K.; Panchapakesan, S. Multiple Decision Procedures: Theory and Methodology of Selection and Ranking Populations; SIAM: Philadelphia, PA, USA, 2002. [Google Scholar]
  19. Buzaianu, E.M.; Chen, P.; Hsu, L. Selecting among treatments with two Bernoulli endpoints. Commun. Stat. Theory Methods 2024, 53, 1964–1984. [Google Scholar] [CrossRef]
Table 1. Classification table.
Table 1. Classification table.
Second Endpoint
First Endpoint12
1 X 11 X 12 X 1
2 X 21 X 22 X 1 ¯
X 2 X 2 ¯ n
Table 2. Design parameters when k = 2 ,   P 1 = 0.80 .
Table 2. Design parameters when k = 2 ,   P 1 = 0.80 .
p 0 , 1 ϕ n c 1 c 2 N E 0 ( N | R ) E 1 ( N | R ) E ( N | R ) RS %
0.40811413243195.96232.41214.1811.86
0.01811413243196.61232.10214.3611.79
0.1781412234190.27222.84206.5511.73
1771412231189.83219.01204.4211.51
2751312225186.30212.83199.5611.3
4751312225187.11212.41199.7611.22
8711311213177.07200.73188.9011.32
100691211207174.27194.16184.2211.01
0.50801413240193.65229.38211.5211.87
0.01801413240193.97229.26211.6111.83
0.1771412231187.35220.16203.7611.79
1761412228186.77216.31201.5411.61
2741312222183.31210.23196.7711.37
4741312222184.22209.72196.9711.27
8701311210174.25198186.1211.37
100671211201169.16188.56178.8611.01
0.60751412225180.99215.14198.0611.97
0.01751412225181.04215.12198.0811.97
0.1731312219177.48208.88193.1811.79
1731312219179.61207.89193.7511.53
2711411213173.74202.21187.9711.75
4691311207170.59195.87183.2311.48
8671211201167.16189.63178.4011.25
100621210186155.61174.58165.1011.24
Note : k = 2 ,   p 0 , 2 = 0.60 ,   δ 0 , 1 = 0.01 ,   δ 0 , 2 = 0.01 ,   δ 1 , 1 = 0.30 ,   δ 1 , 2 = 0.25 ,   P 0 = 0.90 ,   P 1 = 0.80 .
Table 3. Design parameters when k = 2 ,   P 1 = 0.85 .
Table 3. Design parameters when k = 2 ,   P 1 = 0.85 .
p 0 , 1 ϕ n c 1 c 2 N E 0 ( N | R ) E 1 ( N | R ) E ( N | R ) RS ( % )
0.40871513261212.04248.41230.2311.79
0.01871513261212.71248.08230.4011.73
0.10851413255209.71241.68225.7011.49
1851413255212.24240.44226.3411.24
2811412243202.46228.73215.6011.28
4811412243203.30228.21215.7511.21
8811412243204.02227.76215.8911.16
100781312234198.24218.67208.4610.92
0.50861513258209.73245.41227.5711.79
0.01861513258209.97245.28227.6311.77
0.10841413252206.76238.98222.8711.56
1841413252209.18237.80223.4911.31
2801412240199.46226.11212.7911.34
4801412240200.36225.57212.9711.26
8781312234196.99219.37208.1811.03
100731311219185.24204.25194.7511.07
0.60811512243196.90231.23214.0611.91
0.01811512243197.00231.18214.0911.90
0.10811512243197.70230.81214.2511.83
1781412234192.70221.34207.0211.53
2781412234193.54220.89207.2211.45
4771312231193.19217.16205.1811.18
8741411222184.30208.86196.5811.45
100701211210178.72195.91187.3210.80
Note : k = 2 ,   p 0 , 2 = 0.60 ,   δ 0 , 1 = 0.01 ,   δ 0 , 2 = 0.01 ,   δ 1 , 1 = 0.30 ,   δ 1 , 2 = 0.25 ,   P 0 = 0.90 ,   P 1 = 0.85 .
Table 4. Design parameters when k = 2 ,   P 1 = 0.90 .
Table 4. Design parameters when k = 2 ,   P 1 = 0.90 .
p 0 , 1 ϕ n c 1 c 2 N E 0 ( N | R ) E 1 ( N | R ) E ( N | R ) RS ( % )
0.40961514288237.24272.11254.6711.57
0.01961514288237.93271.76254.8511.51
0.10961514288239.33271.17255.2511.37
1921513276231.39258.66245.0211.22
2921513276232.31258.15245.2311.15
4901413270228.93251.92240.4210.95
8901413270229.75251.42240.5810.90
100851412255217.44236.74227.0910.94
0.50951514285234.97269.09252.0311.57
0.01951514285235.24268.96252.1011.54
0.10931613279230.01263.14246.5711.62
1911513273228.35256.03242.1911.29
2911513273229.28255.49242.3911.21
4891413267225.99249.27237.6311.00
8891413267226.93248.71237.8210.93
100821312246211.21227.98219.5910.73
0.60891513267219.34252.27235.8011.68
0.01891513267219.46252.20235.8311.67
0.10891513267220.23251.86236.0411.59
1881413264221.17247.47234.3211.24
2871612261216.51245.11230.8111.57
4851512255213.46238.66226.0611.35
8831412249210.29232.34221.3111.12
100781411234198.57217.13207.8511.17
Note : k = 2 ,   p 0 , 2 = 0.60 ,   δ 0 , 1 = 0.01 ,   δ 0 , 2 = 0.01 ,   δ 1 , 1 = 0.30 ,   δ 1 , 2 = 0.25 ,   P 0 = 0.90 ,   P 1 = 0.90 .
Table 5. Design parameters when k = 3 ,   P 1 = 0.80 .
Table 5. Design parameters when k = 3 ,   P 1 = 0.80 .
p 0 , 1 ϕ n c 1 c 2 N E 0 ( N | R ) E 1 ( N | R ) E ( N | R ) RS ( % )
0.40.00951615380308.89362.28335.5911.69
0.01941714376305.01357.95331.4811.84
0.10931714372303.28354.08328.6811.65
1.00911614364300.94344.40322.6711.35
2.00891514356296.73335.82316.2711.16
4.00881514352293.36331.98312.6711.17
8.00861414344288.88323.93306.4110.93
100.00821413328277.11307.69292.4010.85
0.50.00921714368297.22351.12324.1711.91
0.01921714368297.68351.02324.3511.86
0.10921714368299.18350.45324.8211.73
1.00901614360296.63340.58318.6111.50
2.00881514352292.18332.41312.3011.28
4.00871514348289.80328.60309.2011.15
8.00851613340282.10320.37301.2411.40
100.00801413320270.15300.19285.1710.88
0.60.00881614352284.93335.77310.3511.83
0.01881614352285.05335.56310.3011.85
0.10861514344280.29327.42303.8611.67
1.00861713344279.82326.74303.2811.84
2.00851613340279.78321.82300.8011.53
4.00821513328271.86309.96290.9111.31
8.00801413320267.28301.55284.4211.12
100.00741312296250.71277.28263.9910.81
Note : k = 3 ,   p 0 , 2 = 0.60 ,   δ 0 , 1 = 0.01 ,   δ 0 , 2 = 0.01 ,   δ 1 , 1 = 0.30 ,   δ 1 , 2 = 0.25 ,   P 0 = 0.90 ,   P 1 = 0.80 .
Table 6. Design parameters when k = 3 ,   P 1 = 0.85 .
Table 6. Design parameters when k = 3 ,   P 1 = 0.85 .
p 0 , 1 ϕ n c 1 c 2 N E 0 ( N | R ) E 1 ( N | R ) E ( N | R ) RS ( % )
0.40.001011715404330.13383.44356.7811.69
0.011011715404330.85382.99356.9211.65
0.101011715404333.10382.08357.5911.49
1.00981615392326.96369.43348.1911.18
2.00971714388323.57364.62344.0911.32
4.00951614380319.33356.19337.7611.12
8.00941614376316.11352.35334.2311.11
100.00911514364309.03340.14324.5910.83
0.50.00991715396323.11376.07349.5911.72
0.01991715396323.68375.74349.7111.69
0.10991715396325.23375.14350.1811.57
1.00971615388322.80365.79344.2911.26
2.00961714384319.40361.15340.2811.39
4.00931614372311.32349.43330.3811.19
8.00931614372312.71348.76330.7411.09
100.00861513344291.61320.66306.1311.01
0.60.00941714376306.26357.10331.6811.79
0.01941714376306.41356.78331.6011.81
0.10941714376307.53356.55332.0411.69
1.00921614368305.01347.17326.0911.39
2.00911614364302.20343.25322.7211.34
4.00901514360301.71338.37320.0411.10
8.00871613348290.26326.94308.6011.32
100.00841513336284.41313.77299.0910.98
Note : k = 3 ,   p 0 , 2 = 0.60 ,   δ 0 , 1 = 0.01 ,   δ 0 , 2 = 0.01 ,   δ 1 , 1 = 0.30 ,   δ 1 , 2 = 0.25 ,   P 0 = 0.90 ,   P 1 = 0.85 .
Table 7. Design parameters when k = 3 ,   P 1 = 0.90 .
Table 7. Design parameters when k = 3 ,   P 1 = 0.90 .
p 0 , 1 ϕ n c 1 c 2 N E 0 ( N | R ) E 1 ( N | R ) E ( N | R ) RS ( % )
0.40.001111816444365.66419.29392.4811.60
0.011111816444366.71418.74392.7311.55
0.101101716440366.62413.66390.1411.33
1.001081815432361.99404.50383.2411.29
2.001061715424358.00395.96376.9811.09
4.001061715424359.44395.26377.3511.00
8.001041615416354.66387.29370.9710.82
100.001011714404343.97375.04359.5011.01
0.50.001091915436357.96411.50384.7311.76
0.011091915436358.33411.49384.9111.72
0.101091915436360.19410.60385.4011.61
1.001071815428358.00401.18379.5911.31
2.001051715420354.04392.32373.1811.15
4.001041715416351.23388.19369.7111.13
8.001031615412350.99383.49367.2410.86
100.00971614388332.10359.83345.9710.83
0.60.001041815416341.64392.59367.1111.75
0.011031715412340.19388.10364.1411.62
0.101031715412341.38387.88364.6311.50
1.001021914408337.08383.75360.4111.66
2.001001814400332.66375.32353.9911.50
4.00981714392329.37366.73348.0511.21
8.00971614388328.98361.58345.2811.01
100.00911613364309.59337.64323.6211.09
Note : k = 3 ,   p 0 , 2 = 0.60 ,   δ 0 , 1 = 0.01 ,   δ 0 , 2 = 0.01 ,   δ 1 , 1 = 0.30 ,   δ 1 , 2 = 0.25 ,   P 0 = 0.90 ,   P 1 = 0.90 .
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yin, C.; Buzaianu, E.M.; Chen, P.; Hsu, L. Subset Selection with Curtailment Among Treatments with Two Binary Endpoints in Comparison with a Control. Mathematics 2025, 13, 3067. https://doi.org/10.3390/math13193067

AMA Style

Yin C, Buzaianu EM, Chen P, Hsu L. Subset Selection with Curtailment Among Treatments with Two Binary Endpoints in Comparison with a Control. Mathematics. 2025; 13(19):3067. https://doi.org/10.3390/math13193067

Chicago/Turabian Style

Yin, Chishu, Elena M. Buzaianu, Pinyuen Chen, and Lifang Hsu. 2025. "Subset Selection with Curtailment Among Treatments with Two Binary Endpoints in Comparison with a Control" Mathematics 13, no. 19: 3067. https://doi.org/10.3390/math13193067

APA Style

Yin, C., Buzaianu, E. M., Chen, P., & Hsu, L. (2025). Subset Selection with Curtailment Among Treatments with Two Binary Endpoints in Comparison with a Control. Mathematics, 13(19), 3067. https://doi.org/10.3390/math13193067

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop