Next Article in Journal
Incorporating Clustering Techniques into GAMLSS
Previous Article in Journal
The One Standard Error Rule for Model Selection: Does It Work?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Credibility of Causal Estimates from Regression Discontinuity Designs with Multiple Assignment Variables

1
School of Natural and Applied Sciences, Sol Plaatje University, Kimberley 8301, South Africa
2
School of Statistics and Actuarial Science, University of the Witwatersrand, Johannesburg 2050, South Africa
*
Author to whom correspondence should be addressed.
Stats 2021, 4(4), 893-915; https://doi.org/10.3390/stats4040052
Submission received: 13 August 2021 / Revised: 17 October 2021 / Accepted: 4 November 2021 / Published: 11 November 2021

Abstract

:
In this paper, we determine treatment effects when the treatment assignment is based on two or more cut-off points of covariates rather than on one cut-off point of one assignment variable. using methods that are referred to as multivariate regression discontinuity designs (MRDD). One major finding of this paper is the discovery of new evidence that both matric points and household income have a huge impact on the probability of eligibility for funding from the National Student Financial Aid Scheme (NSFAS) to study for a bachelor’s degree program at universities in South Africa. This evidence will inform policymakers and educational practitioners on the effects of matric points and household income on the eligibility for NSFAS funding. The availability of the NSFAS grant impacts greatly students’ decisions to attend university or seek other opportunities elsewhere. Using the frontier MRDD analytical results, barely scoring matric points greater than or equal to 25 points compared to scoring matric points less than 25 for students whose household income is less than R350,000 (≈US$2500) increases the probability of eligibility for NSFAS funding by a significant 3.75 ( p-value = 0.0001 < 0.05) percentage points. Therefore, we have shown that the frontier MRDD can be employed to determine the causal effects of barely meeting the requirements of one assignment variable, among the subjects that either meet or fail to meet the requirements of the other assignment variable.

1. Introduction

Multivariate regression discontinuity designs (MRDD) raise challenges that are distinct from those identified in traditional RDD [1]. Traditional RDD studies focus on units that are assigned to a treatment based on a single cut-off point and a single continuous assignment variable [2]. In reality, units are usually assigned to a treatment based on more than one continuous assignment variable [3]. Thus, treatment effects may be estimated across a multi-dimensional cut-off frontier, as opposed to a single point on the assignment variable, using methods referred to as multivariate regression discontinuity designs (MRDD) [1,3,4]. For example, using the frontier approach, [1] showed that the MRDD treatment effect estimate, τ M R D D , may be decomposed into a weighted average of two univariate RDD effects, τ x 1 at the X 1 -cut-off and τ x 2 at the X 2 -cut-off. The term frontier means that the average treatment effect estimates for MRDD are only for sub-populations of units located at the cut-off frontier as opposed to the average treatment effect for the overall population under study. Like in the standard RDD, the causal estimates obtained in MRDD have limited external validity as the causal estimates are only identified for observations in the immediate vicinity of the cut-off scores [4]. This paper studies the credibility of estimates of the MRDD. It is important that after estimating τ M R D , a researcher checks for the plausibility of the assumptions of the MRDD estimates. With more credible estimates, inference about causality can reduce the reliance of the causal estimates on the following modelling assumptions [5]:
  • the cut-off scores determining treatment assignment are exogenously set;
  • potential outcomes are continuous functions of the assignment scores at the cut-off scores and
  • the functional form of the model is correctly specified.
Ref. [5] suggest extending the assumption checking for the RDD context to the MRDD. The assumptions will be assessed as they apply to the frontier regression discontinuity using supplementary analyses techniques [6]. Primary analyses in MRDD focus on estimating τ M R D . In contrast, supplementary analyses seek to shed more light on τ M R D from the primary analyses. These supplementary analyses use the fact that assumptions behind the identification strategy often have implications for the data beyond those used in the primary analyses [6]. There are basically four approaches that can be employed for carrying out supplementary analyses, and these are: The McCrary test [7], placebo analysis, robustness and sensitivity, and checking for discontinuity of the assignment variable at the cut-off point.
This paper uses two assignment variables to study the effects of an educational funding intervention in South Africa, the National Student Financial Aid Scheme (NSFAS). The eligibility criteria for NSFAS funding is based on a student’s family annual total household income and the aggregate points they score on their matriculation certificate. In South Africa, a matriculation certificate is a school-leaving diploma or national senior certificate (NSC) that is awarded after completing Grade 12. The total number of scores achieved on seven subjects on the matriculation certificate are referred to in this paper as the matric points. The role that NSFAS funding has on access to post-school education for students from disadvantaged backgrounds in South Africa cannot be ignored. From a Government policy perspective, NSFAS was established to correct apartheid era discrimination in higher education that occurred in South Africa in the past and promote equal access to higher education. In addition, NSFAS is established a sustainable and affordable funding scheme. Therefore, one of the vital objectives of this research is to quantify the causal effect of household income and matric points on the chance of eligibility for NSFAS funding. This will help policy makers to make the right decisions that will ensure that students are adequately funded. For example, there are students who are referred to as the ‘missing middle’ [8]. These students are from households whose incomes are between R350,000 and R600,000. These students do not qualify for NSFAS funding, and yet most of them cannot afford to pay for their higher education. By quantifying the causal effect that matric points and household income has on the ‘missing middle’s eligibility for NSFAS funding, government can implement policy interventions to assist these students to get financial aid.

Related Work

MRDD is a growing approach, and yet there is not enough easily accessible information that can help researchers on how to implement MRDD, and at the same time assess the credibility of the causal estimates using supplementary analysis. MRDD presents opportunities for obtaining unbiased estimates of treatment effects using the same thinking as the single assignment variable RD designs. The MRDD is an extension of the traditional RDD, except that the treatment effects are estimated for multiple cut-offs, as opposed to a single cut-off point. Studies have been carried out with multiple assignment variables and cut-off points for treatment assignment. For example, ref. [4] gave an example of teachers who received a bonus for improving student scores in both Mathematics and English Language Arts (ELA). The authors point out that in some cases, students must pass externally defined standards in several subjects to avoid summer school or to graduate from high school. Multiple regression discontinuity designs (MRDD) have also been used in other disciplines, such as in politics [9]. Authors [2] gave a compilation of over sixty studies, showing where RDD designs have been applied in many different contexts. Ref. [10] estimates the effect of a mandatory summer school program assigned to students who fail to score higher than a preset cut-off in both math and reading exams. Ref. [11] developed a special kind of regression discontinuity design in which the assignment variable is geographic. Their approach, the geographic regression discontinuity (GRD) design, was similar to a standard RD but with two running variables.
MRDD are very appealing, and yet the statistical details regarding their implementation are not as easy as they may seem at first. This is because when the MRDD are applied in other technical journals they require a high degree of technical sophistication to read. Furthermore, the terminology that is used is not well defined and is often used inconsistently. Thus, this paper is filling a void and serving as a guide to practitioners on how to implement MRDD.
The design of this paper builds upon the work by [3,4,12]. Ref. [12]’s paper investigated how the CalGrant program impacted on the college going rates from 1998 to 1999 using a sharp regression discontinuity design. Ref. [3] conducted an extensive simulation study that was closely based on [12]’s work. This paper takes a similar approach to that of [3,12] and considers eligibility criteria for the NSFAS funding program in South Africa to estimate the impact that NSFAS funding has on subsequent enrollment for a bachelor’s degree into SA Universities. Specifically, this paper makes the following contributions:
1
This is the first time that the MRDD has been applied in the context of South Africa to quantify the causal effect that the household income and matric points have on the probability of eligibility for NSFAS funding, and eligibility to study for a bachelor’s degree. By quantifying the causal effect of household and income, policy makers in South Africa can make informed decisions on funding for for students who qualify to study for a bachelor’s degree.
2
This paper provides the first evidence on whether meeting a matric points threshold and a household income threshold increases the chance of eligibility for NSFAS funding.
3
The paper adds to the literature by combining the estimation of causal estimates using MRDD, and some of the supplementary analysis proposed by [6]. The authors noted as a concern that assessing the validity of the assumptions required for interpreting the estimates as causal effects obtained from regression discontinuity analysis is still lagging behind. Therefore, the paper extends the assumption checking of uni-variate RDD to the MRDD using supplementary analyses. These supplementary analyses are carried out to test for discontinuities in average covariate values at the threshold as well as to assess the credibility of the design, and in particular to test for evidence of manipulation of the forcing variable.This is crucial in practice because if the causal effect are not credible, then they are not useful.
4
We have successfully demonstrated that one can use simulated data that closely mimics real world or original data, and still obtain significant and credible causal effect estimates.

2. Literature Review

2.1. Multivariate Regression Discontinuity Design

The basic set-up of RDD is that we are interested in the causal effect of a binary treatment or program denoted by W i , in the presence of an exogenous variable (forcing variable) denoted by X i . There are different RDD methods proposed in the literature that allow us to estimate average treatment effects. Refs. [1,2,4,6,13] provide detailed overviews of these methods. The RDD is framed in the context of the potential outcomes framework, i.e., for a given unit i, there exists two potential outcomes, Y i ( 1 ) and Y i ( 0 ) , and the causal effect is simply the difference, Y i ( 1 ) Y i ( 0 ) . The fundamental problem of causal inference states that we cannot observe Y i ( 1 ) and Y i ( 0 ) at the same time. Because of this problem we usually focus on the average effects of the treatment, that is, averages of Y i ( 1 ) Y i ( 0 ) over all sub-populations rather than on unit level effects [2].
The Multivariate Regression discontinuity designs (MRDD) present opportunities for obtaining unbiased estimates of treatment effects using the same thinking as the single assignment variable RD designs. The MRDD is an extension of the traditional RDD, except that the treatment effects are estimated for multiple cut-offs, as opposed to a single cut-off point. Studies have been carried out with multiple assignment variables and cut-off points for treatment assignment. For example, ref. [4] gave an example of teachers who received a bonus for improving student scores in both Mathematics and English Language Arts (ELA). The authors point out that in some cases students must pass externally defined standards in several subjects to avoid summer school or to graduate from high school. Multiple regression discontinuity designs (MRDD) have also been used in other disciplines, such as in politics [9]. There are multiple strategies for estimating MRDD and these strategies present multiple possible estimands ( τ M R D D ). Ref. [5] presents five strategies for estimating treatment effects in MRDD, namely: binding-score RD, distance based RD, frontier RD, the response surface RD, and fuzzy frontier RD. The authors mention that these four methods have their own advantages and disadvantages that depend on the structure of the data with regards to the correlation between the two assignment variables as well as the locations of the respective cut-points. Ref. [5] indicate that the fuzzy IV methods have limited practical applications as they tend to yield imprecise estimates. Additionally, the response surface RD is sensitive to the functional form mis-specification and this could be a problem when it is implemented using local linear regression, which requires full functional form specification.
In this paper, we apply the frontier RD [5] to estimate the probability of eligibility for NSFAS funding using matric points and household income as the assignment variables. The frontier approach is deemed to be straightforward to implement as the modelling assumptions mentioned earlier are easily assessed by sub-setting the data by frontier [5]. In addition, because the frontier approach reduces the MRDD to at least one single-rating RDDs, they are easily implemented using traditional RDD methods. We implement the frontier approach to obtain the primary analyses estimates. Thus, we adopt the approaches proposed by [4,5] that utilise two assignment variables to assign individuals to a range of different treatment conditions. According to [4], if there are J forcing variables such that W j i = 1 ( X j i c ), ∀j = 1,…,J, then there will be 2 J treatment conditions. For any combination of these 2 J treatment effects, the parameters of interest become the left and right side limits on either side of the cut-off. MRDD with two assignment variables define four different treatment conditions.
For two assignment variables, X 1 and X 2 , and respective cut-offs, c 1 and c 2 , the treatment conditions W 1 i and W 2 i are defined as follows:
W 1 i = 1 , if X 1 i c 1 0 , X 1 i < c 1 and W 2 i = 1 , if X 2 i c 2 0 , X 2 i < c 2
To qualify for NSFAS funding, a student must have a matric certificate with at least 25 matric points and the total family household income should be at most R350,000. MP = 25 and INC = R350,000 are therefore the cut-off points for matric points and household income respectively. Thus, as described by [14], with two assignment variables, there are four treatment conditions that define four distinct regions in a two dimensional space spanned by M P C and I N C C , the centered variables of MP and INC around their respective cut-off points.
1.
Treatment 1: If students score at least 25 matric points and family income is greater than R350,000 (Region 1): W 1 i = 1 and W 2 i = 0
2.
Treatment 2: If students score less than 25 matric points and family income is greater than R350,000 (Region 2): W 1 i = 0 and W 2 i = 0
3.
Treatment 3: If students score less than 25 matric points and family income is at most R350,000 (Region 3): W 1 i = 0 and W 2 i = 1
4.
Treatment 4: If students score at least 25 matric points and family income is at most R350,000 (Region 4): W 1 i = 1 and W 2 i = 1
The parameters that we estimate then become the population conditional mean probabilities of being eligible for NSFAS funding for students in each treatment condition at the appropriate cut-off point. For example, the causal effect of being eligible for NSFAS funding instead of not being eligible for NSFAS funding, for a student scoring on the income (INC) cut-off is then the difference between [2]:
u l ( c ) = lim x 0 E [ N S F i | M P C = x , I N C C = 0 ]
and
u r ( c ) = lim x 0 + E [ N S F i | M P C = x , I N C C = 0 ]

2.2. Multiple Assignment Variables: Estimation Strategies

Equation (2) gives the basic analytic strategy for estimating treatment effects using either single or multiple assignment variables. Observed data is used to estimate the limits of the average potential outcomes at the boundaries of two treatment assignment regions, and then take the difference of the estimated limits. According to [5], these limits must be estimated at the boundary of the observed data for each treatment condition and fitting a regression model of the following form (assuming a sharp regression discontinuity):
Y i = f ( X 1 i , X 2 i , , X j i , W i ) + ϵ i
where X 1 i , X 2 i , , X j i DS and D is the domain of observations used to estimate the model,S is the domain of observations in the full sample, and W i is a zero/one treatment-assignment indicator.
For MRDD with two assignment variables, we have the following:
Y i = f ( X 1 i , X 2 i , W i ) + ϵ i
where X 1 i , X 2 i DS and D is the domain of observations used to estimate the model, S is the domain of observations in the full sample, and W i is a zero/one treatment-assignment indicator [15]. The authors point out that estimators of the MRDD differ in terms of:
(i)
the specification of the function f;
(ii)
the domain (D) of observations used in estimating the model.

3. Materials and Methods

3.1. Data

We focus on estimating the impact that matric points and household income have on the chance that a student qualifies for the NSFAS grant to support his or her bachelor’s degree studies using simulated data for household income (INC) and matric points (MP). There are similar studies that have used simulated data on regression discontinuity designs. For example, [16], developed automated local regression models using synthetic data, and only used real world data to evaluate the models. Ref. [3] simulated a dataset using summary statistics reported in a research paper by [12]. Ref. [3] indicate that they used summary statistics because the original data from [12] was no longer available. In this paper, we took a similar approach and used summary statistics to simulate South Africa’s household income data. The simulated data for household income was generated using the log-normal distribution. The use of the lognormal distribution is motivated by [3,17]. Ref. [17] report that income distribution datasets are usually fit using the lognormal distribution. This finding by [17] is also supported by [18], who state that most literature on income distributions indicate that country income distributions closely resemble the lognormal distribution. Therefore, to simulate income data for South Africa that follows a log-normal distribution, we estimate the mean and standard deviation of total household income from Table 1 of the average household incomes by population group in South Africa (http://www.statssa.gov.za/publications/P0310/P03102014.pdf), (accessed on 11 April 2021).
On the other hand, the use of the truncated beta distribution for the matric points (matric points are similar to GPA) was inspired by [19], who state that the normal assumption is usually employed in mordern educational data. However, the authors state that the normal distribution is not the best choice, and the logitnormal distribution and a scaled, shifted, and truncated beta distribution are better choices for modelling examination scores. Following [3], we chose a scaled, shifted, and truncated beta distribution as it is similar to the empirical distribution reported in [12]. Using Table 2, we consider seven matric subjects and generate matric points that range from 0 to 49. If a student does not achieve any of the subject pass levels that are shown in Table 2 in seven subjects, then the students scores a total of zero matric points. On the other hand, if a student achieves a subject pass level equal to seven in seven subjects, then the student will have a total of 49 matric points.
Different combinations of the scaled, shifted, and truncated beta distribution parameters α and β were employed in simulating the matric points data. We chose α = 5, β = 13 as the preferred parameters for simulating matric points as they give us the density of matric points slightly positively skewed to the right as shown in Figure 1.
After simulating the income and matric points data, we use the data generation process in Equation (5), to generate estimates of the probabilities of eligibility for NSFAS funding P ( N S F i = 1 | I N C c , M P c ) :
P ( N S F i = 1 | I N C c , M P c ) = β 0 + β 1 I N C c + β 2 M P c + I N C c × M P c + ϵ i

3.2. Estimating Causal Effects Using the Frontier Regression Discontinuity Design (FRDD)

The FRDD approach analyses samples of the data based on status (i.e., above or below the cut-off score), on all but one of the assignment variables and then models the discontinuity along the remaining assignment variable using the single assignment variable RDD methods [5]. Focusing on the M P c frontier method entails using sample points where I N C c 0 . Because the assignment variable is centered on its cut-point value, sample points where I N C c 0 are used. For example, when estimating Treatment 1 versus Treatment 2 in Figure 2, we limit the sample only to individuals with I N C c 0 .
Thereafter, we choose a sub-sample from the samples that lie within the optimal bandwidth on the left or right of M P c . This sub-sample will be used in fitting linear probability models of the form:
P ( N S F i = 1 | M P c , T m p i ) = β 0 + β 1 T m p i + β 2 M P i c + β 3 M P i c × T m p i + ϵ i ,
or
P ( N S F i = 1 | M P c , T m p i , I N C c ) = β 0 + β 1 T m p i + β 2 M P i c + β 3 M P i c × T m p i + β 4 I N C i c + ϵ i ,
where P ( N S F i = 1 | M P c , T m p i ) is the probability of eligibility for NSFAS funding given M P c and the treatment assignment variable for matric points, T m p i . Equation (7) includes the covariate I N C c to improve the treatment effect, β 1 ^ ’s precision. The equations are also adapted and expressed as a function of I N C c , T i n c i , and M P c as a covariate, to evaluate the effect of the income variable on the chance of eligibility for NSFAS funding.
We use the linear probability model (LPM) specification, rather than a logistic or probit, to model Equations (6) and (7) because in terms of model fit, LPM is just as good as logistic regression withing a certain range of probabilities to be modelled [20]. Ref. [20] recommends that the probabilities being modelled should be between between 0.20 and 0.80. A plot probability (p) versus the log-odds (Figure 3) illustrates the range of probabilities that was employed.
Figure 3 shows that between 0.20 and 0.80, the relationship between probability (p) and the log odds is almost linear. Therefore, we applied the LPM within this range. We use the linear probability model because according to [14,21], the model becomes more credible when it is applied within range 0.20 and 0.80 for the probability, and a narrow bandwidth for the independent variable(s), using local linear-regression analysis. Furthermore, refs. [14,21] report that the linear probability specification provides consistent and unbiased estimates of the fundamental trends for samples. We used LPM because it is easy to interpret as compared to the logistic model. This means that the regression parameters obtained from fitting LPM are easy to interpret in terms of the population differences in the probability of eligibility for NSFAS funding per unit difference in either matric points or household income. In addition, ref. [14] compared the linear probability model and logistic regression and found out that, within a narrow and optimal bandwidth, these two models produce identical results. Ref. [22] highlight that the linear probability model yields treatment estimates that are just as accurate as those estimated by logistic regression. As we are estimating the effect of matric points and household income on the probability of eligibility for NSFAS funding, the parameter of interest is the coefficient ( β 1 ) of the treatment variable(s) and not the coefficients of the assignment variables. This makes the linear probability model an appropriate analytic procedure for estimating the effect of matric points and household income since the treatment status is a binary variable [22]. Thus, the functional form concerns about the linear probability model may not necessarily apply because all that is required is to estimate the treatment effects, as opposed to estimating the effect of the continuous assignment variables.

4. Experiments

To implement the frontier multiple regression–discontinuity design, we consider three variables, namely, the matric points (MP), household income (INC), and NSFAS funding (NSF). Students apply for NSFAS funding in Grade 12 before their matric results are out and they do not know in advance whether or not they are going to achieve matric points that meet the requirement to study for a bachelor’s degree. This makes MP suitable as an assignment variable, as it cannot be manipulated precisely. MPs around the threshold then become as good as random. The other variable is total household income (INC), which is a measure of the total income of members of the household. The variable NSF indicates the chance of a student being eligible for NSFAS funding to study for a bachelor’s degree within one year after matriculation. Data is simulated that contains MP scores and Income (INC) values with cut-off points of 25 and R350,000 (≈US$25,000), respectively. We create continuous assignment variables (predictors) by centering the normalised MP and INC scores on the values of their respective cut-off points. The centered continuous predictors are labelled M P c and I N C c . Thus, M P c = 0 and I N C c = 0 indicate that the student achieved the minimum requirements for qualifying for NSFAS funding. Following the approach by [14], we create dichotomous versions of the same predictors (MP and INC) labelled, T m p and T i n c , which are the treatment variables that indicate whether a student met the minimum passing standard (MP = 25 points) required in order to study for a bachelor’s degree and also met the maximum income threshold (INC = R350,000) for NSFAS funding. A binary outcome variable Y i was generated such that Y i = 1 if M P i c 0 and I N C i c 0 , i.e., a student receives NSFAS funding and 0 otherwise, i.e, a student does not receive NSFAS funding. The probabilities (between 0 and 1) of eligibility for NSFAS funding are then estimated by fitting a logistic regression to Equation (8), adapted from [4,14], thereby giving the outcome variable as a probability;
P ( N S F i = 1 | M P c , I N C c ) = γ 0 ^ + γ 1 ^ M P i c + γ 2 ^ I N C i c + γ 3 ^ ( M P i c × I N C i c ) + ϵ i
Under normal conditions, the full functional forms of the data generation models are usually not known [5]. Therefore, choosing an estimation method and the best way to implement it could be challenging. Thus, we took a systematic approach where we have relied on programmed algorithms to select the bandwidth that ensures that the chosen functional form approximates our simulated data well. Equations (6) and (7) were used to specify the linear functional forms using an optimal bandwidth that minimises the mean squared error of β 1 ^ . This bandwidth is chosen by utilising a non-parametric density estimation algorithm [23]. The causal effects are then estimated by fitting a linear probability model to Equations (6) and (7), the data that lies within the chosen optimal bandwidths and then derive the estimates of β 1 ^ . Equation (6) specifies the outcomes as a linear function of, say M P c , the treatment variable, T m p , an interaction of M P c and T m p . Equation (7) adds I N C c as a baseline covariate. Besides, we investigate whether adding I N C c 0 as a covariate improves the precision of the estimates of β 1 ^ .
Simulations were conducted based on the values of income, matric points, and Equations (6) and (7). M P c and I N C c were generated as continuous variables. We introduced different values of σ ( σ = 0, 0.05, 0.1 and 0.15) to control for the amount of variation in the assignment variables. The simulated data was initially analysed without adding more variation to the simulated assignment variables, this represented σ = 0. The variation in the assignment variables was gradually increased in steps of 0.05 from 0 to 0.15, and the models were rerun to give the new estimates of the parameters of interest. We will investigate whether or not increasing the variation in the assignment variables affects the treatment assignment mechanism and thus, the treatment effect estimates. Additionally, we considered different samples of sizes 5000, 10,000, 20,000 to determine whether or not increasing the sample size affects the estimated treatment effects. 1000 simulations were performed for each model, by varying the sample sizes and the variability σ that controls the amount of variation in the assignment variables. Overall, the simulation scenarios were composed of three different levels of sample sizes, four error variances and four treatment regions (Figure 2) for each of Equations (6) and (7) giving a total of 96 scenarios. For each scenario we generated 1000 simulated data sets and used Equations (6) and (7) to estimate the the average coefficient β 1 , which gives the estimate of the treatment effect. Because, the true relationship between a binary outcome and continuous explanatory variables is fundamentally not linear, the functional form of the linear probability model is generally not correctly specified. In this paper, we will show through supplementary analyses that within a narrow and optimal bandwidth, the linear specification model is still credible.

5. Results and Analysis

5.1. Estimation of the Causal Effects

The results obtained by fitting the simulated data in each treatment region (Figure 2) to Equation (6) are presented in Table 3 and Table 4. Table 3 shows that for treatment regions R3 vs. R4, the treatment coefficients β 1 ’s are statistically significant (p-value = 0.0119 < 0.05) when N= 20,000 for σ = 0. Additionally, the coefficients are statistically significant (p-value = 0.0349 < 0.05) when N = 20,000 for σ = 0.05. These findings suggest that Equation (6) requires a bigger sample size for it to start yielding significant treatment effects. However, using Equation (7), which includes I N C c as a covariate, yields significant treatment effects for all the levels of σ and for all the different sample sizes as shown in Table 5. Therefore, we will report the results based on the statistically significant estimates obtained by fitting data to Equation (7) (with I N C c as a covariate ) that are shown in Table 5. Table 5 shows that when σ = 0, the treatment effects ( β 1 ’s) for R3 vs. R4 are comparable and statistically significant for the different sample sizes. These results show that when σ = 0 and N= 5000, 10,000 or 20,000, scoring matric points that are greater than or equal to 25 compared to scoring matric points that are less than 25, increases the probability of eligibility for NSFAS funding, i.e., P ( N S F i = 1 ) by 3.75, 3.74, and 3.71 percentage points, respectively, for a unit increase in matric points for students whose households income is less than or equal to R350,000 (≈US$25,000).
The results show strong evidence that scoring matric points that are greater than or equal to 25 and an income less than R350,000 (≈US$25,000) significantly increases the chance of eligibility for NSFAS funding. This makes achieving matric points greater than 25 and household income less than R350,000 important variables to be considered when awarding NSFAS funding. The estimated treatment effects decrease when the variability in the two assignment variables, M P c and I N C c , is varied from σ = 0 to σ = 0.15 in steps of 0.05, while holding the cut-offs constant as shown in Table 5 for R3 vs. R4. Additionally, the treatment effects are all still statistically significant at 5% level of significance. These results suggest that the causal estimates obtained by using the frontier approach may be sensitive to the level of variation in the assignment variables, because increasing the variation in the assignment variables may cause the observations to move away from the cut-off points, thereby decreasing the treatment effects.
As shown in Table 5 for R1 vs. R2, when σ = 0 and N = 10,000, scoring matric points that are greater than or equal to 25 compared to scoring matric points that are less than 25, decreases the probability of eligibility for NSFAS funding, i.e., P ( N S F i = 1 ) by a significant 7.96 (p-value = 0.0111 < 0.05) percentage points for a unit increase in matric points for students whose household income is greater than or equal to R350,000 (≈US$25,000). These students who just meet the matric point threshold but with a household income that is just greater R350,000 (≈US$25,000) do not receive NSFAS funding and yet they are not different from those whose income is just below R350,000 (≈US$25,000). These students are referred to as the “missing middle” [8]. The authors define the “missing middle” as the students who come from households whose incomes are between R350,000 and R600,000. These students do not qualify for NSFAS funding but at the same time they cannot afford to pay for their higher education.
Table 4 and Table 6, explore the effects of having a household income greater than R350,000 compared to having an income less than or equal to R350,000 for students who are either scoring matric points greater than 25 or less than 25. These results show that all the β 1 ’s are small and not statistically significant at 5% level of significance. The implications of these results are that when awarding NSFAS funding, one must first look at whether or not a student has met the matric points threshold and then consider the household income. Considering the matric points first, makes it easier to quantify the number of students that have qualified for university entry. Consequently, NSFAS will then be in a position to quantify those that automatically qualify for NSFAS funding and also, quantify the “missing middle”.

5.2. Supplementary Analysis

The following supplementary analyses are considered in this paper, and they are based on the simulated data when σ = 0 and N = 10,000.

5.2.1. Checking for Continuity of the Conditional Expectation of Exogenous Variables around the Cut-Off/Threshold Value

Graphical analysis is an integral part of any RDD design. It is assumed that the treatment effect or causal effect of interest is measured by the value of the discontinuity in the expected outcome at a particular cut-off point. This is a fundamental assumption that postulates that without an intervention, the outcome variable would have been continuous at the cut-off point. This means that any discontinuity in the outcome is credited only to the treatment or exposure. Graphical analysis provides insights into the RDD results for causal analysis. For example, if the graph has a discontinuity, this would suggest that the intervention had a causal effect on the outcome, whereas if the graph is continuous, then it would suggest that there is no causal effect that can be attributed to the intervention. Current literature that we are aware of on MRDD has limited discussion on a direct extension of the conventional RDD graphs to MRDD. Ref. [3] offers a sub-optimal extension to MRDD, as well as using a different approach, the “slicing” and “sliding window” plots. In this study we deploy two dimensional plots for each of the causal effects as shown in Figure 4 and Figure 5. The graphs compare two treatments at a time, yielding four plots in total (one for each pair of treatments being compared).
Using the function “rdplot()” in the R package “rdrobust” [24], the outcome variable is plotted as a function of the assignment variable, as shown in Figure 4 and Figure 5. The graphs show that there are discontinuities at the cut-off points, thereby providing evidence of a non-zero treatment effect. Additionally, the graphs show that the relationship between the outcome variable and the assignment variables is approximately linear within a very small bandwidth of the assignment variables. Thus, the graphs show that the linear specification model is a plausible and credible model for estimating the treatments effects within a narrow bandwidth.

5.2.2. Manipulation Testing Using Local Polynomial Density Estimation

Ref. [7] introduced the McCrary manipulation test to check for evidence of manipulation near the income or matric points cut-off points. However, in this paper, we employ a newer method proposed by [25] that uses local polynomial density estimators [23] for manipulation testing. The method uses the rddensity() function in the R package ‘rddensity’ to implement manipulation testing procedures. The test is robust to local polynomial order specifications and different bandwidths [25]. Manipulation testing using local polynomial density estimation tests the null hypothesis ( H 0 ) of no discontinuity of density around cut-off point versus the alternative hypothesis that the density is discontinuous around cut-off points. A test statistic, T q ( h l , h r ) , is computed for a given α level of significance [23]. H 0 is rejected if and only if T q ( h l , h r ) ϕ 1 α / 2 . Thus, H 0 indicates that there should not be any difference in the chance of eligibility for NSFAS funding for students on either side of the income or matric points cut-off points. Any differences should be attributed to the treatment effect only. The manipulation test checks whether this is actually true in our data. If students are able to choose their side of the cut-off points in order to influence the outcome, then we might worry that students on either side of the cut-off points are not comparable. The results of the manipulation tests for σ = 0 and N = 10,000 are shown in Table 7.
For a given bandwidth, T q ( h l , h r ) is the final manipulation test statistic. For example, for R3 vs. R4 which represents M P c < 0 vs. M P c 0 when I N C c 0 and cut-off, c= 0, T q ( 0.1307 , 0.1248 ) = 0.4809, with a p-value = 0.6306. This means that we fail to reject H 0 , and conclude that there is no statistical evidence of systematic manipulation of the matric points. Additionally, the manipulation test results for R1 vs. R2, R2 vs. R3 and R1 vs. R4 indicate that there is no evidence of manipulation of the assignment variables.

5.2.3. Sensitivity to Optimal Bandwidth Selection

We investigate whether the causal estimates critically dependent on a particular bandwidth choice, i.e., how sensitive the causal estimates are to outcomes of the units that are located very close to the cut-off points. Ref. [26] describes the implementation of a newer method that employs local polynomial methods to analyse the sensitivity of the causal estimates to the bandwidth choice. The method investigates sensitivity to bandwidth choice by removing or adding units at the end-points of the neighborhood. This method is different to the the continuity-based approach that changes the bandwidths that are used for local polynomial estimation. Additionally, the newer approach uses finite-sample methods to select a bandwidth that is closer to the cut-off points where the local randomization assumption is probably justified. The “rdwinselect()" command contained in the “rdlocrand” R package [27] is employed to check for sensitivity of the causal estimates to the bandwidth that is used. The “rdwinselect()" command automatically selects a window around the cut-off where the treatment can plausibly be assumed to have been as-if randomly assigned [26]. If the causal estimates critically depend on a particular bandwidth, then they are less credible. Table 8 shows the recommended windows around the cut-off points that were selected for each of the treatment region comparisons.
After choosing the appropriate minimum windows around the cut-off points using “rdwinselect()" function that are shown in Table 8, different statistical tests such as the difference-in-means (DM), Kolmogorov–Smirnov (KS), and Wilcoxon rank sum (RS) can be implemented to test whether or not the treatment effects are different from zero or not. In this paper, we implemented the difference-in-means (DM) of the outcomes on either side of the cut-off point using the “rdrandinf()" command in the “rdlocrand” R package. The DM procedure tests the Fisherian null hypothesis [26], that the treatment effect is zero for all units. Ref. [26] points out that the interpretation of the difference-in-means test statistic in the Fisherian framework is different because it tests the sharp null hypothesis of no treatment effect, and it should not be interpreted as an estimated treatment effect. Its main purpose is to test the null hypotheses that are sharp (e.g., no treatment effect), not on point estimation. This is much stronger than testing the hypothesis that the average treatment effect (ATE) is zero. If the treatment effects are not statistically significant using the minimum window when in the first place we obtained statistically significant treatment effects using the optimal bandwidths and vice versa, then it indicates that the treatment effects are sensitive to the bandwidth used. Table 9 shows the results of DM for each treatment region.
The DM for treatment regions R3 vs. R4 and R1 vs. R2 significant at α = 0.05 using the automatically determined minimum bandwidths. Table 5 shows that the treatment effects for R3 vs. R4 and R1 vs. R2 are all significant for σ = 0 and N = 10,000. These results show that the causal estimates do not critically depend on a particular bandwidth. Additionally, Table 9 shows that the DM for treatment regions R2 vs. R3, and R1 vs. R4 are not statistically significant at α = 0.05 using the automatically determined minimum bandwidths. The treatment effects for R2 vs. R3, and R1 vs. R4 are also not statistically significant for σ = 0 and N = 10,000, as shown in Table 6. These results provide evidence that the causal estimates obtained in the simulation studies are credible as they are very robust to bandwidths that are close to the cut-off points for income and matric points.

6. Case Study

6.1. Application of the MRDD to the Graduate Admissions Data Set

The purpose of case study is apply the methods described earlier to a real world data set. The design for this case study is inspired by [28] who used the conventional RDD with grade points average (GPA) as the only running variable. Therefore, we determine whether using two the assignment variables, namely CGPA and GRE scores from the Indian graduate admissions data [29], produce significant and credible causal effects estimates of the chance of student getting admission for a master’s program in the USA. Ref. [30] recommends a total GRE score of 322 for a student to stand a higher chance of being admitted to a graduate program of choice. Thus, we set a GRE cut-off of 322. Similarly, [30] highly recommends CGPA of ≈8.5 in order for a candidate not to be directly rejected due to poor academic performance. Thus, we use 8.5 and 322 as cut-off points for CGPA and GRE, respectively.

6.2. Estimation of the Causal Effects of CGPA and GRE

To estimate the effect of scoring CGPA ≥ 8.5 over CGPA < 8.5 for students having GRE < 322 or GRE ≥ 322, Equation (9) (similar to Equation (7)) is fitted to the data.
P ( A d m i t = 1 | C G P A c , T c g p a , G R E c , T O E F c ) = β 0 + β 1 T c g p a + β 2 C G P A c + β 3 C G P A c × T c g p a + β 4 G R E c + β 5 T O E F L i c + ϵ i ,
The following causal effects are determined using the assignment variables “CGPA” and “GRE”, with respective cut-offs equal to 8.5 and 322. Additionally, the variables are centered around their respective cut-off points to give C G P A c and G R E c . To estimate the effect of GRE, C G P A c , and T c g p a , variables are replaced by G R E c and T g r e in Equation (9), respectively. The data used was of size, N = 500 at the time of writing this paper.
1
Causal Effect 1: C G P A c < 0 vs. C G P A c 0 for G R E c 0
2
Causal Effect 2: C G P A c < 0 vs. C G P A c 0 for G R E C < 0
3
Causal Effect 3: G R E c < 0 vs. G R E c 0 for C G P A c 0
4
Causal Effect 4: G R E c <0 vs. G R E c 0 for C G P A c < 0
The results (Causal Effects 1–4) of fitting to Equation (9) the Graduate Admissions data set are shown in Table 10.
Table 10 shows that Causal Effects 1–4 are significant at 5% level of significance. Causal Effect 1 is equal to the effect of having a CGPA score that is greater than or equal to 8.5 compared to having a CGPA score that is less than 8.5 for all the students whose GRE score is already greater than or equal to 322. The results show that having a having a CGPA score greater than 8.5 compared to having a CGPA that is less than 8.5 increases the chance of graduate college admission by a significant 40.3 (p-value = 0.000 < 0.05) percentage points for students from India who already have a GRE score that is greater than 322. Similarly, Causal Effect 3 shows that having a GRE score that is greater than 322 compared to having a GRE Score that is less than 322 increases the chance of graduate college admission by a significant 60.7 (p-value = 0.000 < 0.05) percentage points for students who already have a CGPA score greater than 8.5. The effect of scoring a CGPA score greater than 8.5 compared to scoring a CGPA less than 8.5 for students who already have a GRE score < 322 (Causal Effects 2) significantly reduces the chance of admission to a graduate master’s program by ≈3.5 ( p-value = 0.025 < 0.05 ) percentage points. On the other hand, obtaining a GRE score > 322 compared to obtaining a GRE score < 322 (Causal Effect 4) increases the chance of college admission by ≈11.8 percentage points for students who already have CGPA < 8.5 for students who already have a CGPA score that is less than 8.5. These results suggests that GRE and CGPA are very important obstacles that must be overcome if one is to have a realistic chance of being admitted to the graduate master’s program. This means that for students from India intending to apply to study for a graduate master’s programs in science and technology in the USA, they must achieve competitive GRE and CGPA of at least 322 and at least 8.5, respectively.
Table 11 shows the manipulation test results of the CGPA and GRE scores. The results show that there is insufficient evidence to reject the null hypothesis ( H 0 ) of no discontinuity of density around the cut-off points for Causal Effects 1–4. This suggests that there is insufficient evidence of manipulation of the CGPA and GRE scores. Thus, the causal estimates are credible. Table 12 shows the difference-in-means (DM) tests for the Fisherian null hypothesis [23], that the treatment effect is zero for all units. We reject the Fisherian null hypothesis that the treatment effect is zero for Causal Effects 1–4 and conclude that the causal effects that are all statistically significant at 5% level of significance since the p-values of the DM statistics (T) shown in Table 12 are all less than 0.05. This shows that causal estimates are not sensitive to the choice of bandwidth because the causal estimate shown in Table 10 are also statistically significant at 5% level. This means that the frontier MRDD produced credible causal estimates.

7. Discussion and Conclusions

7.1. Discussion

This paper has highlighted the importance of using an appropriate regression discontinuity approach when faced with the complexity brought about by having more than one assignment variable. In this study we have used the frontier MRDD because it is easy to implement. It reduces the assumptions checking process to a series of well-defined single assignment variables RDD, whose methods are well-defined in the literature [5]. A data-driven bandwidth selection method was deployed in selecting an optimal bandwidth, thus eliminating the possibility of manipulating the results by choosing an arbitrary bandwidth. Researchers seeking to use local linear regression must ensure that there is a sufficient density of points around the cut-off points. Not having enough density of points at the threshold point may result in less credible causal estimates. The simulation results produced significant causal estimates when the data was random with no induced variability in the assignment variables. For example, based on our analytical results, barely scoring matric points greater than or equal to 25 points compared to scoring matric points less than 25 for students whose household income is less than R350,000 increases the probability of eligibility for NSFAS funding by 3.75 percentage points (Table 5 for σ = 0). When the level of variability in the two assignment variables, M P c and I N C c was increased from σ = 0 to σ = 0.15 in steps of 0.05, while holding the cut-offs constant, the estimated treatment effects decreased respectively for each level of σ . This showed that the frontier MRDD approach was not suitable for handling data with high variability. Thus, for future work we may consider developing methods that can handle variability in the assignment variables better. Besides, the frontier MRDD works very well when σ = 0, i.e., when there is no induced variability in the assignment variables and the causal effect estimates were comparable and credible for the different sample sizes.
We have reported not only the causal effects, but also carried out supplementary analyses to assess the credibility of the primary causal effect estimates. We found evidence of a relationship between the outcome variable and the assignment variable(s) through graphical analysis. Figure 4 and Figure 5 show “visual” evidence of disturbances or discontinuities in the expected smooth relationship between the outcome variable and the assignment variable(s). This gives credibility to the causal estimates as it indicates that the treatments effects actually exist at the cut-offs. The relationship between the probability of eligibility for NSFAS funding and either one of the assignment variables (income or matric points) is assumed to exhibit smooth functions. This assumption of smooth functions is based on the continuity-based framework that is used to explain the required identification assumptions intuitively, and also in developing causal effect estimates. Additionally, we have assessed the existence of the causal effects through an alternative causal framework, making use of the local randomization framework [31]. The local randomisation framework is different from the continuity-based framework in that it formalises the idea of a local randomized experiment near the cutoff by embedding the RD design in a classical, Fisherian causal model, thereby giving interpretation and justification to randomization inference and related classical experimental methods [32]. By using the local randomisation framework, we analysed the units that are in a small window around the cut-off points “as-if” they were randomly assigned to treatment or control. This enabled us to use statistical tools such as the difference-in-means (DM) to test the Fisherian null hypothesis, that the treatment effect is zero for all units. The Fisherian null hypothesis of no treatment effect was applied in both the simulation and the case study. For example, in the case study, we did not apply graphical analysis to visually inspect whether a treatment effect existed or not, but we rejected the Fisherian null hypothesis of no treatment effect, thereby validating the presence of treatment effects for Causal Effects 1–4. The local randomisation assumption near the cut-off has allowed us to use a newer method proposed by [31] in assessing the credibility of causal estimates in a simulation study as well as in a real world data set. Besides, for the continuous assignment variables, M P c and I N C c in the simulation study, we have used graphical analysis (or the continuity-based framework) to detect the presence of a discontinuity or treatment effect and then used the local randomization approach (Fisherian null hypothesis) as a robustness check.
To test for evidence of manipulation of the assignment variables we used a newer intuitive and easy-to-implement non-parametric density estimator that is based on local polynomial techniques [23]. The authors indicate that the estimator is fully automatic and does not require any other transformation of the data. Students must not be able to influence their position relative to the cut-off scores. The local polynomial density manipulation testing results for both the simulation study (Table 7), and the case study (Table 11), show that students were not able to manipulate or choose their preferred side of the assignment variable(s) cut-off point, thus making the causal estimates more credible. Besides, the evidence of no manipulation holds because all students may not precisely manipulate their own matric points. This gives credibility to the causal effects estimated in both the simulation and case study.
Sensitivity to bandwidth choices was performed by examining the robustness of our causal estimates to changing bandwidths. Table 9 shows that when the causal estimates were estimated using narrow windows around the cut-off points, they were all significant at 5% level of significance. This suggests that our average causal estimates results remain largely robust to changing bandwidth choices.
The results of the case study have demonstrated that the frontier MRDD is a valid and it can be applied to a real world data set giving significant and credible causal effects of a treatment. Using the Indian graduate admissions data for master’s programs in science and technology studies in the USA, we obtained significant causal estimates (Table 10). Additionally, the manipulation test results (Table 11) indicated that there was insufficient evidence to reject the null hypothesis H 0 of no discontinuity. This suggests that there was no evidence of manipulation of the GRE or the CGPA in order to gain admission to graduate master’s programs for students from India. Additionally, Table 12 shows that when using the minimum window (bandwidth) around the cut-off points, the Fisherian null hypothesis of no treatment effect is rejected, thereby indicating that the causal estimates were credible and not sensitive to bandwidth choice.
The simulation studies described earlier ensured that we had enough data points on either side of the cutoff points. However, in real-world MRDD applications, the distribution of points around the cut-off points may be sparse and not optimal. Therefore, practitioners wishing to apply MRDD need to take into account the distribution of points around the cutoff points when they are determining the sample size(s) to use. Furthermore, the case study has shown that it essential to perform supplementary analyses such as the manipulation tests or sensitivity to bandwidth choice. This is because failure to perform supplementary analyses may affect the precision and bias of the causal effect estimates.

7.2. Limitations

A limitation of the study was that consolidated data of household income and matric points was not available from the department that handles the National Student Financial Aid Scheme (NSFAS) at the time of writing this paper and we could not directly apply our approach to real income and matric points data. However, we have demonstrated that the approach we have adopted still works for MRDD with two assignment variables and can unearth valuable causal effects of interest. For future work, we strongly recommend that further research should focus on the determination of the minimum sample size required for two or more assignment variables. Additionally, future simulation studies can be employed to examine the effect of variations in the correlation and distribution of the assignment variables. In the simulation studies we used samples of size 5000, 10,000, and 20,000 and in the case study we used data of size 500. Determining the minimum sample required to obtain credible causal estimates will assist those wishing to apply for MRDD. Determining the minimum sample size enables one to have enough density of data points around the cut-off point and thus generate more credible estimates. In generating the outcome variable, the possibility of using different combinations of the data generation distributions could be explored. Future work could also look at studying the effect of NSFAS grants on subsequent enrollment rates for a bachelor’s degree and not just focus of the probability of eligibility for NSFAS funding. One could also look at whether or not NSFAS grants influences the choice of the universities to attend or influence the number of years students spend studying for their bachelor’s degree.

8. Conclusions

In the absence of randomised controlled experiments, regression discontinuity designs offer more opportunities to estimate causal effects by making use of the variation in the treatment assignment induced by a cut-off point. One major finding of this paper is the discovery of new evidence that both matric points and income have a huge impact on the probability of getting NSFAS funding to study at any university in South Africa. This evidence will inform policy makers and educational practitioners on the effects of matric points and income on the chance of eligibility for NSFAS funding. The availability of the NSFAS grant has a huge impact on students’ decisions to attend university or seek other opportunities elsewhere. In summary, this paper makes valuable contributions to the literature on multiple regression discontinuity designs by conducting supplementary analyses that seek to add more credibility to the causal estimates obtained through primary analyses. If one is interested in determining causal effects of barely meeting the requirements of one assignment variable, among the subjects that either meet or fail to meet the requirements of the other assignment variable, then we strongly recommend the use of the frontier multivariate regression discontinuity design as it is easy to implement and it incorporates discontinuities in multiple assignment variables into single regression discontinuity designs along a number of frontiers of the treatment variables.

Author Contributions

Conceptualization, A.W. and C.C.; methodology, A.W.; software, A.W.; validation, A.W., formal analysis, A.W.; investigation, A.W.; resources, A.W.; data curation, A.W.; writing—original draft preparation, A.W.; writing—review and editing, A.W.; visualization, A.W.; supervision, C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

A publicly available dataset was used in the Case Study analysis. This data can be found here: https://www.kaggle.com/mohansacharya/datasets (accessed on 28 August 2021).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wong, V.C.; Steiner, P.M.; Cook, T.D. Analyzing regression-discontinuity designs with multiple assignment variables: A comparative study of four estimation methods. J. Educ. Behav. Stat. 2013, 38, 107–141. [Google Scholar] [CrossRef] [Green Version]
  2. Lee, D.S.; Lemieux, T. Regression discontinuity designs in economics. J. Econ. Lit. 2010, 48, 281–355. [Google Scholar] [CrossRef] [Green Version]
  3. Cheng, Y.A. Regression Discontinuity Designs with Multiple Assignment Variables. 2016. Available online: https://www.econ.berkeley.edu/sites/default/files/EconHonorsThesis_YizhuangAldenCheng.pdf (accessed on 2 April 2021).
  4. Papay, J.P.; Willett, J.B.; Murnane, R.J. Extending the regression-discontinuity approach to multiple assignment variables. J. Econom. 2011, 161, 203–207. [Google Scholar] [CrossRef]
  5. Reardon, S.F.; Robinson, J.P. Regression discontinuity designs with multiple rating-score variables. J. Res. Educ. Eff. 2012, 5, 83–104. [Google Scholar] [CrossRef]
  6. Athey, S.; Imbens, G.W. The state of applied econometrics: Causality and policy evaluation. J. Econ. Perspect. 2017, 31, 3–32. [Google Scholar] [CrossRef] [Green Version]
  7. McCrary, J. Manipulation of the running variable in the regression discontinuity design: A density test. J. Econom. 2008, 142, 698–714. [Google Scholar] [CrossRef] [Green Version]
  8. Garrod, N.; Wildschut, A. How large is the missing middle and what would it cost to fund? Dev. South. Afr. 2021, 38, 484–491. [Google Scholar] [CrossRef]
  9. Cattaneo, M.D.; Titiunik, R.; Vazquez-Bare, G.; Keele, L. Interpreting regression discontinuity designs with multiple cutoffs. J. Politics 2016, 78, 1229–1248. [Google Scholar] [CrossRef] [Green Version]
  10. Matsudaira, J.D. Mandatory summer school and student achievement. J. Econom. 2008, 142, 829–850. [Google Scholar] [CrossRef]
  11. Keele, L.J.; Titiunik, R. Geographic boundaries as regression discontinuities. Political Anal. 2015, 23, 127–155. [Google Scholar] [CrossRef]
  12. Kane, T.J. A Quasi-Experimental Estimate of the Impact of Financial Aid on College-Going; Technical Report; National Bureau of Economic Research: Cambridge, MA, USA, 2003. [Google Scholar]
  13. Imbens, G.W.; Wooldridge, J.M. Recent developments in the econometrics of program evaluation. J. Econ. Lit. 2009, 47, 5–86. [Google Scholar] [CrossRef] [Green Version]
  14. Papay, J.P.; Murnane, R.J.; Willett, J.B. High-school exit examinations and the schooling decisions of teenagers: Evidence from regression-discontinuity approaches. J. Res. Educ. Eff. 2014, 7, 1–27. [Google Scholar] [CrossRef] [PubMed]
  15. Porter, K.E.; Reardon, S.F.; Unlu, F.; Bloom, H.S.; Cimpian, J.R. Estimating causal effects of education interventions using a two-rating regression discontinuity design: Lessons from a simulation study and an application. J. Res. Educ. Eff. 2017, 10, 138–167. [Google Scholar] [CrossRef]
  16. Herlands, W.; McFowland, E., III; Wilson, A.G.; Neill, D.B. Automated local regression discontinuity design discovery. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 1512–1520. [Google Scholar]
  17. Van der Berg, S. Current poverty and income distribution in the context of South African history. Econ. Hist. Dev. Reg. 2011, 26, 120–140. [Google Scholar] [CrossRef] [Green Version]
  18. Sala-i Martin, X. The World Distribution of Income (Estimated from Individual Country Distributions). 2002. Available online: https://www.nber.org/papers/w8933 (accessed on 13 March 2021).
  19. Arthurs, N.; Stenhaug, B.; Karayev, S.; Piech, C. Grades Are Not Normal: Improving Exam Score Models Using the Logit-Normal Distribution. In Proceedings of the 12th International Conference on Educational Data Mining (EDM), Montreal, QC, Canada, 2–5 July 2019; Available online: https://eric.ed.gov/?id=ED599204 (accessed on 17 August 2021).
  20. Von Hippel, P. Linear vs. logistic probability models: Which is better, and when. Stat. Horizons 2015. Available online: https://statisticalhorizons.com/linear-vs-logistic (accessed on 15 September 2021).
  21. Angrist, J.D.; Pischke, J.S. Mostly Harmless Econometrics: An Empiricist’s Companion; Princeton University Press: Princeton, NJ, USA, 2008. [Google Scholar]
  22. Deke, J. Using the Linear Probability Model to Estimate Impacts on Binary Outcomes in Randomized Controlled Trials; Technical Report, Mathematica Policy Research; Department of Health and Human Services, Administration on Children, Youth and Families, Office of Adolescent Health: Washington, DC, USA, 2014. [Google Scholar]
  23. Cattaneo, M.D.; Jansson, M.; Ma, X. Simple local polynomial density estimators. J. Am. Stat. Assoc. 2020, 115, 1449–1455. [Google Scholar] [CrossRef] [Green Version]
  24. Calonico, S.; Cattaneo, M.D.; Farrell, M.H.; Titiunik, R. rdrobust: Software for regression-discontinuity designs. Stata J. 2017, 17, 372–404. [Google Scholar] [CrossRef] [Green Version]
  25. Cattaneo, M.D.; Jansson, M.; Ma, X. Manipulation testing based on density discontinuity. Stata J. 2018, 18, 234–261. [Google Scholar] [CrossRef]
  26. Cattaneo, M.D.; Idrobo, N.; Titiunik, R. A Practical Introduction to Regression Discontinuity Designs: Foundations; Cambridge University Press: Cambridge, UK, 2020. [Google Scholar]
  27. Cattaneo, M.D.; Titiunik, R.; Vazquez-Bare, G. rdlocrand: Local Randomization Methods for RD Designs. R Package Version 0.3. 2018. Available online: https://CRAN.R-project.org/package=rdlocrand (accessed on 15 September 2021).
  28. Goodman, J.; Melkers, J.; Pallais, A. Can online delivery increase access to education? J. Labor Econ. 2019, 37, 1–34. [Google Scholar] [CrossRef] [Green Version]
  29. Acharya, M.S.; Armaan, A.; Antony, A.S. A comparison of regression models for prediction of graduate admissions. In Proceedings of the 2019 International Conference on Computational Intelligence in Data Science (ICCIDS), Chennai, India, 6–7 September 2019; pp. 1–5. Available online: https://ieeexplore.ieee.org/document/8862140 (accessed on 25 August 2021).
  30. Raghunathan, K. Demystifying the American Graduate Admissions Process. 2010. Available online: StudyMode.com (accessed on 25 August 2021).
  31. Cattaneo, M.D.; Frandsen, B.R.; Titiunik, R. Randomization inference in the regression discontinuity design: An application to party advantages in the US Senate. J. Causal Inference 2015, 3, 1–24. [Google Scholar] [CrossRef]
  32. Hill, R.C.; Fomby, T.B.; Escanciano, J.C.; Hillebrand, E.; Jeliazkov, I. Regression Discontinuity Designs: Theory and Applications; Emerald Group Publishing: Bingley, UK; West Yorkshire, UK, 2017. [Google Scholar]
Figure 1. Density of Matric Points.
Figure 1. Density of Matric Points.
Stats 04 00052 g001
Figure 2. Treatment Regions R1 to R4.
Figure 2. Treatment Regions R1 to R4.
Stats 04 00052 g002
Figure 3. Probaility (p) versus log-odds ratio.
Figure 3. Probaility (p) versus log-odds ratio.
Stats 04 00052 g003
Figure 4. (Left): Causal effect 1: Effect of M P c 0 over M P c < 0 for I N C c > 0 (Right): Causal effect 2: Effect of M P c 0 over M P c < 0 for I N C c 0.
Figure 4. (Left): Causal effect 1: Effect of M P c 0 over M P c < 0 for I N C c > 0 (Right): Causal effect 2: Effect of M P c 0 over M P c < 0 for I N C c 0.
Stats 04 00052 g004
Figure 5. (Left): Causal effect 3: Effect of I N C c > 0 over I N C c 0 for M P c < 0 (Right): Causal effect 4: Effect of I N C c 0 over I N C c > 0 for M P c 0 .
Figure 5. (Left): Causal effect 3: Effect of I N C c > 0 over I N C c 0 for M P c < 0 (Right): Causal effect 4: Effect of I N C c 0 over I N C c > 0 for M P c 0 .
Stats 04 00052 g005
Table 1. Average income distribution (Source: STATSSA: Living Conditions Survey 2).
Table 1. Average income distribution (Source: STATSSA: Living Conditions Survey 2).
Population Group
of Household
Average Income% HouseholdsNumber of
Households
Black African92,98380.4218,800
Coloured172,7657.231690
Indian/Asian271,6212.31540
White444,44610.042347
Total981,815 23,377
Table 2. Subject Passes Level System.
Table 2. Subject Passes Level System.
LevelFinal Mark%Achievement
780–100%Outstanding
670–79%Meritorius
560–69%Substantial
450–59%Moderate
340–49%Adequate
230–39%Elementary
10–29%Not Achieved-Fail
Table 3. Top: Simulation results for fitting Equation (6) to compare R3 vs. R4 which represents M P c < 0 vs. M P c ≥ 0 for I N C c < 0, and cut-off c = 0. Bottom: Simulation results for fitting Equation (6) to compare R1 vs. R2 which represents M P c < 0 vs. M P c ≥ for I N C c > 0 and cut-off, c = 0.
Table 3. Top: Simulation results for fitting Equation (6) to compare R3 vs. R4 which represents M P c < 0 vs. M P c ≥ 0 for I N C c < 0, and cut-off c = 0. Bottom: Simulation results for fitting Equation (6) to compare R1 vs. R2 which represents M P c < 0 vs. M P c ≥ for I N C c > 0 and cut-off, c = 0.
R3 vs. R4N h l h r β 1 s.ep-Value
5000−0.13070.12480.03880.01960.1607
σ = 0.0010,000−0.13070.12480.03710.01390.0718
20,000−0.13070.12480.03710.00980.0119
5000−0.13070.12480.02930.01860.2276
σ = 0.0510,000−0.13070.12480.02950.01310.1157
20,000−0.13070.12480.02930.00930.0349
5000−0.13070.12480.01690.01650.3484
σ = 0.1010,000−0.13070.12480.01640.01170.2728
20,000−0.13070.12480.01610.00820.1631
5000−0.13070.12480.00860.01460.4283
σ = 0.1510,000−0.13070.12480.00800.01030.4161
20,000−0.13070.12480.00820.00730.3301
R1 vs. R2
5000−0.19220.2859−0.07630.04070.1061
σ = 0.0010,000−0.19220.2859−0.07900.02870.0171
20,000−0.19220.2859−0.07860.02030.0008
5000−0.19220.2859−0.06800.03930.1388
σ = 0.0510,000−0.19220.2859−0.06890.02760.0334
20,000−0.19220.2859−0.06850.01940.0023
5000−0.19220.2859−0.04740.03730.2668
σ = 0.1010,000−0.19220.2859−0.04750.02580.1139
20,000−0.19220.2859−0.04680.01800.0274
5000−0.19220.2859−0.02720.03460.4550
σ = 0.1510,000−0.19220.2859−0.02750.02430.3304
20,000−0.19220.2859−0.02780.01710.1718
Table 4. Top: Simulation results for fitting Equation (6) to compare R2 vs. R3 which represents I N C c ≤ 0 vs. I N C c > 0 for M P c ≤ 0, and cut-off c= 0. Bottom: Simulation results for fitting Equation (6) to compare R1 vs. R4 which represents I N C c ≤ 0 vs. I N C c > 0 for M P c > 0, and cut-off c= 0.
Table 4. Top: Simulation results for fitting Equation (6) to compare R2 vs. R3 which represents I N C c ≤ 0 vs. I N C c > 0 for M P c ≤ 0, and cut-off c= 0. Bottom: Simulation results for fitting Equation (6) to compare R1 vs. R4 which represents I N C c ≤ 0 vs. I N C c > 0 for M P c > 0, and cut-off c= 0.
R2 vs. R3N h l h r β 1 s.ep-Value
5000−0.50970.5114−0.00030.00400.6029
σ = 0.0010,000−0.50970.5114−0.00030.00280.6055
20,000−0.50970.5114−0.00040.00190.5977
5000−0.50970.5114−0.00040.00550.5924
σ = 0.0510,000−0.50970.5114−0.00040.00380.5874
20,000−0.50970.5114−0.00040.00270.5802
5000−0.50970.5114−0.00070.00980.5685
σ = 0.1010,000−0.50970.5114−0.00040.00680.5667
20,000−0.50970.5114−0.00040.00480.5695
5000−0.50970.5114−0.00020.01560.5554
σ = 0.1510,000−0.50970.5114−0.00030.01080.5637
20,000−0.50970.5114−0.00030.00760.5615
R1 vs. R4
5000−0.42000.2970−0.00140.14250.5046
σ = 0.0010,000−0.42000.2970−0.00170.09960.4979
20,000−0.42000.2970−0.00070.06990.4798
5000−0.42000.2970−0.00330.13650.4917
σ = 0.0510,000−0.42000.2970−0.00080.09600.4957
20,000−0.42000.29700.00070.06750.4890
5000−0.42000.29700.00410.12610.4851
σ = 0.1010,000−0.42000.2970−0.00140.08840.4938
20,000−0.42000.29700.00440.06200.4953
5000−0.42000.29700.00190.11340.4928
σ = 0.1510,000−0.42000.2970−0.00050.07920.4970
20,000−0.42000.29700.00030.05580.4791
Table 5. Top: Simulation results for fitting Equation (7) to compare R3 vs. R4, which represents M P c ≤ 0 vs. M P c > 0 for I N C c 0, and cut-off c = 0. Bottom: Simulation results for fitting Equation (7) to compare R1 vs. R2 which represents M P c ≤ 0 vs. M P c > 0 for I N C c > 0, and cut-off c = 0.
Table 5. Top: Simulation results for fitting Equation (7) to compare R3 vs. R4, which represents M P c ≤ 0 vs. M P c > 0 for I N C c 0, and cut-off c = 0. Bottom: Simulation results for fitting Equation (7) to compare R1 vs. R2 which represents M P c ≤ 0 vs. M P c > 0 for I N C c > 0, and cut-off c = 0.
R3 vs. R4N h l h r β 1 s.ep-Value
5000−0.13070.12480.03750.00670.0001
σ = 0.0010,000−0.13070.12480.03740.00470.0000
20,000−0.13070.12480.03710.00330.0000
5000−0.13070.12480.02960.00560.0002
σ = 0.0510,000−0.13070.12480.02920.00390.0000
20,000−0.13070.12480.02910.00280.0000
5000−0.13070.12480.01630.00370.0019
σ = 0.1010,000−0.13070.12480.01620.00260.0000
20,000−0.13070.12480.01610.00180.0000
5000−0.13070.12480.00800.00220.0084
σ = 0.1510,000−0.13070.12480.00800.00150.0001
20,000−0.13070.12480.00800.00110.0000
R1 vs. R2
5000−0.19220.2859−0.07780.03620.0843
σ = 0.0010,000−0.19220.2859−0.07960.02560.0111
20,000−0.19220.2859−0.07820.01810.0003
5000−0.19220.2859−0.06820.03410.1078
σ = 0.0510,000−0.19220.2859−0.06920.02410.0200
20,000−0.19220.2859−0.06850.01690.0012
5000−0.19220.2859−0.04690.02980.1961
σ = 0.1010,000−0.19220.2859−0.04730.02100.0675
20,000−0.19220.2859−0.04680.01470.0101
5000−0.19220.2859−0.02730.02490.3434
σ = 0.1510,000−0.19220.2859−0.02750.01760.2041
20,000−0.19220.2859−0.02780.01260.0746
Table 6. Top: Simulation results for fitting Equation (7) to compare R2 vs. R3 which represents I N C c ≤ 0 vs. I N C c > 0 for M P c 0, and cut-off c= 0. Bottom: Simulation results for fitting Equation (7) to compare R1 vs. R4 which represents I N C c ≤ 0 vs. I N C c > 0 for M P c > 0 , and cut-off c= 0.
Table 6. Top: Simulation results for fitting Equation (7) to compare R2 vs. R3 which represents I N C c ≤ 0 vs. I N C c > 0 for M P c 0, and cut-off c= 0. Bottom: Simulation results for fitting Equation (7) to compare R1 vs. R4 which represents I N C c ≤ 0 vs. I N C c > 0 for M P c > 0 , and cut-off c= 0.
R2 vs. R3N h l h r β 1 s.ep-Value
5000−0.50970.5114−0.00040.00310.5879
σ = 0.0010,000−0.50970.5114−0.00040.00210.6032
20,000−0.50970.5114−0.00040.00150.5758
5000−0.50970.5114−0.00050.00410.5817
σ = 0.0510,000−0.50970.5114−0.00040.00290.5836
20,000−0.50970.5114−0.00040.00200.5763
5000−0.50970.5114−0.00040.00670.5835
σ = 0.1010,000−0.50970.5114−0.00050.00460.5589
20,000−0.50970.5114−0.00050.00330.5558
5000−0.50970.5114−0.00060.00930.5558
σ = 0.1510,000−0.50970.5114−0.00080.00650.5368
20,000−0.50970.5114−0.00010.00450.5427
R1 vs. R4
5000−0.42000.2970−0.00140.08420.5050
σ = 0.0010,000−0.42000.2970−0.00160.05910.4956
20,000−0.42000.2970−0.00070.04140.5179
5000−0.42000.2970−0.00090.07870.4877
σ = 0.0510,000−0.42000.2970−0.00030.05520.5116
20,000−0.42000.2970−0.00060.03890.5114
5000−0.42000.29700.00110.06680.4903
σ = 0.1010,000−0.42000.29700.00050.04700.4963
20,000−0.42000.29700.00350.03300.5022
5000−0.42000.29700.00370.05370.5147
σ = 0.1510,000−0.42000.2970−0.00050.03770.4982
20,000−0.42000.29700.00110.02650.4886
Table 7. Examining Manipulation at the Income and Matric points cut-off points.
Table 7. Examining Manipulation at the Income and Matric points cut-off points.
Causal EffectBandwidth T q ( h l , h r ) p-Value
R3 vs. R4 (−0.1307, 0.1248)0.48090.6306
R1 vs. R2 (−0.1922, 0.2859 )0.23000.8181
R2 vs. R3 (−0.5097, 0.5114)0.91220.3617
R1 vs. R4 (−0.4200, 0.2970)0.35710.721
Table 8. Minimum window around the cut-off points where the treatment can plausibly be assumed to have been randomly assigned.
Table 8. Minimum window around the cut-off points where the treatment can plausibly be assumed to have been randomly assigned.
Treatment RegionWindow
R3 vs. R4(−0.0037, 0.0029)
R1 vs. R2(−0.0455, 0.0541)
R2 vs. R3(−0.095, 0.1336)
R1 vs. R4(−0.0973, 0.1237)
Table 9. Difference-in-means (DM) test statistics for the treatment regions under investigation.
Table 9. Difference-in-means (DM) test statistics for the treatment regions under investigation.
Treatment RegionDM Test Statistic (T)p-Value
R3 vs. R40.3830.000
R1 vs. R20.1380.000
R2 vs. R3−0.0030.103
R1 vs. R4−0.0780.132
Table 10. The results of fitting to Equation (9) the Graduate Admissions Data Set.
Table 10. The results of fitting to Equation (9) the Graduate Admissions Data Set.
Causal Effect h l h r β 1 s.ep-Value
Causal Effect 1−0.4000.5500.4030.1040.000
Causal Effect 2−0.2000.480−0.0350.0160.025
Causal Effect 3−24.018.00.6070.0280.000
Causal Effect 4−23.013.00.1180.0180.000
Table 11. Examining manipulation at the CGPA and GRE cut-off points.
Table 11. Examining manipulation at the CGPA and GRE cut-off points.
Causal Effect h l h r T q ( h l , h r ) p-Value
Causal Effect 1−0.4000.550−1.1100.267
Causal Effect 2−0.2000.480−0.7010.483
Causal Effect 3−24.018.00.9210.357
Causal Effect 4−23.013.01.5050.133
Table 12. Tests for difference in means (DM) using the minimum window (bandwidth).
Table 12. Tests for difference in means (DM) using the minimum window (bandwidth).
Causal EffectMinimum WindowDM Testp-Value
(Bandwidth)Statistic (T)
Causal Effect 1(−0.4, 0.42)0.3470.001
Causal Effect 2(−0.2, 0.19)0.0050.035
Causal Effect 3(−2, 1)0.6650.000
Causal Effect 4(−4, 5)0.3100.000
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Whata, A.; Chimedza, C. Credibility of Causal Estimates from Regression Discontinuity Designs with Multiple Assignment Variables. Stats 2021, 4, 893-915. https://doi.org/10.3390/stats4040052

AMA Style

Whata A, Chimedza C. Credibility of Causal Estimates from Regression Discontinuity Designs with Multiple Assignment Variables. Stats. 2021; 4(4):893-915. https://doi.org/10.3390/stats4040052

Chicago/Turabian Style

Whata, Albert, and Charles Chimedza. 2021. "Credibility of Causal Estimates from Regression Discontinuity Designs with Multiple Assignment Variables" Stats 4, no. 4: 893-915. https://doi.org/10.3390/stats4040052

APA Style

Whata, A., & Chimedza, C. (2021). Credibility of Causal Estimates from Regression Discontinuity Designs with Multiple Assignment Variables. Stats, 4(4), 893-915. https://doi.org/10.3390/stats4040052

Article Metrics

Back to TopTop