Next Article in Journal
Generalized Fuzzy Graph Connectivity Parameters with Application to Human Trafficking
Next Article in Special Issue
On a Periodic Capital Injection and Barrier Dividend Strategy in the Compound Poisson Risk Model
Previous Article in Journal
Global Mittag–Leffler Stability and Stabilization Analysis of Fractional-Order Quaternion-Valued Memristive Neural Networks
Previous Article in Special Issue
Asymptotic Approximations of Ratio Moments Based on Dependent Sequences
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Treating Nonresponse in Probability-Based Online Panels through Calibration: Empirical Evidence from a Survey of Political Decision-Making Procedures

by
Antonio Arcos 
1,*,
Maria del Mar Rueda 
1 and
Sara Pasadas-del-Amo
2
1
Department of Statistics and Operational Research, University of Granada, 18071 Granada, Spain
2
Institute for Advanced Social Studies, 14004 Cordoba, Spain
*
Author to whom correspondence should be addressed.
Mathematics 2020, 8(3), 423; https://doi.org/10.3390/math8030423
Submission received: 14 February 2020 / Revised: 11 March 2020 / Accepted: 11 March 2020 / Published: 15 March 2020
(This article belongs to the Special Issue Probability, Statistics and Their Applications)

Abstract

:
The use of probability-based panels that collect data via online or mixed-mode surveys has increased in the last few years as an answer to the growing concern with the quality of the data obtained with traditional survey modes. However, in order to adequately represent the general population, these tools must address the same sources of bias that affect other survey-based designs: namely under coverage and non-response. In this work, we test several approaches to produce calibration estimators that are suitable for survey data affected by non response where auxiliary information exists at both the panel level and the population level. The first approach adjusts the results obtained in the cross-sectional survey to the population totals, while, in the second, the weights are the result of two-step process where different adjusts on the sample, panel, and population are done. A simulation on the properties of these estimators is performed. In light of theory and simulation results, we conclude that weighting by calibration is an effective technique for the treatment of non-response bias when the response mechanism is missing at random. These techniques have also been applied to real data from the survey Andalusian Citizen Preferences for Political Decision-Making Procedures.

1. Introduction

In the last few years, we are witnessing a strong development of online research methods in general and web surveys specifically. According to the World Association of Opinion and Marketing Research professionals (ESOMAR) data referring to 2016, 56% of the global revenues in the market research field was made using different kinds of quantitative and qualitative online methods. Online is also the preferred mode for survey-based research such as web or smartphone surveys accounts for two-thirds of the revenues on this chapter [1].
The growth of online research methods goes in parallel with the increasing difficulties that traditional modes, such as face to face or telephone interviews, are facing to obtain samples that achieve the highest quality standards [2,3,4]. Changes in the patterns of the distribution of working and leisure time, the incorporation of women to the workforce, and the increase in residential and geographic mobility have hampered the location of survey respondents. On the other hand, respondents are less willing to participate in surveys as they are more concerned with privacy and confidentiality issues and tired of the growing number of unsolicited requests from the survey and the marketing industry [5,6,7]. As a result of these changes, the production of data that is fully representative of the general population is more and more costly both in terms of money and time [8].
However, at the same time, we have never been more connected and willing to share our thoughts and opinions in the comments section of online newspapers or in our social networks. In this context, there are more and more examples of research that extract public opinion information from the net and social media on issues such as hospitality and tourism management [9], urban mobility [10], health [11], or the social perception of energy [12], among others. In addition, new methods and techniques are being developed in order to take advantage of the enormous amount of information available from a wide range of online sources [13].
From the start of the modern Internet, in the 1990s, survey research practitioners decidedly embraced this technology as a way to contact and interview respondents and web surveys have become the predominant mode of survey data collection nowadays [14]. Web surveys are a very attractive option because of its cost-effectiveness, compared with other modes such as mail, telephone, and face to face, the speed of data collection, and the advantages associated with the computerization of the questionnaire (prevention of branching-errors, data validation, adaptive design, etc.) and self-administration (less intrusive and convenient for respondents) [15].
However, currently the web survey mode has some limitations to adequately represent the general population. In spite of the fast adoption of the Internet in the last few decades, the number of non-users is still important in most countries and their profiles differ significantly from internet users. Moreover, new divides are appearing among internet users related to the knowledge and the type of usage that they do on the Internet [16]. As a result, web surveys that fail to include non-internet or the less intensive internet users are at a high risk of incurring coverage bias [5,17]. A second problem that hinders the use of probability sampling in web surveys of the general population is the lack of a proper sampling frame of email addresses [14].
The main strategy developed in order to overcome those limitations is to use probability-based online panels, where the participants are recruited via survey modes that allow probabilistic selection (CATI, Computer-Assisted Telephone Interviewing, or CAPI, Computer-Assisted Personal Interviewing) and non-internet users are either offered a device and internet connection in order to be able to respond to the surveys, or interviewed with an alternative mode (mail, CATI or CAPI) [18,19]. Some examples of these tools are the Longitudinal Internet Studies for the Social Sciences (LISS) in the Netherlands, the Étude Longitudinal par Internet pour les Sciences Sociales (ELIPSS) in France or the German Internet Panel (GIP) and the GESIS Panel in Germany, among others [20].
Coverage or self-selection are not a problem for probability-based online panels as compared to access panels, but, as with any other survey, they suffer from non-response even if the use of responsive or adaptive design features (incentives, switch in modes, refusals conversion, etc.) account for participation rates that are significantly higher than the average web or CATI survey [19]. Nonresponse may be an important source of bias if respondents and non-respondents differ in key characteristics [21]. To minimize non-response bias and improve panel representativeness, panel data must be corrected by using different weighting adjustment techniques [17].
The two most common methods to weight online panels are post-stratification and propensity score weighting (see [19]). However, other more advanced weighting techniques, such as calibration or pseudo-empirical likelihood, can be used. The application of calibration for the treatment of non-response has been treated by many authors (e.g., [22,23,24,25,26,27,28], ...). These papers consider samples with a single sample design and a single non-response mechanism and a double calibration is used to treat the two types of auxiliary information that are available (at the sample level and at the population level). The extension of the calibration method to the case of probability panels is not straightforward. In the case of panels, non-response can occur at two different moments of the process: during recruitment and during each of the subsequent surveys [29], and their behavior patterns may be different. However, in the probability panels, there is an additional problem as there are two sources of randomness derived from the two sample designs (the first by which the panel is obtained, and the second by which the sample is selected from the panel) that must be taken into consideration.
In this paper, we will adapt the calibration methodology to the case of probabilistic panels for the treatment of non-response. To the best of our knowledge, this is the first time that this methodology is studied in this context. We will propose different calibration methods in one and two steps and compare the efficiency of these estimators by means of a simulation study. Finally, we will apply the proposed techniques to a real survey conducted using the Citizen Panel for Social Research (PACIS), a probability-based mixed-modes panel of the Andalusian population recruited and maintained by the Institute for Advanced Social Studies(IESA-CSIC).

2. Proposed Calibration Strategies in Panel Surveys

Let y be the interest variable related to a finite population U = 1 , , N consisting of N units. For this population U, a panel P of size N P is drawn with probability p d 1 ( P ) according to d 1 . For this sampling design, the first order probabilities are given by π i ( 1 ) = P i p d 1 ( P ) for i U , where P i denotes the sum of the panels containing the unit i. We denote by d i ( 1 ) = 1 / π i ( 1 ) the panel weight for unit i.
In a second-phase, a probability sample s of size n is selected from the panel P with conditional probability p ( s P ) . In this sampling design d 2 , the first-order inclusion probabilities are denoted by π i P . For each unit in the panel P, we denote by d i ( 2 ) = 1 / π i P the design weight from the panel.
The final sampling design p ( · ) defined in U has first-order inclusion probabilities π i = P i p d 1 ( P ) π i P , i , U . Let d i = 1 / π i denote the sampling design-basic weight for unit i U .
We denote by y i the value of the variable y associated with the i-th individual. We consider the linear parameter Y = i U N y i , the population total. We assume missing data on the sample s for the variable y, which can be divided into the disjoint sets
s r = { i s / i   responds   to   main   variable } s m = { i s / i   do   not   responds   to   main   variable } ,
where s r is the respondent sample is of size r, and  s m is the size n r .
The natural candidate to estimate the population total is the Horvitz–Thompson estimator Y ^ H T = i s r d i y i ; however, it may lead to biased estimates because certain specific groups can be substantially under-represented. These errors can be overcome by the use of reweighting techniques. A standard weighting procedure is to rake or poststratify weights to external population control totals. Raking and post-stratification are special cases of calibration adjustment.
Calibration adjustments [30,31] can be used to extrapolate the estimations of a survey. Calibration estimates reduce non-coverage, non-response, and other biases [23,32]. We will adapt this methodology in our context.
To do so, we assume the existence of auxiliary information relative to several variables related to the main variable y. Auxiliary information can exist simultaneously at the population and the panel level, thus this auxiliary information is given as two vectors: x k * and x k o . Auxiliary information at population level is denoted by x k * , and the population vector total U x k * is known. Auxiliary information at panel level is noted by x k o , and these values are known for k P . The total panel X P o = P x k o is thus known and the total population of these variables can be estimated by P d k ( 1 ) x k o . We denote by x k the auxiliary vector constructed by combining the two vectors:
x k = x k * x k o
and
X = U x k * P d k ( 1 ) x k o
We will use these auxiliary variables in different ways in the calibration process.

2.1. Calibration in One Step

Now, we use all auxiliary information to calibrate at population level.
We define a one step calibration estimator as:
Y ^ CAL 1 = s r w k y k , w k = d k · g k
where the weights w k , for k s r , modify as little as possible, in an average sense, the original sampling weights d k , while respecting the calibration equations
s r w k x k * = U x k * and s r w k x k o = P d k ( 1 ) x k o
or equivalently
k s r w k x k = X
Then, given a distance measure G ( w k , d k ) , the calibration process consists of finding the solution of the following minimization problem:
min w k s r G ( w k , d k )
while respecting the calibration Equation (5).
This calibration problem results in final calibrated weights w k = d k F ( q k x k λ ) , where F ( · ) is the inverse of G ( w k , d k ) / w k , where λ is the Lagrange multiplier and q k are known positive constants used for scaling the calibrated weights (in many situations the value of q k = 1 , k s ).
Since Equation (6) depends on the chosen distance measure G ( w k , d k ) , each different distance measure leads to a specific weighting system and thereby to a new estimator. Many distance measures have been proposed for calibration. The authors in [33] consider several distance measures with desirable properties.
Under mild regularity assumptions, the authors in [33] also state the important result that all the calibration estimators given by (6) are asymptotically equivalent to the regression estimator. Thus, the choice of the distance measure G ( w k , d k ) has only a modest impact on important properties of estimators as the variance.
The raking ratio method (also known as the multiplicative method) uses the Kullback–Leibler pseudo-distance
G ( w k , d k ) = 1 q k ( w k l o g ( w k d k ) + d k w k ) .
The calibrated weights are obtained as w k = d k e x p ( q k x k λ ) . The existence of solutions for this problem is not guaranteed (see [33,34]).
In linear calibration (the case of the Euclidean distance), the weights can be positive or negative, but raking calibration guarantees positive weights, with the same unbiasedness properties as generalized regression estimator; however, the model used for the adjustment of non-response has a slightly more reasonable form than the generalized regression estimator [28]. Finally, the raking method is an available procedure for calculating calibration weights in the best known statistical packages.
Note 1. The authors in [35] introduced an alternative calibration approach called the functional form calibration that gives the same estimators. In this paper, we consider the usual calibration approach because it is simpler to formulate.
Note 2. In some practical situations, the  π i P for all possible panels P are not known. For these cases, one can use (following the idea of [36]) the π 🟉 -estimator changing π i by π 🟉 = π i ( 1 ) π i P as the base estimator and thus calculate a similar π 🟉 -calibrated estimator.

2.2. Calibration in Two Steps

Calibration is a very flexible technique to incorporate varied information at various stages. Next, we will explore several ways of doing two-stage calibration:
  • A two-step calibration method can be defined in this manner:
    Step 1: Adjusting the representativeness of the panel in the population.
    The calibration on population auxiliary variables whose population totals U x k * are known yields the calibrations weights w k ( 1 ) = g k ( 1 ) · d k ( 1 ) , k = 1 , , P , where d k ( 1 ) are the panel weights under the restrictions
    P w k ( 1 ) x k * = U x k * .
    Then, each unit in the panel has a weight that summarizes the auxiliary information obtained from the population.
    Step 2: Adjusting the non-response of the sample in the panel.
    The panel auxiliary information is incorporated through a calibration estimator whose calibrated weights w k ( 2 ) = g k ( 2 ) · d k ( 2 ) (here the starting weights for calibration are the design weights of the units in the panel) verified
    s r w k ( 2 ) x k o = P x k o .
    The final weights are obtained by multiplying these weights and the two-step calibration estimator proposed is
    Y ^ CAL 2 = s r w k y k = s r w k ( 1 ) · w k ( 2 ) · y k = s r d k ( 1 ) · g k ( 1 ) · d k ( 2 ) · g k ( 2 ) · y k .
  • By using the same idea as in [22], we can consider a second procedure:
    Step 1. Computing first intermediate weights by calibrating the survey response set to the panel. Thus, we calculate the weights by min w k s r G ( w k , w k ( 2 ) ) subject to
    s r w k x k o = P x k o .
    The calibrated weights obtained in this first step are noted by w P c k .
    Step 2. Then, using these intermediate weights w P c k for calibration to the population. The problem is now min w k s r G ( w k , w P c k ) subject to
    s r w k x k * = U x k *
    We denote the resulting calibration estimator by Y ^ CAL 3 .
  • We consider other alternative by calibrating the survey response set to the sample that is:
    Step 1. Compare the response set with the sample by min w k s r G ( w k , w k ( 2 ) ) subject to
    s r w k x k o = s x k o .
    The calibrated weights obtained calibrating the survey response set are noted by w S c k .
    Step 2. Compare, using these intermediate weights, the sample with the population by min w k s r G ( w k , w S c k ) subject to
    s r w k x k * = U x k * .
    We denote the resulting calibration estimator by Y ^ CAL 4 .
Note that one particular auxiliary variable can be used in both steps (see the next Section 3).
Previously discussed calibration methods are defendable procedures. They may be comparable in terms of computational burden. In addition, the effectiveness of each method can be assessed by the theoretical variance of the estimator that cannot be expressed simply. We can apply automated linearization to each of the calibration estimators to find the approximate variance (see [37]).

3. Simulation Study

We select for this simulation study the adults of eusilcP data set (population size 47,123), a population available from the R-packages simFrame [38] and simPop [39]. As a panel, we draw a sample of 1500 households stratified by region and select all members in each household. This process is repeated 1000 times (average panel size over simulations 2857.607). For each panel, we select a simple random sample of size 1000 (this process is repeated 1000 times).
As target variable, we use eqIncome, a numeric variable about household income. As  x * -variables (for which are known the population total), we use age, gender, and ecoStat, a factor showing the person’s economic status. As x 0 -variable (for which only are known the panel total), we select hy040n, a numeric income from rental of a property or land (cor(eqIncome, hy040n) = 0.387).
For each panel, we generated missing values due to non-response in the sample using two mechanisms: Missing at random (MAR) with age as auxiliar variable and missing not at random (MNAR), with an auxiliar variable namely benefits, obtained as a sum of the variables referred to different types of subsidies (cor(benefits,eqIncome) = 0.530737). Missing completely at random (MCAR) is not included in this simulation study because it is not a practical situation.
Uncalibrated estimator, ht, and alternative calibration estimators are computed in each simulation runs: cal1 (one step calibration) and the three versions of calibration using two steps: cal2, cal3, and cal4.
Figure 1 and Figure 2 show the boxplots of our results. Green line is the true population mean, and blue square is the average over 1000 × 1000 simulation runs for each method compared.
From Figure 1 and Figure 2, some empirical evidence can be highlighted:
  • The Horvitz–Thompson estimator is a biased estimator and the non-response bias increases with the non-response rate. The best behavior is taken by estimators cal1 and cal2: they correct bias very well even for the largest non-response rates and have the least variability. In this situation of MAR non-response, the calibration estimators cal3 and cal4 are not able to correct the non-response bias: they are even more biased than the HT estimator. Note that the bias of the Horvitz–Thompson estimator is negative, while the cal3 and cal4 estimators have a positive bias.
  • The behavior of the estimators in the second scenario is very different. None of the estimators can correct the non-response bias (note that the green line is off the graph when the non-response rate is 60%) especially when the non-response rate is large. The estimators that are best profiled to reduce them are the cal3 and cal4 estimators although the bias is very large for high response rates. This result was expected since the MNAR non-response is very complex to deal with.
  • It is noteworthy that the estimators cal1 and cal2 have a similar behavior to each other as well as cal3 and cal4. It is also observed that the estimators cal1 and cal2 have less variability in general than cal3 and cal4.

4. Application to the Pacis Survey on Andalusian Citizen Preferences for Political Decision-Making Processes

4.1. The Dataset

The survey Andalusian Citizen Preferences for Political Decision-Making Processes is part of a wider project titled Why do we hate politics? (PRY079/14). This project was leaded by two IESA/CSIC researchers and funded by the public foundation Centro de Estudios Andaluces [40] with the aim of analyzing the perception that lay citizens have on democracy and its limits. Fieldwork was conducted from 30 November to 19 December 2015, the day before the Spanish general election took place. This survey was the second survey conducted using the PACIS.
The PACIS is a probability-based, mixed-modes panel of the Andalusian general population. It was recruited offline following full probability sampling methods in the first semester of 2015. The target population was the general population with 16 years and older residing in independent private households in Andalusia. As obtaining a random sample of specifically named persons from the Spanish Continuous Population Register is restricted to official statistics and international surveys, we conducted address-based sampling using as a sampling frame the Spanish Cadaster, the administrative inventory of real estate that includes all the properties segmented by use (residence, rural, industrial, etc.). The cadaster includes the address of each property and geospatial data that was used to ascribe the addresses to a census section and to design the maps that helped the recruiters to locate the households selected to be part of the initial sample. Census sections are operational partitions of the territory that are defined by easily identifiable boundaries and group 1000 to 2500 residents each. From this sampling frame, households were selected using two-stage stratified cluster sampling where the primary sampling unit was the census section and the secondary sampling unit the household address. Census sections were stratified by province using proportional allocation. Once the household was located and contacted, the aim was to try to register all the persons residing in the household. Recruitment was made in two sequential phases using postal invitations first and face to face visits to households in the second place.
PACIS is a mixed-mode panel because, although web surveys are the main interviewing mode, CATI is used to interview offline population and online panel members that did not answer to online invitation to participate in the survey. In order to maximize response rates and lower attrition, participation is rewarded with 5 euros that the respondent may receive or donate to a charity. It must be noted that PACIS was not conceived to study changes over time, as it is the case with longitudinal panels, but as a valid pool of respondents for extracting cross-sectional samples.
The method used for selecting those survey samples is stratified random sampling with proportional allocation to the population distribution by sex and age.

4.2. Nonresponse in PACIS Surveys

As it has been explained, non-response occurs at two different moments in the case of panels: the recruitment stage and the subsequent surveys processes. Table 1 and Table 2 show the final disposition of case codes and the outcome rates for the recruitment phase of the panel and the PACIS survey on Andalusian Citizen Preferences for Political Decision-Making Processes.
The overall response rate at panel registration was 21.6% (American Association for Public Opinion Research AAPOR RR3), a figure that is in line with the outcome rates obtained by the German Internet Panel (18.1%, RR3) and somewhat lower than more resourced panels like ELIPSS (27.3% AAPOR RR3) or the GESIS panel (25.1% AAPOR RR5) [20].
Refusals at the respondent level are also the most important reason for not participating in the second wave PACIS survey. In addition, 2398 panelists were invited to participate in the survey and 1081 completed the questionnaire, which accounts for a response rate of 48.1%.
Table 3 compares the distribution of the panel members and the respondents to the second wave with the general Andalusian population according to continuous population register statistics. Given that, in both cases, sampling selection has followed a probabilistic procedure and that the recruitment and the interviewing modes are not subjected to undercoverage, we assume that the biggest part of the observed differences is caused by non-response.
We find that the PACIS underrepresents men, the oldest age group (over 60 y.o.), inactive population, people with a basic education level, and non-nationals. Residents in Malaga, a coastal area with a large proportion of secondary residences and holiday rentals, are also less likely to have registered as members in the panel.

4.3. Main and Auxiliary Variables in the PACIS

The data at the population level were obtained from the Spanish National Institute of Statistics selecting 2015 as the reference year. In our case, the recruitment of the panel and the survey were conducted the same year. However, we would like to note that an important aim for computing the weighting adjustments is to correct for the increasing mismatches between the panel and the population associated as the panel ages (panel attrition, changes in the population, etc.). To fullfill this objective in the case of cross-sectional panels, we must take the time of the survey as the reference for selecting the auxiliary data at the population level [41].
Some demographics variables are described in Table 4.
Three main variables in this study are analyzed.
The first of them is related to the “interest in politics". The variable is the answer to the following question: How interested would you say you are in politics?: Very interested, Quite interested, Somewhat interested, Hardly interested, Not at all interested.
The second is about the preferences that the interviewees have about the way in which political decisions should be made. The variable is the answer to the question: How would you like to be made political decisions in Spain, Andalusia, your municipality? The answer to this question was offered in the form of a average score of a numerical scale from 0 to 10, where 0 means that citizens should make all decisions on their own, and 10 that politicians should make all decisions on their own.
In addition, the last one is related to the concrete procedures of decision-making preferred by the interviewees. The question is: We would like you to rate the following ways of making political decisions? The answer to this question was offered in the form of scores in a numerical scale from 0 to 10, where 0 means that the procedure in question does not help anything in decision-making and the 10 that is the best way to make decisions. Six procedures are analyzed: Elections are the way for citizens to intervene in decisions (Elections), Take policy decisions through consultation to experts (Consult to experts), Organize assemblies and meetings so that people can take decisions by themselves (Assemblies), Allow experts to take important policy decisions (Allow experts), Organize referenda frequently (Referenda), and Let the government take the decisions (Government).
There are several auxiliary variables that can be used for calibration at the panel level. We have followed the criterion proposed by [25] based on the H 3 bias indicator, to select the variables that, included in the auxiliary vector, were the most effective for bias reduction. The final auxiliary variables selected are described in Table 4 and Table 5.

4.4. Results

Table 6, Table 7 and Table 8 show the estimated proportions (in percent, %) for the three variables by using the two proposed calibration methods. Root mean squared errors ( R M S E ) are estimated using jackknife techniques and are also shown in Table 6, Table 7 and Table 8.
By examining the estimators and the RMSE in the above tables, we are led to the following observations:
  • Survey results after applying calibration weights are different from unweighted results.
  • Differences in survey results after applying one-step or two-step calibration estimators are small in estimates and in errors.
  • Calibration estimator in one step cal1 come close in RMSE to cal2. The two-step calibration estimators cal3 and cal4 estimators also give similar results to each other.
  • There is no estimator that always has the lowest RMSE errors.
We would like to note that the results obtained by applying these methods to the survey indicate that 45% of Andalusians are interested in politics, a figure that is somewhat higher than that of the Spanish Citizens (40% according to 2016 European Social Survey data) but still far from the 47% of EU citizens that are interested in politics [42]. Regarding their preferences on who should make the political decisions, most Andalusians choose central positions, with a few more preferring the participatory to the representative pole and that for the three governance levels analyzed (state, region, and municipality). This preference is coherent with the way Andalusian people evaluate different decision-making procedures, with the elections obtaining the highest score (7.27) just before consultation with experts (7.05) and citizen assemblies (6.7). The same pattern is observed in a nationally representative survey conducted in Spain in 2011 (Centro de Investigaciones Sociológicas, CIS, Study number 2860) [43].

5. Discussion

In this work, we examined probability-based panel survey settings where there are missing values.
We use raking calibration weighting to remove selection biases resulting from unit non-response. Although this technique was developed to reduce standard errors, many authors have shown that calibration can also be used to handle non response [22,23,24].... In Section 2, we consider four methods to use the different auxiliary information in the calibration although other alternatives are possible. In each method, a set of calibrated weights is computed by modifying the design weights through the available auxiliary information. This single set of weights will then be used to calculate the calibrated estimates of all the population parameters of interest since the values of the weights do not depend on y variables. This is a great advantage of the calibration technique in multipurpose surveys.
An important problem is the selection of auxiliary vector for the calibration. The authors in [44,45] point out that the selection of the variables is more important than the method used to calibrate. The role of the explanatory variables has been demonstrated to be crucial at the post-adjustment stage. The advantage of having a probability-based panel is that more auxiliary information is available for the calibration. Many different vectors can be composed. The authors in [25] provide computational tools for obtaining a preference ordering of potential x-vectors, with the objective to reduce as much as possible the bias remaining in the calibration estimator. The selection of variables for the calibration is a very interesting problem but quite complex to deal with in this situation, which is different for each method used. This is a problem to deal with in future research.
We conduct a simulation study in which it is observed that the estimators cal1 and cal2 have a similar behavior to each other as well as cal3 and cal4. The simulation study also shows that calibration can significantly reduce bias for MAR mechanisms while it is not very effective for NMAR mechanisms.
We use data from one survey conducted using the Citizen Panel for Social Research to test the behavior of proposed estimators. Although the same auxiliary information goes into the final weights, the resulting calibration estimators are not identical. In the same way, the differences for the estimators of the variances can sometimes be significant depending on the relationship between the variables.

6. Conclusions

The main conclusion of this work is that weighting by calibration is an effective technique for reducing non-response bias in probability-based panels when data are missing at random (MAR). Regarding the calibration method, in one or two steps, there are not theoretical reasons to prefer one over the other. However, the two-steps method has some practical advantages. Knowing the weights of each unit in the sampling frame allows for incorporating this information in the selection of the subsequent cross-sectional samples and maximizing the number of auxiliary variables when computing the weights for each specific survey. Another benefit of this approach is that it allows for isolating the sources of error produced at each step (i.e., coverage error in the recruitment of the panel and attrition bias in the subsequent surveys). It must be taken into account that panel weights have to be updated taking the time of the survey as the reference for selecting the auxiliary data at the population level when using this approach to calibrate cross-sectional panels. In this case, the calibration in one step may be more straightforward, particularly if auxiliary variables at the survey level do not add much value to the quality of the estimates.
In all, the choice of the calibration method is a complex decision where different aspects come into play: the technical features of the panel and the survey, the mechanism of non-response affecting our data, and the availability of auxiliary information.

Author Contributions

The authors contributed equally to this work in conceptualization, methodology, software, and original draft preparation. All authors have read and agreed to the published version of the manuscript.

Funding

The work was supported by the Ministerio de Economia, Industria y Competitividad, Spain, under Grant MTM2015-63609-R.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. ESOMAR; Global Market Research: Amsterdam, The Netherlands, 2016.
  2. Peytchev, A.; Carley-Baxter, L.R.; Black, M.C. Multiple Sources of Nonobservation Error in Telephone Surveys: Coverage and Nonresponse. Sociol. Methods Res. 2011, 40, 138–168. [Google Scholar] [CrossRef]
  3. Eckman, S.; Kreuter, F. Undercoverage Rates and Undercoverage Bias in Traditional Housing Unit Listing. Sociol. Methods Res. 2013, 42, 264–293. [Google Scholar] [CrossRef]
  4. Gummer, T. Assessing Trends and Decomposing Change in Nonresponse Bias: The Case of Bias in Cohort Distributions. Sociol. Methods Res. 2017. [Google Scholar] [CrossRef]
  5. De Leeuw, E. Counting and Measuring Online: The Quality of Internet Surveys. Bull. Sociol. Methodol. Methodol. Sociol. 2012, 114, 68–78. [Google Scholar] [CrossRef]
  6. Dixon, J.; Tucker, C. Survey Nonresponse. In Handbook of Survey Research; Marsden, P., Wright, J., Eds.; Emerald Group Publishing: Bingley, UK, 2010; pp. 593–630. [Google Scholar]
  7. Tucker, C.; Lepkowski, J. Telephone Survey Methods: Adapting to Change. In Advances in Telephone Survey Methodology; Lepkowski, J., Tucker, C., Brick, J., de Leeuw, E., Japec, L., Lavrakas, P., Sangster, R., Eds.; Wiley-Interscience: Hoboken, NJ, USA, 2008; pp. 3–26. [Google Scholar]
  8. Vehovar, V.; Berzelak, N.; Lozar Manfreda, K. Mobile Phones in an Environment of Competing Survey Modes: Applying Metric for Evaluation of Costs and Errors. Soc. Sci. Comput. Rev. 2010, 28, 303–318. [Google Scholar] [CrossRef]
  9. Lu, Y.; Chen, Z.; Law, R. Mapping the progress of social media research in hospitality and tourism management from 2004 to 2014. J. Travel Tour. Mark. 2018, 35, 102–118. [Google Scholar] [CrossRef]
  10. Manca, M.; Boratto, L.; Roman, V.M.; i Gallissà, O.M.; Kaltenbrunner, A. Using social media to characterize urban mobility patterns: State-of-the-art survey and case-study. Online Soc. Netw. Media 2017, 1, 56–69. [Google Scholar] [CrossRef]
  11. Sinnenberg, L.; Buttenheim, A.M.; Padrez, K.; Mancheno, C.; Ungar, L.; Merchant, R.M. Twitter as a tool for health research: A systematic review. Am. J. Public Health 2017, 107, e1–e8. [Google Scholar] [CrossRef]
  12. Li, R.; Crowe, J.; Leifer, D.; Zou, L.; Schoof, J. Beyond big data: Social media challenges and opportunities for understanding social perception of energy. Energy Res. Soc. Sci. 2019, 56, 101217. [Google Scholar] [CrossRef]
  13. Japec, L.; Kreuter, F.; Berg, M.; Biemer, P.; Decker, P.; Lampe, C.; Usher, A. Big data in survey research: AAPOR task force report. Public Opin. Q. 2015, 79, 839–880. [Google Scholar] [CrossRef]
  14. Callegaro, M.; Manfreda, K.L.; Vehovar, V. Web Survey Methodology; SAGE: Thousand Oaks, CA, USA, 2015. [Google Scholar]
  15. Fricker, R.D.; Schonlau, M. Advantages and Disadvantages of Internet Research Surveys: Evidence from the Literature. Field Methods 2002, 14, 347–367. [Google Scholar] [CrossRef] [Green Version]
  16. Tsetsi, E.; Rains, S.A. Smartphone Internet access and use: Extending the digital divide and usage gap. Mob. Media Commun. 2017, 5, 239–255. [Google Scholar] [CrossRef]
  17. Lee, M.H. Statistical Methods for Reducing Bias in Web Surveys. Ph.D. Thesis, Simon Fraser University, Burnaby, BC, Canada, 2011. Available online: http://summit.sfu.ca/item/11783 (accessed on 15 November 2019).
  18. DiSogra, C.; Callegaro, M. Metrics and Design Tool for Building and Evaluating Probability-Based Online Panels. Soc. Sci. Comput. Rev. 2016, 34, 26–40. [Google Scholar] [CrossRef]
  19. Callegaro, M.; Baker, R.; Bethlehem, J.; Göritz, A.S.; Krosnick, J.A.; Lavrakas, P.J. Online panel research: History, concepts, applications and a look at the future, Chapter 1. In Online Panel Research: A Data Quality Perspective; Callegaro, M., Baker, R., Bethlehem, J., Göritz, A.S., Krosnick, J.A., Lavrakas, P.J., Eds.; John Wiley & Sons, Ltd: Chichester, UK, 2014; pp. 1–22. [Google Scholar]
  20. Blom, A.G.; Bosnjak, M.; Cornilleau, A.; Cousteaux, A.-S.; Das, M.; Douhou, S.; Krieger, U. A Comparison of Four Probability-Based Online and Mixed-Mode Panels in Europe. Soc. Sci. Comput. Rev. 2016, 34, 8–25. [Google Scholar] [CrossRef]
  21. Cheng, A.; Zamarro, G.; Orriens, B. Personality as a predictor of unit non-response in an internet panel. Sociol. Methods Res. 2018. [Google Scholar] [CrossRef]
  22. Andersson, P.; Särndal, C.E. Calibration for non-response treatment: In one or two stepsf. Stat. J. IAOS 2014, 32, 375–381. [Google Scholar] [CrossRef]
  23. Särndal, C.E.; Lundström, S. Estimation in Surveys with Nonresponse; John Wiley and Sons: New York, NY, USA, 2005. [Google Scholar]
  24. Särndal, C.E.; Lundström, S. Assessing auxiliary vectors for control of non-response bias in the calibration estimator. J. Off. Stat. 2008, 4, 251–260. [Google Scholar]
  25. Särndal, C.E.; Lundström, S. Design for estimation: Identifying auxiliary vectors to reduce non-response bias. Surv. Methodol. 2010, 36, 131–144. [Google Scholar]
  26. Kott, P.S. Using Calibration Weighting to Adjust for Nonresponse and Coverage Errors. Surv. Methodol. 2006, 32, 133–142. [Google Scholar]
  27. Kott, P.S.; Liao, D. Providing double protection for unit non-response with a nonlinear calibration-weighting routine. Surv. Res. Methods 2012, 6, 105–111. [Google Scholar]
  28. Kott, P.S.; Liao, D. One step or two? Calibration weihting form a complete list frame with non-response. Surv. Methodolgy 2015, 41, 165–181. [Google Scholar]
  29. Bethlehem, J.; Callegaro, M. Introduction to Part IV. In Online Panel Research: A Data Quality Perspective; Callegaro, M., Baker, R., Bethlehem, J., Göritz, A.S., Krosnick, J.A., Lavrakas, P.J., Eds.; John Wiley & Sons, Ltd: Chichester, UK, 2014; pp. 263–272. [Google Scholar]
  30. Särndal, C.E. The calibration approach in survey theory and practice. Surv. Methol. 2007, 33, 99–119. [Google Scholar]
  31. Arcos, A.; Rueda, M.; Trujillo, M.; Molina, D. Review of estimation methods for landline and cell phone surveys. Sociol. Methods Res. 2015, 44, 458–485. [Google Scholar] [CrossRef]
  32. Arcos, A.; Contreras, J.M.; Rueda, M. A novel calibration estimator in social surveys. Sociol. Methods Res. 2014, 43, 465–489. [Google Scholar] [CrossRef]
  33. Deville, J.C.; Särndal, C.E. Calibration estimators in survey sampling. J. Am. Stat. Assoc. 1992, 87, 376–382. [Google Scholar]
  34. Devaud, D.; Tillé, Y. Deville and Särndal’s calibration: Revisiting a 25-years-old successful optimization problem. TEST 2019. [Google Scholar] [CrossRef]
  35. Estevao, V.M.; Särndal, C.E. A functional form approach to calibration. J. Off. Stat. 2000, 16, 379–399. [Google Scholar]
  36. Särndal, C.E. Methods for estimating the precision of survey estimates when imputation has been used. Surv. Methodol. 1992, 18, 241–252. [Google Scholar]
  37. Estevao, V.M.; Särndal, C.E. Survey Estimates by Calibration on Complex Auxiliary Information. Internat. Statist. Rev. 2006, 74, 127–147. [Google Scholar] [CrossRef]
  38. Alfons, A.; Templ, M.; Filzmoser, P. An Object-Oriented Framework for Statistical Simulation: The R Package simFrame. J. Stat. Softw. 2010, 37, 1–36. [Google Scholar] [CrossRef] [Green Version]
  39. Templ, M.; Meindl, B.; Kowarik, A.; Dupriez, O. Simulation of Synthetic Complex Data: The R Package simPop. J. Stat. Softw. 2017, 79, 1–38. [Google Scholar] [CrossRef] [Green Version]
  40. Alarcón, P.; Font, J.; Fernéz, J. Tell me what you trust and what you think about political actors and I will tell you what democracy you prefer. In XIII AECPA Conference. “La fortaleza de Europa: Vallas y puentes”; Universidad de Santiago de Compostela: Santiago de Compostela, Galicia, 2017. [Google Scholar]
  41. Verma, V.; Betti, G.; Ghellini, G. Cross-sectional and longitudinal weighting in a rotational household panel: Applications to EU-SILC. Stat. Transit. 2007, 8, 5–50. [Google Scholar]
  42. ESS Round 8: European Social Survey Round 8 Data; Data file edition 2.1; Data- Archive and distributor of ESS data for ESS ERIC; Norwegian Centre for Research Data for ESS ERIC: Bergen, Norway, 2016.
  43. Font, J.; Wojcieszak, M.; Navarro, C.J. Participation, Representation and Expertise: Citizen Preferences for Political Decision-Making Processes. Political Stud. 2015, 63, 153–172. [Google Scholar] [CrossRef]
  44. Rizzo, L.; Kalton, G.; Brick, J.M. A comparison of some weighting adjustment methods for panel non-response. Surv. Methodol. 1996, 22, 43–53. [Google Scholar]
  45. Mercer, A.; Lau, A.; Kennedy, C. For Weighting Online Opt-In Samples, What Matters Most? Pew Res. Center. 2018, 2018, 28. [Google Scholar]
Figure 1. Missing at Random case. Boxplots of compared estimators for missing rates 20%, 40% and 60%. Green line is the true population mean. Blue square is the average over simulations runs.
Figure 1. Missing at Random case. Boxplots of compared estimators for missing rates 20%, 40% and 60%. Green line is the true population mean. Blue square is the average over simulations runs.
Mathematics 08 00423 g001
Figure 2. Missing Not at Random case. Boxplots of compared estimators for missing rates 20%, 40%, and 60%. Green line is the true population mean. Blue square is the average over simulations runs.
Figure 2. Missing Not at Random case. Boxplots of compared estimators for missing rates 20%, 40%, and 60%. Green line is the true population mean. Blue square is the average over simulations runs.
Mathematics 08 00423 g002
Table 1. Final disposition and outcome rates for the PACIS recruitment phase (face to face mostly).
Table 1. Final disposition and outcome rates for the PACIS recruitment phase (face to face mostly).
n
Interview 2803
Valid registration after postal
contact (online or hotline phone)186
Valid registration after F2F
contact (visits to households)2617
Eligible, non-interview 5886
Refusals5474
Contacted, visit postponed378
Language barriers34
Unknown eligibility 6123
No answer5690
Unable to reach / Unsafe area308
Unable to locate address125
Not eligible 2345
Non residence866
Vacant housing unit1479
Total of units in the sample 17,517
Response Rate RR118.9%
Response Rate RR321.6%
Refusal Rate RF139.7%
Cooperation Rate COOP232.3%
Contact Rate CON264.3%
Table 2. Final disposition and Outcome rates for PACIS 2nd Wave Survey (Web + CATI).
Table 2. Final disposition and Outcome rates for PACIS 2nd Wave Survey (Web + CATI).
n
Interview 1081
Conducted online623
Conducted by CATI458
Eligible, non-interview 911
Refusals812
Break off/ Implicit refusal (internet surveys)64
Logged on to survey, did not complete any item33
Telephone answering device (confirming HH)2
Unknown eligibility 254
No answer254
Not eligible 152
Fax/Data line2
Phone number doesn’t match with panelist150
Total of units in the sample 2398
Response Rate RR148.1%
Response Rate RR348.5%
Refusal Rate RF140.5%
Cooperation Rate COOP254.3%
Contact Rate CON289.3%
Table 3. Sample percentages for population, panel members, and respondents to PACIS Second Wave Survey.
Table 3. Sample percentages for population, panel members, and respondents to PACIS Second Wave Survey.
Sample DistributionPanelDiff. with2nd WaveDiff. withPopulation
MembersPopulationRespondentsPopulation
Women57.1%+6.0 %49.2%− 1.9 %51.1%
Over 60 years old16.9%−9.0 %19.1%−6.8 %25.9%
Residence
 Malaga14.7%−4.7 %15.6%−3.8 %19.4%
 Granada14.5%+3.6 %12.3%+1.4 %10.9%
Education level
 Up to primary19.9%−8.1 %13.3%−14.7 %28.0%
 Secondary58.0%+8.4 %59.5%+9.9 %49.6%
 Tertiary22.2%−0.2 %27.2%+4.8 %22.4%
Labor status
 Employed37.4%−1.6 %38.2%−0.8 %39.0%
 Unemployed30.6%+10.3 %33.7%+13.4 %20.3%
 Economically inactive32.0%−8.7 %28.1%−12.6 %40.7%
Foreign nationality3.8%−2.9 %2.6%−4.1 %7.9%
Table 4. Education level by sex and age at population and panel levels.
Table 4. Education level by sex and age at population and panel levels.
PrimarySecondaryHigher PrimarySecondaryHigher
Men
16–2971,968537,81459,433 3521750
30–4491,673702,591194,410 40310143
45–59197,901586,726126,083 5823977
>60471,729231,98897,686 1199452
Women
16–2944,609487,020112,122 4421188
30–4478,176619,471272,774 45466210
45–59194,971572,750149,620 12132294
>60681,508224,65172,158 2048228
Those estimates at population level can be found on the INE website: http://www.ine.es/.
Table 5. Auxiliary variables at panel level (percentages).
Table 5. Auxiliary variables at panel level (percentages).
Laboral Status
WorkingRetiredUnemployedStudentHousework
37.4116.1230.587.468.42
Province
AlmeríaCádizCórdobaGranadaHuelvaJaénMálagaSevilla
7.6714.729.6714.518.338.3314.6622.10
PCNetEmailLand phoneCellular phone
% of owners
78.0579.157.9637.1278.59
Table 6. Point estimation of proportions (in percent, %) using several methods: unweighted (ht), calibration in one (cal1) or two stages (cal2, cal3 and cal4). Main variable: “Interest in politics”. Estimated Root Mean Squared Error using jackknife (in percent, %).
Table 6. Point estimation of proportions (in percent, %) using several methods: unweighted (ht), calibration in one (cal1) or two stages (cal2, cal3 and cal4). Main variable: “Interest in politics”. Estimated Root Mean Squared Error using jackknife (in percent, %).
How Interested Would You Say You Are in Politics?
Est. RMSE(%)
htcal1cal2cal3cal4 htcal1cal2cal3cal4
Extremely17.0214.6315.0514.8814.71 1.531.281.391.281.27
Very33.3030.0430.0331.3931.05 2.542.162.182.052.03
Moderately27.3427.0226.8527.0427.07 1.951.851.841.771.78
Slightly14.7917.5017.2016.5516.84 1.961.911.761.591.61
Not at all7.5510.8210.8610.1310.33 1.861.551.501.361.39
Table 7. Point estimation of proportions (in percent, %) using several methods: unweighted (ht), calibration in one (cal1) or two stages (cal2, cal3 and cal4). Main variable:“Preferences for political decisions”. Estimated Root Mean Squared Error using jackknife.
Table 7. Point estimation of proportions (in percent, %) using several methods: unweighted (ht), calibration in one (cal1) or two stages (cal2, cal3 and cal4). Main variable:“Preferences for political decisions”. Estimated Root Mean Squared Error using jackknife.
How Would You Like to Be Made Political Decisions in ...?
Spain Andalusia Municipality
Est.RMSE Est.RMSE Est.RMSE
ht4.590.19 4.350.19 4.280.19
cal14.530.17 4.410.17 4.270.17
cal24.530.17 4.450.17 4.300.17
cal34.540.15 4.450.15 4.320.16
cal44.550.15 4.460.15 4.320.16
Table 8. Point estimation of proportions (in percent, %) using several methods: Calibration in one (cal1) or two stages (cal2, cal3 and cal4). Main variable:“Procedures of decision-making”. Estimated Root Mean Squared Error using jackknife.
Table 8. Point estimation of proportions (in percent, %) using several methods: Calibration in one (cal1) or two stages (cal2, cal3 and cal4). Main variable:“Procedures of decision-making”. Estimated Root Mean Squared Error using jackknife.
We Would Like You to Rate the Following
Ways of Making Political Decisions in Andalusia?
ElectionsConsult
to Experts
AssembliesAllow
Experts
ReferendumsPolitical
Leaders
ht E s t . 7.347.106.406.445.884.30
R M S E 0.300.290.250.290.230.19
cal1 E s t . 7.337.036.706.515.904.28
R M S E 0.240.220.240.220.220.18
cal2 E s t . 7.277.056.706.575.874.29
R M S E 0.230.220.230.230.210.18
cal3 E s t . 7.277.076.636.505.844.25
R M S E 0.210.200.200.190.190.16
cal4 E s t . 7.277.076.646.515.844.26
R M S E 0.210.200.200.200.190.16

Share and Cite

MDPI and ACS Style

Arcos , A.; Rueda , M.d.M.; Pasadas-del-Amo, S. Treating Nonresponse in Probability-Based Online Panels through Calibration: Empirical Evidence from a Survey of Political Decision-Making Procedures. Mathematics 2020, 8, 423. https://doi.org/10.3390/math8030423

AMA Style

Arcos  A, Rueda  MdM, Pasadas-del-Amo S. Treating Nonresponse in Probability-Based Online Panels through Calibration: Empirical Evidence from a Survey of Political Decision-Making Procedures. Mathematics. 2020; 8(3):423. https://doi.org/10.3390/math8030423

Chicago/Turabian Style

Arcos , Antonio, Maria del Mar Rueda , and Sara Pasadas-del-Amo. 2020. "Treating Nonresponse in Probability-Based Online Panels through Calibration: Empirical Evidence from a Survey of Political Decision-Making Procedures" Mathematics 8, no. 3: 423. https://doi.org/10.3390/math8030423

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop