ESRA logo

Tuesday 16th July       Wednesday 17th July       Thursday 18th July       Friday 19th July      

Download the conference book

Download the program





Wednesday 17th July 2013, 16:00 - 17:30, Room: No. 14

Investigating non respondents: how to get reliable data and how to use them

Convenor Mrs Michele Ernst Staehli (FORS)
Coordinator 1Mrs Hideko Matsuo (KU Leuven)
Coordinator 2Mr Dominique Joye (UNIL)

Session Details

Since response rates are not a sufficient indicator of data quality, as they are not directly linked to nonresponse bias, survey researchers have increased effort to access reliable data about nonrespondents. The purpose is to detect, quantify and ideally to adjust for nonresponse bias. But collecting such data presents several challenges. Besides the problem of their accessibility, the cost of collection and the burden for interviewers as well as for reluctant/refusing sample units, the quality of the data is not simple to achieve. There either are very basic data on the whole sample frame which mostly don't explain much of nonresponse bias, or such data are based on rough evaluation i.e. by interviewers or very short questions raising problems of data quality, or they are collected by distinct surveys raising problems of comparability with the data of respondents.
This session proposes to gather and discuss experiences about the collection of such data with, in particular, the perspective of optimizing nonresponse follow-up surveys.
The main questions we want to address are:
- which kind of data are most useful (type of collection)
- which variables/items best detect nonresponse bias (content of information)
- which is or are the best way(s) to collect reliable data
- how can the quality of such data be assured and improved and
- eventually, how can these data best be analyzed?

We especially welcome papers that compare
- different sources of data (e.g. paradata, observable data, data from nonrespondent follow-ups or surveys)
- different types of information about the nonrespondents (e.g. contextual, socio-demographic, attitudinal)
- different designs for surveying nonrespondents
- different methods of analysis and bias adjustment.
We also look forward to discover unconventional items and innovative designs and methods.


Paper Details

1. Assessing non-responses biases in the French "End of life" survey

Dr Nicolas Razafindratsima (INED)
Dr Sophie Pennec (INED)
Dr Stephane Legleye (INED, Inserm)

In 2010, the French national institute of demographic studies (INED) conducted a survey on end-of-life in France. The survey covers a random sample of 15 000 deaths. For each death, the physician who filled in the death certificate was asked to fill/complete a questionnaire on the circumstances of the patient's end of life, whether online or by paper and postal mode. Response rate was about 40%. A follow-up survey, with a very short questionnaire with some socio-demographic questions, was then conducted by telephone among a sample of non-respondent physicians. Its response rate was fairly good (80%).

The objective of our presentation is to study non-response biases in this survey, using different sources and strategies:

- data from the follow-up survey on non-respondents;
- data issued from the fieldwork management software (which is disconnected from the response file);
- by analysing the characteristics of those physicians who responded by internet with those who responded by paper;
- by using different weighting strategies: while calibration (Deville et al. 1993) is the most appropriate way to compute weights in such a survey, there are several methodological possibilities. One is to use only benchmark statistics on the deceased, another is to introduce also the characteristics of the physicians. We will compare these two strategies.

References

Deville, J.-C., Särndal, C.-E. and Sautory, O. (1993). "Generalized raking procedures in survey sampling", Journal of the American Statistical Association, vol 88, n°423, pp. 1013-1020


2. Building a final response set

Dr Peter Lundquist (Statistician Stats Sweden)
Professor Carl-erik Särndal

In recent literature on survey nonresponse, new indicators of the quality of the data collection have been proposed. These include indicators of imbalance and representativity (of the set of respondents) and distance (between respondents and non-respondents), computed on available auxiliary variabels. The analysis of imbalance could be obtained as coefficients of variation of the response propensity scores. This finding is used in the data collection. At suitable points in the data collection, we may wish to reduce the number of contact attempts for selected sample elements, or to stop the attempts altogether for specified units. The response propensity scores help identify the units that should be targeted in these interventions.
Both theoretical and empirical resultat are presented. We use propensity scores in conjunction with paradata from the Swedish CATI-system to examine the inflow of data (as a function of the call attempt number) for the Swedish Living Conditions Survey (LCS). Cost savings realized by fewer calls can be redirected to enhance quality of other aspects of the survey design.


3. Use of a non-response follow-up survey to assess the validity of R-indicators as a measure of the risk of bias

Dr Caroline Vandenplas (UNIL)
Dr Caroline Roberts (UNIL)

In the search for alternatives to response rates as indicators of nonresponse bias, representativity indicators, or 'R-indicators' are among the most promising. The use of R-indicators depends on the availability of auxiliary data that correlate both with the response propensity and key survey variables. Their suitability for social surveys measuring a variety of subjective phenomena may therefore be limited. We investigate this problem using data from the Swiss European Social Survey (2010), including new sample data based on population registers, and data from a nonresponse follow-up survey. The R-indicators are computed with sample data and the risk of bias on a hypothetical survey variable due to differential nonresponse across population subgroups (Maximal Absolute Bias) estimated. Data from the nonresponse follow-up are used to investigate the presence of bias on selected survey variables.
Examining changes in the R-indicator and Maximal Absolute Bias finds that the risk of bias is reduced as a result of fieldwork efforts. However, the difference between estimates from the main survey and the nonresponse follow-up suggests no improvement in biases resulting from response rate increases. We find only a weak correlation between the register variables used to construct the R-indicator and variables in the nonresponse follow-up suggesting that the constructed R-indicator may not be informative about the real extent of bias from nonresponse.
We discuss the strengths and the limitations of R-indicators and nonresponse follow-ups as evidence of the presence of bias in such surveys.



4. Can we make better use of non-response surveys? A case study from ESS Round 5 Swiss data

Ms Ahu Alanya (KU Leuven)
Professor Joye Dominique
Dr Michelle Stähli
Professor Jaak Billiet (KU Leuven)

Nonresponse follow-up surveys (NRS) provide information on a subsample of non-respondents. Previous research on NRS data from the ESS showed that nonresponse adjustment based on NRS makes little difference in survey estimates between weighted and unweighted respondent sample. In order to strengthen the adjustment and gain more insight into the remainder of nonrespondents, the current study uses multiple imputation to first complete the NRS for all non-respondents based on rich frame data and paradata available for all sample units. Then, we model response propensity weights using the multiply imputed NRS variables. In the final step, non-response weights for respondents are computed by averaging propensity scores from the multiple imputed data sets and applying subclassification on propensity scores.