Web Survey Bibliography
Web surveys are increasingly considered as cost-effective mode of data collection for large-scale social surveys. Existing face-to-face surveys introduce mixed-mode approaches including web, and a number of probability-based online panels have recently been established in the United States and Europe. In contrast to interviewer-administered surveys, however, survey researchers lose control over the environment in which respondents complete the survey. Web respondents can decide when and where to fill in the questionnaire, and might be exposed to various sources of distractions or might choose to get involved in other activities while filling in the questionnaire. In particular, respondents who use a mobile device might be in distracting environments, where other people are present. Distractions and multi-tasking are potential threats to data quality as respondents might not be able to fully concentrate on the survey task but might rely on cognitive shortcuts.
This paper reports on results from a laboratory experiment that is being conducted in November 2016 to examine how distractions during web survey completion influence data quality, and to identify if the environment of survey completion is a potential source of measurement error.
Subjects (N = 276) are randomly assigned to experimental groups using a 3 (form of distraction) x 2 (device type) design and are asked to complete an online survey. The three forms of distraction are music versus conversation between other people in the room versus no distraction, and the two levels for device type are PC versus tablet. Distractions were chosen to represent two sources of distractions that are likely to occur in web survey settings.
I will examine the effects of distraction and device type on various data quality measures, including item-nonresponse, straight-lining, extreme response styles, response consistency, survey duration, and responses to an Instructional Manipulation Check.
This paper adds to research on how the environment in which respondents fill in questionnaires affects response quality in web surveys.
Web survey bibliography - European survey research associaton conference 2017, ESRA, Lisbon (26)
- Effects of sampling procedure on data quality in a web survey; 2017; Rimac, I.; Ogresta, J.
- Paradata as an aide to questionnaire design: Improving quality and reducing burden; 2017; Timm, E.; Stewart, J.; Sidney, I.
- Fieldwork monitoring and managing with time-related paradata; 2017; Vandenplas, C.
- Interviewer effects on onliner and offliner participation in the German Internet Panel; 2017; Herzing, J. M. E.; Blom, A. G.; Meuleman, B.
- Interviewer Gender and Survey Responses: The Effects of Humanizing Cues Variations; 2017; Jablonski, W.; Krzewinska, A.; Grzeszkiewicz-Radulska, K.
- Millennials and emojis in Spain and Mexico.; 2017; Bosch Jover, O.; Revilla, M.
- Where, When, How and with What Do Panel Interviews Take Place and Is the Quality of Answers Affected...; 2017; Niebruegge, S.
- Comparing the same Questionnaire between five Online Panels: A Study of the Effect of Recruitment Strategy...; 2017; Schnell, R.; Panreck, L.
- Nonresponses as context-sensitive response behaviour of participants in online-surveys and their relevance...; 2017; Wetzlehuetter, D.
- Do distractions during web survey completion affect data quality? Findings from a laboratory experiment...; 2017; Wenz, A.
- Predicting Breakoffs in Web Surveys; 2017; Mittereder, F.; West, B. T.
- Measuring Subjective Health and Life Satisfaction with U.S. Hispanics; 2017; Lee, S.; Davis, R.
- Humanizing Cues in Internet Surveys: Investigating Respondent Cognitive Processes; 2017; Jablonski, W.; Grzeszkiewicz-Radulska, K.; Krzewinska, A.
- A Comparison of Emerging Pretesting Methods for Evaluating “Modern” Surveys; 2017; Geisen, E., Murphy, J.
- The Effect of Respondent Commitment on Response Quality in Two Online Surveys; 2017; Cibelli Hibben, K.
- Pushing to web in the ISSP; 2017; Jonsdottir, G. A.; Dofradottir, A. G.; Einarsson, H. B.
- The 2016 Canadian Census: An Innovative Wave Collection Methodology to Maximize Self-Response and Internet...; 2017; Mathieu, P.
- Push2web or less is more? Experimental evidence from a mixed-mode population survey at the community...; 2017; Neumann, R.; Haeder, M.; Brust, O.; Dittrich, E.; von Hermanni, H.
- In search of best practices; 2017; Kappelhof, J. W. S.; Steijn, S.
- Redirected Inbound Call Sampling (RICS); A New Methodology ; 2017; Krotki, K.; Bobashev, G.; Levine, B.; Richards, S.
- An Empirical Process for Using Non-probability Survey for Inference; 2017; Tortora, R.; Iachan, R.
- The perils of non-probability sampling; 2017; Bethlehem, J.
- A Comparison of Two Nonprobability Samples with Probability Samples; 2017; Zack, E. S.; Kennedy, J. M.
- A test of sample matching using a pseudo-web sample; 2017; Chatrchi, G., Gambino, J.
- A Partially Successful Attempt to Integrate a Web-Recruited Cohort into an Address-Based Sample; 2017; Kott, P. S., Farrelly, M., Kamyab, K.
- Nonprobability sampling as model construction; 2017; Mercer, A. W.