Nowadays, internet-based surveys are increasingly
used for data collection, because their usage is simple and cheap. Also they
give fast access to a large group of respondents. There are many factors
affecting internet surveys, such as measurement, survey design and sampling
selection bias. The sampling has an important place in selection bias in
internet survey. In terms of sample selection, the type of access to internet
surveys has several limitations. There are internet surveys based on restricted
access and on voluntary participation, and these are characterized by their
implementation according to the type of survey. It can be used probability and
non-probability sampling, both of which may lead to biased estimates. There are
different ways to correct for selection biases; poststratification or weighting
class adjustments, raking or rim weighting, generalized regression modeling and
propensity score adjustments. This paper aims to describe methodological
problems about selection bias issues and to give a review in internet surveys.
Also the objective of this study is to show the effect of various correction
techniques for reducing selection bias.
References
[1]
Schonlau, M., Soest, A., Kapteyn, A. and Couper, M. (2009) Selection Bias in Web Surveys and the Use of Propensity Score. Social Methods and Research, 37, 291-318. http://dx.doi.org/10.1177/0049124108327128
[2]
Internet World Stat. www.internetworldstats.com
[3]
Web Survey Methodology. www.websm.org
[4]
Bethlelem, J. (2008) How Accurate Are Self-Selection Web Surveys? Discussion Paper, University Amsterdam.
[5]
Asan, Z. and Ayhan, H.Ö. (2013) Sampling Frame Coverage and Domain Adjustment Procedures for Internet Surveys. Quality and Quantity, 47, 3031-3042. http://dx.doi.org/10.1007/s11135-012-9701-8
[6]
Ayhan, H.Ö. (2000) Estimators of Vital Events in Dual—Record Systems. Journal of Applied Statistics, 27, 157-169.
http://dx.doi.org/10.1080/02664760021691
[7]
Ayhan, H.Ö. (2003) Combined Weighting Procedures for Post-Survey Adjustment in Complex Sample Surveys. Bulletin of the International Statistical Institute, 60, 53-54.
[8]
Bethlelem, J. (2008) Applied Survey Methods a Statistical Perspective. John Wiley & Sons, Hoboken.
[9]
Couper, M.P. (2011) Web Survey Methodology: Interface Design, Sampling and Statistical Inference. Instituto Vasco de Estadística (EUSTAT).
[10]
Bethlelem, J. (2010) Selection Bias in Web Survey. International Statistical Review, 78, 161-188.
http://dx.doi.org/10.1111/j.1751-5823.2010.00112.x
[11]
Lee, M.H. (2011) Statistical Methods for Reducing Bias in Web Surveys.
https://www.stat.sfu.ca/content/dam/sfu/stat/alumnitheses/2011/MyoungLee_
[12]
Luth, L. (2008) An Emprical Approach to Correct Self-Selection Bias of Online Panel Research. CASRO Panel Conference. https://luthresearch.com/wp-content/uploads/2015/12/Luth_CASRO_Paper_b08.pdf
[13]
Bethlehem, J. and Biffignandi, S. (2012) Handbook of Web Surveys. John Wiley & Sons, Hoboken.
[14]
Kalton, G. and Flores-Cervantes, I. (2003) Weighting Methods. Journal of Official Statistics, 19, 81-97.
[15]
Tourangeau, R., Conrad, F. and Couper, M.P. (2013) The Science of Web Surveys. Oxford University Press, Oxford.
http://dx.doi.org/10.1093/acprof:oso/9780199747047.001.0001
[16]
Lee, S. (2006) Propensity Score Adjustment as a Weghting Scheme for Volunteer Panel Web Surveys. Journal of Offical Statistics, 22, 329-349.
[17]
Steinmetz, S., Tijdens, K. and Pedrazade, P. (2009) Comparing Different Weighting Procedures for Volunteer Web Surveys. Working paper-09-76, University of Amsterdam.
[18]
Callegaro, M., Baker, R., Bethlehem, J., Göritz, A.S., Krosnick, J.A. and Lavrakas, P.L. (2014) Online Panel Research a Data Quality Perspective. John Wiley, Hoboken. http://dx.doi.org/10.1002/9781118763520
[19]
Lee, S. (2006). An Evaluation Nonresponse and Coverage Errors in a Prerecruited Probability Web Panel Survey. Social Science Computer Review, 24\, 460-475. http://dx.doi.org/10.1177/0894439306288085