Elisa Flüß 1, Christine M Bond 2, Gareth T Jones 1 and Gary J Macfarlane 1*

Similar documents
Appendix 2 protocol addendum (added 15/08/2012)

Effects of pre-notification, invitation length, questionnaire length and reminder on participation rate: a quasi-randomised controlled trial

Clinical audit for occupational therapy intervention for children with autism spectrum disorder: sampling steps and sample size calculation

Sara Rodgers, Illary Sbizzera, Sarah Cockayne, Caroline Fairhurst, Sarah E. Lamb, Wesley Vernon, Judith Watson, Catherine Hewitt, David Torgerson 1

The validity of the diagnosis of inflammatory arthritis in a large population-based primary care database

Relationship between incident types and impact on patients in drug name errors: a correlational study

The Canadian contribution to the otolaryngology literature: a five year bibliometric analysis

A new score predicting the survival of patients with spinal cord compression from myeloma

Background. Increasing Response Rates to Postal Questionnaires. Non-response. Loss of power. Example of Bias. The Roberts Study

TRIALS. Arthur D Wong 1, John Kirby 2, Gordon H Guyatt 1,3, Paul Moayyedi 1, Parag Vora 2 and John J You 1,3*

Increasing Response to Mailed Questionnaires by Including a Pencil/Pen

SPARRA Mental Disorder: Scottish Patients at Risk of Readmission and Admission (to psychiatric hospitals or units)

Dichotomizing partial compliance and increased participant burden in factorial designs: the performance of four noncompliance methods

GLOSSARY OF GENERAL TERMS

Clinical Trials: Questions and Answers

The updated incidences and mortalities of major cancers in China, 2011

1 von , 04:42

Hae Won KIM. KIM Reproductive Health (2015) 12:91 DOI /s x

Cover Page. The handle holds various files of this Leiden University dissertation

TeesRep - Teesside's Research Repository

Increasing bystander CPR: potential of a one question telecommunicator identification algorithm

What do letters to the editor publish about randomized controlled trials? A cross-sectional study

CHAMP: CHecklist for the Appraisal of Moderators and Predictors

The prognostic value of blood ph and lactate and metformin concentrations in severe metformin-associated lactic acidosis

Downloaded from:

Pollyanna Hardy 1*, Jennifer L. Bell 1, Peter Brocklehurst 2 and on behalf of The Epidural and Position Trial Collaborative Group

Strategies for Recruiting Low- Income Minority Fathers into an Adverse Childhood Experiences (ACE) Study

Citation Characteristics of Research Published in Emergency Medicine Versus Other Scientific Journals

Title: Cutaneous Adverse Drug Reactions in Indian population: A systematic review

Mode of delivery, but not questionnaire length, affected response in an epidemiological study of eating-disordered behavior

Prevalence of Marijuana Use among University Students in Bolivia, Colombia, Ecuador, and Peru

Dr Edward KoKoAung Master of Clinical Science School of Translational Science Faculty of Health Sciences The University of Adelaide South Australia.

Methodology for the VoicesDMV Survey

Aalborg Universitet. Statistical analysis plan Riis, Allan; Karran, E. L. ; Jørgensen, Anette; Holst, S.; Rolving, N. Publication date: 2017

Improving Response to the American Community Survey (ACS)

EFFECTIVENESS OF PHONE AND LIFE- STYLE COUNSELING FOR LONG TERM WEIGHT CONTROL AMONG OVERWEIGHT EMPLOYEES

A household survey on screening practices of household contacts of smear positive tuberculosis patients in Vietnam

Internet Dependency among University Entrants: A Pilot Study

A randomized comparison of three data collection models for the measurement of parent experiences with diabetes outpatient care

Strategies to increase the uptake of the influenza vaccine by healthcare workers: A summary of the evidence

Systematic reviews and meta-analyses of observational studies (MOOSE): Checklist.

Re-growth of an incomplete discoid lateral meniscus after arthroscopic partial resection in an 11 year-old boy: a case report

Author s response to reviews

Role of respondents education as a mediator and moderator in the association between childhood socio-economic status and later health and wellbeing

Eva Nohlert 1*, John Öhrvik 1,2 and Ásgeir R. Helgason 3,4

Clinical Epidemiology for the uninitiated

Author's response to reviews

Biomedical journal speed and efficiency: a cross-sectional pilot survey of author experiences

The Nottingham eprints service makes this work by researchers of the University of Nottingham available open access under the following conditions.

Title: Who does not participate in a follow-up postal study? A survey of infertile couples treated by in vitro fertilization

Hospital tests and patient related factors influencing time-to-theatre in 1000 cases of suspected appendicitis: a cohort study

Comparison of complications in one-stage bilateral total knee arthroplasty with and without drainage

Citation for published version (APA): Weert, E. V. (2007). Cancer rehabilitation: effects and mechanisms s.n.

Lars Jacobsson 1,2,3* and Jan Lexell 1,3,4

Surveillance report Published: 8 June 2017 nice.org.uk. NICE All rights reserved.

Couples Information Leaflet

Identifying best practice in actions on tobacco smoking to reduce health inequalities

Pre-operative intraocular pressure does not influence outcome of trabeculectomy surgery: a retrospective cohort study

The Impact of Advance Letters on Cellphone Response in a Statewide Dual-Frame Survey

Takashi Yagisawa 1,2*, Makiko Mieno 1,3, Norio Yoshimura 1,4, Kenji Yuzawa 1,5 and Shiro Takahara 1,6

Cover Page. The handle holds various files of this Leiden University dissertation

Research Article Opioid Use Is Not Associated with Incomplete Wireless Capsule Endoscopy for Inpatient or Outpatient Procedures

NHS public health functions agreement Service specification No.11 Human papillomavirus (HPV) programme

PROSPERO International prospective register of systematic reviews

Following a Telephone Survey with a Mail Survey

Outcomes Monitoring System Iowa Project

Bereavement round up: Support after suicide bereavement

The EuroQol and Medical Outcome Survey 36-item shortform

Electronic Cigarette Use and Respiratory Symptoms in Chinese Adolescents in Hong Kong. Title. Wang, MP; Ho, DSY; Leung, LT; Lam, TH

A cluster RCT of enhanced pharmacy services (EPS) to improve outcomes for patients on methadone maintenance therapy (MMT)

CONSORT 2010 checklist of information to include when reporting a randomised trial*

Abdominal Aortic Aneurysm (AAA) Screening. Date: 7 March 2017 Version: 1.0

Contents. Version 1.0: 01/02/2010 Protocol# ISRCTN Page 1 of 7

Catriona Matheson, Christine M. Bond, Michela Tinelli. Introduction

Dehiscence of detached internal limiting membrane in eyes with myopic traction maculopathy with spontaneous resolution

Effectiveness of the surgical torque limiter: a model comparing drill- and hand-based screw insertion into locking plates

Implementing Lung Cancer Screening in Federally Qualified Health Centers: Concerns about Resources and Quality

Risk of selection bias in randomised trials

Changes in paracetamol, antidepressants and opioid poisoning in Scotland during the 1990s

Survey Methods in Relationship Research

GATE CAT Intervention RCT/Cohort Studies

Cochrane Pregnancy and Childbirth Group Methodological Guidelines

CASE STUDY: Measles Mumps & Rubella vaccination. Health Equity Audit

John Evans 3/4 October 2013 Workshop on Methods for Research Synthesis Cambridge, MA

Saudi dental students perceptions of pediatric behavior guidance techniques

Mammographic density and risk of breast cancer by tumor characteristics: a casecontrol

Research Article Self-Monitoring of Blood Pressure in Hypertension: AUKPrimaryCareSurvey

Diagnostic aid to rule out pneumonia in adults with cough and feeling of fever. A validation study in the primary care setting

Within-Household Selection in Mail Surveys: Explicit Questions Are Better Than Cover Letter Instructions

Effectiveness of inhaled furosemide for acute asthma exacerbation: a meta-analysis

Replications of Forecasting Research

Cost-effectiveness of brief intervention and referral for smoking cessation

Madison Bartenders Baseline Survey

Critical Appraisal Series

Women s use of complementary and alternative medicine for the treatment of menopause-related symptoms: A health services research study.

Evolving patterns of tobacco use in northern Sweden

Wellness Coaching for People with Prediabetes

Vinyl Chloride Exposure Study of the Community and Local Emergency Responders

Excerpt from Interim Report Health of Houston Survey 2010

Predictors of seasonal influenza vaccination in chronic asthma

Transcription:

Flüß et al. BMC Medical Research Methodology 2014, 14:104 RESEARCH ARTICLE Open Access The effect of an internet option and single-sided printing format to increase the response rate to a population-based study: a randomized controlled trial Elisa Flüß 1, Christine M Bond 2, Gareth T Jones 1 and Gary J Macfarlane 1* Abstract Background: Paper questionnaires are a common means to collect self-reported information in population-based epidemiological studies. Over the past decades, the response rates to epidemiological studies have been decreasing which can affect the selection process of eligible subjects and lead to non-response bias. Hence, research into strategies to increase questionnaire response rates is crucial. The aim of this study was therefore to explore the effectiveness of single-sided questionnaires and an internet option for response in increasing response rates to a population-based study. Methods: A 2 2 factorial experiment was embedded within a large population-based study of pain and pain management. Persons in the study sample were 4600 residents in Grampian (north of Scotland) aged 25 years and over who were randomly selected from health board records. Sampled persons were randomly assigned to either receive a single-sided or double-sided questionnaire with or without an internet option to respond. The study questionnaire was distributed via post. Results: The overall study response rate was 36.3%. When compared to the reference group that received no intervention (response rate = 35.5%), the response rate changed only marginally when single-sided questionnaires were distributed (35.8%) or when an option to reply via the internet was provided (34.3%). A somewhat higher increase in response rates was achieved when both strategies were employed (39.6%). Overall, no significant effect on response rate was determined for each strategy or their interaction. Conclusions: Evidence from this study suggests that neither single-sided questionnaires nor the option to reply via the internet resulted in a significant increase in response rates to population-based studies. Keywords: Response rate, Randomized controlled trial, Postal questionnaires, Health surveys, Data collection, Internet Background Questionnaires to collect self-reported information are one of the key elements in population-based studies. A common means to conduct these studies is the use of postally distributed questionnaires which are a convenient way to reach a large number of people in a short period of time [1]. While population studies are less prone to * Correspondence: g.j.macfarlane@abdn.ac.uk 1 Epidemiology Group, University of Aberdeen, Polwarth Building, Foresterhill, Aberdeen AB25 2ZD, UK Full list of author information is available at the end of the article selection bias, low study response rates and selective participation can result in non-response bias which can threaten a study s validity [2]. It has been widely observed that response rates to epidemiological studies have been decreasing over the past decades [3,4]. A retrospective review to investigate the change in study participation over time, reported that between 1970 and 2003 the response rates to cohort and cross-sectional studies decreased by 0.54% (95 CI: 1.33, 0.24) and 0.67% (95% CI: 1.91, 0.56) per year, respectively [4]. According to current predictions, response rates 2014 Flüß et al.; licensee BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly credited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

Flüß et al. BMC Medical Research Methodology 2014, 14:104 Page 2 of 6 are likely to decline further in the future [3]. While statistical methods, like weighting or imputation, are useful to substitute missing responses, it remains questionable whether they are applicable to studies in which respondents only represent a small fraction of the sample population. It is therefore crucial to explore methods that improve questionnaire response rates to population studies. Much research has already been done to explore methods to improve response rates to postal questionnaires [5,6]. To date, the most effective strategies are known to be incentives, the use of special or certified delivery services or shorter questionnaires [5]. Pre-notification of receiving a questionnaire, reminder mailings and personalization methods can also increase questionnaire response rates significantly. However there are limitations to the existing literature on response rates. It is not clear to what extent methods that have proven to be effective for specific groups, for example, physicians or students are effective for general populations. Secondly, evidence collected previously for approaches based on newer technologies such as the internet may no longer be valid due to the wider use of those technologies within society. Finally the cost-effectiveness of different methods should also be assessed to allow researchers to choose the most appropriate overall approach. A method that can easily be applied in populationbased studies is to change the printing format of the questionnaire. As yet, four randomized controlled trials (RCTs) have examined the effect on response rates when questionnaires were printed single-sided versus double-sided [5]. Their findings suggest that the distribution of single-sided questionnaires leads to a small but significant effect on response rates (OR = 1.22; 95% CI: 1.01, 1.47). This may seem counter-intuitive since shorter questionnaires have shown to increase response rates. However, it may be speculated that questionnaire respondents felt more encouraged to complete the instrument after noticing that only the first side of the questionnaire pages was printed. They may have believed that its completion would not take up too much of their time after all, hence filled it in and returned it. None of the four trials were conducted within a general population sample. It would therefore be interesting to test this strategy in a population-based study. The provision of an internet option to respond to a paper questionnaire may be another potential method to increase the response rate to population-based studies if prospective respondents prefer one administration mode over the other. The effect on response rates when participants received an option to reply via the internet was summarized in a recent meta-analysis [7]. Only four out of twelve peer-reviewed studies were RCTs that were conducted within general population samples [8-11]. The pooled Odds Ratio suggests that an additional web option does not improve response rates to populationbased studies (OR = 0.90; 0.73, 1.11). Two out of the four trials were carried out more than five years ago. In the context of rapidly growing internet penetration rates (in Scotland: 62.7% in 2007 and 77.4% in 2012; [12]), it is possible that an additional response option could become more acceptable and in this way increase questionnaire response rates. The aims of the current study were to explore the effectiveness of printing the questionnaires single-sided and providing an internet option to respond and to assess the cost-effectiveness of both methods in a population-based study. It was hypothesized that the distribution of single-sided questionnaires and the provisionofaninternetoptionhaveanindependent effect on questionnaire response rates, while no interaction effect between methods was expected. A secondary aim was to use additional data collected in this study to update an earlier meta-analysis of the evidence on the effect of an internet option in population studies [7]. Methods Study design A 2 2 factorial experiment was embedded within a cross-sectional population-based study of pain and pain management in 2012/2013. Using an electronic randomization program, persons in the study sample were randomly allocated to either receive (1) a singlesided or double-sided questionnaire and (2) an option or no option to reply via the internet. Study procedure A random sample of 4600 residents in Grampian (north of Scotland, UK) aged 25 years and over was selected from health board records. Upon randomization, selected persons were sent a notification letter that they had been selected for the study. One week later they were sent a survey pack comprising an invitation letter, an information sheet, the questionnaire and a pre-paid reply envelope. Potential participants were advised to read the information sheet and to complete and return the questionnaire if they wished to take part. The invitation letter for those in the web option groups contained, in addition, the URL link to the electronic questionnaire and their individual ID number and password for its access. Non-respondents were sent a second survey pack appropriate to their randomization group, three weeks after the first contact (i.e. two weeks after the questionnaire distribution) (Figure 1).

Flüß et al. BMC Medical Research Methodology 2014, 14:104 Page 3 of 6 Figure 1 Flowchart of the study procedure. Study questionnaire The 20-page study questionnaire included questions on participants demographic characteristics (gender, age, and educational background), their health, pain and pain management. Validated standard questionnaires were used to collect information on participants health (SF-36 [13]) and pain status (set of pain questions with manikins and Chronic Pain Grade [14]). The questionnaire to collect pain management information was validated within a small population sample before the conduct of the study (Results to be published). Depending on their randomization, persons in the study sample either received a 10-sheet (double-sided) or a 20-sheet (single-sided) questionnaire. Statistical analysis The main outcome of interest was questionnaire response rate; defined as the percentage of completed and partially completed questionnaires returned after excluding from the denominator those which were invalid (e.g. change of address or death). With questionnaire response as the dependent variable, logistic regression models were performed in order to determine predictors for questionnaire response. Odds Ratios and 95% Confidence Intervals were calculated in order to quantify the effect of the strategies on the questionnaire response rate. A meta-analysis was performed to summarize the current evidence on the effectiveness of an internet option in increasing response rates to population-based studies. Additionally, the cost-effectiveness of single-sided questionnaires and an internet option was investigated by determining the cost per additional response. Ethical approval The study was approved by the North of Scotland National Health Service Research Ethics Committee (REC reference: 12/NS/0079). Results Response rate The response rate before the reminder surveys were distributed to non-respondents was 21.4%. At the end of data collection, 183 questionnaires were found invalid (for example, due to an invited participant being deceased or no longer resident at the given address) and 1604 completed questionnaires were returned giving an overall study response rate of 36.3% (1604/4417). Out of the respondents who were offered an internet option, 60 completed the electronic questionnaire (7.3%). Time to respond Completed study questionnaires were returned a median of 10 days (Interquartile range (IQR): 7, 20) after their distribution. Electronic questionnaires were received earlier than paper questionnaires (median, web: 7 days; median, paper: 10 days; Mann Whitney test: p < 0.001). Respondents demographic characteristics Females represented 55.3% of respondents. The median age was 55 years (IQR: 44, 65) with the oldest respondent aged 94 years. The majority of respondents reported secondary school (29.6%), a vocational qualification (20.0%) or a professional qualification (16.7%) as their highest level of education. When compared to nonrespondents, respondents were significantly more likely to be female and of older age (Chi-square test: p < 0.001,

Flüß et al. BMC Medical Research Methodology 2014, 14:104 Page 4 of 6 both tests). There was a significant interaction between gender and age, showing that the positive effect of female gender on response reduces with older age. Electronic respondents(n=60)weresignificantlymorelikelytobe younger and more educated when compared to paper respondents (Chi-square test: p < 0.001, both tests). Response rate by study group There were small differences in response rates between the four study groups. The reference group achieved a response rate of 35.5%. When sampled persons were provided with an internet option, this dropped to 34.3%. In contrast, the response rate increased to 35.8% when participants were sent a single-sided questionnaire with no option to respond via the internet. A greater increase in response rate of 4.1% was determined when participants received both interventions (39.6%). The effectiveness of the strategies in increasing response rates From the logistic model there was no main effect on response rates when an internet option was provided (OR = 0.95, 95% CI = 0.80, 1.13) or when single-sided questionnaires were distributed (OR = 1.01, 95% CI = 0.85, 1.20). Moreover, the interaction between both strategies led to a larger, albeit not statistically significant, effect (OR = 1.24, 95% CI = 0.97, 1.59, Table 1). When comparing the response rates of those who were randomized to receive an internet option and those who were not, the strategy of providing an internet option increased the response rate non-significantly by 1.3% (OR = 1.06; 95% CI: 0.94, 1.20). A somewhat larger, yet non-significant effect was determined when the response rates of those who were allocated to receive a single-sided questionnaire and those who were allocated to be sent a double-sided questionnaire were compared: the response rate increased by 2.8% when participants received a single-sided questionnaire (OR = 1.13; 95% CI: 0.998, 1.28). Internet option summary of evidence When the results of the current study were added to those reported in previous population studies (Figure 2), there no effect on response rates when an internet option versus no internet option was provided for participants (OR = 0.95; 0.83, 1.10). There was no obvious Table 1 The effect of the strategies on response rates Logistic Regression, Model term Crude OR (95% CI) No intervention 1.00 Web (main effect) 0.95 (0.80, 1.13) Single-sided (main effect) 1.01 (0.85, 1.20) Web + Single-sided (interaction term) 1.24 (0.97, 1.59) pattern of an improvement in the effectiveness of an internet option over time. There was significant heterogeneity between studies. Cost-effectiveness The employment of the strategies to increase response rates was associated with additional cost. The use of single-sided questionnaires led to increased cost for printing (since the required number of pages doubled) and postage (because the survey packs were heavier). The implementation of an electronic version of the study questionnaire resulted in extra costs due to programming time and set up cost. Due to the low uptake rate of the internet option, differences in data entry costs were minimal. It was therefore decided not to include the data entry costs in the cost-effectiveness analysis. As outlined in Table 2, the cost per response was the lowest in the reference group that did not receive either intervention ( 6.64). An additional cost of at least 1.02 per response was spent for the employment of the interventions. The internet option resulted in lower response at greater cost when double-sided questionnaires were distributed. The use of single-sided rather than doublesided questionnaires cost 119.19 per additional response. The combination of both methods was more cost-effective: Distributing single-sided questionnaire with an option to reply via the internet cost 25.20 per extra response. Discussion Main findings Overall, neither the use of single-sided questionnaires nor the provision to reply via the internet significantly improved the questionnaire response rate in the current study. A small but non-significant interaction effect was determined when both strategies were used in combination. Thecurrentstudywasthefirststudythatexploredtheeffectiveness of single-sided questionnaires in a populationbased sample. We found no effect on response rates when questionnaires were printed single-sided (OR = 1.13; 95% CI: 0.998, 1.28). Likewise, no effect on response rates was determined when an internet option was provided for the sampled persons (OR = 1.06; 95% CI: 0.94, 1.20). An updated meta-analysis confirmed that based on the current evidence of five studies, an option to reply via the internet does not improve response rates to a population study (OR = 0.95; 0.83, 1.10). Strengths and limitations The current study had several strengths. Firstly, we used a randomized controlled study design. Secondly, the use of a two by two factorial design enabled us to test the independentandthecombinedeffectoftwodifferentmethodstoincrease questionnaire response rates. Thirdly, we conducted a subsequent cost-effectiveness analysis which provided

Flüß et al. BMC Medical Research Methodology 2014, 14:104 Page 5 of 6 Figure 2 Comparison of an internet option vs. no internet option to increase response to population studies. insight into the expenses that were associated with the respective strategies under investigation. A limitation of the study was that it was a sub-study within a large cross-sectional study. The study sample size was therefore powered on the main outcome variables and not on the difference in questionnaire response rates. Nevertheless, the main effects of both strategies on questionnaire response rates were very small and insufficient power to detect a meaningful difference is unlikely to have been a reason for observed lack of effect. It is probable however, that there was a lack of power to detect a significant interaction effect between both studies. Secondly, due to a small number of electronic responses the interpretation of some results was limited. It was determined that electronic respondents were significantly younger and more highly educated when compared to paper respondents. While a higher proportion of electronic respondents were male and living in an urban area (information provided by the sampling frame), these differences were not statistically significant most probably due to a lack of power. Findings in relation to previous studies Out of the 824 respondents who were randomized to the internet option groups, only 60 (7.3%) used the electronic version to respond to the questionnaire. The low uptake rate of the internet option was similar to that observed in the population study by Ziegenfuss et al. [9], in which only 8.0% of the web option respondents used the electronic questionnaire. While a lack of internet coverage was not a concern in the current study, it is likely that the initial mode of contact caused the low internet uptake rate. Since all prospective respondents were sent a paper questionnaire, they may have been more inclined to complete and return the paper version. A couple of hypotheses have been discussed in the literature to explain why an internet option is ineffective in increasing response rates [7]. Firstly, when approached by mail, the completion of an electronic questionnaire requires additional effort for participants since they have to switch task mode. It can be hypothesized that a number of people mayinitiallyhavedecidedtousetheelectronicquestionnaire but never actually logged on to complete and submit. Secondly, long and complicated URLs (e.g. those containing numbers and case sensitive letters) may make it more difficult for participants to access the online questionnaire and appear off-putting. As a result, fewer participants may consider completing the online version. Nevertheless, the URL in the current study was reasonably short and only contained lowercase letters). Thirdly, people may be less likely torespondwhenprovidedwithaninternetoptionbecause they are required to make a choice at the very beginning. According to Schwartz [15], offering choices has negative effects on people s decision-making. Since every choice is associated with opportunity costs, one must consider the trade-offs concerned with each option they are offered. This,inturn,makesitlessappealingtochooseeitherone. It remains uncertain why the use of single-sided questionnaires was ineffective in the current study. The questionnaire in the current study consisted of 20 pages and appeared very bulky when printed single-sided. It can be assumed that people might have felt discouraged when they received the questionnaire. Surprisingly, we established that although not statistically significant, there was an effect on questionnaire response rate that laid in the interaction of both strategies (OR = 1.24, 95% CI = 0.97, 1.59). It remains questionable why the response rates were only increased Table 2 Strategies to improve questionnaire response: Cost-effectiveness No internet option & Double-sided Internet option & Double-sided No internet option & Single-sided Internet option & Single-sided Response rate (in %) 35.5 34.3 35.8 39.4 Nr. of respondents 393 382 387 442 Nr. of eligible participants 1106 1082 1113 1116 Cost per response ( ) 6.64 8.59 7.67 8.36 Cost per additional response ( ) 119.19 25.20

Flüß et al. BMC Medical Research Methodology 2014, 14:104 Page 6 of 6 when sampled persons received both interventions. It can be speculated that fewer people considered using the internet option when they were sent a single-sided questionnaire. People may have believed that its completion would take up less of their time using the printed version and hence, were more likely to complete and return it straight away. We added our findings to the results of previous population studies that were summarized in a recent meta-analysis by Medway and Fulton [7]. In order to account for any studies that were published after 2011, we conducted a comprehensive literature search. However, no additional papers were identified. Conclusions Taken as a whole, neither an option to reply via the internet nor the use of single-sided questionnaires were effective methods to significantly increase the response rates to a population-based sample. The results derived from the web option comparison were in agreement with the pooled effect size of the current evidence. Furthermore, it has been demonstrated that the cost per additional response was high, even when both methods were combined. As result of the current findings, researchers should be aware that neither method is effective in increasing response rates to future population studies. Besides, researchers should be alerted to select the methods to increase questionnaire response rates in consideration of their effectiveness and the costs concerned. Due to the current trend of decreasing response rates, further research into strategies to increase those remains important. Future studies should always provide sufficient information on the cost-effectiveness of each employed strategy in order to allow an appropriate assessment of these. Received: 21 May 2014 Accepted: 12 August 2014 Published: 9 September 2014 References 1. Silman AJ, Macfarlane GJ: Epidemiological studies. A practical Guide. 2nd edition. Cambridge: Cambridge University Press; 2005. 2. Galea S, Tracy M: Participation rates in epidemiologic studies. Ann Epidemiol 2007, 17(9):643 653. 3. Galea S: Participation rates in epidemiologic studies. Ann Epidemiol 2007, 17(9):643. 4. Morton LM: Reporting participation in epidemiologic studies: a survey of practice. Am J Epidemiol 2006, 163(3):197. 5. Edwards PJ, Roberts I, Clarke MJ, Diguiseppi C, Wentz R, Kwan I, Cooper R, Felix LM, Pratap S: Methods to increase response to postal and electronic questionnaires. Cochrane Database Syst Rev 2009, 3. MR000008. 6. Nakash R: Maximising response to postal questionnaires a systematic review of randomised trials in health research. BMC Med Res Methodol 2006, 6(1):5. 7. Medway RL, Fulton J: When more gets you less: a meta-analysis of the effect of concurrent web options on mail survey response rates. Public Opin Q 2012, 76(4):733 746. 8. Brogger J, Nystad W, Cappelen I, Bakke P: No increase in response rate by adding a web response option to a postal population survey: a randomized trial. J Med Internet Res 2007, 9(5):e40. 9. Ziegenfuss JY, Beebe TJ, Rey E, Schleck C, Locke GR, Talley NJ: Internet option in a mail survey: More harm than good? Epidemiology 2010, 21(4):585 586. 10. Schneider SJ, Cantor D, Malakhoff L, Arieira C, Segel P, Nguyen K, Tancreto JG: Telephone, internet, and paper data collection modes for the Census 2000 short form. J Official Stat 2005, 21(1):89 101. 11. Smyth JD, Dillman DA, Christian LM, O'Neill AC: Using the internet to survey small towns and communities: limitations and possibilities in the early 21st century. Am Behav Sci 2010, 53(9):1423 1448. 12. Widen use of the internet; http://www.scotland.gov.uk/about/performance/ scotperforms/indicator/internet. 13. Ware JE Jr, Sherbourne CD: The MOS 36-item short-form health survey (SF-36): I. Conceptual framework and item selection. Med Care 1992, 30(6):473 483. 14. Von Korff M, Dworkin SF, Le Resche L: Graded chronic pain status: an epidemiologic evaluation. Pain 1990, 40(3):279 291. 15. Schwartz B: The paradox of choice. New York: Harper Perennial; 2004. doi:10.1186/1471-2288-14-104 Cite this article as: Flüß et al.: The effect of an internet option and single-sided printing format to increase the response rate to a population-based study: a randomized controlled trial. BMC Medical Research Methodology 2014 14:104. Competing interests The authors declare that they have no competing interests. Authors contributions This study was conceived by GJM, GTJ and CMB. It was planned by all the authors and EF conducted the study. EF drafted the manuscript and all other authors critically reviewed the manuscript and provided important intellectual content. All authors read and approved the final manuscript. Acknowledgements We would like to thank the Institute of Applied Health Sciences (IAHS) at the University of Aberdeen for funding the PhD studentship of EF. Furthermore, we would like to thank everyone who was involved in the study, including Professor Sir Lewis Ritchie (Director of Public Health, NHS Grampian), John Lemon (University of Aberdeen), Dr. Fiona Garton (University of Aberdeen) and the Aberdeen Service User Group. Lastly, we would like to acknowledge all data entry clerks (Maxx Livingstone, Rory Macfarlane, Georgia Mannion-Krase and Hazel Reilly) and participants of the study. Author details 1 Epidemiology Group, University of Aberdeen, Polwarth Building, Foresterhill, Aberdeen AB25 2ZD, UK. 2 Centre for Academic Primary Care, University of Aberdeen, Polwarth Building, Foresterhill, Aberdeen AB25 2ZD, UK. Submit your next manuscript to BioMed Central and take full advantage of: Convenient online submission Thorough peer review No space constraints or color figure charges Immediate publication on acceptance Inclusion in PubMed, CAS, Scopus and Google Scholar Research which is freely available for redistribution Submit your manuscript at www.biomedcentral.com/submit