Jump to content

Response rate (survey)

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by Josvebot (talk | contribs) at 22:23, 28 May 2018 (v1.43b - WP:WCW project (Unicode control characters)). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

A U.S. National Agricultural Statistics Service statistician explains response rate data at a 2017 briefing to clarify the context of crop production data.

In survey research, response rate, also known as completion rate or return rate, is the number of people who answered the survey divided by the number of people in the sample. It is usually expressed in the form of a percentage. The term is also used in direct marketing to refer to the number of people who responded to an offer.

The general consensus in academic surveys is to choose one of the six definitions summarized by the American Association for Public Opinion Research (AAPOR).[1] These definitions are endorsed by the National Research Council and the Journal of the American Medical Association, among other well recognized institutions.[citation needed] They are:

  1. Response Rate 1 (RR1) – or the minimum response rate, is the number of complete interviews divided by the number of interviews (complete plus partial) plus the number of non-interviews (refusal and break-off plus non-contacts plus others) plus all cases of unknown eligibility (unknown if housing unit, plus unknown, other).
  2. Response Rate 2 (RR2) – RR1 + counting partial interviews as respondents.
  3. Response Rate 3 (RR3) – estimates what proportion of cases of unknown eligibility is actually eligible. Those respondents estimated to be ineligible are excluded from the denominator. The method of estimation *must* be explicitly stated with RR3.
  4. Response Rate 4 (RR4) – allocates cases of unknown eligibility as in RR3, but also includes partial interviews as respondents as in RR2.
  5. Response Rate 5 (RR5) – is either a special case of RR3 in that it assumes that there are no eligible cases among the cases of unknown eligibility or the rare case in which there are no cases of unknown eligibility. RR5 is only appropriate when it is valid to assume that none of the unknown cases are eligible ones, or when there are no unknown cases.
  6. Response Rate 6 (RR6) – makes that same assumption as RR5 and also includes partial interviews as respondents. RR6 represents the maximum response rate.

The six AAPOR definitions vary with respect to whether or not the surveys are partially or entirely completed and how researchers deal with unknown nonrespondents. Definition #1, for example, does NOT include partially completed surveys in the numerator, while definition #2 does. Definitions 3–6 deal with the unknown eligibility of potential respondents who could not be contacted. For example, there is no answer at the doors of 10 houses you attempted to survey. Maybe 5 of those you already know house people who qualify for your survey based on neighbors telling you whom lived there, but the other 5 are completely unknown. Maybe the dwellers fit your target population, maybe they don't. This may or may not be considered in your response rate, depending on which definition you use.

Example: if 1,000 surveys were sent by mail, and 257 were successfully completed (entirely) and returned, then the response rate would be 25.7%.

Importance

A survey’s response rate is the result of dividing the number of people who were interviewed by the total number of people in the sample who were eligible to participate and should have been interviewed.[2] A low response rate can give rise to sampling bias if the nonresponse is unequal among the participants regarding exposure and/or outcome. Such bias is known as nonresponse bias.

For many years, a survey's response rate was viewed as an important indicator of survey quality. Many observers presumed that higher response rates assure more accurate survey results (Aday 1996; Babbie 1990; Backstrom and Hursh 1963; Rea and Parker 1997). But because measuring the relation between nonresponse and the accuracy of a survey statistic is complex and expensive, few rigorously designed studies provided empirical evidence to document the consequences of lower response rates until recently.

Such studies have finally been conducted in recent years, and several conclude that the expense of increasing the response rate frequently is not justified given the difference in survey accuracy.

One early example of a finding was reported by Visser, Krosnick, Marquette and Curtin (1996) who showed that surveys with lower response rates (near 20%) yielded more accurate measurements than did surveys with higher response rates (near 60 or 70%).[3] In another study, Keeter et al. (2006) compared results of a 5-day survey employing the Pew Research Center’s usual methodology (with a 25% response rate) with results from a more rigorous survey conducted over a much longer field period and achieving a higher response rate of 50%. In 77 out of 84 comparisons, the two surveys yielded results that were statistically indistinguishable. Among the items that manifested significant differences across the two surveys, the differences in proportions of people giving a particular answer ranged from 4 percentage points to 8 percentage points.[4]

A study by Curtin et al. (2000) tested the effect of lower response rates on estimates of the Index of Consumer Sentiment (ICS). They assessed the impact of excluding respondents who initially refused to cooperate (which reduces the response rate 5–10 percentage points), respondents who required more than five calls to complete the interview (reducing the response rate about 25 percentage points), and those who required more than two calls (a reduction of about 50 percentage points). They found no effect of excluding these respondent groups on estimates of the ICS using monthly samples of hundreds of respondents. For yearly estimates, based on thousands of respondents, the exclusion of people who required more calls (though not of initial refusers) had a very small one.[5]

Holbrook et al. (2005) assessed whether lower response rates are associated with less unweighted demographic representativeness of a sample. By examining the results of 81 national surveys with response rates varying from 5 percent to 54 percent, they found that surveys with much lower response rates decreased demographic representativeness within the range examined, but not by much.[6]

Choung et al. (2013) looked at community response rate to a mailed questionnaire about functional gastrointestinal disorders. The response rate to their community survey was 52%. Then, they took a random sample of 428 responders and 295 nonresponders for medical record abstraction, and compared nonresponders against responders. They found that respondents had a significantly higher body mass index and more health care seeking behavior for non-GI problems. However, except for diverticulosis and skin diseases, there was no significant difference between responders and nonresponders in terms of any gastrointestinal symptoms or specific medical diagnosis.[7]

Dvir and Gafni (2018) examined if consumer response rate is influenced by the amount of information provided. In a series of large-scale web experiments (n= 535 and n= 27,900), they compared variants of marketing web pages (also called Landing page), focusing on how changed to content amount impact users’ willingness to provide their e-mail address (a behavior called Conversion rate in marketing terms). The results showed significant higher response rate on the shorter pages, which indicates that contrary to earlier work, not all response rate theories are effective online [8].

Nevertheless, in spite of these recent research studies, a higher response rate is preferable because the missing data is not random.[9] There is no satisfactory statistical solution to deal with missing data that may not be random. Assuming an extreme bias in the responders is one suggested method of dealing with low survey response rates. A high response rate (>80%) from a small, random sample is preferable to a low response rate from a large sample.[10]

See also

References

  1. ^ Standard Definitiopns – AAPOR. AAPOR http://www.aapor.org/Standards-Ethics/Standard-Definitions-%281%29.aspx. Retrieved 3 March 2016. {{cite web}}: Missing or empty |title= (help)
  2. ^ "Response Rates – An Overview." American Association for Public Opinion Research (AAPOR). 29 Sept 2008. http://www.aapor.org/Education-Resources/For-Researchers/Poll-Survey-FAQ/Response-Rates-An-Overview.aspx
  3. ^ Visser, Penny S.; Krosnick, Jon A.; Marquette, Jesse; Curtin, Michael (1996). "Mail Surveys for Election Forecasting? An Evaluation of the Colombia Dispatch Poll". Public Opinion Quarterly. 60: 181–227. doi:10.1086/297748.
  4. ^ Keeter, Scott, Courtney Kennedy, Michael Dimock, Jonathan Best and Peyton Craighill. 2006. "Gauging the Impact of Growing Nonresponse on Estimates from a National RDD Telephone Survey." Public Opinion Quarterly. 70(5): 759–779.
  5. ^ Curtin, Richard; Presser, Stanley; Singer, Eleanor (2000). "The Effects of Response Rate Changes on the Index of Consumer Sentiment". Public Opinion Quarterly. 64 (4): 413–428. doi:10.1086/318638.
  6. ^ Holbrook, Allyson, Jon Krosnick, and Alison Pfent. 2007. "The Causes and Consequences of Response Rates in Surveys by the News Media and Government Contractor Survey Research Firms." In Advances in telephone survey methodology, ed. James M. Lepkowski, N. Clyde Tucker, J. Michael Brick, Edith D. De Leeuw, Lilli Japec, Paul J. Lavrakas, Michael W. Link, and Roberta L. Sangster. New York: Wiley. https://pprg.stanford.edu/wp-content/uploads/2007-TSMII-chapter-proof.pdf
  7. ^ Seon Choung, Rok; Richard Locke, III; Schleck, Cathy D.; Ziegenfuss, Jeanette Y.; Beebe, Timothy J.; Zinsmeister, Alan R.; Talley, Nicholas J. (2013). "A low response rate does not necessarily indicate non-response bias in gastroenterology survey research: a population-based study". Journal of Public Health. 21 (1): 87–95. doi:10.1007/s10389-012-0513-z.
  8. ^ Dvir, Nim; Gafni, Ruti (2018). "When Less Is More: Empirical Study of the Relation Between Consumer Behavior and Information Provision on Commercial Landing Pages". Informing Science: The International Journal of an Emerging Transdiscipline. 21: 019–039. doi:10.28945/4015. ISSN 1547-9684.
  9. ^ Altman, DG; Bland, JM (Feb 2007). "Missing data". BMJ. 334 (7590): 424. doi:10.1136/bmj.38977.682025.2C. PMC 1804157. PMID 17322261.
  10. ^ Evans, SJ (Feb 1991). "Good surveys guide" (PDF). BMJ. 302 (6772): 302–3. doi:10.1136/bmj.302.6772.302. PMC 1669002. PMID 2001503.