38
SAMPLING ISSUES: PART I LEONIE HUDDY, STONY BROOK UNIVERSITY [email protected] MATTHEW BAUM HARVARD UNIVERSITY

Sampling Issues: part I

  • Upload
    izzy

  • View
    34

  • Download
    1

Embed Size (px)

DESCRIPTION

Sampling Issues: part I. Leonie huddy , Stony brook university [email protected] matthew baum harvard university. outline. Major Sources of Survey Error II. Coverage Error 1. Coverage Problems in US (Phone & Web) 2. Coverage Issues in Sweden 3. Implications for Experiments - PowerPoint PPT Presentation

Citation preview

Page 1: Sampling Issues: part I

SAMPLING ISSUES:PART ILEONIE HUDDY, STONY BROOK [email protected]

MATTHEW BAUM HARVARD UNIVERSITY

Page 2: Sampling Issues: part I

OUTLINEI. Major Sources of Survey ErrorII. Coverage Error

1. Coverage Problems in US (Phone & Web)

2. Coverage Issues in Sweden

3. Implications for Experiments

III. Non-Response Error1. Rates in the US2. Factors Influencing Response Rates3. Response rates in Sweden

IV. Survey Mode Errors

Page 3: Sampling Issues: part I

I. MAJOR SOURCES OF SURVEY ERROR (ALWIN/GROVES)1. Coverage error: Error due to failure to include some elements of the population in the sampling frame (e.g, cell phones in RDD landline study in the US, non-computer households in a web survey)2. Sampling error: Errors due to sampling a subset rather than the entire population. 3. Non-response error: Error due to failure to obtain data from all selected population elements (young males harder to reach; Latinos reluctant) 4. Measurement error: Error that occurs when observed value is different from the true value (higher reports of voter turnout in ANES)These errors also apply to survey experiments

Page 4: Sampling Issues: part I

II. COVERAGE ERROR (GROVES)DEFINITIONSSampling frame: set of lists or procedures intended to identify all elements of the target population; e.g., RDD, national registry (SPAR), US Postal Mail Delivery SystemCoverage: Undercoverage - some population elements are missing from the sample frame (e.g., cell phone users who are disproportionately young and less affluent in an RDD landline study; older respondents who lack computers or broadband in a web survey)Ineligible units (non-working phone #s)Clustering of elements at a single frame element (several people with 1 phone number)Duplication: single target element linked to multiple frame units (a person listed more than once in a national registry)

Page 5: Sampling Issues: part I

2. NON-COVERAGE PROBLEMS IN TELEPHONE SAMPLESNON-LANDLINE HOUSEHOLDS• younger, more mobile, less affluent• more ethnic and racial minorities • live in rural areas, south, central cities• Reaching 25% of the US population Solution?• Base sample on a mix of cell phone-only and

landline households and eliminate those with landlines from the cell phone sample

• Post-stratification weights based on demographic factors

Page 6: Sampling Issues: part I

THE GROWING CELL-ONLY POPULATION BY AGE (PEW 2010)

Page 7: Sampling Issues: part I

AGE COMPOSITION OF LANDLINE PHONE SAMPLES (PEW, 2010)

Page 8: Sampling Issues: part I

3. Non-Coverage on Web: % Broadband at Home (Pew)

2005 2006 2007 2009All adult Americans 30% 42% 47% 60%GenderMale 31 45 50 61Female 27 38 44 58Age18-29 38 55 63 7630-49 36 50 59 6750-64 27 38 40 5665+ 8 13 15 26Race/EthnicityWhite (not Hispanic) 31 42 48 63Black (not Hispanic) 14 31 40 52EducationLess than high school 10 17 21 24High school grad 20 31 34 46Some college 35 47 58 73College + 47 62 70 83IncomeUnder $30K 15 21 30 42$30K-50K 27 43 46 62$50K-$75K 35 48 58 73Over $75K 57 68 76 83

Page 9: Sampling Issues: part I

4. NON-COVERAGE ISSUES IN SWEDEN Telephone & Mail• Sweden has a low-rate of cell-only households; ≤ 5% (Hecke & Weise 2012; in

Telephone telephone surveys in Europe, ed. Häider, Häider, & Kϋhne; Springer, Heidelberg ) Non-coveraqe is far less of a problem in Sweden because samples are drawn from the national SPAR registry

• No sample frame is ever 100% so there may still be minor non-coverage issues

Internet • “Sweden has a unique position in the world when it comes to Internet use, not

only because it is one of the countries with highest share of Internet users in the world but also because Internet use is more widely spread in Swedish society compared to other countries, in terms of age and educational level (Findahl 2007; 2008b). Among younger Swedes 16–25 years old almost all (97%) use the Internet at least once a month; among older Swedes 56–65 years old Internet use is currently as high as 75%. The corresponding figure among individuals 66–75 years old is lower, however, at 51% (Findahl 2008a).” (quoted in Kallmen et al )

Page 10: Sampling Issues: part I

SPAR – NATIONAL SWEDISH POPULATION REGISTRYStatens personadressregister, SPAR includes all persons who are registered as resident in Sweden. • The data in SPAR is updated each day with data from the Swedish

Population Register.• SPAR is specifically regulated in Swedish Law by the Act of (1998:527)

statens personadressregister and by the Regulation (1998:1234) of statens personadressregister and the Swedish Tax Agency Regulation on handing out data from SPAR (SKVFS 2011:06).

The aim of SPAR is clear from the purposes set out in article 3 of the Act. It states that personal data in SPAR may be processed to:• update, supplement and verify personal information or• select names and addresses for direct marketing, public service

announcements or other comparable activities.Processing data in this respect is the same as handing out the data electronically. Data in SPAR are, after decision by the Swedish Tax Agency, electronically handed out at cost price.

Page 11: Sampling Issues: part I

III. NON-RESPONSE ERRORTwo key types of non-response:• Non-contact: the failure to reach the chosen respondent• Refusal: chosen respondent does not cooperate

Rates have declined precipitously in the US over the last 2 decades; • Non-contact rates by telephone dropped dramatically after

2000 and the introduction of caller ID• Refusal rates are higher in urban areas

Page 12: Sampling Issues: part I

RESPONSE RATES IN THE U.S. Response Rate = Number of people who completed an interview/total number of eligible respondents contacted (including not at home, refused, etc.)

• Household CAPI or IN-PERSON surveys: in the U.S. these are around 50-60% in university research centers.

• Telephone surveys: In the US, 40-50% at university centers using very stringent and expensive methods; lower for typical phone surveys at university centers (25-35%) much lower for marketing and media (6-20%)

• Mail surveys: very variable; possible to get 15-20% RR with follow up; but depends on the population.

• Web Surveys: Depends on the population. Could be as high as 50-70% within an organization with a known email list and organizational support, or <1% with a random group (e.g., banner ad recruitment).

Page 13: Sampling Issues: part I

WHAT INFLUENCES RESPONSE RATES (RR)?

1.SURVEY MODE : highest for household in-person interview, generally lowest on the web. 2. RESPONDENT SUBGROUP. Non-response is often higher in cities; can also vary with age (young are harder to contact), and gender (men are harder to contact). 3. TYPE OF SURVEY ORGANIZATION: academic polls vs commercial. RR Typically higher when conducted by an academic or non-profit organization. 4. UNIT OF INTERVIEW: Higher RR if anyone in the home or a surrogate can be interviewed. • US National Health Interview Survey (NHIS): non-response rate is xx%• Sweden Census (SCB) contacts relatives of respondents to increase RR (from

Jacob Sohlberg)5. EFFORT TO REACH NON-RESPONDDENTS : Greater number of contact attempts, use of financial incentives, refusal conversion, longer interviewing period, all increase costs. 6. SURVEY TOPIC AND RESPONDENT INTEREST : Slightly higher RR on topical surveys in the news or those in which respondent is very involved and interested.

Page 14: Sampling Issues: part I

WHO IS MISSING?—US NON-RESPONDENTS1. Age: Underestimate the young.

• Largely due to non-contact2. Gender: Under represent men

• More difficult to contact and refuse more3. Race/Ethnicity:

• Oversample blacks by phone• Undersample blacks in household in-person

Typical Solution: Weight respondents to demographic population benchmarks

Page 15: Sampling Issues: part I

NON-RESPONSE RATES IN SWEDEN• Varies by sample mode (mail, phone, web, IVR)• RR Remains high in mail surveys

Page 16: Sampling Issues: part I

RESPONSE RATES WEB VS. MAIL: SWEDENKÄLLMÉN ET AL 2011Two random samples of 1250 individuals each were drawn from the same national register (DAFA-SPAR) over all Swedish individuals (aged 17-71) having a registered address.1. Electronic, web-based response group, received a postcard with the

same introductory text, an URLlink and a log-in code to the electronic version of the questionnaire

2. Paper-and-pen response group, two reminders were sent, three and six weeks after the main mailing.

After the first mailing, 314 individuals (25%) responded to the AUDIT paper version and 167 (13%) responded to the web-based version. Following the first reminder, the total number of responses was 483 (39%) in the paper group and 230 (18%) in the web-based group. After the second reminder the final number of responses for the paper version was 663 (53.6%), 276 men and 344 women (43 did not disclose their gender). For the web-based version of the AUDIT, the final number of responses was 324 (26.2%), 140 men and 184 women.

Page 17: Sampling Issues: part I

Web IVR Respondent Total choice Internet/IVR

Initial sampling pool 2 000 2 000 1 000 5 000 Incorrect addresses 24 40 14 78 Final sampling pool 1 976 1 960 986 4 922 Cumulative n Response no reminder 393 312 226 931 After 1 reminder 639 557 380 1 576 After 2 reminders 753 665 443 1 861 Response rate 38.1% 33.9% 44.9% 37.8%

WEB VS. INTERACTIVE VOICE RESPONSE (IVR)SINADINOVIC ET AL, 2011

Page 18: Sampling Issues: part I

5. SO WHAT? IMPLICATIONS FOR SURVEY EXPERIMENTSMajor problem with coverage and non-response errors is sample bias – overly educated, too sophisticated, older, etc. Does this matter when running an experiment with random assignment?It depends on :1. Heterogeneous experimental treatment effect2. Well-theorized and well-measured sources of experimental

treatment heterogeneity (an issue to which we will return when discussing measurement issues)

The following slides cover 2 examples concerning heterogeneous experimental treatment effects that depend on level of political sophistication (involvement or partisanship).

Page 19: Sampling Issues: part I

FROM DRUCKMAN AND KAM, 2011The external validity of a single experimental study must be assessed in light of an entire research agenda, and in light of the goal of the study (e.g., testing a theory or searching for facts).

Assessment of external validity involves multiple-dimensions including the sample, context,

time, and conceptual operationalization. There is no reason per se to prioritize the sample as the source of an inferential problem.

The nature of the sample—and the use of students—matters in certain cases. However, a necessary condition is: a heterogeneous (or moderated) treatment effect. Then the impact depends on:

o If the heterogeneous effect is theorized, the sample only matters if there is virtually no variance on the moderator.

The range of heterogeneous, non-theorized cases may be much smaller than often thought. Indeed, when it comes to a host of politically relevant variables, student samples do not significantly differ from non-student samples.

There are cases where student samples are desirable since they facilitate causal tests or make for more challenging assessments.

Page 20: Sampling Issues: part I

SOURCE OF SURVEY ERROR (ALWIN)Non-observed (bias) Non-observed (variance)-coverage bias -coverage error variance-sampling bias -sampling error variance-nonresponse bias -nonresponse error variance Observed (bias) Observed (variance)-interviewer bias -interviewer error variance-respondent bias -respondent error variance-instrument bias -instrument error variance-mode bias -mode error variance

Page 21: Sampling Issues: part I

EFFECTS OF SAMPLE BIAS (COVERAGE AND / OR NON-RESPONSE): UNPREDICTABLE EFFECTS IN EXPERIMENTS Bias can either enhance, dampen, or have no effect on the experimental outcomeExample 1: From The Ambivalent Partisan (Lavine et al)Most sophisticated LEAST affected by ideology in presence of a partisan cueIn the following example, researchers are interested in whether partisan labels would override ideological content in support of a policy. The answer varies with the mix of ambivalent vs. towards strong, univalent partisans in the partisan. Bias in the sample towards strong partisans would lead to stronger overall effects of a partisan cue.

Page 22: Sampling Issues: part I

KNOWLEDGE NETWORKS POLICY STUDY; THE AMBIVALENT PARTISAN (LAVINE, JOHNSON, STEENBERGEN IN PRESS) Policy Only Condition: Congress has recently debated two policy measures dealing with benefits to social welfare recipients.

 The first policy, POLICY 1, calls for $1000 per month for a family of one child, with an additional $200 dollars for each additional child. These benefits are intended to last 7 years. Recipients would also receive $2,000 a year in food stamps and extra subsidies for housing and child care. (Generous)

 The second policy, POLICY 2, calls for $400 per month for a family of one child, with an additional $50 dollars for one additional child. These benefits are intended to last for 3 years. Recipients would also receive $500 a year in food stamps but no extra subsidies for housing or child care. (Less Generous)

 Policy + Cue Condition: Democrats and Republicans in Congress have recently debated two policy measures dealing with benefits to social welfare recipients.

 The first policy, POLICY 1, proposed by Republicans, calls for $1000 per month for a family of one child, with an additional $200 dollars for each additional child. These benefits are intended to last 7 years. Under this Republican plan, recipients would also receive $2,000 a year in food stamps and extra subsidies for housing and child care. Generous

 The second policy, POLICY 2, proposed by Democrats, calls for $400 per month for a family of one child, with an additional $50 dollars for one additional child. These benefits are intended to last for 3 years. Under this Democratic plan, recipients would also receive $500 a year in food stamps but no extra subsidies for housing or child care. Less Generous

Page 23: Sampling Issues: part I

PREDICTED MARGINAL EFFECT OF LIBERAL VS. CONSERVATIVE POLITICAL ORIENTATION ON PREFERENCE FOR THE MORE GENEROUS POLICY PROPOSAL: KNOWLEDGE NETWORKS PANEL

Page 24: Sampling Issues: part I

EFFECTS OF SAMPLE BIAS: EXAMPLE 2Policy Support and Emotive Visual Imagery (Huddy & Gunthosdottir, 2000)Highly involved MOST affected by visual cueIn this example, the goal was to understand the impact on policy support of a positive or negative image of an animal that would be saved by an environmental policyThe effects varied with one’s position on environmental issues and so the findings would be stronger in a sample with a bias towards pro-environment views

Page 25: Sampling Issues: part I

STIMULUS MATERIALSThe design of this study is a 2 (pro or anti-environment message) times 5 (no animal, cute mammal, ugly mammal, cute insect, ugly insect) between subjects factorial design. The stimulus material consisted of flyers emulating pro and anti-environment fundraising letters. All flyers, whether pro or anti-environment, were about the same fictitious environmental dilemma, in which mining would assist an impoverished population living in the Guatemalan rainforest but would destroy the habitat of a geographically restricted animal. The pro-environment flyer argued for the protection of the animal; the anti-environment flyer argued that human needs outweigh environmental concerns. Both the name of the fictitious animal, Guatemalan Cobyx, and the fictitious organization, Club Berneaud International (CBI), were held constant

Page 26: Sampling Issues: part I

U g l y I n s e c t U g l y M a m m a l

F i g u r e 1 E m o t i v e V i s u a l I m a g e r y : C u t e a n d U g l y A n i m a l s

i n t h e P r o a n d A n t i - E n v i r o n m e n t F l y e r

C u t e I n s e c t C u t e M a m m a l

Page 27: Sampling Issues: part I

 Predicted Levels of Action for a Pro-Environment Organization Among Strongest Environment Supporters

 

0

1

2

3

4

5

6

7

8

9

No Picture Monkey Butterfly Bat BugEmotive Image

Note: Predicted levels of action calculated at a value of .25 on the pro-environment scale.

High InvolvementLow Involvement

Page 28: Sampling Issues: part I

IV. SURVEY MODE ERRORS: NON-RESPONSE, NON-COVERAGE, AND MEASUREMENT ERROR1. Survey Mode Errors Can Conflate Several Sources of ErrorIn practice mode effects can reflect a different sample population, non-coverage, and non-response errors. Can eliminate population differences by randomly assigning respondents to mode from within the same population (e.g., SPAR)Still get large differences in response rate by mode in Sweden; e.g., Kallmen et al.

Page 29: Sampling Issues: part I

DIFFERENCES DUE TO RESPONSE RATE & SURVEY MODE:AUDIT SCORES TO IDENTIFY PROBLEM DRINKING, > 8 FOR MEN; > 6 FOR WOMEN (KÄLLMÉN ET AL 2011)

Gender Response Mode

n Mean-audit score

Std. Dev. Size of difference

Men Electronic 140 5.80 4.77 .25

  Paper 239 4.73 4.20  

Women Electronic 184 4.12 4.29 .21

  Paper 294 3.39 2.59  

Page 30: Sampling Issues: part I

2. MODE & MEASUREMENT ERROROrigins of measurement differences by mode(1)Interviewers affect responses (e.g., telephone vs. web),Get decreased reporting of undesirable attitudes and behavior in personal interiviews

(2) Comprehension affected by aural (phone) vs. visual (web) mode• Get visual layout effects, primacy, recency• Typically get a primacy effect on paper, recency on phone• More positive responses to scales on phone (when do not see the scale)

(3) Ask different types of questions in different modes. • On the web use different kinds of responses for multiple vs. single responses (not

comparable to phone) e.g., checklists and grids• Show cards in personal interviews• Can include longer lists of response options in person, mail, or web

Page 31: Sampling Issues: part I

MODE BIAS ALTERS LINK BETWEEN GENDER & # OF SEXUAL PARTNERS,

TOURANGEAU ET AL 2000

Page 32: Sampling Issues: part I

3. MOVE TO MIXED MODE SURVEY DESIGNS (DILLMAN)Benefits of Mixed Mode Deisgns:• Lower Cost; Start with least expensive method

• Improve Timeliness• In 2003 NSF earned degrees survey, asked which mode best and used it in 2006.

Improved response time.

• Reduce Coverage Error; • Access to different kinds of people

• Easier to Provide Incentives in some Modes• By mail in an initial mailing

• Improve RR and Reduce Non-response Error• Do it in sequence

• Reduce Measurement Error on sensitive questionsBut creates numerous complications for survey experiments

Page 33: Sampling Issues: part I

SPECIALIZED POPULATIONS ON THE WEB• On occasion, may need to seek out special populations

which are readily accessible on the web.

Page 34: Sampling Issues: part I

Mediator and Participant Recruitment Details, SMIS Studies

1 Culture

Wars (2006)

2 Partisan Identity

(2007)

3 Partisan Identity

(2008)

4 Campaign

Ads (2007)

5 Political

Metaphors (2007)

Blog Average

6 Political

Metaphors (2008)

Data Collection Dates

6/6-7/31, 2006

5/16-6/4, 2007

3/17-5/2, 2008

3/10-5/5, 2007

6/23-7/15, 2007

-- 4/15-5/13, 2008

Mediator Type

Blogs/ Forums

Blogs/ Forums

Blogs/ Forums

Blogs/ Forums

Blogs/ Forums1

Blogs/ Forums

RAs3

Mediators Contacted

100 100 178 198 50 125.5 4

Mediators Participated

24 4 23 18 6 15 4

Mediator Response Rate

24% 4% 13% 9% 12% 12.4% 100%

Participants (N)

2248 630 3219 1452 2972 1569.2 141

Yield: Particip. / # Mediators

93.7 157.5 140.0 80.7 49.5 104.3 35.3

Page 35: Sampling Issues: part I

1 Culture Wars,

2006

2 Partisan Identity, 2007

3 Partisan Identity, 2008

ANES, 2008

Political Participation

Attend political meetings, rallies 37.7 --- --- 7.8 Campaign button, ticker 49.3 --- --- 16.2 Persuade other voters 76.5 --- --- 41.2 Candidate donation – 2004/ 2008a 40.1 39.9 54.7 10.3 Party donation – 2004/ 2008b 33.4 35.6 37.3 7.2 Volunteer for pres. candidatec --- 29.7 30.8 3.7

Volunteer for party/ organizationc --- 27.2 20.7 Vote – 2004, 2008d

91.7 96.5 98.3 70.6

Constraintd

PID & Ideology .74** .90** .65** .56** PID & Democratic Vote Choice .78** .92** 66** .75** Ideology & Democratic Vote .83** .92** .65** .58** Church Attendance & Dem. Vote -.45** -.35** -.25** -.16** Biblical Orthodox& Dem Vote -.61** -.58** -.35** -.18**

Page 36: Sampling Issues: part I

SPECIALIZED POPULATIONS S15.3 European MSM Internet Survey (EMIS): differences in sexually transmissible infection testing in European countriesU Marcus1, et al. Sex Transm Infect 2011;87:A19 doi:10.1136/sextrans-2011-050102.64Methods From June through August 2010, the European MSM Internet Survey (EMIS) mobilised more than 180 000 respondents from 38 European countries to complete an online questionnaire in one of 25 languages. The questionnaire covered sexual happiness, HIV and STI-testing and diagnoses, unmet prevention needs, intervention performance, HIV-related stigma and gay-related discrimination. Recruitment was organised predominantly online, through gay social media, and links and banners on more than 100 websites for MSM all over Europe.

Page 37: Sampling Issues: part I

REFERENCESDruckman, James N. and Cindy D, Kam. 2011. “Students as Experimental Participants: A Defense of the ‘Narrow Data Base.’” In James N. Druckman, Donald P. Green, James H. Kuklinski, and Arthur Lupia, eds., Handbook of Experimental Political Science.Cassese, Huddy, Hartman, Mason & Weber. 2012. Socially-Mediated Internet Surveys (SMIS): Recruiting Participants for Online Experiments, under review.  Don A. Dillman. 2009. Internet, Mail and Mixed Mode Surveys: The Tailored Design Method. 3rd ed. Hoboken, NJ: Wiley. ISBN: 9780471698685 (cloth)Håkan Källmén & Kristina Sinadinovic & Anne H. Berman & Peter Wennberg; NORDIC STUDIES ON ALCOHOL AND DRUGS V O L . 28. 2011Groves, Robert M. et al. 2009. Survey Methodology. 2nd edition., Hoboken, NJ: John Wiley & Sons.Hecke & Weise, 2012. In Telephone Surveys in Europe, ed. Häider, Häider, & Kϋhne; Springer, Heidelberg. \

Page 38: Sampling Issues: part I

REFERENCES Kristina Sinadinovic, Peter Wennberg, Anne H. BermanDrug and Alcohol Dependence, 2011, 114:55-60Lavine, Johnson, Steenberge. In press. The Ambivalent Partisan.Tourangeau, Roger, Lance Rips and Kenneth Rasinski. 2000. The Psychology of Survey Response. New York: Cambridge University Press. ISBN: 0521576296.Huddy, Leonie and Anna Gunthorsdottir. 2000. The Persuasive Effects of Emotive Visual Imagery: Superficial Manipulation or A Deepening of Conviction? Political Psychology. 21:745-778.