Q5 Surveys Flashcards

1
Q

Two ways that surveys breed skepticism

A
  • Aren’t surveys often wrong?
  • How can a small sample represent an entire nation?
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

How can a small sample be fine?

A

If the sample is representative and diverse, since size is less important than diversity when it comes to a representative sample

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Is a survey different from a poll?

A

To us, surveys and polls are identical because if they’re good then they’re done the same way

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Four key principles to lessen interpretation errors

A
  • Surveys are snapshots, not predictions (can only caption a person’s current thinking)
  • Opinions can change
  • A recent, high-profile event can influence survey responses
  • Survey results are in a range, not a point
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

How are survey results in a range?

A

The percentages given as survey results are the center of the range, and extend + or - however many percentage points in the sampling error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What do presidential polls measure?

A

National preferences, not electoral votes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Why aren’t presidential polls done for each state?

A
  • It’s too expensive
  • We use Electoral College so small states can have more clout than they would have if each person’s vote was counted equally
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How to get a higher confidence level

A

The range must be wider

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How to get a smaller range

A

Lower the confidence level

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Confidence level

A

The expectation that repeated surveys will be within the sampling error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Sampling error

A

The expected variance between a population and a representative sample

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

If you double the size of a sample of 1,200, and the sampling error is + or - 3 percentage points, what would the new sampling error be?

A
  • Only 1 percentage point higher because of standard deviation
  • Reducing sampling error by 1 percentage point requires a big, expensive increase in sample size
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Sample subgroups

A
  • Carry larger error margins
  • A larger sample is needed if a subgroup analysis is important
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Questions we should always ask

A
  • Who did the survey?
  • Who paid for the survey?
  • Which survey mode was used?
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How can you tell which organizations doing surveys are good?

A
  • Should be transparent about its method and survey details (response rate, why some respondents were removed)
  • Should publicly disclose all details about questions and answers (number of responses, whether it was randomized)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

When does the issue of payment surface?

A

Whenever an advocacy group pays for a survey

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Does the involvement of an advocacy group automatically invalidate the survey?

A

No, because if it’s done properly, then the survey can still be reliable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

How are advocacy groups often selective?

A

Tend to release only results that support their position

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

How to make sure an advocacy groups’ survey is valid

A

See the entire survey and all the details about the methodology, the questions and answers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Survey modes

A

Written
- Mail
- In person
- Online panel
Interview
- Telephone

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Mail survey mode

A

Cheap, but few respond

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

In-person survey mode

A

Reach is hard to find, and costly

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Online panel survey mode

A
  • Efficient, but answers could be insincere and have straightlining
  • Not available to offline people
24
Q

Telephone interview survey mode

A
  • Effective, but hard to get responses
  • Proven to be reliable and likely to elicit more sincere answers
  • Cost is an issue
25
Does written or interview surveys have an advantage for sensitive subjects?
- Written has an advantage because it's best for subjective subjects - People are more candid in writing and more guarded when speaking with a human being
26
Random digit dial
Draws a representative sample by dialing phone numbers at random
27
What is the problem with telephone interviews today?
Fewer people these days are willing to answer telephone surveys (lower response rate)
28
Response rate
The percentage of valid potential respondents who agree to take a survey
29
What is the problem with lower response rate?
- Raises costs because more people must be called to get 1 response - Invokes non-response bias
30
Non-response bias
- Occurs when those who choose to respond to a survey differ in a meaningful way from those who decline - Potential risk, but not yet an issue
31
Best alternative for telephone surveys
Online panels
32
Some organizations offer polls on their websites or social media feeds
- Don't represent the people who visit the websites, let alone the public - Rely on non-representative samples, so they don't matter
33
How does a valid online panel survey draw a representative sample?
Randomly chooses among panelists
34
Types of panels
- Volunteer, opt-in panel - Invitation
35
Volunteer panel
Most common because it's the cheapest
36
Invitation panel
Better to invite panelists at random by sending letters or dialing phones
37
How are survey panelist participants categorized?
By demographics to reflect the population, then assigned at random to different surveys
38
Straightlining
- When respondents speed through a survey without giving sincere answers - Greater for opt-in panels and lower for invitation panels
39
What to do if the sample rate is only half the population rate
Assign the survey answers from those respondents twice the weight
40
Answers from overrepresented groups
Would be given proportionally less weight
41
Weighting
Adjusting the value of answers if the sample demographics differ substantially from the population
42
Social desirability bias
Tendency to give the "correct" answer to questions involving civic or social responsibility
43
Instead of asking "Do you plan to vote in the upcoming election?" what is better?
- "Identify your polling place" - If they name it quickly, its evidence they do vote - Does not work in states like FL with voting by mail
44
Pollsters use lists of registered voters to know if they're likely to vote in the upcoming election
Does not remove social desirability bias, but it at least narrows the population by excluding those not registered
45
A slight change in wording in questions
Can change results, such as saying "forbid" versus "allow"
46
How does offering a choice in a survey matter?
- We are wired to say yes and want to be agreeable when given a "yes" or "no" question - Asking for a preference produces a more valid answer
47
Acquiescence bias
A tendency to select a favorable response or agree when in doubt
48
How can different surveys, done properly, produce different results with different question order?
Asking other questions before can make people change their opinion on the core question
49
Priming
- When exposure to a stimulus unconsciously influences a response - Not neutral!
50
Answer order in writing
The first one written will be chosen the most often
51
Answer order spoken
The last one spoken will be chosen most often
52
Primacy effect
- A tendency to choose or remember items offered first - Advantage seen in written surveys
53
Recency effect
- A tendency to choose or remember items offered last - Advantage seen in oral surveys
54
How can a telephone oral survey make sure no one answer choice has an advantage?
On the next call, move the last question to be the first, and so forth for the next calls
55
Likert scale
A consistent sequence that always begins with Agree option, a Neutral center, and ends with Disagree
56
Closed-ended vs. open-ended answer choices
Open-ended answers better reflects true public opinion but cannot make all surveys open-ended because it's slow and costly
57
Adding in an answer choice saying the respondent "doesn't know enough"
- When giving the option to admit lacking information, social desirability bias drops out - Majority usually chooses it