Lesson 3: Characteristics of Good Sample Surveys and Comparative Studies


Learning Objectives

Chapters 4 & 5

After successfully completing this lesson, you should be able to:

Terms to Know

Chapter 4

From Chapter 5



In this lesson, we will add to our knowledge base by explaining ways to obtain appropriate samples for statistical studies.

3.1 Common Research Strategies

Chapter 4 Section 4.1

The following research strategies are described in this section of the textbook.

  1. Sample Surveys
  2. Experiments
  3. Observational Studies
  4. Meta-Analyses (also covered in Chapter 25--not required for the course)
  5. Case Studies

Terms Used with Sample Surveys (Chapter 4 Section 4.2 in Textbook)

It is first necessary to distinguish between a census and a sample survey.  A census is a collection of data from every member of the population, while a sample survey is a collection of data from a subset of the population.  A sample survey is a type of observational study. Obviously, it is much easier to conduct a sample survey than a census.  The remaining sections of this lesson (Chapter 4) will discuss issues about sample surveys.

Of the many terms that are used with sample surveys, the following four need the most clarification because of how they are connected to each other.

The graph illustrates the relationship between population, sampling frame and sample. The population characteristics can be estimated by observed sample characteristics.

Figure 3.1 Relationship between Population, Sampling Frame and Sample

Example 3.1. Who are those angry women?

(Streitfield, D., 1988 and Wallis, 1987)

Recalling some of the information from Example 2.1 in Lesson 2, in 1987, Shere Hite published a best-selling book called Women and Love: A Cultural Revolution in Progress. This 7-year research project produced a controversial 922-page publication that summarized the results from a survey that was designed to examine how American women felt about their relationships with men. Hite mailed out 100,000 fifteen-page questionnaires to women who were members of a wide variety of organizations across the U.S.   Questionnaires were actually sent to the leader of each organization. The leader was asked to distribute questionnaires to all members. Each questionnaire contained 127 open-ended questions with many parts and follow-ups. Part of Hite's directions read as follows: "Feel free to skip around and answer only those questions you choose." Approximately 4500 questionnaires were returned.

In Lesson 2, we determined that the

It is also easy to identify that the sampling unit was an American woman.  So, the key question is "What is the sampling frame?" Some might think that the sampling frame was the 100,000 women who received the questionnaires (that's the intended sample).  However, this answer is not correct because the sampling frame was the list from which the 100,000 who were sent the survey was obtained.  In this instance, the sampling frame included all American women who had some affiliation with an organization because those are the women that had some possibility of being contacted.  If the response rate had been 100%, the sample would have been the 100,000 women who responded to the survey.

You should also remember that ideally the sampling frame should include the entire population. If this is not possible, the sampling frame should appropriately represent the desired population. In this case, the sampling frame of all American women who were "affiliated with some organization" did not appropriately represent the population of all American women.  In Lesson 2, we called this problem selection bias.

Chapter 4 of your text also lists three difficulties that are possible when samples are obtained for surveys. These three difficulties, which happen to be possible with this example, include:

  1. Using the wrong sampling frame. We just discussed this problem in the
    preceding paragraph. This problem is also called selection bias.
  2. Not reaching the individuals selected.   Because the questionnaire was sent to leaders of organizations, there is no guarantee that these questionnaires actually reached the women who were supposed to be in the sample.
  3. Getting a low response rate.   In Lesson 2, we learned that this survey has a problem with nonresponse bias because of the low response rate. This problem can will create bias if the people who respond have different views than those who do not.

Summary: Focusing on these distinctions is meant to help you think carefully about the process of creating a sample so you can identify issues that might arise in interpreting the results of a sample survey.

The process:  You want to know about a POPULATION but you only really have access to a SAMPLING FRAME that you can draw an INTENDED SAMPLE from; but in the end you only get observations from the actual SAMPLE.

When you read about a sample survey always try to break down the process used into these component parts.

When a report says that a random sample was used, that usually means that the intended sample was randomly selected from the sampling frame.  You must then judge whether the sampling frame was really representative of the population and whether the sample was really representative of the intended sample.  When you read the methodology used in high quality sample surveys, you will find that they go to great lengths to make adjustments to avoid bias from these issues.  If no such adjustments are made, survey results can be quite misleading.

3.2 The Beauty of Sampling

Sample surveys are generally used to estimate the percentage of people in the population that have a certain characteristic or opinion.  If you follow the news, you will probably recall that most of these polls are based on samples of size 1000 to 1500 people.  So, why is a sample size of around 1000 people commonly used in surveying?  The answer is based on understanding what is called the margin of error.

The margin of error:

For a sample size of n = 1000, the margin of error for a sample proportion is around \(\frac {1}{\sqrt{n}}=\frac{1}{\sqrt{1000}}≈0.03\) , or about 3%.

As you can see the margin of error formula depends only on the square root of the size of the sample (n).  For example, if you have four times as many people in your sample, your margin of error will be cut in half and your survey will be twice as reliable.  The size of the population is not used in the calculation of the margin of error.  So, a percentage estimated by a selected sample size will have the same margin of error (reliability), regardless of whether the population size is 50,000 or 5 billion.   If a survey is conducted using an unbiased methodology then the margin of eror tells us directly about the accuracy of the poll at estimating a population parameter.  It also helps that pollsters believe that an accuracy of ± 3% is reasonable with surveys.

So what does the margin of error represent?  The following statement represents the generic interpretation of a margin of error.

Generic Interpretation:  If one obtains many unbiased samples of the same size from a defined population, the difference between the sample percent and the true population percent will be within the margin of error, at least 95% of the time.

Key Features of the Interpretation of the Margin of Error

Example 3.2. Margin of Error

Suppose a recent random poll based on 1000 American adults finds that 55% approve of the president's current education plan.  Since the sample size is 1000, the margin of error is about 3%.  These poll results suggest that 55% ± 3% of all Americans approve of the president's current economic plan. What is the correct interpretation of this margin of error?

Margin of Error Interpretation

The difference between our sample percent and the true population percent will be within 3%, at least 95% of the time.  This means that we are almost certain that 55% ± 3% or (52% to 58%) of all Americans approve of the president's current education plan.   Because the range of possible values from this poll all fall above 50%, we can also say that we are pretty sure that a majority of Americans support the President's current education plan.  If any of the range of possible values would have been 50% or less, then we would not have been able to say that the majority supported the plan.  The range of values (52% to 58%) is called a 95% confidence interval.   We will go into further detail about confidence intervals in Lesson 9.

3.3 Relationship between Sample Size and Margin of Error

As seen from the formula in the previous section, there is a predictable square root relationship between sample size and margin of error. The numbers found in Table 3.1 help to explain this relationship.

Table 3.1. Calculated Margins of Error for Selected Sample Sizes

Sample Size (n) Margin of Error (M.E.)
200 7.1%
400 5.0%
700 3.8%
1000 3.2%
1200 2.9%
1500 2.6%
2000 2.2%
3000 1.8%
4000 1.6%
5000 1.4%

From this table, one can clearly see that as sample size increases, the margin of error decreases. To cut the margin of error in half, like from 3.2% down to 1.6%, you need four times as big of a sample, like going from 1000 to 4000 respondants.  In order to add additional clarity to this finding, the information from Table 3.1 is also displayed in Figure 3.2.

The graph shows the relationship between sample size and margin of error. Margin of error decreases as the sample size increases.

Figure 3.2 Relationship Between Sample Size and Margin of Error

In Figure 3.2, you again find that as the sample size increases, the margin of error decreases.  However, you should also notice that there is a diminishing return from taking larger and larger samples.  in the table and graph, the amount by which the margin of error decreases is most substantial between samples sizes of 200 and 1500.  This implies that the reliability of the estimate is more strongly affected by the size of the sample in that range.  In contrast, the margin of error does not substantially decrease at sample sizes above 1500 (since it is already below 3%).  It is rarely worth it for pollsters to spend additional time and money to bring the margin of error down below 3% or so.  After that point, it is probably better to spend additional resources on reducing sources of bias that might be on the same order as the margin of error.  An obvious exception would be in a government survey, like the one used to estimate the unemployment rate, where even tenths of a percent matter.  

3.4 Simple Random Sampling and Other Sampling Methods

Sampling Methods can be classified into one of two categories:

Probability Sampling

In probability sampling it is possible to both determine which sampling units belong to which sample and the probability that each sample will be selected. The following sampling methods, which are listed in Chapter 4, are types of probability sampling:

  1. Simple Random Sampling (SRS)
  2. Stratified Sampling
  3. Cluster Sampling
  4. Systematic Sampling
  5. Multistage Sampling (in which some of the methods above are combined in stages)

Of the five methods listed above, students have the most trouble distinguishing between stratified sampling and cluster sampling.

Stratified Sampling is possible when it makes sense to partition the population into groups based on a factor that may influence the variable that is being measured.   These groups are then called strata.  An individual group is called a stratum.  With stratified sampling one should:

Stratified sampling works best when a heterogeneous population is split into fairly homogeneous groups.  Under these conditions, stratification generally produces more precise estimates of the population percents than estimates that would be found from a simple random sample. Table 3.2 shows some examples of ways to obtain a stratified sample.

Table 3.2. Examples of Stratified Samples

  Example 1 Example 2 Example 3
Population All people in U.S. All PSU intercollegiate athletes All elementary students in the local school district
Groups (Strata)

4 Time Zones in the U.S. (Eastern,Central, Mountain,Pacific)

26 PSU intercollegiate teams 11 different elementary schools in the local school district
Obtain a Simple Random Sample 500 people from each of the 4 time zones 5 athletes from each of the 26 PSU teams 20 students from each of the 11 elementary schools
Sample 4 × 500 = 2000 selected people 26 × 5 = 130 selected athletes 11 × 20 = 220 selected students

Cluster Sampling is very different from Stratified Sampling. With cluster sampling one should

It is important to note that, unlike with the strata in stratified sampling, the clusters should be microcosms, rather than subsections, of the population.   Each cluster should be heterogeneous. Additionally, the statistical analysis used with cluster sampling is not only different, but also more complicated than that used with stratified sampling.

Table 3.3. Examples of Cluster Samples

  Example 1 Example 2 Example 3
Population All people in U.S. All PSU intercollegiate athletes All elementary students in a local school district
Groups (Clusters) 4 Time Zones in the U.S. (Eastern,Central, Mountain,Pacific.) 26 PSU intercollegiate teams 11 different elementary schools in the local school district
Obtain a Simple Random Sample 2 time zones from the 4 possible time zones 8 teams from the 26 possible teams 4 elementary schools from the l1 possible elementary schools
Sample every person in the 2 selected time zones every athlete on the 8 selected teams every student in the 4 selected elementary schools

Each of the three examples that are found in Tables 3.2 and 3.3 were used to illustrate how both stratified and cluster sampling could be accomplished. However, there are obviously times when one sampling method is preferred over the other. The following explanations add some clarification about when to use which method.

The most common method of carrying out a poll today is using Random Digit Dialing in which a machine random dials phone numbers.  Some polls go even farther and have a machine conduct the interview itself rather than just dialing the number!  Such "robo call polls" can be very biased because they have extremely low response rates (most people don't like speaking to a machine) and because federal law prevents such calls to cell phones.  Since the people who have landline phone service tend to be older than people who have cell phone service only, another potential source of bias is introduced.  National polling organizations that use random digit dialing in conducting interviewer based polls are very careful to match the number of landline versus cell phones to the population they are trying to survey.

Non-probability Sampling

The following sampling methods that are listed in your text are types of non-probability sampling that should be avoided:

  1. volunteer samples
  2. haphazard (convenience) samples

Since such non-probability sampling methods are based on human choice rather than random selection, statistical theory cannot explain how they might behave and potential sources of bias are rampant.   In your textbook, the two types of non-probability samples listed above are called "sampling disasters."

Read the article: "How Polls are Conducted" by the Gallup organization available on ANGEL.

The article provides great insight into how major polls are conducted. When you are finished reading this article you may want to go to the Gallup Poll Web site, http://www.gallup.com, and see the results from recent Gallup polls.  Another excellent source of public opinion polls on a wide variety of topics using solid sampling methodology is the Pew Reserach Center website at http://www.pewresearch.org  When you read one of the summary reports on the Pew site, there is a link (in the upper right corner) to the complete report giving more detailed results and a full description of their methodology as well as a link to the actual questionnaire used in the survey so you can judge whether their might be bias in the wording of their survey.

It is important to be mindful of margin or error as discussed in this article. We all need to remember that public opinion on a given topic cannot be appropriately measured with one question that is only asked on one poll.  Such results only provide a snapshot at that moment under certain conditions.  The concept of repeating procedures over different conditions and times leads to more valuable and durable results. Within this section of the Gallup article, there is also an error: "in 95 out of those 100 polls, his rating would be between 46% and 54%." This should instead say that in an expected 95 out of those 100 polls, the true population percent would be within the confidence interval calculated. In 5 of those surveys, the confidence interval would not contain the population percent.

3.5 Defining a Common Language for Comparative Studies

Overview: We've learned some of the very basics about research studies that compare two or more samples of one variable.  Chapter 5 allows us to explore this topic in more detail.  We first need to learn a few terms. These include:

  1. experimental unit
  2. explanatory variable
  3. treatment
  4. response (outcome) variable
  5. confounding variable

The experimental unit  is the smallest basic object to which one can assign different conditions (treatments.)  In research studies, the experimental unit does not always have to be a person. In fact, the statistical terminology that is associated with research studies actually came from studies done in agriculture. Examples of an  experimental unit include:

The explanatory variable is the variable used to form or define the different samples.  In randomized experiments, one explanatory variable is the variable that is used to explain differences in the groups. In this instance, the explanatory variable can also be called a treatment when each experimental unit is randomly assigned a certain condition. Examples of explanatory variables include:

You should note that gender and type of plant cannot be called treatments because one cannot randomly assign gender or type of plant.

The response (outcome) variable is the outcome of the study that is either measured or counted. We have seen the response (outcome) variable in previous lessons. Examples of response variables include:

Of course some variables may play different roles in different studies. For example, in an experiment to see whether a new diet might held in reducing your weight; weight is the response variable and whether you used the new diet or not would be the explanatory variable.  On the other hand, in an observational study to examine how a person's weight might affect their heart rate; weight would play the role of an explanatory variable and heart rate would be the response variable.

A confounding variable is a variable that affects the response variable and is also related to the explanatory variable. The effect of a confounding variable on the response variable cannot be separated from the effect of the explanatory variable.  Therefore, we cannot clearly determine that the explanatory variable is solely responsible for any effect on the response or outcome variable when a confounding variable is present.   Confounding variables are problematic in observational studies.

Image of Lab MiceExample 3.3.  Laboratory experiments conducted in the 1980s showed that pregnant mice exposed to high does of ultrasound gave birth to lower weight infant mice than unexposed mice (in fact the higher the dose the greater the effect on birthweights). This worried obstetricians who feared that sonograms given to women during pregnancy might cause lower weights in their children.  Researchers at Johns Hopkins University Hospital then examined the birthweights of infants of mothers who had sonograms versus those whose mothers had no such exposure.  They found that the 1598 infants who had been exposed averaged a couple of ounces lower in weight than the 944 infants whose mothers did not have a sonogram.  However, the women who got sonograms were more likely to have had twins in the past and were more likely to be over 40 years old.  Having twins or being over 40 are examples of confounding variables in this study since they provide an alternate expanation for the data.  You can not tell whether it was the sonogram that caused the lower birthweights or just the confounding medical reasons for getting the sonagram in the first place.  Later experimental evidence in humans did not show sonograms to have any affect (see Abramowicz et al, 2008 for a review).

3.6 Types of Research Studies

So far we have discussed two basic types of comparative statistical research studies:

With a randomized experiment, the researcher

With an observational study, the researcher

Example 3.4. Randomized Experiment (Two Independent Samples)

An educator wants to compare the effectiveness of computer software that teaches reading versus a standard curriculum used to teach reading. The educator tests the reading ability of a group of 60 students and then randomly divides them into classes of 30 students each. One class uses the computer regularly while the other class uses a standard curriculum. At the end the semester, the educator retests the students and compares the mean increase in reading ability for the two groups.

This example is a randomized experiment because the students were randomly assigned to one of two methods to learn reading.  Also in this example:

The randomization that is used in this example cancels out other factors (confounding variables) that could also affect a change in reading ability. Specifically, the randomization will cancel out factors that may result from either self-selected or haphazardly-formed groups. With self-selection, students might base their decision on whether or not they like the computer or whether or not their friends will be in the class. This is no longer a problem when the groups are randomly formed.  Consequently, "cause and effect" statements can be used if statistical significance is found and other precautions are used to treat each group the same except for the different treatments assigned.

In statistics, we also say that the two samples in this study are independent. The label of independent samples is used when the results for the one sample have no impact on the results found with the second sample.  In this instance, each student provided a measurement for only one treatment.  The results from students in one group will not impact the results of students in the other group, so the results from the two samples are independent.

Example 3.5. Observational Study (Two Independent Samples)

A medical researcher conjectures that smoking can result in wrinkled skin around the eyes. The researcher obtained a sample of smokers and a sample of nonsmokers. Each person was classified as either having or not having prominent wrinkles. The study compared the percent of prominent wrinkles for the two groups.

This example cannot be a randomized experiment because it would be both unrealistic and unethical to randomly assign who would be the smoker and who would be the nonsmoker. Also in this example:

Because this example is an observational study, it is possible that confounding variables may also be responsible for whether or not a person has prominent wrinkles. Possible confounding variables include (1) how much time the person spends outside, (2) whether or not the person wears sun screen, and (3) other variables that revolve around health and nutrition (especially those that could be related to smoking status). Because we can't separate the impact that these variables may have on the response variable, "cause and effect" conclusions are never possible. The researcher would be limited to saying either that there is an association between smoking status and wrinkle status or that there is a difference in the two percents when comparing smokers to nonsmokers.

This is also an example where the two samples are independent. The individuals in this study were classified as being either smokers or nonsmokers. The results from the smoking group had no impact on the results from the nonsmoking group.

Example 3.6. Randomized Experiment (Two Dependent Samples or Matched Pairs)

Is the right hand stronger than the left hand for those who are right-handed? An instrument has been developed to measure the force exerted (in pounds) when squeezed by one hand. The subjects for this study include 10 right-handed people.  How can we best answer this question?

What would happen if we tried to implement what was done in Example 3.4?  This would mean that we would randomly assign five people to use their right and five people to use their left hand. The results from the two groups would then be compared. Hopefully you see that even though randomization is being used with this approach, the results may not be the best because it is possible that - just by the luck of the draw with so few people - the one group could be comprised of strong people while the other group could be comprised of weak people.  If this happened, one could erroneously conclude that one hand is stronger for reasons other than that there is a difference in the two hands.

A better approach would be to have each person use both hands and then compare the results for the two hands. With this approach, the

The design used in the example is called a block design because the results from each person form a block. Specifically, this block design is called a matched pairs (block) design because each person provides two data observations that can be paired together (i.e. left and right hands of the same person form the pairs). Consequently, we can say that we have two dependent samples.  Table 3.3 shows how a spreadsheet for the data in the matched pairs design might look.

Table 3.3. Spreadsheet of Matched Pairs (Block) Design for Example 3.6

Person Force from Right Hand Force from Left Hand

In Table 3.3, one sees that the results from each person form a block. The reason that this design is used is so that unwanted or extraneous variation can be removed from the data.  In order to accomplish this goal, the data analysis is based on the differences rather than on the original data.    By using the differences, we are comparing the two data observations each person provides to each other which distinguishes matched pairs from independent samples.     Table 3.4 shows some data that could have been collected in this study.

Table 3.4. Picture Data of Matched Pairs (Block) Design for Example 3.6

Person Force from Right Hand (pounds) Force from Left Hand (pounds) Difference = (Right Hand Force) - (Left Hand Force)
1473847-38 = 9 pounds
2201520-15 = 5 pounds
3332633-26 = 7 pounds
10282728-27 = 1 pound

As you examine the results from Table 3.4, you should see that there are innate differences in strength when comparing the people who participated in the study. For example, Person 1 is much stronger than Person 2.  However, the variation from person to person is no longer a factor when the differences are used in the data analysis rather than using the original data.

Also, as you examine Table 3.4, you should see why we classify the two measurements for each experimental unit as dependent.   A higher value in one hand is usually followed by a higher value in the other hand.   The values are more similar for each pair of measurements for each experimental unit than the values are between experimental units.

Even though the matched paired design is critical in this example, this study would also benefit from randomization. Since each person is doing both things or providing two measurements, the randomization could be used to determine the order in which the treatments are done. Why would this enhance the study? Problems can exist with block designs, including matched pair designs, when what happens with the first measurement "carries over" to the second measurement. This "carryover" effect is a type of confounding that is found with block designs.  

For example,  "carryover" effect could possibly occur if complicated equipment was used to measure force exerted by a hand. If everyone used their right hand first, they might not do so well with the right hand because of not understanding the equipment, but do much better with their left hand because they learned how the equipment worked.  In statistics, this is called a training effect. The opposite, however, could also take place. Suppose everyone was asked to first exert force with their right hand for 15 minutes and then repeat this task with their left hand.  Participants might do okay with their right hand but become either bored or fatigued or sore when asked to repeat with this task with the left hand. So again, what happened with the first measurement would "carryover" and affect the second measurement. One may conclude that one hand is stronger than the other, not because this is really true, but because the "carryover" effect allowed this to happen.

The overall conclusion is that if you randomly assign the order of treatment, some people will use their right hand first and other people will use their left hand first. This randomization should cancel out the possibility of a "carryover" effect.  In statistics, we call this a randomized block design, as shown in Table 3.5. Randomizing the order of treatment makes this a randomized experiment.

Table 3.5. Randomized Matched Pairs (Block) Design for Example 3.6

Person Hand Used First Hand Used Second
1Right HandLeft Hand
2Left HandRight Hand
3Right HandLeft Hand
10Left HandRight Hand

Example 3.7. Observational Study (Two Dependent Samples or Matched Pairs)

An owner of a theater wants to determine if the time of the showing affects attendance at a "scary" movie. In order to check this claim, a sample of five nights from all possible  nights over the past month was obtained. The attendance (total number of tickets sold) for both the 7:00 PM and the 9:30 PM showings was determined for each of the five nights.

In this example:

This example also uses a matched pair (block) design because there are two measurements made on each night. A picture of this matched pair block design is found in Table 3.3.

Table 3.6. Matched Paired (Block) Design for Example 3.7

Night Attendance at 7:00 PM Showing Attendance at 9:30 PM Showing

Again, why is the matched pairs design preferred over two independent samples? In this example, our goal is determine whether or not time of showing affects attendance at the "scary" movie. We do not want any extraneous or other unwanted variation to explain the differences in attendance. In this example, the potential unwanted variation would be the variation that would exist from night to night. Some of the selected nights may fall on a weekend while other nights may fall on a weekday. This factor could affect attendance. However, this will no longer be a problem when both measurements are made on the same night.

This example, however, cannot be a randomized experiment because it would be impossible to randomly assign time of showing. The 7:00 PM showing will always take place before the 9:30 PM showing. Consequently, there is a possibility that what happens at the 7:00 PM showing may "carryover" and affect attendance at the 9:30 PM.  A possible "carryover" effect could be the fact there is a limited amount of parking near the theatre. If this were true, perhaps those at the 7:00 PM showing take all the available spots. Then people planning to attend that 9:30 PM showing may not attend because of not being able to find a parking spot. However, this problem may not exist if there is sufficient time between the two showings so that those who attended the 7:00 PM showing had time to leave before those who arrived for the 9:30 PM showing.  In any event, because this is an observational study, confounding variables are possible. "Cause and effect" conclusions may not be used if statistical significance is found.

Table 3.7. Summary of the Four Examples

ExamplesType of StudyType of SamplesRandomization UsedIs Confounding Possible?
3.4ExperimentTwo IndependentRandomize type of treatmentNo, randomization cancels out confounding
3.5ObservationalTwo IndependentNoneYes
3.6ExperimentTwo Dependent (Matched Pairs)Randomize order of treatmentNo, randomization cancels out confounding
3.7ObservationalTwo Dependent (Matched Pairs)NoneYes

3.7 Designing a Better Observational Study

While reading Chapter 5 you should pay particular attention to review the "Difficulties and Disasters in Experiments" that are listed in Section 5.3. You should also pay attention to the information found in Section 5.4 about ways to conduct a good observational study. There are ways to design observational studies so that problems with confounding can be minimized. A case-control study is the best example of where confounding can be minimized.

In a Case-Control study people with the response of interest form a group of "cases" and are compared to a group of "controls" who are in similar circumstances except for the fact that they have the response.  This type of study is very common in the study of factors that might be associated with uncommon diseases.

Image of a man talking on a cell phoneExample 3.8. In order to study whether the longterm use of cell phones might be associated with a greater risk of brain tumors, researchers in France conducted a case-control study (see Coureau et al, 2013).  In their study 253 glioma patients (a type of brain tumor) were compared as to their cell phone use with 892 matched controls of similar age who lived in the same areas (according to electoral rolls).  In this way, the researchers hoped that other environmental factors associated with the differing brain tumor rates seen in differing parts of the country would be eliminated as confounders.

The difficulty in performing a case-control study comes with finding a good group of controls (e.g. similar life circumstances, same gender, similar age, similar similar family histories of disease, etc...).  

Notice that case-control studies are done retrospectively - they compare patients who have the disease today with those who don't and ask them about past exposures or behaviors.  But people's memories might be faulty or affected by the fact that they have the disease.  For example in a study of people who have arthritis, patients will often "remember" that their parents also suffered from arthritis indicating a strong genetic component.  However when their non-suffering siblings with the same parents are asked the same question, no genetic component is found.

One alternative is to condct a Prospective study in which people with different exposures or behaviors (the explanatory variables) are followed over time to see how many in each situation get the disease (the response variable).

Example 3.9.  447,357 non-hispanic white cancer free members of the AARP who were 50 to 71 years old in 1995-1996.  These subjects werefollowed for about ten years until the end of 2006 and during that decade 2,904 of them developed melanoma skin cancer.  All of the subjects filled out a series of questionnaires at the start of the study and explanatory factors identified at that time could be examined to see if they might be associated with the onset of melanoma later in life.  In one study based on this very large cohort of AARP memers, Loftfield et al, 2015 found a relationship between heavy coffee drinking (4 cups a day or more) and a modest increase in the risk of getting melanoma.  Of course, this observed association can not be viewed as causal since other variables associated with melanoma risk (like exposure to sunlight) might also be different between the heavy coffee drinkers and the non coffee drinkers.

Lesson 3 - Test Yourself!

Now it's time to test yourself to check your understanding of the material in this lesson.  Be sure to also try the Practice quiz on ANGEL (you can take that multiple times to see more practice problems). 

Think About It!

Come up with an answer to these questions by yourself and then click the icon on the left to reveal the answer.

1. Which of the following is not an example of probability sampling?

a. simple random sampling
b. cluster sampling
c. convenience sampling
d. stratified sampling

2. Which of the following surveys would have the smallest margin of error?

a. a sample size of n = 1,600 from a population of 50 million
b. a sample size of n = 500 from a population of 5 billion
c. a sample size of n = 100 from a population of 10 million

3. Suppose a recent survey finds that 80% of Penn State students prefer that fall semester begins after Labor Day. The results of this survey were based on opinions expressed by 200 Penn State students. Which of the following represents the calculation of the margin of error for this survey?

a. 200
b. 1/200
c. \(1/ \sqrt{200}\)
d. \(\sqrt{200}\)

4. Suppose a margin of error for a poll is 4%. What is the correct interpretation of the margin of error for this poll? In about 95% of all samples of this size, the ________________.

a. difference between the sample percent and the population percent will be within 4%.
b. probability that the sample percent does not equal the population percent is 4%.
c. probability that the sample percent does equal the population percent is 4%.
d. difference between the sample percent and the population percent will exceed 4%.

5. In order to survey the opinions of its customers, a restaurant chain obtained a random sample of 30 customers from each restaurant in the chain. Each selected customer was asked to fill out a survey. Which one of the following sampling plans was used in this survey?

a. cluster sampling
b. stratified sampling

6. Eighty individuals who wish to lose weight are randomly divided into two groups of 40. One group is given an exercise program to follow while the other group follows a special diet. After three months, the researcher compares mean weight losses in the two groups. What type of study is this?

a. randomized experiment 
b. observational study 

7. Which type of randomization plays an important role in experiments?

a. randomizing the type of treatment 
b. randomizing the order of treatment 
c. both a and b 
d. neither a nor b 

8. A researcher conducts a study to determine whether or not smoking causes lung cancer. What is the explanatory variable in this study?

a. whether or not someone smokes 
b. whether or not someone has cancer