EDR725
 StartSyllabusClassLibraryCommunicate
Help EDR725 : The Class : Strategies : Sampling : Electronic Text

"It's All in Who You Know:" Qualitative Population and Sampling Procedures

Dear Cyber-scholars! We've established the rationale for the qualitative approach, and also compared it with more traditional quantitative/experimental models. In addition, we've skipped ahead to the data analysis and reporting to examine some strategies for compiling qualitative findings and results.

Now, let's backtrack to the remaining elements of the research plan. Today, we'll concern ourselves with the "who/what" part: namely, population and sampling methods.

At this point, I'd like to suggest that the "more traditional" labels for population and sampling procedures also apply to qualitative research. For those of you who didn't take Intro to Research with me, I'm going to provide you with a link to the related Module #5 below.

http://jan.ucc.nau.edu/~mid/edr610/class/sampling/procedures/lesson5-1-1.html
Population and Sampling Procedures (EDR 610)


Sampling

In this chapter, we'll first take a look at some important distinctions in the nature of the population for qualitative vs. quantitative studies. In doing so, we'll visit Robert K. Yin's distinguishing terminology of "statistical generalization" vs. "analytic generalization." Then, we'll conclude with some specialized sampling terminology that is particularly appropriate for qualitative research studies.

I. The Nature of the Population: Some Key Differences in Thinking (between the quantitative vs. qualitative approaches)

As a reminder to our Intro to Research friends, we learned that the population is that totality (usually persons, but could be 'things' like curricular materials, schools, clinics, etc.) to which wish to generalize or project our sample findings.

As you'll see (or will recall) from related discussion you may have had in Intro to Statistics:

  1. We define our totality of interest; e.g., the population;
  2. We then proceed to select or draw a sample from that population - the sample represents the "cases" (again, usually persons, but not always) that we actually include in our research study; and
  3. Finally, based upon the findings and results from our sample, we then project back or generalize these sample findings to our population. Say, for instance, that we find an average difference in science achievement score between the 6th-grade boys and girls of our sample. How confident can we then be that this gender-based average difference in science achievement would be expected to also hold true for the entire population of 6th-graders from which we selected or drew our study sample?


Now ... in both of our "Intros," research and statistics, we learned that if we select the sample randomly or probabilistically from that population, we can then go on to generalize the sample findings, if they are quantitative, with a certain degree of confidence to that entire population. We learned to make statements such as "The researcher can therefore be 95% confident that the average science achievement scores will differ for the 6th-grade boys and girls in the population at large."

We learned too that a probabilistic sampling procedure means that every sampling unit (again, usually persons, but not always) had some known, positive, non-zero (not necessarily equal) chance of being selected into our sample. For this to happen, everything hinges on a detailed, clear, operational definition of our population. This is so that we can be able to list all 'elements' of that population; e.g., we'll know "who's in and who's out." We didn't miss any eligible sampling unit, nor did we double-count anyone. This, then, is the basis of that "everyone had some known chance of being drawn."

Conversely, we learned that there will always be a margin of error, called our alpha or Type I error, that the "quantitative effect" (e.g., mean difference; correlation) that we observed in our sample is just a 'sampling fluke,' but in reality there is no such "effect" (e.g., mean difference; correlation) in the population at large. But we learned that we can bound this "risk of wrongly assuming an effect for the whole population," or Type I error. Commonly accepted Type I error risk factors are 5% or 1%.

Well -- this type of classic, quantitative inferential generalization is fine and dandy -- but it may not be what you are looking for in qualitative research! Qualitative design methodology expert Robert K. Yin refers to the above probabilistic selection of subjects and generalization of quantitative results as statistical generalization.

Yin has identified a totally different rationale and purpose for qualitative research. He calls it analytic generalization. In analytic generalization, according to Yin, you are looking not to generalize a single, limited quantitative finding across many cases. Rather, you are seeking an in-depth, rich, contextual understanding of a phenomenon.

-- > Think "fox vs. hedgehog" back from our Module #1 and you'll have the right idea!!!

For analytic generalization, then, you are not seeking to project a given limited finding across hundreds or thousands of cases with a specified level of confidence. In fact -- you may not even know yet for sure what the "theory," model, or finding is -- that is why you are doing the exploratory qualitative work in the first place!

Thus, issues like "with 95% confidence," "Type I error," "p-values" and the like are of little or no importance - yet. We are not that far in our understanding of the key variables, factors or phenomenon of interest.

  1. As a result, then, for analytic generalization, probabilistic, random-sampling schemes may be quite undesirable and counter-productive!

    That "average" or "middle-ness" may be quite uninteresting to you! You may need, and want, to pre-target the extreme cases! e.g., schools with "poor" interpersonal climate as compared to schools with "good" climate so that you can "max out" the factors that seem to be producing the difference and smoke them out more easily!

  2. Yet another related, key difference: you may not need, or want, "large" samples for analytic generalization!

    Again, your focus, like the 'territory' of the hedgehog, may be to 'stay in a single spot' and acquire a detailed understanding of the factors that drive that particular school, clinic, office, vocational rehabilitation training center, and so forth. Your own goal for the time being is not to identify factors that will generalize or apply to other such schools, clinics, etc., for now. You may be doing the study to solve an immediate and pressing problem at that school, clinic, etc. If what you find is helpful in other similar settings, that is of course "gravy", and what Yin refers to as our next step: cross-case analysis. But it may not be your immediate need or focus for now. So for that reason, you do not need to worry about large samples of certain minimum sizes.

    Our Intermediate Statistics partners will also recall that we learned some rules of thumb for desired minimum sample sizes in connection with many inferential or analytic statistics, such as the independent-samples and matched-pairs t-tests. You'll recall that one reason for such minimum sample sizes is to help ensure "a good mix on all potential key intervening variables" so that we can generalize the quantitative findings confidently to our target population. Here, too, such rationale is much more applicable to statistical generalization of limited quantitative findings across large target populations, than it would be to gather masses of "often messy" exploratory-type qualitative data and use it to begin to make some sense of a broad phenomenon of interest. So the larger sample sizes may not be needed for the latter.

-- > Think of the Donald Duck addendum in our Module #1! More information times fewer subjects (qualitative) balances out against Less information times more subjects (quantitative!)!


We might summarize the two major themes and points of distinction as follows:






II. Some Specialized Qualitative Sampling Terminology

Please remember that qualitative procedures can be combined with quantitative procedures! These are known as multimethod designs. Also, please recall that qualitative approaches may on occasion be applied in a "theory-testing" pseudo-experimental sense -- e.g., when the researcher is working with established, well-validated models of behavior and phenomena and wishes to collect non-numeric evidence on these models. Thus, as mentioned earlier, we can apply any and all of the "more traditional" sampling terminology (please see related http://jan.ucc.nau.edu/~mid/edr610/class/sampling/procedures/lesson5-1-1.html Intro to Research Module 5) to qualitative studies, as well!

Nonetheless, given the key distinctions between "statistical" and "analytic" generalization, some specialized sampling terms have come into use with qualitative designs. The following come to us courtesy of Michael Quinn Patton - not only a true "evaluation research guru," but someone with a genuine "zest" and "passion" for all things qualitative! I can't think of a finer source! Here we go for some add'l. qualitative sample-selection buzzwords:

  1. Extreme or deviant-case sampling. As indicated earlier, the "middle-ness" may actually not be all that interesting or revealing for what we want to know from our qualitative data. Instead, the extremities may be much more useful to us. Suppose we focus on a clinic that has a reputation for "efficient" service of clients, and compare it to one with long waiting lists. We would study each one intensively and then kind of reason "bass-ackwards" to discern those factors or variables that seem to be responsible for the differences we observe. This, then, is the idea behind such pretargeting of "extreme" cases. By maxing out differences, we hope to have a better handle on discovering the 'causes' (in a non-experimental sense, of course, but maybe even more valid!) of such differences.

  2. Maximum-variation sampling. This strategy is sort of trying to "have your cake and eat it too." In other words, remember the idea behind statistical generalization? You want to get 'as good a mix on as many other factors as possible' so that you can more confidently project or generalize your findings regarding the single phenomenon of interest back to a target population. That is, you want to say, "there was a good mix of gender, ethnicity, aptitude level, etc., etc., across both groups. So if I still see greater learning going on with the hands-on peer-assisted teaching method, I can rule out these other factors and more confidently attribute the difference to the teaching method itself."

    Well, normally a probabilistic (random) sampling procedure of "lots" of cases from an also-large target population is a good way to help ensure such a "mix." But what if you don't have the luxury of "large" samples and/or the desire or ability to choose them at random? Well, you can try and "purposefully build in" as much diversity as possible into your selection of your more limited number of cases for your qualitative sample. For example, if you are selecting clinics from a centralized statewide web or network of such clinics, you might make sure you draw from rural, urban and suburban settings. In addition, you might try to 'sort of control for' experience by picking a relatively new clinic for every "old," established one, however you define "age" of clinic to be. And you could continue the scenario by trying to match on still other key variables: e.g., funding sources, type of clientele served, and so forth.

    This, then, is the general idea behind maximum-variation sampling. You are attempting to build in diversity on certain key factors or variables in order to hopefully give you at least some generalizing power over those factors, as well as to look at the impact they may have on your phenomenon of interest.

  3. Homogeneous samples. This is in a way the exact opposite of the preceding (#2, in the preceding discussion). You are pre-targeting a subgroup as "alike" as possible in order to study it intensively. You would thus be staying within rural new clinics of X funding level, etc., etc. This strategy is particularly useful when you are doing focus group interviewing - a data collection procedure we'll be talking more about in future lesson packets.

  4. u. This strategy is particularly useful when your goal is to characterize for outside readers the classic "typical" subject, case, etc. For our Intro to Statistics friends, then, it might involve determining the modal (most frequently occurring) category and then hand-picking (perhaps with the help of "cultural insiders" or "key informants") a subject who fits "that typical" profile. For example, I suppose that records could be used to determine the "typical" NAU undergraduate student -- which would be the modal major? socioeconomic status? gender? ethnicity? place of residence? etc., etc.? Such a student could be selected and then interviewed, surveyed, etc. Keep in mind, though: "typical" for purposes of informing outsiders "what is typical?" does not necessarily imply "therefore, generalize what you find from this 'typical' case, or cases, to all NAU undergraduate students." So, once again, there is a key difference between "statistical" vs. "analytic" generalization.

  5. Critical case sampling. While this one may also at times be equivalent to the "extreme" or "deviant" case sampling that we first encountered in # 1, above, this is not necessarily the case. For critical case sampling, we pretarget a "key" individual, subgroup, setting, etc., that is particularly relevant to our study. For example, media experts may draft some sample TV and/or print advertising regarding a candidate for political office. They may target the "well-educated" citizens of a community as the focus of their pilot study of the understandability of these test ads. Members of this "well-educated" subgroup might be shown the ads and then invited to participate in a focus group or individual interview session designed to gauge these subjects' reactions to these ads. For if the "well-educated" citizens of that community, however operationally defined, cannot understand the message or basic premise of these campaign TV and print ads, then it's a likely bet that they will not go over too well with the general populace, either. Thus, one can consider these critical cases as somewhat akin to the "purposive" or "judgment" sampling procedures discussed in the Intro to Research population and sampling packet addendum. As with typical case sampling in # 4, pg. 9, our goal is not necessarily to generalize to all subjects, but rather to obtain the in-depth information and understanding regarding these key cases, subgroups, etc.

  6. Snowball or chain sampling. This is pretty much as discussed in our related Intro to Research lesson packet. By using subjects to help locate other target subjects, we may gain better access to such specialized, information-rich, limited subpopulations than we would relying on our own resources alone.

  7. Criterion sampling. This too may be considered synonymous with "judgment" or "purposive" sampling, as per our Intro to Research population and sampling discussion Module #5, EDR 610 learning materials. It involves needing to pre-target and include those subjects, cases, etc., which meet certain key selection criteria. These could be, for instance, "Career Ladder teacher incentive programs in rural Arizona communities." Such inclusion criteria may be critical with regard to the phenomenon that we are trying to study qualitatively. Thus, rather than take the chance that they will come up in a random draw, we need to "explicitly go after" them in our selection of qualitative sample subjects, subgroups, settings, cases, etc.

  8. Confirmatory and disconfirmatory cases. This involves the first "baby steps" of trying to generalize analytically. It is akin to Robert Yin's 'cross-case analysis,' with one slight twist.

    Suppose we have studied a school system, as one of our former EDL doctoral candidates Dr. Dee Dee Nevelle did for her dissertation research, to identify the factors that appear to contribute to "positive school climate." The next step for Dee Dee, and/or others who wish to pursue her line of research, is to:

    1. Locate similar types of schools that also appear to be "positive climate" schools and check to see if those factors that Dee Dee identified also appear to be present in these other "positive climate" schools -- e.g., "confirmatory cases:" but also:

    2. Play "devil's advocate" with yourself! and really road-test your findings from the other extreme. That is -- now locate schools that have been identified as "poor" or "negative" climate schools. Are those factors that were identified by Dee Dee Nevelle absent in these negative climate schools? or on the other hand, are they also present here -- in which case we have some contradictory evidence to deal with regarding their original, presumed linkage to "good climate"? -- e.g., "disconfirmatory cases."

      By replicating (confirming) and looking at what happens at the other extreme (disconfirming), we are approaching the beginnings of theory testing and validation of our initial qualitative findings. This lends validity to these initial findings and results.

  9. Convenience sampling. This too is as discussed in our Intro to Research population and sampling Module #5. It involves pre-targeting particularly accessible sites, subjects, etc. So much of qualitative research is highly dependent on gaining access to the site or field. Thus you may be at the mercy of "gatekeepers" or "key informants" that will allow such access.

As you've probably figured out, there is some overlap in the preceding list. Also, as with the more traditional population and sampling labels from our Intro to Research Module #5, it is very likely that a given research study will utilize more than one of these procedures.


* * *


We'll continue our qualitative sojourn by proceeding into some strategies for collecting qualitative data! 'Till next time, dear friends, remember that access is everything .. !!!

Once you have finished you should:

Go on to Assignment 1
or
Go back to Qualitative Population and Sampling Procedures

E-mail M. Dereshiwsky at statcatmd@aol.com
Call M. Dereshiwsky at (520) 523-1892


NAU

Copyright © 1999 Northern Arizona University
ALL RIGHTS RESERVED