Professional Documents
Culture Documents
Gerardo L. Munck
University of Southern California, Los Angeles, California, USA
Jay Verkuilen
University of Illinois, Champaign-Urbana, Champaign,
Illinois, USA
Glossary
case study A study in which one unit is analyzed, typically in
an intensive manner that is attentive to time and process.
Although, strictly speaking, in a case study the N 1,
frequently the effective number of observations is considerably higher.
cross-sectional design A study in which observations on
a variable or multiple variables are collected across units at
the same point in time.
experimental design A study in which the treatment is
consciously manipulated by the researcher and in which
units are randomly assigned to treatment and control groups.
To distinguish experimental from quasi-experimental
designs, the former are sometimes called randomized
experiments.
external validity Concept originally introduced by Donald T.
Campbell to refer to the generalizability of the finding of
a causal relationship between variables beyond the domain
of the actual units, spatial and temporal setting, and specific
treatments that are examined.
internal validity Concept originally introduced by Donald T.
Campbell to refer to a causal relationship between variables
in the actual units, spatial and temporal setting, and specific
treatments that are examined.
large N study A study in which observations are made across
a large number of units. Such studies, however, vary
significantly in terms of their N, with typical cross-national
studies in the field of comparative politics and international
relations oscillating between 30 and 100 and those using
opinion surveys reaching into the thousands.
longitudinal design A study in which multiple observations
on variables are collected across time for the same unit.
Also known as time series design.
Encyclopedia of Social Measurement, Volume 3 2005, Elsevier Inc. All Rights Reserved.
385
386
Research Designs
387
Research Designs
Assumption
Potential problem
Assumption
Potential problem
Direction of causality
Reverse causation
Unit 1 at time 1
IV
DV
IV
DV
DV/IV
DV
DV
IV
IV
ii
DV
iii
iv
DV
Units 2, 3, etc.
at time 1
IV
Spuriousness
IV
DV
IV
DV
Units 2, 3, etc.
at time 2
IV
IV
DV
DV
DV
DV
IV
DV
Unit 1
Unit 1
IV1
DV
IV2
IV1
DV
IV2
Units 2, 3, etc.
IV1
Units 2, 3, etc.
DV
IV2
IV3
DV
IV4
process and to consider how design options have an indirect but important impact on causal inference as a result
of their requirements and contributions vis-a`-vis theory
and data.
Third, Campbell et al.s discussion of design options is
fairly limited and their assessment biased toward experimental designs. Indeed, they place such strong emphasis
on certain design optionsthe ability of researcher to
consciously manipulate independent variables and randomly assign units to treatment and control groupsthat
they offer a narrow and unbalanced optic on questions of
research design. They do highlight the difficulties of
conducting experiments and emphasize the ways in
which experiments are likely to never guarantee that
all threats to validity are eliminated. Moreover, occasionally they offer an exemplary display of pluralism.
However, they tend to overlook some significant shortcomings associated with experimental data and downplay
the significant potential virtues of nonexperimental data,
and they ignore both the role played by design choices
other than those that are defining elements of experiments and the way in which design choices have an indirect impact on causal inference. Thus, to assess the
Theory
Research
design
Causal inference
Data
388
Table I
Research Designs
Research Designs: Classification Criteria and Options
Classification criteria
How is the value of the independent
variable(s) assigned?
Main options
Disaggregate designs
Random sample
Purposive (deliberate or intentional) sample
Entire population
Representative
Typical (mode, mean, or median)
Heterogeneous (extreme and typical)
Census
Many
Few
One
Large-N
Small-N
Case study
Across units
Across time
Cross-sectional
Longitudinal
many units are selected? and How are comparisons organized (i.e., whether temporally and/or spatially) (Table I)?
Moreover, it is necessary to address the manner in which
design choices have both a direct and an indirect impact
on causal inference.
Recognizing the seminal nature of the work by
Campbell et al. but also seeking to overcome its limitations,
this article provides an overview of the current state of
knowledge on research design. The number of alternative
research designs is, in principle, very large. Indeed, focusing only on the main options related to the four criteria
presented in Table I, it is clear that the range of possible
designs is much greater than usually believed, although not
all will be sensible. However, the following discussion is
organized in terms of research traditions, given that this is
how much research and discussion about design options
is carried out, focusing on experimental and quasiexperimental designs first, turning next to large N or
quantitative designs, and completing the discussion with
case study and small N or qualitative designs. In each case,
the focus is on the prototypical designs associated with
each tradition and an assessment and comparison of the
strengths and weaknesses of these designs. The need for
choices about research design to be explicitly addressed
and justified, and the need to actively construct bridges
across research traditions, is emphasized.
Research Designs
389
The units are randomly divided into two groupsthe treatment group and the control groupand a treatment is applied to the
treatment group and not to the control group.
Measures are taken of the values of the dependent or outcome variable for the treatment and control groups. The generated data are
organized as follows:
Treatment group
Control group
Units 1, 2, 3, . . . , n
Units 1, 2, 3, . . . , n
Independent
Yes
No
variable
(treatment)
Dependent
Value
Value
variable
____________________________
Alternatively, after the units have been randomly divided into two groups, a pretest measure of the value of the dependent variable
for both the treatment group and the control group is made. The generated data are organized as follows:
Treatment group
Control group
Units 1, 2, 3, . . . , n
Units 1, 2, 3, . . . , n
T1
T2
T1
T2
Independent
No
Yes
No
No
variable
(treatment)
Dependent
Value
Value
Value
Value
variable
____________________________
In both instances, the analysis of the data focuses on the difference in values on the dependent variable between treatment and
control groups (the italicized values), a difference that can be interpreted as a measure of the causal effect of the treatment.
When a pretest measure on the dependent variable is obtained in addition to a posttest measure, a comparison of the values of units
of the treatment and control groups at T1 (the underlined values) serves to double-check that all units are equivalent, given that
this goal is already ensured by the random assignment of units.
Figure 4 Experimental designs: Defining features.
Table II Experimental Designs: An Assessment
Design elements
Strengths
Conscious manipulation of
independent variables
Establishes equivalence of
units, helps guard against
third, extraneous variables
(internal validity)
Weaknesses
Lack of viability, due to practical and/or ethical reasons, to
study many important questions
Unsuitable for the study of action and lack of attention to
causal mechanisms (internal validity)
Unsuitable for an assessment of complex causes
Requires a priori knowledge of plausible independent
variables and well-specified causal model; not useful in
theory generation
Tends to generate obtrusive, reactive measurements
Lack of viability, due to practical and/or ethical reasons, to
study many important questions
390
Research Designs
Research Designs
Observational Designs
The distinction between experimental and observational
studies is a fundamental and deep one. The key difference
is that in observational studies, control of possible third
variables is not attained automatically through random
assignment. Rather, in observational studies third
variables have to be formulated and measured explicitly,
and control is sought through the analysis of the data.
However, the distinction between large N studies, on
the one hand, and case and small N studies, on the
other hand, is equally profound and probably more pervasive. This second distinction is not unique to observational studies. Indeed, the quantitative vs qualitative
distinction runs through both the experimental and the
observational research communities. However, the discussion of this distinction is developed only in the context
of observational studies and focuses primarily on the prototypical quantitative and qualitative studies: the large N,
cross-sectional study and the small N study based on the
longitudinal case study, respectively.
Large N Studies
A large N study has some considerable strengths that
make it, in some regards, superior to an experimental
study. First, because it uses data generated through the
natural course of events, it is a viable design for studying
Table III
391
Design elements
Assignment of value of the
independent variable(s) by nature
Random selection of units or selection
of an entire population
Selection of many units that are
compared cross-sectionally and/or
longitudinally
Strengths
Viability of studying important questions
that involve nonmanipulable variables
Establishes generalizability (external
validity)
Establishes patterns of association with
a high degree of precision and
confidence
Constitutes a tool for theory generation
Weaknesses
392
Research Designs
Research Designs
Table IV
393
Design elements
Assignment of value of the
independent variable(s) by nature
Selection of one or a few units that are
compared longitudinally and/or
cross-sectionally
Strengths
Weaknesses
394
Research Designs
due to their requirements and contributions vis-a`-vis theory and data, of these choices.
A second point is the need to creatively construct
bridges across research traditions. Although it is common
for certain traditions to be presented as inherently superior to others and the standard against which other traditions should be measured, this article has shown that all
traditions are characterized by certain strengths and
weaknesses and that it is thus more accurate and useful
to think in terms of the tradeoffs involved in working
within different traditions. An implication of this assessment, then, is that greater effort should be made to capitalize on what are clearly the complementary strengths of
different traditions (compare Tables IIIV).
Efforts at bridging, whether carried out through multiple studies on the same question or mixed designs that
combine multiple designs within a single study, are very
demanding. Thus, although it is common to point out that
multiple studies in the context of a shared research program offer a way of combining different designs, such
combinations are only effective inasmuch as research
programs are organized around clearly specified concepts
and questions and are advanced, at least to a certain extent, through explicitly coordinated teamwork. In turn,
the effective use of mixed designs requires a level of
methodological sophistication, as well as theoretical and
substantive knowledge, that is rare. Nonetheless, the high
payoffs associated with the use of mixed methods make
these options strongly recommendable.
Further Reading
Blalock, H. (1991). Are there really any constructive alternatives to causal modeling? Sociol. Methodol. 21, 325335.
Bollen, K. A. (1989). Structural Equations with Latent
Variables. Wiley, New York.
Brady, H. E., and Collier, D. (eds.) (2004). Rethinking Social
Inquiry: Diverse Tools, Shared Standards. Rowman &
Littlefield/Berkeley Public Policy Press, Lanham, MD.
Campbell, D. T. (1988a). Factors relevant to the validity of
experiments in social settings. In Methodology and
Epistemology for Social Science: Selected Papers. Donald
T. Campbell (E. Samuel Overman, ed.), pp. 151166.
University of Chicago Press, Chicago [Original work
published 1957].
Campbell, D. T. (1988b). Degrees of freedom and the case
study. In Methodology and Epistemology for Social Science:
Selected Papers. Donald T. Campbell (E. Samuel Overman,
Research Designs
ed.), pp. 377388. University of Chicago Press, Chicago
[Original work published 1975].
Campbell, D. T. (1999). Relabeling internal and external validity for applied social scientists. In Social Experimentation
(D. T. Campbell and M. J. Russo, eds.), pp. 111122. Sage,
Thousand Oaks, CA [Original work published 1986].
Collier, D. (1993). The comparative method. In Political
Science: The State of the Discipline II (A. W. Finifter, ed.),
pp. 105119. American Political Science Association,
Washington, DC.
Cronbach, L. J. (1982). Designing Evaluations of Educational
and Social Programs. Jossey-Bass, San Francisco.
Freedman, D. A. (1991). Statistical analysis and shoe leather.
Sociol. Methodol. 21, 291313.
Goldthorpe, J. H. (2000). On Sociology: Numbers, Narratives,
and the Integration of Research and Theory. Oxford
University Press, Oxford, UK.
Good, P. I. (2000). Permutation Tests: A Practical Guide to
Resampling Methods for Testing Hypotheses. Springer,
Berlin.
Kagel, J. H., and Roth, A. E. (eds.) (1995). The Handbook
of Experimental Economics. Princeton University Press,
Princeton, NJ.
395
King, G., Keohane, R. O., and Verba, S. (1994). Designing Social Inquiry. Scientific Inference in Qualitative
Research. Princeton University Press, Princeton, NJ.
Mahoney, J. (2000). Strategies of causal inference in small-N
research. Sociol. Methods Res. 28(4), 387424.
McDermott, R. (2002). Experimental methods in political
science. Annu. Rev. Political Sci. 5, 3161.
Oehlert, G. W. (2000). A First Course in Design and Analysis
of Experiments. Freeman, New York.
Shadish, W. R., Cook, T. D., and Campbell, D. T. (2002).
Experimental and Quasi-Experimental Designs for Generalized Causal Inference. Houghton Mifflin, Boston.
Smith, H. L. (1990). Specification problems in experimental and
nonexperimental social research. Sociol. Methodol. 20, 5991.
Tashakkori, A., and Teddlie, C. (1998). Mixed Methodology:
Combining Qualitative and Quantitative Approaches. Sage,
Thousand Oaks, CA.
Webb, E. J., Campbell, D. T., Schwartz R. D., and Sechrest, L.
(2000). Unobtrusive Measures. Revised Edition. Sage,
Thousand Oaks, CA.
Western, B., and Jackman, S. (1994). Bayesian inference
for comparative research. Am. Political Sci. Rev. 88(2),
412423.