Professional Documents
Culture Documents
Tutorial
Purpose: Single-subject experimental designs (SSEDs) SSEDs are included. Specific exemplars of how SSEDs
represent an important tool in the development and imple- have been used in speech-language pathology research are
mentation of evidence-based practice in communication provided throughout.
sciences and disorders. The purpose of this article is to review Conclusion: SSED studies provide a flexible alternative to
the strategies and tactics of SSEDs and their application in traditional group designs in the development and identification
speech-language pathology research. of evidence-based practice in the field of communication
Method: The authors discuss the requirements of each sciences and disorders.
design, followed by advantages and disadvantages. The logic
and methods for evaluating effects in SSED are reviewed
as well as contemporary issues regarding data analysis Key Words: single-subject experimental designs, tutorial,
with SSED data sets. Examples of challenges in executing research methods, evidence-based practice
T
he use of single-subject experimental designs (SSEDs)
has a rich history in communication sciences and high-quality research evidence is integrated with practitioner
disorders (CSD) research. A number of important experience and client preferences and values into the process
studies dating back to the 1960s and 1970s investigated of making clinical decisions.” The focus on the individual
fluency treatments using SSED approaches (e.g., Hanson, client afforded by SSEDs makes them ideal for clinical
1978; Haroldson, Martin, & Starr, 1968; Martin & Siegel, applications. The potential strength of the internal validity of
1966; Reed & Godden, 1977). Several reviews, tutorials, and SSEDs allows researchers, clinicians, and educators to ask
textbooks describing and promoting the use of SSEDs questions that might not be feasible or possible to answer
in CSD were published subsequently in the 1980s and with traditional group designs. Because of these strengths,
1990s (e.g., Connell, & Thompson, 1986; Fukkink, 1996; both clinicians and researchers should be familiar with the
Kearns, 1986; McReynolds & Kearns, 1983; McReynolds application, interpretation, and relationship between SSEDs
& Thompson, 1986; Robey, Schultz, Crawford, & Sinner, and evidence-based practice.
1999). Despite their history of use within CSD, SSEDs The goal of this tutorial is to familiarize readers with
are sometimes overlooked in contemporary discussions of the logic of SSEDs and how they can be used to establish
evidence-based practice. This article provides a compre- evidence-based practice. The basics of SSED methodology
hensive overview of SSEDs specific to evidence-based are described, followed by descriptions of several commonly
practice issues in CSD that, in turn, could be used to inform implemented SSEDs, including their benefits and limita-
disciplinary research as well as clinical practice. tions, and a discussion of SSED analysis and evaluation
In the current climate of evidence-based practice, the tools issues. A set of standards for the assessment of evidence
provided by SSEDs are relevant for researchers and prac- quality in SSEDs is then reviewed. Examples of how
titioners alike. The American Speech-Language-Hearing SSEDs have been used in CSD research are provided
Association (ASHA; 2005) promotes the incorporation of throughout. Finally, a number of current issues in SSEDs,
evidence-based practice into clinical practice, defining including effect size calculations and the use of statistical
techniques in the analysis of SSED data, are considered.
a
University of Minnesota, Minneapolis
Correspondence to Breanne Byiers: byier001@umn.edu The Role of SSEDs in Evidence-Based Practice
Editor: Carol Scheffner Hammer Numerous criteria have been developed to identify best
Associate Editor: Patrick Finn educational and clinical practices that are supported by re-
Received April 5, 2011 search in psychology, education, speech-language science,
Revision received September 21, 2011 and related rehabilitation disciplines. Some of the guidelines
Accepted May 20, 2012 include SSEDs as one experimental design that can help
DOI: 10.1044/1058-0360(2012/11-0036) identify the effectiveness of specific treatments (e.g.,
American Journal of Speech-Language Pathology • Vol. 21 • 397–414 • November 2012 • A American Speech-Language-Hearing Association 397
Chambless et al., 1998; Horner et al., 2005; Yorkston et al., continued increase during the intervention phase constitutes
2001). Many research communities, however, hold the a treatment effect is likely to be compromised. By con-
position that randomized control trials (RCTs) represent the vention, a minimum of three baseline data points are required
“gold standard” for research methodology aimed at validat- to establish dependent measure stability (Kazdin, 2010), with
ing best intervention practices; therefore, RCTs de facto more being preferable. If stability is not established in the
become the only valid research methodology that is necessary initial sessions, additional measurements should be obtained
for establishing evidence-based practice. until stability is achieved. Alternatively, steps can be taken
RCTs do have many specific advantages related to under- to introduce additional controls (strengthening internal
standing causal relations by addressing methodological validity) into the baseline sessions that may contribute to
issues that may compromise the internal validity of research variability.
studies. Kazdin (2010), however, compellingly argued that
certain characteristics of SSEDs make them an important
addition and alternative to large-group designs. He argued Visual Data Inspection as a Data Reduction
that RCTs may not be feasible with many types of inter- Strategy: Changes in Level, Trend,
ventions, as resources for such large-scale studies may not be and Variability
available to test the thousands of treatments likely in use in
Once the data in all conditions have been obtained, they
any given field. In addition, the carefully controlled con-
are examined for changes in one or more of three parameters:
ditions in which RCTs must be conducted to ensure that
the results are interpretable may not be comparable and/or level, trend (slope), and variability. Level refers to the av-
erage rate of performance during a phase. Panel A of Figure 1
possible to implement in real-life (i.e., uncontrolled) con-
shows hypothetical data demonstrating a change in level.
ditions. SSEDs are an ideal tool for establishing the viability
In this case, the average rate of performance during the base-
of treatments in real-life settings before attempts are made
line phase is lower than the average rate of performance
to implement them at the large scale needed for RCTs (i.e.,
during the intervention phase. Figure 1 also illustrates that
scaling up). Ideally, several studies using a variety of meth-
the change in level occurred immediately following the
odologies will be conducted to establish an intervention as
change in phase. The change in level is evident, in part,
evidence-based practice. When a treatment is established
because there is no overlap between the phases, meaning
as evidence based using RCTs, it is often interpreted as
meaning that the intervention is effective with most or all
individuals who participated. Unfortunately, this may not be
the case (i.e., there are responders and nonresponders). Thus, FIGURE 1. Hypothetical data demonstrating unambiguous
systematic evaluation of the effects of a treatment at an changes in level (Panel A), trend (Panel B), and variability
(Panel C).
individual level may be needed, especially within the context
of educational or clinical practice. SSEDs can be helpful
in identifying the optimal treatment for a specific client and
in describing individual-level effects.
design (see Table 1) to determine whether it meets the stan- specific experimental design implemented. For many designs,
dards, with or without reservations. If the design is not found each time the intervention is implemented (or withdrawn
to be adequate, no further steps are needed. If the design following an initial intervention phase), an opportunity to pro-
meets the standards, the second step is to conduct a visual vide an instance of effect replication is created. This within-
analysis of the results to determine whether the data suggest study replication is the basis of internal validity for SSEDs.
an experimental effect. If the visual analysis supports the By replicating an investigation across different partici-
presence of an effect, the data should be examined for dem- pants, or different types of participants, researchers and
onstrations of noneffect, such as those depicted in Figure 2. If clinicians can examine the generality of the treatment effects
no evidence of an experimental effect is found, the process and thus potentially enhance external validity. Kazdin (2010)
is terminated. If the visual analysis suggests that the results distinguished between two types of replication. Direct rep-
support the effectiveness of the intervention, the reviewer lication refers to the application of an intervention to new
can move on to the third step: assessing the overall state of participants under exactly, or nearly exactly, the same
the evidence in favor of an intervention by examining the conditions as those included in the original study. This type
number of times its effectiveness has been demonstrated, both of replication allows the researcher or clinician to determine
within and across participants. The importance of replication whether the findings of the initial study were specific to
in SSEDs is discussed in more detail in the next section. the participant(s) who were involved. Systematic replication
If the design meets the standards and the visual analysis involves the repetition of the investigation while systemat-
indicates that there is an effect, with no demonstrations of ically varying one or more aspects of the original study. This
noneffect, the study would be considered one that pro- might include applying the intervention to participants with
vides strong evidence. If it meets the standards and there is more heterogeneous characteristics, conducting the inter-
evidence of an effect, but the results include at least one vention in a different setting with different dependent var-
demonstration of noneffect, then the study would be con- iables, and so forth. The variation inherent to systematic
sidered one that provides moderate evidence. The results replication allows the researcher, educator, or clinician to
of all studies that reported the effects of a particular interven- determine the extent to which the findings will generalize
tion can then be examined for overall level of evidence in across different types of participants, settings, or target
favor of the treatment. behaviors. As noted by Johnston and Pennypacker (2009),
conducting direct replications of an effect tells us about the
certainty of our knowledge, whereas conducting systematic
Replication for Internal and External Validity replications can expand the extent of our knowledge.
Replication is one of the hallmarks of SSEDs. Experi- An intervention or treatment cannot be considered evidence
mental control is demonstrated when the effects of the inter- based following the results of a single study. The WWCH
vention are repeatedly and reliably demonstrated within a panel recommended that an intervention have a minimum of
single participant or across a small number of participants. five supporting SSED studies meeting the evidence standards
The way in which the effects are replicated depends on the if the studies are to be combined into a single summary rating
Pre-experimental Does outcome X change from • Quick and efficient to implement. • Does not control for threats
(AB) baseline levels with the introduction • Appropriate for low-stakes to internal validity; not an
of intervention B? decision making. experimental design.
Withdrawal Does outcome X covary with • Easy to implement, strong • There are ethical considerations
(ABA/ABAB) introduction and withdrawal of experimental control when regarding withdrawing or reversing
intervention B? effects are immediate and large. a potentially effective intervention.
• Not all behaviors are “reversible.”
Multiple- Does outcome X change from • Does not require withdrawal • Ethical considerations regarding
baseline/ baseline levels with the introduction of intervention. keeping individuals/behaviors
multiple-probe of intervention B over multiple • Appropriate for nonreversible in baseline conditions for a long
participants, responses, settings, behaviors. period.
etc.? • Requires multiple individuals, responses,
settings, etc., that are comparable in
order to replicate effects.
Changing- Do changes in the level • Does not require reversal. • Change must take place in graduated
criterion of outcome X correspond • Appropriate for behaviors that steps; not appropriate for behaviors
to changes in the intervention can be changed gradually. that require immediate change.
criteria? • Useful for consequence-based • Requires the use of incentive- or
interventions. consequence-based interventions.
Multiple-treatment What are the relative effects • Can be extended to compare • Behaviors should be reversible to
of interventions A and B any number of interventions or demonstrate relative effects.
(and C, D, etc.) on outcome variables. • Only comparisons between adjacent
X compared to each other • Can extend a withdrawal study conditions are appropriate.
and/or baseline levels? when effects of initial intervention • Can be time consuming and
are not as pronounced as expected. complicated to implement when
• Can be used to conduct the number of interventions being
component analyses of compared increases.
necessary and sufficient • Results are susceptible to multiple
intervention components. treatment interference.
Alternating What are the relative effects • Can be extended to compare • Behaviors must be readily
treatments of interventions A and B any number of interventions reversible to obtain differentiation
(and C, D, etc.) on outcome X or variables. between conditions.
compared with each other • Can provide strong experimental • Results are susceptible to multiple
and/or baseline levels? evidence in relatively few sessions. treatment interference.
Adapted What are the relative effects of • Less prone to multiple • Set of behaviors or stimuli must
alternating intervention A on outcome X and treatment interference. be directly comparable for effects
treatments intervention B on outcome Y? • Can provide strong experimental to be meaningful.
evidence in relatively few sessions. • Potential generalization across
• Does not require reversal. behaviors must be considered.
convincingly that the intervention is responsible for changes must be carefully selected to ensure that any learning that
in the behavior across all the conditions. When implementing takes place in one will not transfer to the next. Similarly,
the intervention across individuals, it may be necessary— contexts or stimuli must be sufficiently dissimilar so as to
to avoid diffusion of the treatment—to ensure that the minimize the likelihood of effect generalization.
participants do not interact with one another. When the Although an assumption of independence suggests that
intervention is implemented across behaviors, the behaviors researchers should select conditions that are clearly dissimilar
FIGURE 5. Speech volume during a token reinforcement intervention and follow-up using a changing-
criterion design. From “A controlled single-case treatment of severe long-term selective mutism in a child
with mental retardation,” by Facon, Sahiri, and Riviere, (2008), Behavior Therapy, 39, p. 313. Copyright 2008
by Elsevier. Reprinted with permission.
and time-out was implemented. By the end of the phase, the Intervention 1. In essence, this creates the potential for an
target behavior was eliminated, prompting had decreased, order effect (or a carryover effect). Alternatively, Interven-
and signing had increased. To identify the active components tion 1 may have measurable but delayed effects on the
of the treatment package, a dropout component analysis dependent variable, making it appear that Intervention 2 is
was conducted. First, the time-out component of the inter- effective when the results should be attributed to Interven-
vention was removed, leaving the FCT component alone. tion 1. Such possibilities should be considered when multi-
A decreasing trend in signing and an increasing trend in treatment studies are being planned (see Hains & Baer, 1989,
hand biting were observed. This was reversed when the full for a comprehensive discussion of multiple-treatment inter-
treatment packaged was reimplemented. In the third phase ference). A final, longer phase in which the final “winning”
of the component analysis, the FCT component was removed, treatment is implemented for an extended time can help
leaving time-out and differential reinforcement of other be- alleviate some of the concerns regarding multiple-treatment
havior (DRO). Again, a decreasing trend in signing and an interference.
increasing trend in hand biting were observed, which were Advantages and disadvantages of multiple-treatment
again reversed when the full treatment package was applied. designs. Designs such as ABCABC and ABCBCA can be
Overall, visual inspection of these data provides a strong very useful when a researcher wants to examine the effects
argument for the necessity of both the FCT and time-out of two interventions. These designs provide strong internal
components in the effectiveness of the treatment package, validity evidence regarding the effectiveness of the inter-
and no indications of noneffect are present in the data. The ventions. External validity, however, may be compromised
design, however, does not meet the standards set forth by by the threat of multiple-treatment interference. Additionally,
the WWCH panel. This is because (a) the final two final the same advantages and disadvantages of ABAB designs
treatment phases do not include the minimum of three data apply, including issues related to the reversibility of the
points and (b) the individual treatment component phases target behavior. Despite their limitations, these designs can
(FCT only and time-out/DRO) were implemented only once provide strong empirical data upon which to base decisions
each. As a result, the data from this study could not be used regarding the selection of treatments for an individual client.
to support the treatment package as an evidence-based Although, in theory, these types of designs can be extended
practice by the IES standards. Additional data points to compare any number of interventions or conditions, doing
within each phase, as well as replications of the phases, so beyond two becomes excessively cumbersome; there-
would strengthen the study results. fore, the alternating treatments design should be considered.
One disadvantage of all designs that involve two or more
interventions or independent variables is the potential for
Alternating Treatments and Adapted Alternating
multiple-treatment interference. This occurs when the same
participant receives two or more treatments whose effects Treatments Designs
may not be independent. As a result, it is possible that the Alternating treatments design (ATD). The logic of the
order in which the interventions are given will affect the ATD is similar to that of multiple-treatment designs, and the
results. For example, the effects of two interventions may be types of research questions that it can address are also com-
additive, so that the effects of Intervention 2 are enhanced parable. The major distinction is that the ATD involves the
beyond what they should be because Intervention 2 followed rapid alternation of two or more interventions or conditions
FIGURE 7. Number of correct responses and tongue clicks during discrete trial training sessions in Spanish (Sp.)
and English (Eng.) using an alternating treatments design. From “Effects of language instruction on response
accuracy and challenging behavior in a child with autism,” by Lang et al., 2011, Journal of Behavioral Education,
20, p. 256. Copyright 2001 by Springer Science+Business Media, LLC. Reprinted with permission.
The references for this article include 10 HighWire-hosted articles which you can
access for free at: http://ajslp.asha.org/cgi/content/full/21/4/397#BIBL